Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add BibleNLP dataset #583

Merged
merged 9 commits into from
Apr 28, 2024
Merged

Conversation

davidstap
Copy link
Contributor

@davidstap davidstap commented Apr 26, 2024

Checklist for adding MMTEB dataset

This dataset has an extremely large coverage of 829 languages, and includes numerous low-resource languages. For this reason I only considered English-centric directions, resulting in a reasonable 1656 pairs. I will add performance results later, but initial testing suggests that while quality is not good for most (low-resource) languages, it is far better than random guesses, so IMO a good addition to the benchmark.

  • I have tested that the dataset runs with the mteb package.
  • I have run the following models on the task (adding the results to the pr). These can be run using the mteb run -m {model_name} -t {task_name} command.
    • sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
    • intfloat/multilingual-e5-small
  • I have checked that the performance is neither trivial (both models gain close to perfect scores) nor random (both models gain close to random scores).
  • If the dataset is too big (e.g. >2048 examples), considering using self.stratified_subsampling() under dataset_transform()
  • I have filled out the metadata object in the dataset file (find documentation on it here).
  • Run tests locally to make sure nothing is broken using make test.
  • Run the formatter to format the code using make lint.
  • I have added points for my submission to the points folder using the PR number as the filename (e.g. 438.jsonl).

@davidstap
Copy link
Contributor Author

davidstap commented Apr 26, 2024

Could you review this PR @KennethEnevoldsen @imenelydiaker? I'm personally quite excited by the large language coverage.

Copy link
Collaborator

@isaac-chung isaac-chung left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is awesome! Just a few comments.

Copy link
Collaborator

@dokato dokato left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Really interesting dataset :) find some of my suggestions below

@davidstap
Copy link
Contributor Author

I have addressed your comments @isaac-chung @dokato, thanks for your suggestions!

Results are included now as well. intfloat/multilingual-e5-small performs best, but like I pointed out the performance is bad for most low-resource languages. However, almost all directions are way better than random chance (1/256), so this seems like a valuable addition to evaluate (future) multilingual models.

I think this is ready to merge now, do you agree? Let me know, then I'll calculate the points. Thanks!

@davidstap davidstap marked this pull request as ready for review April 28, 2024 10:55
Copy link
Collaborator

@isaac-chung isaac-chung left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for iterating! I think it's good enough to merge. In terms of points, the dataset is 2 + 4 * number of new languages added to the BitextMining task. Please check if any of the languages have already been covered in that folder and the multilingual subfolder as well. Then 2 pts each to @dokato and I for the reviews.

@davidstap
Copy link
Contributor Author

davidstap commented Apr 28, 2024

Thanks for your quick response! I calculated the number of new languages using this script: 756 new languages (out of 829 languages used in this task).

I'll update the points and merge.

@davidstap
Copy link
Contributor Author

Ready to merge now @isaac-chung (I don't have write access)

@isaac-chung isaac-chung merged commit 593fc8f into embeddings-benchmark:main Apr 28, 2024
7 checks passed
@KennethEnevoldsen KennethEnevoldsen mentioned this pull request May 10, 2024
10 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants