Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Added XLMRobertaModel for Reranking #2686

Merged
merged 3 commits into from
Dec 30, 2024

Conversation

akshayballal95
Copy link
Contributor

Some of the good reranker models are based on this architecture and I think its a good addition to the current repository of models.

The architecture is not very different than Bert but the position embedding technique is a bit different.

@akshayballal95 akshayballal95 changed the title Added XLMRobertaModels for Reranking Added XLMRobertaModel for Reranking Dec 29, 2024
akshayballal95 and others added 2 commits December 30, 2024 10:16
…te README and fix attention mask dimensions

- Introduced a new `Task` enum to replace string task identifiers in the xlm-roberta example.
- Updated the logic in `main.rs` to handle tasks using the new enum.
- Enhanced README with example output for fill-mask task.
- Fixed dimension retrieval in `prepare_4d_attention_mask` function for better clarity and safety.
@LaurentMazare LaurentMazare merged commit 91f1f01 into huggingface:main Dec 30, 2024
9 of 10 checks passed
@LaurentMazare
Copy link
Collaborator

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants