Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bidirectional model? #99

Open
fransilvionGenomica opened this issue Jan 10, 2024 · 2 comments
Open

Bidirectional model? #99

fransilvionGenomica opened this issue Jan 10, 2024 · 2 comments

Comments

@fransilvionGenomica
Copy link

Hello,

Thanks for the great paper! If I understand correctly, Mamba model is similar to a one directional LSTM. Is there a way to implement it in not causal but bidirectional way, so the model can see information from both sequence ends? It would be similar to BERT encoder architecture in that sense I guess.

@albertfgu
Copy link
Contributor

There are many approaches to this. You can implement a variety of naive methods to incorporate information from both ends, such as in this PR: #52.

There are some slightly more clever methods that are in progress. Stay tuned!

@AxelElaldi
Copy link

Hi! Thank you for your awesome work! I am also interested in using Mamba with different bidirectional methods. Besides PR #52, is there any new features related to the more advanced methods you mentioned? Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants