Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] Create embedding module #86

Draft
wants to merge 6 commits into
base: main
Choose a base branch
from
Draft

Conversation

shavit
Copy link
Contributor

@shavit shavit commented Mar 30, 2024

Closes #22

@shavit shavit mentioned this pull request Apr 20, 2024
Create layers with weights
shavit added 4 commits July 9, 2024 15:09
* Test addition of MLMultiArrays and broadcasting
* Create generic number array from MLMultiArray
* Add double type to config
@bryan1anderson
Copy link

Any updates on this one?

@shavit
Copy link
Contributor Author

shavit commented Oct 18, 2024

Right, I’ve got the BERT embeddings implemented, but this pull request relies on another one for reading safe tensors. That said, having the BERT tokenizer implemented should be sufficient, and I've already done that in a different branch 4-pretokenizer-bert-unsupported, and merged both into another project.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Support for embedding models (BGE, GTE etc)
2 participants