Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Tracking][WebLLM] Function calling (beta) and Embeddings #526

Open
5 of 7 tasks
CharlieFRuan opened this issue Aug 4, 2024 · 1 comment
Open
5 of 7 tasks

[Tracking][WebLLM] Function calling (beta) and Embeddings #526

CharlieFRuan opened this issue Aug 4, 2024 · 1 comment

Comments

@CharlieFRuan
Copy link
Contributor

CharlieFRuan commented Aug 4, 2024

This issue tracks various action items we would like to complete with regard to the features function calling and embeddings.

Function calling (beta)

We are calling it beta because multiple iterations may be needed for function calling. It may be hard to conform different open-source models' function calling formats to OpenAI API. We will try to make each iteration non-breaking.

Embedding, Multi-model Engine, Concurrency

@CharlieFRuan CharlieFRuan changed the title [Tracking][WebLLM] Function calling and Embeddings [Tracking][WebLLM] Function calling (beta) and Embeddings Aug 4, 2024
@CharlieFRuan
Copy link
Contributor Author

Some future TODOs for embeddings:

  • Support nomic-v1.5 for longer context (may not simply return first token logits like snowflake-arctic, hence requiring changes in EmbeddingPipeline)
  • Support matryoshka models, hence support dimension field for those models
  • Implement prefill chunking (currently not needed since we only support 512 window size)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: In Progress
Development

No branches or pull requests

1 participant