Skip to content

Commit

Permalink
Update can-llama-3-listen.mdx (#58)
Browse files Browse the repository at this point in the history
Updated the Hugging Face links.
  • Loading branch information
eckartal authored Aug 8, 2024
1 parent d938464 commit 28250d9
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions src/pages/blog/can-llama-3-listen.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ import ResearchCTABlog from '@/components/Blog/ResearchCTABlog'
<BlogAuthors authors={["Alan Dao", "Rex Ha", "Bach Vu", "Phong Tran"]}/>

<Callout emoji="πŸ“š">
We invite you to join [llama3-s](https://discord.gg/Q7dXd6b9Zb): an open and ongoing research experiment focused on teaching llama3 to listen. This blog post is a quick project update and an open call to join to us to build a speech adapter for open source models.
We invite you to join [llama3-s](https://discord.gg/Q7dXd6b9Zb): an open and ongoing research experiment focused on teaching llama3 to listen. This blog post is a quick project update and an open call to join us in building a speech adapter for open-source models.
</Callout>

Inspired by the [Chameleon](https://arxiv.org/pdf/2405.09818) and [Llama Herd](https://arxiv.org/pdf/2407.21783) papers, llama3-s is an early-fusion, audio and text, multimodal model. We're conducting this research entirely in the open, with an open-source [codebase](https://github.com/homebrewltd/llama3-s), [open data](https://huggingface.co/datasets/homebrewltd/instruction-speech-v1.5) and [open weights](https://huggingface.co/homebrewltd/llama3-s-2024-07-19).
Expand Down Expand Up @@ -80,9 +80,9 @@ You can find the datasets here:

| Date | HF Checkpoint | Tokens |
| --- | --- | --- |
| πŸ“… 2024-07-19 | πŸ”—Β https://huggingface.co/homebrew-research | πŸ”’ 1.35B |
| πŸ“… 2024-07-18 | πŸ”—Β https://huggingface.co/datasets/homebrew-research/instruction-speech-v1.5 | πŸ”’ 800M |
| πŸ“… 2024-06-30 | πŸ”—Β https://huggingface.co/datasets/homebrew-research/instruction-speech-v1 | πŸ”’ 450M |
| πŸ“… 2024-07-19 | πŸ”—Β https://huggingface.co/homebrewltd | πŸ”’ 1.35B |
| πŸ“… 2024-07-18 | πŸ”—Β https://huggingface.co/datasets/homebrewltd/instruction-speech-v1.55 | πŸ”’ 800M |
| πŸ“… 2024-06-30 | πŸ”—Β https://huggingface.co/datasets/homebrewltd/instruction-speech-v1 | πŸ”’ 450M |

**Training**: The instruct tuning was done with fsdp2 ([Torchtune](https://github.com/pytorch/torchtune)) mixed-precision, on a [llama3 8b](https://huggingface.co/meta-llama/Meta-Llama-3-8B) base model, with the final weights in bf16. We used the [AdamMini](https://arxiv.org/abs/2406.16793) optimizer, a global batchsize of 128 (mini-batches of 2-4), a 3e-4 learning rate, and a slightly longer warm up ratio. You can find the full steps to reproduce our training here on [Hugging Face](https://huggingface.co/homebrewltd/llama3-s-2024-07-19).

Expand Down

0 comments on commit 28250d9

Please sign in to comment.