Skip to content

Commit

Permalink
[Bugfix] MLPSpeculator: Use ParallelLMHead in tie_weights=False case. (
Browse files Browse the repository at this point in the history
…vllm-project#6303)

Signed-off-by: Thomas Parnell <[email protected]>
  • Loading branch information
tdoublep authored and jimpang committed Jul 24, 2024
1 parent 445987d commit 836ae08
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion vllm/model_executor/models/mlp_speculator.py
Original file line number Diff line number Diff line change
Expand Up @@ -110,7 +110,7 @@ def __init__(self, config: MLPSpeculatorConfig, **kwargs) -> None:
])

self.head = nn.ModuleList([
nn.Linear(self.inner_dim, self.vocab_size, bias=False)
ParallelLMHead(self.vocab_size, self.inner_dim, bias=False)
for _ in range(self.max_speculative_tokens)
])
self.ln = nn.ModuleList([
Expand Down

0 comments on commit 836ae08

Please sign in to comment.