We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b1bb2c4 commit a039040Copy full SHA for a039040
vllm/model_executor/models/llama_eagle3.py
@@ -97,12 +97,12 @@ def forward(
97
residual: Optional[torch.Tensor],
98
) -> tuple[torch.Tensor, torch.Tensor]:
99
if self.layer_idx == 0:
100
+ # First layer: concatenate embeds with hidden_states
101
embeds = self.input_layernorm(embeds)
102
hidden_states, residual = self._residual_norm(hidden_states=hidden_states)
103
hidden_states = torch.cat([embeds, hidden_states], dim=-1)
104
else:
- # Subsequent layers: only process hidden_states
105
- # and residuals
+ # Subsequent layers: process hidden_states and residuals only
106
hidden_states, residual = self.input_layernorm(hidden_states, residual)
107
108
# Self Attention
0 commit comments