Skip to content

Commit

Permalink
Update llama.py
Browse files Browse the repository at this point in the history
  • Loading branch information
danielhanchen committed Jan 26, 2024
1 parent bd2ff90 commit a208ec4
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions unsloth/models/llama.py
Original file line number Diff line number Diff line change
Expand Up @@ -234,7 +234,7 @@ def LlamaAttention_fast_forward(
bsz, q_len, _ = hidden_states.size()

# Check for inference
if past_key_value is not None and q_len == 1 and bsz == 1:
if False:#past_key_value is not None and q_len == 1 and bsz == 1:
A, past_key_value = LlamaAttention_fast_forward_inference(
self,
hidden_states,
Expand Down Expand Up @@ -350,7 +350,7 @@ def LlamaDecoderLayer_fast_forward(
past_key_value (`Tuple(torch.FloatTensor)`, *optional*): cached past key and value projection states
"""
bsz, q_len, hd = hidden_states.size()
if (past_key_value is not None and q_len == 1 and bsz == 1):
if False:#(past_key_value is not None and q_len == 1 and bsz == 1):
# Self Attention
residual = hidden_states
hidden_states = fast_rms_layernorm_inference(self.input_layernorm, hidden_states)
Expand Down

0 comments on commit a208ec4

Please sign in to comment.