Skip to content

Commit

Permalink
update
Browse files Browse the repository at this point in the history
  • Loading branch information
songhappy committed Sep 4, 2024
1 parent f325660 commit 428e62b
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion python/llm/src/ipex_llm/transformers/models/llama.py
Original file line number Diff line number Diff line change
Expand Up @@ -1579,7 +1579,7 @@ def llama_attention_forward_4_41_original(
past_key_value.key_cache[self.layer_idx] = key_states
past_key_value.value_cache[self.layer_idx] = value_states

if cache_position is not None:
if attention_mask is not None:
new_attention_mask = attention_mask[:, :, :, 0:kv_seq_len]
else:
new_attention_mask = attention_mask
Expand Down

0 comments on commit 428e62b

Please sign in to comment.