Skip to content

Commit

Permalink
fix style
Browse files Browse the repository at this point in the history
  • Loading branch information
rnwang04 committed Feb 7, 2025
1 parent e26dc9a commit 7da6608
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -460,7 +460,7 @@ def convert_fused_minicpm_layer(model, fused_layers, n_splits_linear, n_splits_d
np_dtype = np.int8 if weights[0][0].dtype == torch.int8 else np.uint8
else: # FP16 Linear
np_dtype = np.float16

if not const_parameter:
input_layer_norm_weights = None
post_attn_layernorm_weights = None
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -324,7 +324,7 @@ def convert_fused_qwen_layer(model, fused_layers, n_splits_linear, n_splits_down
np_dtype = np.int8 if weights[0][0].dtype == torch.int8 else np.uint8
else: # FP16 Linear
np_dtype = np.float16

if not const_parameter:
input_layer_norm_weights = None
post_attn_layernorm_weights = None
Expand Down

0 comments on commit 7da6608

Please sign in to comment.