Skip to content

Commit

Permalink
Clang format
Browse files Browse the repository at this point in the history
  • Loading branch information
BruceXcluding committed Jan 2, 2025
1 parent 4bb3332 commit b10c089
Showing 1 changed file with 8 additions and 4 deletions.
12 changes: 8 additions & 4 deletions python/sglang/srt/layers/moe/fused_moe_triton/fused_moe.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,15 +11,19 @@
import torch
import triton
import triton.language as tl
from sgl_kernel import moe_align_block_size as sgl_moe_align_block_size
from vllm import _custom_ops as ops

from sglang.srt.layers.moe.topk import select_experts
from sglang.srt.layers.quantization.fp8_kernel import per_token_group_quant_fp8
from sglang.srt.utils import direct_register_custom_op, get_device_name, is_hip
from sgl_kernel import moe_align_block_size as sgl_moe_align_block_size

is_hip_flag = True if is_hip() else False
is_hip_flag = False
if not is_hip():
from sgl_kernel import moe_align_block_size as sgl_moe_align_block_size

is_hip_flag = False
else:
is_hip_flag = True

logger = logging.getLogger(__name__)
padding_size = 128 if bool(int(os.getenv("MOE_PADDING", "0"))) else 0
Expand Down Expand Up @@ -408,7 +412,7 @@ def moe_align_block_size(
)
num_tokens_post_pad = torch.empty((1), dtype=torch.int32, device=topk_ids.device)
if num_experts >= 224:
if enable_moe_align_block_size_triton:
if enable_moe_align_block_size_triton or is_hip_flag:
moe_align_block_size_triton(
topk_ids,
num_experts,
Expand Down

0 comments on commit b10c089

Please sign in to comment.