Skip to content

Commit

Permalink
Apply suggestions from code review
Browse files Browse the repository at this point in the history
Co-authored-by: Mohit Sharma <[email protected]>
  • Loading branch information
Narsil and mht-sharma authored Jan 30, 2025
1 parent 9376066 commit f56e24b
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 4 deletions.
4 changes: 2 additions & 2 deletions Dockerfile_amd
Original file line number Diff line number Diff line change
Expand Up @@ -279,9 +279,9 @@ RUN git clone https://github.com/danieldk/marlin-kernels.git && \

FROM kernel-builder AS moe-kernels
WORKDIR /usr/src
ENV MOE_KERNELS_BRANCH=9fb72ccb00b8da96a41f9b9c48f305c7dbd266bc
ENV MOE_KERNELS_BRANCH=d7e042bf9f7aff10c631212fc71b24895d66eb59
ENV VLLM_TARGET_DEVICE=rocm
RUN git clone https://github.com/mht-sharma/moe-kernels.git && \
RUN git clone https://github.com/danieldk/moe-kernels.git && \
cd moe-kernels && \
git checkout ${MOE_KERNELS_BRANCH} && \
python setup.py install
Expand Down
2 changes: 0 additions & 2 deletions server/text_generation_server/layers/moe/unquantized.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,8 +73,6 @@ def forward(self, x: torch.Tensor, *, gating_output: torch.Tensor) -> torch.Tens
num_expert_group=self.n_expert_group,
topk_group=self.topk_group,
)
# from loguru import logger
# logger.info("Fused MoE is used here")
return fused_moe(
x,
w1=self.gate_up_proj,
Expand Down

0 comments on commit f56e24b

Please sign in to comment.