Skip to content

Commit

Permalink
use fused moe for all arches by default
Browse files Browse the repository at this point in the history
  • Loading branch information
lzhangzz committed Nov 4, 2024
1 parent aaaad76 commit 4f67317
Showing 1 changed file with 0 additions and 6 deletions.
6 changes: 0 additions & 6 deletions src/turbomind/triton_backend/llama/LlamaTritonModel.cc
Original file line number Diff line number Diff line change
Expand Up @@ -336,12 +336,6 @@ LlamaTritonModel<T>::LlamaTritonModel(size_t tensor_para_size,
}
else {
moe_param_.method = ft::MoeParam::kFused;
// Note: This will fail when GPUs of different SMs are mixed
if (weight_type_ != ft::WeightType::kINT4 && ft::getSMVersion() >= 90) {
// On sm90 the cuBLAS method may be faster as our grouped GEMM is not
// optimized for GMMA yet
moe_param_.method = ft::MoeParam::kNaive;
}
}

TM_LOG_INFO("%s", toString().c_str());
Expand Down

0 comments on commit 4f67317

Please sign in to comment.