From 4dee78d553fb29477ec1cfa3df0b7b91a0c6b527 Mon Sep 17 00:00:00 2001 From: lvhan028 Date: Wed, 28 Aug 2024 13:18:01 +0800 Subject: [PATCH] fix lint --- src/turbomind/triton_backend/llama/LlamaTritonModel.cc | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/src/turbomind/triton_backend/llama/LlamaTritonModel.cc b/src/turbomind/triton_backend/llama/LlamaTritonModel.cc index 2e8750b0c..ee09600b0 100644 --- a/src/turbomind/triton_backend/llama/LlamaTritonModel.cc +++ b/src/turbomind/triton_backend/llama/LlamaTritonModel.cc @@ -44,7 +44,7 @@ std::shared_ptr AbstractTransformerModel::createLlamaM ft::FT_CHECK(false); } - auto ft_instance_hyperparameter = reader["ft_instance_hyperparameter"]; + auto ft_instance_hyperparameter = reader["ft_instance_hyperparameter"]; const std::string data_type = ft_instance_hyperparameter["data_type"].as(); int tensor_para_size = ft_instance_hyperparameter["tensor_para_size"].as(); std::string model_dir = ft_instance_hyperparameter["model_dir"].as(); @@ -242,7 +242,8 @@ LlamaTritonModel::LlamaTritonModel(size_t tensor_para_size, attn_param_.use_dynamic_ntk = reader["attention_config"]["use_dynamic_ntk"].as(0); attn_param_.use_logn_attn = reader["attention_config"]["use_logn_attn"].as(0); - attn_param_.original_max_position_embeddings = reader["attention_config"]["original_max_position_embeddings"].as(0); + attn_param_.original_max_position_embeddings = + reader["attention_config"]["original_max_position_embeddings"].as(0); engine_param_.max_batch_size = reader["engine_config"]["max_batch_size"].as(0); engine_param_.max_prefill_token_num = reader["engine_config"]["max_prefill_token_num"].as(0);