diff --git a/ark/ops/ops_matmul.cc b/ark/ops/ops_matmul.cc index b9c12b65e..40a2aafc2 100644 --- a/ark/ops/ops_matmul.cc +++ b/ark/ops/ops_matmul.cc @@ -237,9 +237,8 @@ Tensor *Model::matmul(Tensor *mat_a, Tensor *mat_b, Tensor *mat_y, Dims split_output_shape = output_shape; split_output_shape[0] *= split_k; Tensor *output_buffer = this->tensor(split_output_shape, mat_y->type); - vector mat_y_shards = - this->sharding(output_buffer, 0, output_shape[0], - name + "/sharding_mat_y"); + vector mat_y_shards = this->sharding( + output_buffer, 0, output_shape[0], name + "/sharding_mat_y"); int axis_a; int axis_b; diff --git a/ark/ops/ops_reduce.cc b/ark/ops/ops_reduce.cc index f64da0fd3..afbfa4cbc 100644 --- a/ark/ops/ops_reduce.cc +++ b/ark/ops/ops_reduce.cc @@ -138,7 +138,7 @@ std::string ReduceEMeanOp::function_name(const OpConfig &cfg) const template Tensor *Model::reduce(Tensor *input, int axis, Tensor *output, - const std::string &name) + const std::string &name) { assert(input != nullptr); OpPrecType pt = OP_PREC_NONE; @@ -158,12 +158,13 @@ Tensor *Model::reduce(Tensor *input, int axis, Tensor *output, output = this->tensor(reduced_shape, input->type); } else { if (output->shape != reduced_shape) { - LOG(ERROR, "invalid output shape ", output->shape, " with input " - "shape ", input->shape, " and reduction axis ", axis); + LOG(ERROR, "invalid output shape ", output->shape, + " with input shape ", input->shape, " and reduction axis ", + axis); } if (output == input) { LOG(ERROR, "output tensor cannot be the same as input tensor for " - "reduce_sum op"); + "reduce_sum op"); } } ReduceOpType op{pt, input, output, axis, name};