From 617c5943e47f27022968dc6d70551372486e632f Mon Sep 17 00:00:00 2001 From: nihuini Date: Wed, 8 May 2024 19:09:12 +0800 Subject: [PATCH] clean --- src/layer/arm/gru_int8.h | 6 ------ src/layer/arm/lstm_int8.h | 6 ------ src/layer/arm/rnn_int8.h | 6 ------ 3 files changed, 18 deletions(-) diff --git a/src/layer/arm/gru_int8.h b/src/layer/arm/gru_int8.h index 40fd17ece23..8d0c720bb46 100644 --- a/src/layer/arm/gru_int8.h +++ b/src/layer/arm/gru_int8.h @@ -23,9 +23,6 @@ void gru_int8_gate_output_vfpv4(const Mat& gates, Mat& hidden_state, Mat& top_bl static void gru_transform_weight_int8(const Mat& weight_xc, const Mat& weight_xc_int8_scales, const Mat& weight_hc, const Mat& weight_hc_int8_scales, const Mat& bias_c, Mat& weight_data_tm, Mat& weight_data_tm_int8_descales, Mat& bias_c_tm, int size, int num_output, int num_directions, const Option& opt) { - // TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC - // TODO dispatch for __ARM_FEATURE_DOTPROD - #if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD if (ncnn::cpu_support_arm_asimddp()) { @@ -643,9 +640,6 @@ static void gru_int8_gate_output(const Mat& gates, Mat& hidden_state, Mat& top_b static void gru_int8(const Mat& bottom_blob_int8, const Mat& bottom_blob_int8_descales, Mat& top_blob, int elemtype, int reverse, const Mat& weight_data_tm, const Mat& weight_data_tm_int8_descales, const Mat& bias_c, Mat& hidden_state, const Option& opt) { - // TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC - // TODO dispatch for __ARM_FEATURE_DOTPROD - #if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD if (ncnn::cpu_support_arm_asimddp()) { diff --git a/src/layer/arm/lstm_int8.h b/src/layer/arm/lstm_int8.h index 6c5e681b0ba..8e7a4ed348a 100644 --- a/src/layer/arm/lstm_int8.h +++ b/src/layer/arm/lstm_int8.h @@ -23,9 +23,6 @@ void lstm_int8_gate_output_vfpv4(const Mat& gates, const Mat& weight_hr, Mat& hi static void lstm_transform_weight_int8(const Mat& weight_xc, const Mat& weight_xc_int8_scales, const Mat& weight_hc, const Mat& weight_hc_int8_scales, const Mat& bias_c, Mat& weight_data_tm, Mat& weight_data_tm_int8_descales, Mat& bias_c_tm, int size, int num_output, int num_directions, int hidden_size, const Option& opt) { - // TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC - // TODO dispatch for __ARM_FEATURE_DOTPROD - #if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD if (ncnn::cpu_support_arm_asimddp()) { @@ -392,9 +389,6 @@ static void lstm_int8_gate_output(const Mat& gates, const Mat& weight_hr, Mat& h static void lstm_int8(const Mat& bottom_blob_int8, const Mat& bottom_blob_int8_descales, Mat& top_blob, int elemtype, int reverse, const Mat& weight_data_tm, const Mat& weight_data_tm_int8_descales, const Mat& bias_c, const Mat& weight_hr, Mat& hidden_state, Mat& cell_state, const Option& opt) { - // TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC - // TODO dispatch for __ARM_FEATURE_DOTPROD - #if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD if (ncnn::cpu_support_arm_asimddp()) { diff --git a/src/layer/arm/rnn_int8.h b/src/layer/arm/rnn_int8.h index 7f5c6a99686..2850fe9cc98 100644 --- a/src/layer/arm/rnn_int8.h +++ b/src/layer/arm/rnn_int8.h @@ -23,9 +23,6 @@ void rnn_int8_gate_output_vfpv4(const Mat& gates, Mat& hidden_state, Mat& top_bl static void rnn_transform_weight_int8(const Mat& weight_xc, const Mat& weight_xc_int8_scales, const Mat& weight_hc, const Mat& weight_hc_int8_scales, const Mat& bias_c, Mat& weight_data_tm, Mat& weight_data_tm_int8_descales, Mat& bias_c_tm, int size, int num_output, int num_directions, const Option& opt) { - // TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC - // TODO dispatch for __ARM_FEATURE_DOTPROD - #if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD if (ncnn::cpu_support_arm_asimddp()) { @@ -332,9 +329,6 @@ static void rnn_int8_gate_output(const Mat& gates, Mat& hidden_state, Mat& top_b static void rnn_int8(const Mat& bottom_blob_int8, const Mat& bottom_blob_int8_descales, Mat& top_blob, int elemtype, int reverse, const Mat& weight_data_tm, const Mat& weight_data_tm_int8_descales, const Mat& bias_c, Mat& hidden_state, const Option& opt) { - // TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC - // TODO dispatch for __ARM_FEATURE_DOTPROD - #if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD if (ncnn::cpu_support_arm_asimddp()) {