Skip to content

Commit

Permalink
clean
Browse files Browse the repository at this point in the history
  • Loading branch information
nihui committed May 8, 2024
1 parent 779fba0 commit 617c594
Show file tree
Hide file tree
Showing 3 changed files with 0 additions and 18 deletions.
6 changes: 0 additions & 6 deletions src/layer/arm/gru_int8.h
Original file line number Diff line number Diff line change
Expand Up @@ -23,9 +23,6 @@ void gru_int8_gate_output_vfpv4(const Mat& gates, Mat& hidden_state, Mat& top_bl

static void gru_transform_weight_int8(const Mat& weight_xc, const Mat& weight_xc_int8_scales, const Mat& weight_hc, const Mat& weight_hc_int8_scales, const Mat& bias_c, Mat& weight_data_tm, Mat& weight_data_tm_int8_descales, Mat& bias_c_tm, int size, int num_output, int num_directions, const Option& opt)
{
// TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
// TODO dispatch for __ARM_FEATURE_DOTPROD

#if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD
if (ncnn::cpu_support_arm_asimddp())
{
Expand Down Expand Up @@ -643,9 +640,6 @@ static void gru_int8_gate_output(const Mat& gates, Mat& hidden_state, Mat& top_b

static void gru_int8(const Mat& bottom_blob_int8, const Mat& bottom_blob_int8_descales, Mat& top_blob, int elemtype, int reverse, const Mat& weight_data_tm, const Mat& weight_data_tm_int8_descales, const Mat& bias_c, Mat& hidden_state, const Option& opt)
{
// TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
// TODO dispatch for __ARM_FEATURE_DOTPROD

#if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD
if (ncnn::cpu_support_arm_asimddp())
{
Expand Down
6 changes: 0 additions & 6 deletions src/layer/arm/lstm_int8.h
Original file line number Diff line number Diff line change
Expand Up @@ -23,9 +23,6 @@ void lstm_int8_gate_output_vfpv4(const Mat& gates, const Mat& weight_hr, Mat& hi

static void lstm_transform_weight_int8(const Mat& weight_xc, const Mat& weight_xc_int8_scales, const Mat& weight_hc, const Mat& weight_hc_int8_scales, const Mat& bias_c, Mat& weight_data_tm, Mat& weight_data_tm_int8_descales, Mat& bias_c_tm, int size, int num_output, int num_directions, int hidden_size, const Option& opt)
{
// TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
// TODO dispatch for __ARM_FEATURE_DOTPROD

#if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD
if (ncnn::cpu_support_arm_asimddp())
{
Expand Down Expand Up @@ -392,9 +389,6 @@ static void lstm_int8_gate_output(const Mat& gates, const Mat& weight_hr, Mat& h

static void lstm_int8(const Mat& bottom_blob_int8, const Mat& bottom_blob_int8_descales, Mat& top_blob, int elemtype, int reverse, const Mat& weight_data_tm, const Mat& weight_data_tm_int8_descales, const Mat& bias_c, const Mat& weight_hr, Mat& hidden_state, Mat& cell_state, const Option& opt)
{
// TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
// TODO dispatch for __ARM_FEATURE_DOTPROD

#if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD
if (ncnn::cpu_support_arm_asimddp())
{
Expand Down
6 changes: 0 additions & 6 deletions src/layer/arm/rnn_int8.h
Original file line number Diff line number Diff line change
Expand Up @@ -23,9 +23,6 @@ void rnn_int8_gate_output_vfpv4(const Mat& gates, Mat& hidden_state, Mat& top_bl

static void rnn_transform_weight_int8(const Mat& weight_xc, const Mat& weight_xc_int8_scales, const Mat& weight_hc, const Mat& weight_hc_int8_scales, const Mat& bias_c, Mat& weight_data_tm, Mat& weight_data_tm_int8_descales, Mat& bias_c_tm, int size, int num_output, int num_directions, const Option& opt)
{
// TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
// TODO dispatch for __ARM_FEATURE_DOTPROD

#if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD
if (ncnn::cpu_support_arm_asimddp())
{
Expand Down Expand Up @@ -332,9 +329,6 @@ static void rnn_int8_gate_output(const Mat& gates, Mat& hidden_state, Mat& top_b

static void rnn_int8(const Mat& bottom_blob_int8, const Mat& bottom_blob_int8_descales, Mat& top_blob, int elemtype, int reverse, const Mat& weight_data_tm, const Mat& weight_data_tm_int8_descales, const Mat& bias_c, Mat& hidden_state, const Option& opt)
{
// TODO dispatch for __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
// TODO dispatch for __ARM_FEATURE_DOTPROD

#if NCNN_RUNTIME_CPU && NCNN_ARM82DOT && __aarch64__ && !__ARM_FEATURE_DOTPROD
if (ncnn::cpu_support_arm_asimddp())
{
Expand Down

0 comments on commit 617c594

Please sign in to comment.