diff options
Diffstat (limited to 'sbc/sbc_primitives_neon.c')
| -rw-r--r-- | sbc/sbc_primitives_neon.c | 299 | 
1 files changed, 299 insertions, 0 deletions
diff --git a/sbc/sbc_primitives_neon.c b/sbc/sbc_primitives_neon.c new file mode 100644 index 00000000..ea8446f1 --- /dev/null +++ b/sbc/sbc_primitives_neon.c @@ -0,0 +1,299 @@ +/* + * + *  Bluetooth low-complexity, subband codec (SBC) library + * + *  Copyright (C) 2004-2009  Marcel Holtmann <marcel@holtmann.org> + *  Copyright (C) 2004-2005  Henryk Ploetz <henryk@ploetzli.ch> + *  Copyright (C) 2005-2006  Brad Midgley <bmidgley@xmission.com> + * + * + *  This library is free software; you can redistribute it and/or + *  modify it under the terms of the GNU Lesser General Public + *  License as published by the Free Software Foundation; either + *  version 2.1 of the License, or (at your option) any later version. + * + *  This library is distributed in the hope that it will be useful, + *  but WITHOUT ANY WARRANTY; without even the implied warranty of + *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU + *  Lesser General Public License for more details. + * + *  You should have received a copy of the GNU Lesser General Public + *  License along with this library; if not, write to the Free Software + *  Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA + * + */ + +#include <stdint.h> +#include <limits.h> +#include "sbc.h" +#include "sbc_math.h" +#include "sbc_tables.h" + +#include "sbc_primitives_neon.h" + +/* + * ARM NEON optimizations + */ + +#ifdef SBC_BUILD_WITH_NEON_SUPPORT + +static inline void _sbc_analyze_four_neon(const int16_t *in, int32_t *out, +					 const FIXED_T *consts) +{ +	/* TODO: merge even and odd cases (or even merge all four calls to this +		 function) in order to have only aligned reads from 'in' array +		 and reduce number of load instructions */ +	asm volatile ( +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmull.s16  q0, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmull.s16  q1, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" + +		"vmlal.s16  q0, d6, d10\n" +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vmlal.s16  q1, d7, d11\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmlal.s16  q0, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmlal.s16  q1, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" + +		"vmlal.s16  q0, d6, d10\n" +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vmlal.s16  q1, d7, d11\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmlal.s16  q0, d4, d8\n" +		"vmlal.s16  q1, d5, d9\n" + +		"vpadd.s32  d0, d0, d1\n" +		"vpadd.s32  d1, d2, d3\n" + +		"vrshrn.s32 d0, q0, %3\n" + +		"vld1.16    {d2, d3, d4, d5}, [%1, :128]!\n" + +		"vdup.i32   d1, d0[1]\n"  /* TODO: can be eliminated */ +		"vdup.i32   d0, d0[0]\n"  /* TODO: can be eliminated */ + +		"vmull.s16  q3, d2, d0\n" +		"vmull.s16  q4, d3, d0\n" +		"vmlal.s16  q3, d4, d1\n" +		"vmlal.s16  q4, d5, d1\n" + +		"vpadd.s32  d0, d6, d7\n" /* TODO: can be eliminated */ +		"vpadd.s32  d1, d8, d9\n" /* TODO: can be eliminated */ + +		"vst1.32    {d0, d1}, [%2, :128]\n" +		: "+r" (in), "+r" (consts) +		: "r" (out), +		  "i" (SBC_PROTO_FIXED4_SCALE) +		: "memory", +		  "d0", "d1", "d2", "d3", "d4", "d5", +		  "d6", "d7", "d8", "d9", "d10", "d11"); +} + +static inline void _sbc_analyze_eight_neon(const int16_t *in, int32_t *out, +					 const FIXED_T *consts) +{ +	/* TODO: merge even and odd cases (or even merge all four calls to this +		 function) in order to have only aligned reads from 'in' array +		 and reduce number of load instructions */ +	asm volatile ( +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmull.s16  q6, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmull.s16  q7, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" +		"vmull.s16  q8, d6, d10\n" +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vmull.s16  q9, d7, d11\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmlal.s16  q6, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmlal.s16  q7, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" +		"vmlal.s16  q8, d6, d10\n" +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vmlal.s16  q9, d7, d11\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmlal.s16  q6, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmlal.s16  q7, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" +		"vmlal.s16  q8, d6, d10\n" +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vmlal.s16  q9, d7, d11\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmlal.s16  q6, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmlal.s16  q7, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" +		"vmlal.s16  q8, d6, d10\n" +		"vld1.16    {d4, d5}, [%0, :64]!\n" +		"vmlal.s16  q9, d7, d11\n" +		"vld1.16    {d8, d9}, [%1, :128]!\n" + +		"vmlal.s16  q6, d4, d8\n" +		"vld1.16    {d6,  d7}, [%0, :64]!\n" +		"vmlal.s16  q7, d5, d9\n" +		"vld1.16    {d10, d11}, [%1, :128]!\n" + +		"vmlal.s16  q8, d6, d10\n" +		"vmlal.s16  q9, d7, d11\n" + +		"vpadd.s32  d0, d12, d13\n" +		"vpadd.s32  d1, d14, d15\n" +		"vpadd.s32  d2, d16, d17\n" +		"vpadd.s32  d3, d18, d19\n" + +		"vrshr.s32 q0, q0, %3\n" +		"vrshr.s32 q1, q1, %3\n" +		"vmovn.s32 d0, q0\n" +		"vmovn.s32 d1, q1\n" + +		"vdup.i32   d3, d1[1]\n"  /* TODO: can be eliminated */ +		"vdup.i32   d2, d1[0]\n"  /* TODO: can be eliminated */ +		"vdup.i32   d1, d0[1]\n"  /* TODO: can be eliminated */ +		"vdup.i32   d0, d0[0]\n"  /* TODO: can be eliminated */ + +		"vld1.16    {d4, d5}, [%1, :128]!\n" +		"vmull.s16  q6, d4, d0\n" +		"vld1.16    {d6, d7}, [%1, :128]!\n" +		"vmull.s16  q7, d5, d0\n" +		"vmull.s16  q8, d6, d0\n" +		"vmull.s16  q9, d7, d0\n" + +		"vld1.16    {d4, d5}, [%1, :128]!\n" +		"vmlal.s16  q6, d4, d1\n" +		"vld1.16    {d6, d7}, [%1, :128]!\n" +		"vmlal.s16  q7, d5, d1\n" +		"vmlal.s16  q8, d6, d1\n" +		"vmlal.s16  q9, d7, d1\n" + +		"vld1.16    {d4, d5}, [%1, :128]!\n" +		"vmlal.s16  q6, d4, d2\n" +		"vld1.16    {d6, d7}, [%1, :128]!\n" +		"vmlal.s16  q7, d5, d2\n" +		"vmlal.s16  q8, d6, d2\n" +		"vmlal.s16  q9, d7, d2\n" + +		"vld1.16    {d4, d5}, [%1, :128]!\n" +		"vmlal.s16  q6, d4, d3\n" +		"vld1.16    {d6, d7}, [%1, :128]!\n" +		"vmlal.s16  q7, d5, d3\n" +		"vmlal.s16  q8, d6, d3\n" +		"vmlal.s16  q9, d7, d3\n" + +		"vpadd.s32  d0, d12, d13\n" /* TODO: can be eliminated */ +		"vpadd.s32  d1, d14, d15\n" /* TODO: can be eliminated */ +		"vpadd.s32  d2, d16, d17\n" /* TODO: can be eliminated */ +		"vpadd.s32  d3, d18, d19\n" /* TODO: can be eliminated */ + +		"vst1.32    {d0, d1, d2, d3}, [%2, :128]\n" +		: "+r" (in), "+r" (consts) +		: "r" (out), +		  "i" (SBC_PROTO_FIXED8_SCALE) +		: "memory", +		  "d0", "d1", "d2", "d3", "d4", "d5", +		  "d6", "d7", "d8", "d9", "d10", "d11", +		  "d12", "d13", "d14", "d15", "d16", "d17", +		  "d18", "d19"); +} + +static inline void sbc_analyze_4b_4s_neon(int16_t *pcm, int16_t *x, +					 int32_t *out, int out_stride) +{ +	/* Fetch audio samples and do input data reordering for SIMD */ +	x[64] = x[0]  = pcm[8 + 7]; +	x[65] = x[1]  = pcm[8 + 3]; +	x[66] = x[2]  = pcm[8 + 6]; +	x[67] = x[3]  = pcm[8 + 4]; +	x[68] = x[4]  = pcm[8 + 0]; +	x[69] = x[5]  = pcm[8 + 2]; +	x[70] = x[6]  = pcm[8 + 1]; +	x[71] = x[7]  = pcm[8 + 5]; + +	x[72] = x[8]  = pcm[0 + 7]; +	x[73] = x[9]  = pcm[0 + 3]; +	x[74] = x[10] = pcm[0 + 6]; +	x[75] = x[11] = pcm[0 + 4]; +	x[76] = x[12] = pcm[0 + 0]; +	x[77] = x[13] = pcm[0 + 2]; +	x[78] = x[14] = pcm[0 + 1]; +	x[79] = x[15] = pcm[0 + 5]; + +	/* Analyze blocks */ +	_sbc_analyze_four_neon(x + 12, out, analysis_consts_fixed4_simd_odd); +	out += out_stride; +	_sbc_analyze_four_neon(x + 8, out, analysis_consts_fixed4_simd_even); +	out += out_stride; +	_sbc_analyze_four_neon(x + 4, out, analysis_consts_fixed4_simd_odd); +	out += out_stride; +	_sbc_analyze_four_neon(x + 0, out, analysis_consts_fixed4_simd_even); +} + +static inline void sbc_analyze_4b_8s_neon(int16_t *pcm, int16_t *x, +					  int32_t *out, int out_stride) +{ +	/* Fetch audio samples and do input data reordering for SIMD */ +	x[128] = x[0]  = pcm[16 + 15]; +	x[129] = x[1]  = pcm[16 + 7]; +	x[130] = x[2]  = pcm[16 + 14]; +	x[131] = x[3]  = pcm[16 + 8]; +	x[132] = x[4]  = pcm[16 + 13]; +	x[133] = x[5]  = pcm[16 + 9]; +	x[134] = x[6]  = pcm[16 + 12]; +	x[135] = x[7]  = pcm[16 + 10]; +	x[136] = x[8]  = pcm[16 + 11]; +	x[137] = x[9]  = pcm[16 + 3]; +	x[138] = x[10] = pcm[16 + 6]; +	x[139] = x[11] = pcm[16 + 0]; +	x[140] = x[12] = pcm[16 + 5]; +	x[141] = x[13] = pcm[16 + 1]; +	x[142] = x[14] = pcm[16 + 4]; +	x[143] = x[15] = pcm[16 + 2]; + +	x[144] = x[16] = pcm[0 + 15]; +	x[145] = x[17] = pcm[0 + 7]; +	x[146] = x[18] = pcm[0 + 14]; +	x[147] = x[19] = pcm[0 + 8]; +	x[148] = x[20] = pcm[0 + 13]; +	x[149] = x[21] = pcm[0 + 9]; +	x[150] = x[22] = pcm[0 + 12]; +	x[151] = x[23] = pcm[0 + 10]; +	x[152] = x[24] = pcm[0 + 11]; +	x[153] = x[25] = pcm[0 + 3]; +	x[154] = x[26] = pcm[0 + 6]; +	x[155] = x[27] = pcm[0 + 0]; +	x[156] = x[28] = pcm[0 + 5]; +	x[157] = x[29] = pcm[0 + 1]; +	x[158] = x[30] = pcm[0 + 4]; +	x[159] = x[31] = pcm[0 + 2]; + +	/* Analyze blocks */ +	_sbc_analyze_eight_neon(x + 24, out, analysis_consts_fixed8_simd_odd); +	out += out_stride; +	_sbc_analyze_eight_neon(x + 16, out, analysis_consts_fixed8_simd_even); +	out += out_stride; +	_sbc_analyze_eight_neon(x + 8, out, analysis_consts_fixed8_simd_odd); +	out += out_stride; +	_sbc_analyze_eight_neon(x + 0, out, analysis_consts_fixed8_simd_even); +} + +void sbc_init_primitives_neon(struct sbc_encoder_state *state) +{ +	state->sbc_analyze_4b_4s = sbc_analyze_4b_4s_neon; +	state->sbc_analyze_4b_8s = sbc_analyze_4b_8s_neon; +} + +#endif  | 
