blob: f4ab2f798916e4ffcecd31b464af7a7d9c375e21 [file] [log] [blame]
/* ------------------------------------------------------------------
* Copyright (C) 1998-2009 PacketVideo
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
* express or implied.
* See the License for the specific language governing permissions
* and limitations under the License.
* -------------------------------------------------------------------
*/
/*
Pathname: fxp_mul32_arm_v4_gcc.h
------------------------------------------------------------------------------
REVISION HISTORY
Who: Date:
Description:
------------------------------------------------------------------------------
INCLUDE DESCRIPTION
------------------------------------------------------------------------------
*/
#ifndef FXP_MUL32_V4_ARM_GCC
#define FXP_MUL32_V4_ARM_GCC
#ifdef __cplusplus
extern "C"
{
#endif
#include "pv_audio_type_defs.h"
#if defined (_ARM_V4_GCC) /* ARM_V4 GNU COMPILER */
#define preload_cache( a)
static inline Int32 shft_lft_1(Int32 L_var1)
{
Int32 x;
register Int32 ra = L_var1;
Int32 z = INT32_MAX;
asm volatile(
"mov %0, %1, asl #1\n\t"
"teq %1, %0, asr #1\n\t"
"eorne %0, %2, %1, asr #31"
: "=&r*i"(x)
: "r"(ra),
"r"(z));
return(x);
}
static inline Int32 fxp_mul_16_by_16bb(Int32 L_var1, Int32 L_var2)
{
Int32 tmp1;
Int32 tmp2;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
asm volatile(
"mov %0, %3, asl #16\n\t"
"mov %0, %0, asr #16\n\t"
"mov %1, %2, asl #16\n\t"
"mov %1, %1, asr #16\n\t"
"mul %0, %1, %0"
: "=&r*i"(tmp1),
"=&r*i"(tmp2)
: "r"(ra),
"r"(rb));
return (tmp1);
}
#define fxp_mul_16_by_16(a, b) fxp_mul_16_by_16bb( a, b)
static inline Int32 fxp_mul_16_by_16tb(Int32 L_var1, Int32 L_var2)
{
Int32 tmp1;
Int32 tmp2;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
asm volatile(
"mov %0, %3, asl #16\n\t"
"mov %0, %0, asr #16\n\t"
"mov %1, %2, asr #16\n\t"
"mul %0, %1, %0"
: "=&r*i"(tmp1),
"=&r*i"(tmp2)
: "r"(ra),
"r"(rb));
return (tmp1);
}
static inline Int32 fxp_mul_16_by_16bt(Int32 L_var1, Int32 L_var2)
{
Int32 tmp1;
Int32 tmp2;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
asm volatile(
"mov %0, %3, asr #16\n\t"
"mov %1, %2, asl #16\n\t"
"mov %1, %1, asr #16\n\t"
"mul %0, %1, %0"
: "=&r*i"(tmp1),
"=&r*i"(tmp2)
: "r"(ra),
"r"(rb));
return (tmp1);
}
static inline Int32 fxp_mul_16_by_16tt(Int32 L_var1, Int32 L_var2)
{
Int32 tmp1;
Int32 tmp2;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
asm volatile(
"mov %0, %3, asr #16\n\t"
"mov %1, %2, asr #16\n\t"
"mul %0, %1, %0"
: "=&r*i"(tmp1),
"=&r*i"(tmp2)
: "r"(ra),
"r"(rb));
return (tmp1);
}
static inline Int32 fxp_mac_16_by_16(Int16 L_var1, Int16 L_var2, Int32 L_add)
{
Int32 tmp;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
register Int32 rc = (Int32)L_add;
asm volatile(
"mla %0, %1, %2, %3"
: "=&r*i"(tmp)
: "r"(ra),
"r"(rb),
"r"(rc));
return (tmp);
}
static inline Int32 fxp_mac_16_by_16_bb(Int16 L_var1, Int32 L_var2, Int32 L_add)
{
Int32 tmp1;
Int32 tmp2;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
register Int32 rc = (Int32)L_add;
asm volatile(
"mov %0, %3, asl #16\n\t"
"mov %0, %0, asr #16\n\t"
"mla %1, %0, %2, %4"
: "=&r*i"(tmp1),
"=&r*i"(tmp2)
: "r"(ra),
"r"(rb),
"r"(rc));
return (tmp2);
}
static inline Int32 fxp_mac_16_by_16_bt(Int16 L_var1, Int32 L_var2, Int32 L_add)
{
Int32 tmp1;
Int32 tmp2;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
register Int32 rc = (Int32)L_add;
asm volatile(
"mov %0, %3, asr #16\n\t"
"mla %1, %0, %2, %4"
: "=&r*i"(tmp1),
"=&r*i"(tmp2)
: "r"(ra),
"r"(rb),
"r"(rc));
return (tmp2);
}
static inline Int32 cmplx_mul32_by_16(Int32 x, Int32 y, Int32 exp_jw)
{
Int32 rTmp0;
Int32 iTmp0;
Int32 result64_hi;
register Int32 ra = (Int32)x;
register Int32 rb = (Int32)y;
register Int32 rc = (Int32)exp_jw;
asm volatile(
"mov %0, %5, asr #16\n\t"
"mov %1, %5, asl #16\n\t"
"mov %0, %0, asl #16\n\t"
: "=&r*i"(rTmp0),
"=&r*i"(iTmp0),
"=&r*i"(result64_hi)
: "r"(ra),
"r"(rb),
"r"(rc));
asm volatile(
"smull %0, %2, %3, %0\n\t"
"smlal %1, %2, %4, %1"
: "=&r*i"(rTmp0),
"=&r*i"(iTmp0),
"=&r*i"(result64_hi)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline Int32 fxp_mul32_by_16(Int32 L_var1, Int32 L_var2)
{
Int32 rTmp0;
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
asm volatile(
"mov %0, %4, asl #16\n\t"
"smull %2, %1, %0, %3"
: "=&r*i"(rTmp0),
"=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
#define fxp_mul32_by_16b( a, b) fxp_mul32_by_16( a, b)
static inline Int32 fxp_mul32_by_16t(Int32 L_var1, Int32 L_var2)
{
Int32 rTmp0;
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
asm volatile(
"mov %0, %4, asr #16\n\t"
"mov %0, %0, asl #16\n\t"
"smull %2, %1, %0, %3"
: "=&r*i"(rTmp0),
"=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mac32_by_16(Int32 L_var1, Int32 L_var2, Int32 L_add)
{
Int32 rTmp0;
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)L_var1;
register Int32 rb = (Int32)L_var2;
register Int32 rc = (Int32)L_add;
asm volatile(
"mov %0, %4, asl #16\n\t"
"mov %1, %5\n\t"
"smlal %2, %1, %0, %3"
: "=&r*i"(rTmp0),
"=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline int64 fxp_mac64_Q31(int64 sum, const Int32 L_var1, const Int32 L_var2)
{
sum += (int64)L_var1 * L_var2;
return (sum);
}
static inline Int32 fxp_mac32_Q30(const Int32 a, const Int32 b, Int32 L_add)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
register Int32 rc = (Int32)L_add;
asm volatile("smull %1, %0, %2, %3\n\t"
"add %4, %4, %0, asl #2\n\t"
"add %0, %4, %1, lsr #30"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline Int32 fxp_mac32_Q31(Int32 L_add, const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
register Int32 rc = (Int32)L_add;
asm volatile("smull %1, %0, %2, %3\n\t"
"add %0, %0, %4"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline Int32 fxp_msu32_Q31(Int32 L_sub, const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
register Int32 rc = (Int32)L_sub;
asm volatile("smull %1, %0, %2, %3\n\t"
"sub %0, %4, %0"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q31(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile(
"smull %1, %0, %2, %3"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q30(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #2\n\t"
"orr %0, %0, %1, lsr #30"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mac32_Q29(const Int32 a, const Int32 b, Int32 L_add)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
register Int32 rc = (Int32)L_add;
asm volatile("smull %1, %0, %2, %3\n\t"
"add %4, %4, %0, lsl #3\n\t"
"add %0, %4, %1, lsr #29"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline Int32 fxp_msu32_Q29(const Int32 a, const Int32 b, Int32 L_sub)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
register Int32 rc = (Int32)L_sub;
asm volatile("smull %1, %0, %2, %3\n\t"
"sub %4, %4, %0, lsl #3\n\t"
"sub %0, %4, %1, lsr #29"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb),
"r"(rc));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q29(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #3\n\t"
"orr %0, %0, %1, lsr #29"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q28(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #4\n\t"
"orr %0, %0, %1, lsr #28"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q27(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #5\n\t"
"orr %0, %0, %1, lsr #27"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q26(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #6\n\t"
"orr %0, %0, %1, lsr #26"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q20(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #12\n\t"
"orr %0, %0, %1, lsr #20"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q15(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #17\n\t"
"orr %0, %0, %1, lsr #15"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
static inline Int32 fxp_mul32_Q14(const Int32 a, const Int32 b)
{
Int32 result64_hi;
Int32 result64_lo;
register Int32 ra = (Int32)a;
register Int32 rb = (Int32)b;
asm volatile("smull %1, %0, %2, %3\n\t"
"mov %0, %0, lsl #18\n\t"
"orr %0, %0, %1, lsr #14"
: "=&r*i"(result64_hi),
"=&r*i"(result64_lo)
: "r"(ra),
"r"(rb));
return (result64_hi);
}
#endif
#ifdef __cplusplus
}
#endif
#endif /* FXP_MUL32_V4_ARM_GCC */