Chris@69: /* Copyright (C) 2013 Xiph.Org Foundation and contributors */ Chris@69: /* Chris@69: Redistribution and use in source and binary forms, with or without Chris@69: modification, are permitted provided that the following conditions Chris@69: are met: Chris@69: Chris@69: - Redistributions of source code must retain the above copyright Chris@69: notice, this list of conditions and the following disclaimer. Chris@69: Chris@69: - Redistributions in binary form must reproduce the above copyright Chris@69: notice, this list of conditions and the following disclaimer in the Chris@69: documentation and/or other materials provided with the distribution. Chris@69: Chris@69: THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS Chris@69: ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT Chris@69: LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR Chris@69: A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER Chris@69: OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, Chris@69: EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, Chris@69: PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR Chris@69: PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF Chris@69: LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING Chris@69: NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS Chris@69: SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. Chris@69: */ Chris@69: Chris@69: #ifndef FIXED_ARMv4_H Chris@69: #define FIXED_ARMv4_H Chris@69: Chris@69: /** 16x32 multiplication, followed by a 16-bit shift right. Results fits in 32 bits */ Chris@69: #undef MULT16_32_Q16 Chris@69: static OPUS_INLINE opus_val32 MULT16_32_Q16_armv4(opus_val16 a, opus_val32 b) Chris@69: { Chris@69: unsigned rd_lo; Chris@69: int rd_hi; Chris@69: __asm__( Chris@69: "#MULT16_32_Q16\n\t" Chris@69: "smull %0, %1, %2, %3\n\t" Chris@69: : "=&r"(rd_lo), "=&r"(rd_hi) Chris@69: : "%r"(b),"r"(SHL32(a,16)) Chris@69: ); Chris@69: return rd_hi; Chris@69: } Chris@69: #define MULT16_32_Q16(a, b) (MULT16_32_Q16_armv4(a, b)) Chris@69: Chris@69: Chris@69: /** 16x32 multiplication, followed by a 15-bit shift right. Results fits in 32 bits */ Chris@69: #undef MULT16_32_Q15 Chris@69: static OPUS_INLINE opus_val32 MULT16_32_Q15_armv4(opus_val16 a, opus_val32 b) Chris@69: { Chris@69: unsigned rd_lo; Chris@69: int rd_hi; Chris@69: __asm__( Chris@69: "#MULT16_32_Q15\n\t" Chris@69: "smull %0, %1, %2, %3\n\t" Chris@69: : "=&r"(rd_lo), "=&r"(rd_hi) Chris@69: : "%r"(b), "r"(SHL32(a,16)) Chris@69: ); Chris@69: /*We intentionally don't OR in the high bit of rd_lo for speed.*/ Chris@69: return SHL32(rd_hi,1); Chris@69: } Chris@69: #define MULT16_32_Q15(a, b) (MULT16_32_Q15_armv4(a, b)) Chris@69: Chris@69: Chris@69: /** 16x32 multiply, followed by a 15-bit shift right and 32-bit add. Chris@69: b must fit in 31 bits. Chris@69: Result fits in 32 bits. */ Chris@69: #undef MAC16_32_Q15 Chris@69: #define MAC16_32_Q15(c, a, b) ADD32(c, MULT16_32_Q15(a, b)) Chris@69: Chris@69: /** 16x32 multiply, followed by a 16-bit shift right and 32-bit add. Chris@69: Result fits in 32 bits. */ Chris@69: #undef MAC16_32_Q16 Chris@69: #define MAC16_32_Q16(c, a, b) ADD32(c, MULT16_32_Q16(a, b)) Chris@69: Chris@69: /** 32x32 multiplication, followed by a 31-bit shift right. Results fits in 32 bits */ Chris@69: #undef MULT32_32_Q31 Chris@69: #define MULT32_32_Q31(a,b) (opus_val32)((((opus_int64)(a)) * ((opus_int64)(b)))>>31) Chris@69: Chris@69: #endif