/* * Copyright 2011 The LibYuv Project Authors. All rights reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "source/row.h" #ifdef __cplusplus namespace libyuv { extern "C" { #endif // This module is for GCC Neon #if !defined(YUV_DISABLE_ASM) && defined(__ARM_NEON__) #define YUV422TORGB \ "vld1.u8 {d0}, [%0]! \n" \ "vld1.u32 {d2[0]}, [%1]! \n" \ "vld1.u32 {d2[1]}, [%2]! \n" \ "veor.u8 d2, d26 \n"/*subtract 128 from u and v*/\ "vmull.s8 q8, d2, d24 \n"/* u/v B/R component */\ "vmull.s8 q9, d2, d25 \n"/* u/v G component */\ "vmov.u8 d1, #0 \n"/* split odd/even y apart */\ "vtrn.u8 d0, d1 \n" \ "vsub.s16 q0, q0, q15 \n"/* offset y */\ "vmul.s16 q0, q0, q14 \n" \ "vadd.s16 d18, d19 \n" \ "vqadd.s16 d20, d0, d16 \n" \ "vqadd.s16 d21, d1, d16 \n" \ "vqadd.s16 d22, d0, d17 \n" \ "vqadd.s16 d23, d1, d17 \n" \ "vqadd.s16 d16, d0, d18 \n" \ "vqadd.s16 d17, d1, d18 \n" \ "vqrshrun.s16 d0, q10, #6 \n" \ "vqrshrun.s16 d1, q11, #6 \n" \ "vqrshrun.s16 d2, q8, #6 \n" \ "vmovl.u8 q10, d0 \n"/* set up for reinterleave*/\ "vmovl.u8 q11, d1 \n" \ "vmovl.u8 q8, d2 \n" \ "vtrn.u8 d20, d21 \n" \ "vtrn.u8 d22, d23 \n" \ "vtrn.u8 d16, d17 \n" \ #if defined(HAS_I422TOARGBROW_NEON) || \ defined(HAS_I422TOBGRAROW_NEON) || \ defined(HAS_I422TOABGRROW_NEON) static const vec8 kUVToRB = { 127, 127, 127, 127, 102, 102, 102, 102, 0, 0, 0, 0, 0, 0, 0, 0 }; static const vec8 kUVToG = { -25, -25, -25, -25, -52, -52, -52, -52, 0, 0, 0, 0, 0, 0, 0, 0 }; #endif #ifdef HAS_I422TOARGBROW_NEON void I422ToARGBRow_NEON(const uint8* y_buf, const uint8* u_buf, const uint8* v_buf, uint8* rgb_buf, int width) { asm volatile ( "vld1.u8 {d24}, [%5] \n" "vld1.u8 {d25}, [%6] \n" "vmov.u8 d26, #128 \n" "vmov.u16 q14, #74 \n" "vmov.u16 q15, #16 \n" "1: \n" YUV422TORGB "vmov.u8 d21, d16 \n" "vmov.u8 d23, #255 \n" "vst4.u8 {d20, d21, d22, d23}, [%3]! \n" "subs %4, %4, #8 \n" "bgt 1b \n" : "+r"(y_buf), // %0 "+r"(u_buf), // %1 "+r"(v_buf), // %2 "+r"(rgb_buf), // %3 "+r"(width) // %4 : "r"(&kUVToRB), // %5 "r"(&kUVToG) // %6 : "cc", "memory", "q0", "q1", "q2", "q3", "q8", "q9", "q10", "q11", "q12", "q13", "q14", "q15" ); } #endif // HAS_I422TOARGBROW_NEON #ifdef HAS_I422TOBGRAROW_NEON void I422ToBGRARow_NEON(const uint8* y_buf, const uint8* u_buf, const uint8* v_buf, uint8* rgb_buf, int width) { asm volatile ( "vld1.u8 {d24}, [%5] \n" "vld1.u8 {d25}, [%6] \n" "vmov.u8 d26, #128 \n" "vmov.u16 q14, #74 \n" "vmov.u16 q15, #16 \n" "1: \n" YUV422TORGB "vswp.u8 d20, d22 \n" "vmov.u8 d21, d16 \n" "vmov.u8 d19, #255 \n" "vst4.u8 {d19, d20, d21, d22}, [%3]! \n" "subs %4, %4, #8 \n" "bgt 1b \n" : "+r"(y_buf), // %0 "+r"(u_buf), // %1 "+r"(v_buf), // %2 "+r"(rgb_buf), // %3 "+r"(width) // %4 : "r"(&kUVToRB), // %5 "r"(&kUVToG) // %6 : "cc", "memory", "q0", "q1", "q2", "q3", "q8", "q9", "q10", "q11", "q12", "q13", "q14", "q15" ); } #endif // HAS_I422TOBGRAROW_NEON #ifdef HAS_I422TOABGRROW_NEON void I422ToABGRRow_NEON(const uint8* y_buf, const uint8* u_buf, const uint8* v_buf, uint8* rgb_buf, int width) { asm volatile ( "vld1.u8 {d24}, [%5] \n" "vld1.u8 {d25}, [%6] \n" "vmov.u8 d26, #128 \n" "vmov.u16 q14, #74 \n" "vmov.u16 q15, #16 \n" "1: \n" YUV422TORGB "vswp.u8 d20, d22 \n" "vmov.u8 d21, d16 \n" "vmov.u8 d23, #255 \n" "vst4.u8 {d20, d21, d22, d23}, [%3]! \n" "subs %4, %4, #8 \n" "bgt 1b \n" : "+r"(y_buf), // %0 "+r"(u_buf), // %1 "+r"(v_buf), // %2 "+r"(rgb_buf), // %3 "+r"(width) // %4 : "r"(&kUVToRB), // %5 "r"(&kUVToG) // %6 : "cc", "memory", "q0", "q1", "q2", "q3", "q8", "q9", "q10", "q11", "q12", "q13", "q14", "q15" ); } #endif // HAS_I422TOABGRROW_NEON #ifdef HAS_SPLITUV_NEON // Reads 16 pairs of UV and write even values to dst_u and odd to dst_v // Alignment requirement: 16 bytes for pointers, and multiple of 16 pixels. void SplitUV_NEON(const uint8* src_uv, uint8* dst_u, uint8* dst_v, int width) { asm volatile ( "1: \n" "vld2.u8 {q0,q1}, [%0]! \n" // load 16 pairs of UV "subs %3, %3, #16 \n" // 16 processed per loop "vst1.u8 {q0}, [%1]! \n" // store U "vst1.u8 {q1}, [%2]! \n" // Store V "bgt 1b \n" : "+r"(src_uv), // %0 "+r"(dst_u), // %1 "+r"(dst_v), // %2 "+r"(width) // %3 // Output registers : // Input registers : "memory", "cc", "q0", "q1" // Clobber List ); } #endif // HAS_SPLITUV_NEON #ifdef HAS_COPYROW_NEON // Copy multiple of 64 void CopyRow_NEON(const uint8* src, uint8* dst, int count) { asm volatile ( "1: \n" "pld [%0, #0xC0] \n" // preload "vldm %0!,{q0,q1,q2,q3} \n" // load 64 "subs %2, %2, #64 \n" // 64 processed per loop "vstm %1!,{q0,q1,q2,q3} \n" // store 64 "bgt 1b \n" : "+r"(src), // %0 "+r"(dst), // %1 "+r"(count) // %2 // Output registers : // Input registers : "memory", "cc", "q0", "q1", "q2", "q3" // Clobber List ); } #endif // HAS_COPYROW_NEON #ifdef HAS_MIRRORROW_NEON void MirrorRow_NEON(const uint8* src, uint8* dst, int width) { asm volatile ( // compute where to start writing destination "add %1, %2 \n" // work on segments that are multiples of 16 "lsrs r3, %2, #4 \n" // the output is written in two block. 8 bytes followed // by another 8. reading is done sequentially, from left to // right. writing is done from right to left in block sizes // %1, the destination pointer is incremented after writing // the first of the two blocks. need to subtract that 8 off // along with 16 to get the next location. "mov r3, #-24 \n" "beq 2f \n" // back of destination by the size of the register that is // going to be mirrored "sub %1, #16 \n" // the loop needs to run on blocks of 16. what will be left // over is either a negative number, the residuals that need // to be done, or 0. if this isn't subtracted off here the // loop will run one extra time. "sub %2, #16 \n" // mirror the bytes in the 64 bit segments. unable to mirror // the bytes in the entire 128 bits in one go. // because of the inability to mirror the entire 128 bits // mirror the writing out of the two 64 bit segments. "1: \n" "vld1.8 {q0}, [%0]! \n" // src += 16 "vrev64.8 q0, q0 \n" "vst1.8 {d1}, [%1]! \n" "vst1.8 {d0}, [%1], r3 \n" // dst -= 16 "subs %2, #16 \n" "bge 1b \n" // add 16 back to the counter. if the result is 0 there is no // residuals so jump past "adds %2, #16 \n" "beq 5f \n" "add %1, #16 \n" "2: \n" "mov r3, #-3 \n" "sub %1, #2 \n" "subs %2, #2 \n" // check for 16*n+1 scenarios where segments_of_2 should not // be run, but there is something left over. "blt 4f \n" // do this in neon registers as per // http://blogs.arm.com/software-enablement/196-coding-for-neon-part-2-dealing-with-leftovers/ "3: \n" "vld2.8 {d0[0], d1[0]}, [%0]! \n" // src += 2 "vst1.8 {d1[0]}, [%1]! \n" "vst1.8 {d0[0]}, [%1], r3 \n" // dst -= 2 "subs %2, #2 \n" "bge 3b \n" "adds %2, #2 \n" "beq 5f \n" "4: \n" "add %1, #1 \n" "vld1.8 {d0[0]}, [%0] \n" "vst1.8 {d0[0]}, [%1] \n" "5: \n" : "+r"(src), // %0 "+r"(dst), // %1 "+r"(width) // %2 : : "memory", "cc", "r3", "q0" ); } #endif // HAS_MIRRORROW_NEON #ifdef HAS_MIRRORROWUV_NEON void MirrorRowUV_NEON(const uint8* src, uint8* dst_a, uint8* dst_b, int width) { asm volatile ( // compute where to start writing destination "add %1, %3 \n" // dst_a + width "add %2, %3 \n" // dst_b + width // work on input segments that are multiples of 16, but // width that has been passed is output segments, half // the size of input. "lsrs r12, %3, #3 \n" "beq 2f \n" // the output is written in to two blocks. "mov r12, #-8 \n" // back of destination by the size of the register that is // going to be mirrord "sub %1, #8 \n" "sub %2, #8 \n" // the loop needs to run on blocks of 8. what will be left // over is either a negative number, the residuals that need // to be done, or 0. if this isn't subtracted off here the // loop will run one extra time. "sub %3, #8 \n" // mirror the bytes in the 64 bit segments "1: \n" "vld2.8 {d0, d1}, [%0]! \n" // src += 16 "vrev64.8 q0, q0 \n" "vst1.8 {d0}, [%1], r12 \n" // dst_a -= 8 "vst1.8 {d1}, [%2], r12 \n" // dst_b -= 8 "subs %3, #8 \n" "bge 1b \n" // add 8 back to the counter. if the result is 0 there is no // residuals so return "adds %3, #8 \n" "beq 4f \n" "add %1, #8 \n" "add %2, #8 \n" "2: \n" "mov r12, #-1 \n" "sub %1, #1 \n" "sub %2, #1 \n" "3: \n" "vld2.8 {d0[0], d1[0]}, [%0]! \n" // src += 2 "vst1.8 {d0[0]}, [%1], r12 \n" // dst_a -= 1 "vst1.8 {d1[0]}, [%2], r12 \n" // dst_b -= 1 "subs %3, %3, #1 \n" "bgt 3b \n" "4: \n" : "+r"(src), // %0 "+r"(dst_a), // %1 "+r"(dst_b), // %2 "+r"(width) // %3 : : "memory", "cc", "r12", "q0" ); } #endif // HAS_MIRRORROWUV_NEON #endif // __ARM_NEON__ #ifdef __cplusplus } // extern "C" } // namespace libyuv #endif