/* * Copyright (c) 2020 Martin Storsjo * Copyright (c) 2024 Ramiro Polla * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "libavutil/aarch64/asm.S" #define RGB2YUV_COEFFS 16*4+16*32 #define BY v0.h[0] #define GY v0.h[1] #define RY v0.h[2] #define BU v1.h[0] #define GU v1.h[1] #define RU v1.h[2] #define BV v2.h[0] #define GV v2.h[1] #define RV v2.h[2] #define Y_OFFSET v22 #define UV_OFFSET v23 // convert rgb to 16-bit y, u, or v // uses v3 and v4 .macro rgbconv16 dst, b, g, r, bc, gc, rc smull v3.4s, \b\().4h, \bc smlal v3.4s, \g\().4h, \gc smlal v3.4s, \r\().4h, \rc smull2 v4.4s, \b\().8h, \bc smlal2 v4.4s, \g\().8h, \gc smlal2 v4.4s, \r\().8h, \rc // v3:v4 = b * bc + g * gc + r * rc (32-bit) shrn \dst\().4h, v3.4s, #7 shrn2 \dst\().8h, v4.4s, #7 // dst = b * bc + g * gc + r * rc (16-bit) .endm // void ff_rgb24toyv12_neon(const uint8_t *src, uint8_t *ydst, uint8_t *udst, // uint8_t *vdst, int width, int height, int lumStride, // int chromStride, int srcStride, int32_t *rgb2yuv); function ff_rgb24toyv12_neon, export=1 // x0 const uint8_t *src // x1 uint8_t *ydst // x2 uint8_t *udst // x3 uint8_t *vdst // w4 int width // w5 int height // w6 int lumStride // w7 int chromStride ldrsw x14, [sp] ldr x15, [sp, #8] // x14 int srcStride // x15 int32_t *rgb2yuv // extend width and stride parameters uxtw x4, w4 sxtw x6, w6 sxtw x7, w7 // src1 = x0 // src2 = x10 add x10, x0, x14 // x10 = src + srcStride lsl x14, x14, #1 // srcStride *= 2 add x11, x4, x4, lsl #1 // x11 = 3 * width sub x14, x14, x11 // srcPadding = (2 * srcStride) - (3 * width) // ydst1 = x1 // ydst2 = x11 add x11, x1, x6 // x11 = ydst + lumStride lsl x6, x6, #1 // lumStride *= 2 sub x6, x6, x4 // lumPadding = (2 * lumStride) - width sub x7, x7, x4, lsr #1 // chromPadding = chromStride - (width / 2) // load rgb2yuv coefficients into v0, v1, and v2 add x15, x15, #RGB2YUV_COEFFS ld1 {v0.8h-v2.8h}, [x15] // load 24 values // load offset constants movi Y_OFFSET.8h, #0x10, lsl #8 movi UV_OFFSET.8h, #0x80, lsl #8 1: mov w15, w4 // w15 = width 2: // load first line ld3 {v26.16b, v27.16b, v28.16b}, [x0], #48 // widen first line to 16-bit uxtl v16.8h, v26.8b // v16 = B11 uxtl v17.8h, v27.8b // v17 = G11 uxtl v18.8h, v28.8b // v18 = R11 uxtl2 v19.8h, v26.16b // v19 = B12 uxtl2 v20.8h, v27.16b // v20 = G12 uxtl2 v21.8h, v28.16b // v21 = R12 // calculate Y values for first line rgbconv16 v24, v16, v17, v18, BY, GY, RY // v24 = Y11 rgbconv16 v25, v19, v20, v21, BY, GY, RY // v25 = Y12 // load second line ld3 {v26.16b, v27.16b, v28.16b}, [x10], #48 // pairwise add and save rgb values to calculate average addp v5.8h, v16.8h, v19.8h addp v6.8h, v17.8h, v20.8h addp v7.8h, v18.8h, v21.8h // widen second line to 16-bit uxtl v16.8h, v26.8b // v16 = B21 uxtl v17.8h, v27.8b // v17 = G21 uxtl v18.8h, v28.8b // v18 = R21 uxtl2 v19.8h, v26.16b // v19 = B22 uxtl2 v20.8h, v27.16b // v20 = G22 uxtl2 v21.8h, v28.16b // v21 = R22 // calculate Y values for second line rgbconv16 v26, v16, v17, v18, BY, GY, RY // v26 = Y21 rgbconv16 v27, v19, v20, v21, BY, GY, RY // v27 = Y22 // pairwise add rgb values to calculate average addp v16.8h, v16.8h, v19.8h addp v17.8h, v17.8h, v20.8h addp v18.8h, v18.8h, v21.8h // calculate average add v16.8h, v16.8h, v5.8h add v17.8h, v17.8h, v6.8h add v18.8h, v18.8h, v7.8h ushr v16.8h, v16.8h, #2 ushr v17.8h, v17.8h, #2 ushr v18.8h, v18.8h, #2 // calculate U and V values rgbconv16 v28, v16, v17, v18, BU, GU, RU // v28 = U rgbconv16 v29, v16, v17, v18, BV, GV, RV // v29 = V // add offsets and narrow all values addhn v24.8b, v24.8h, Y_OFFSET.8h addhn v25.8b, v25.8h, Y_OFFSET.8h addhn v26.8b, v26.8h, Y_OFFSET.8h addhn v27.8b, v27.8h, Y_OFFSET.8h addhn v28.8b, v28.8h, UV_OFFSET.8h addhn v29.8b, v29.8h, UV_OFFSET.8h subs w15, w15, #16 // store output st1 {v24.8b, v25.8b}, [x1], #16 // store ydst1 st1 {v26.8b, v27.8b}, [x11], #16 // store ydst2 st1 {v28.8b}, [x2], #8 // store udst st1 {v29.8b}, [x3], #8 // store vdst b.gt 2b subs w5, w5, #2 // row += 2 add x0, x0, x14 // src1 += srcPadding add x10, x10, x14 // src2 += srcPadding add x1, x1, x6 // ydst1 += lumPadding add x11, x11, x6 // ydst2 += lumPadding add x2, x2, x7 // udst += chromPadding add x3, x3, x7 // vdst += chromPadding b.gt 1b ret endfunc // void ff_interleave_bytes_neon(const uint8_t *src1, const uint8_t *src2, // uint8_t *dest, int width, int height, // int src1Stride, int src2Stride, int dstStride); function ff_interleave_bytes_neon, export=1 sub w5, w5, w3 sub w6, w6, w3 sub w7, w7, w3, lsl #1 1: ands w8, w3, #0xfffffff0 // & ~15 b.eq 3f 2: ld1 {v0.16b}, [x0], #16 ld1 {v1.16b}, [x1], #16 subs w8, w8, #16 st2 {v0.16b, v1.16b}, [x2], #32 b.gt 2b tst w3, #15 b.eq 9f 3: tst w3, #8 b.eq 4f ld1 {v0.8b}, [x0], #8 ld1 {v1.8b}, [x1], #8 st2 {v0.8b, v1.8b}, [x2], #16 4: tst w3, #4 b.eq 5f ld1 {v0.s}[0], [x0], #4 ld1 {v1.s}[0], [x1], #4 zip1 v0.8b, v0.8b, v1.8b st1 {v0.8b}, [x2], #8 5: ands w8, w3, #3 b.eq 9f 6: ldrb w9, [x0], #1 ldrb w10, [x1], #1 subs w8, w8, #1 bfi w9, w10, #8, #8 strh w9, [x2], #2 b.gt 6b 9: subs w4, w4, #1 b.eq 0f add x0, x0, w5, sxtw add x1, x1, w6, sxtw add x2, x2, w7, sxtw b 1b 0: ret endfunc // void ff_deinterleave_bytes_neon(const uint8_t *src, uint8_t *dst1, uint8_t *dst2, // int width, int height, int srcStride, // int dst1Stride, int dst2Stride); function ff_deinterleave_bytes_neon, export=1 sub w5, w5, w3, lsl #1 sub w6, w6, w3 sub w7, w7, w3 1: ands w8, w3, #0xfffffff0 // & ~15 b.eq 3f 2: ld2 {v0.16b, v1.16b}, [x0], #32 subs w8, w8, #16 st1 {v0.16b}, [x1], #16 st1 {v1.16b}, [x2], #16 b.gt 2b tst w3, #15 b.eq 9f 3: tst w3, #8 b.eq 4f ld2 {v0.8b, v1.8b}, [x0], #16 st1 {v0.8b}, [x1], #8 st1 {v1.8b}, [x2], #8 4: tst w3, #4 b.eq 5f ld1 {v0.8b}, [x0], #8 shrn v1.8b, v0.8h, #8 xtn v0.8b, v0.8h st1 {v0.s}[0], [x1], #4 st1 {v1.s}[0], [x2], #4 5: ands w8, w3, #3 b.eq 9f 6: ldrh w9, [x0], #2 subs w8, w8, #1 ubfx w10, w9, #8, #8 strb w9, [x1], #1 strb w10, [x2], #1 b.gt 6b 9: subs w4, w4, #1 b.eq 0f add x0, x0, w5, sxtw add x1, x1, w6, sxtw add x2, x2, w7, sxtw b 1b 0: ret endfunc