mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-09-19 21:06:42 +00:00
bbe95f7353
From x86inc: > On AMD cpus <=K10, an ordinary ret is slow if it immediately follows either > a branch or a branch target. So switch to a 2-byte form of ret in that case. > We can automatically detect "follows a branch", but not a branch target. > (SSSE3 is a sufficient condition to know that your cpu doesn't have this problem.) x86inc can automatically determine whether to use REP_RET rather than REP in most of these cases, so impact is minimal. Additionally, a few REP_RETs were used unnecessary, despite the return being nowhere near a branch. The only CPUs affected were AMD K10s, made between 2007 and 2011, 16 years ago and 12 years ago, respectively. In the future, everyone involved with x86inc should consider dropping REP_RETs altogether.
217 lines
7.3 KiB
NASM
217 lines
7.3 KiB
NASM
;*****************************************************************************
|
|
;* x86-optimized functions for stereo3d filter
|
|
;*
|
|
;* Copyright (C) 2015 Paul B Mahol
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;*****************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA
|
|
|
|
; rgbrgbrgbrgb
|
|
; rrrrggggbbbb
|
|
|
|
shuf: db 0, 4, 8, 1,5, 9, 2, 6,10,3, 7,11,-1,-1,-1,-1
|
|
ex_r: db 0,-1,-1,-1,3,-1,-1,-1,6,-1,-1,-1, 9,-1,-1,-1
|
|
ex_g: db 1,-1,-1,-1,4,-1,-1,-1,7,-1,-1,-1,10,-1,-1,-1
|
|
ex_b: db 2,-1,-1,-1,5,-1,-1,-1,8,-1,-1,-1,11,-1,-1,-1
|
|
|
|
SECTION .text
|
|
|
|
INIT_XMM sse4
|
|
%if ARCH_X86_64
|
|
cglobal anaglyph, 6, 10, 14, 2*6*mmsize, dst, lsrc, rsrc, dst_linesize, l_linesize, r_linesize, width, height, o, cnt
|
|
%define ana_matrix_rq r6q
|
|
%define ana_matrix_gq r7q
|
|
%define ana_matrix_bq r8q
|
|
|
|
%else ; ARCH_X86_32
|
|
%if HAVE_ALIGNED_STACK
|
|
cglobal anaglyph, 3, 7, 8, 2*9*mmsize, dst, lsrc, rsrc, dst_linesize, l_linesize, o, cnt
|
|
%else
|
|
cglobal anaglyph, 3, 6, 8, 2*9*mmsize, dst, lsrc, rsrc, dst_linesize, o, cnt
|
|
%define l_linesizeq r4mp
|
|
%endif ; HAVE_ALIGNED_STACK
|
|
%define ana_matrix_rq r3q
|
|
%define ana_matrix_gq r4q
|
|
%define ana_matrix_bq r5q
|
|
%define r_linesizeq r5mp
|
|
%define widthd r6mp
|
|
%define heightd r7mp
|
|
%define m8 [rsp+mmsize*12]
|
|
%define m9 [rsp+mmsize*13]
|
|
%define m10 [rsp+mmsize*14]
|
|
%define m11 [rsp+mmsize*15]
|
|
%define m12 [rsp+mmsize*16]
|
|
%define m13 [rsp+mmsize*17]
|
|
%endif ; ARCH
|
|
|
|
mov ana_matrix_rq, r8m
|
|
mov ana_matrix_gq, r9m
|
|
mov ana_matrix_bq, r10m
|
|
movu m3, [ana_matrix_rq+ 0]
|
|
movq m5, [ana_matrix_rq+16]
|
|
pshufd m0, m3, q0000
|
|
pshufd m1, m3, q1111
|
|
pshufd m2, m3, q2222
|
|
pshufd m3, m3, q3333
|
|
pshufd m4, m5, q0000
|
|
pshufd m5, m5, q1111
|
|
mova [rsp+mmsize*0], m0
|
|
mova [rsp+mmsize*1], m1
|
|
mova [rsp+mmsize*2], m2
|
|
mova [rsp+mmsize*3], m3
|
|
mova [rsp+mmsize*4], m4
|
|
mova [rsp+mmsize*5], m5
|
|
|
|
movu m3, [ana_matrix_gq+ 0]
|
|
movq m5, [ana_matrix_gq+16]
|
|
pshufd m0, m3, q0000
|
|
pshufd m1, m3, q1111
|
|
pshufd m2, m3, q2222
|
|
pshufd m3, m3, q3333
|
|
pshufd m4, m5, q0000
|
|
pshufd m5, m5, q1111
|
|
mova [rsp+mmsize*6 ], m0
|
|
mova [rsp+mmsize*7 ], m1
|
|
mova [rsp+mmsize*8 ], m2
|
|
mova [rsp+mmsize*9 ], m3
|
|
mova [rsp+mmsize*10], m4
|
|
mova [rsp+mmsize*11], m5
|
|
|
|
%if ARCH_X86_64
|
|
movu m11, [ana_matrix_bq+ 0]
|
|
movq m13, [ana_matrix_bq+16]
|
|
pshufd m8, m11, q0000
|
|
pshufd m9, m11, q1111
|
|
pshufd m10, m11, q2222
|
|
pshufd m11, m11, q3333
|
|
pshufd m12, m13, q0000
|
|
pshufd m13, m13, q1111
|
|
mov widthd, dword widthm
|
|
mov heightd, dword heightm
|
|
%else
|
|
movu m3, [ana_matrix_bq+ 0]
|
|
movq m5, [ana_matrix_bq+16]
|
|
pshufd m0, m3, q0000
|
|
pshufd m1, m3, q1111
|
|
pshufd m2, m3, q2222
|
|
pshufd m3, m3, q3333
|
|
pshufd m4, m5, q0000
|
|
pshufd m5, m5, q1111
|
|
mova [rsp+mmsize*12], m0
|
|
mova [rsp+mmsize*13], m1
|
|
mova [rsp+mmsize*14], m2
|
|
mova [rsp+mmsize*15], m3
|
|
mova [rsp+mmsize*16], m4
|
|
mova [rsp+mmsize*17], m5
|
|
mov dst_linesizeq, r3m
|
|
%if HAVE_ALIGNED_STACK
|
|
mov l_linesizeq, r4m
|
|
%endif
|
|
%endif ; ARCH
|
|
|
|
.nextrow:
|
|
mov od, widthd
|
|
xor cntd, cntd
|
|
|
|
.loop:
|
|
movu m3, [lsrcq+cntq]
|
|
pshufb m1, m3, [ex_r]
|
|
pshufb m2, m3, [ex_g]
|
|
pshufb m3, [ex_b]
|
|
movu m0, [rsrcq+cntq]
|
|
pshufb m4, m0, [ex_r]
|
|
pshufb m5, m0, [ex_g]
|
|
pshufb m0, [ex_b]
|
|
pmulld m1, [rsp+mmsize*0]
|
|
pmulld m2, [rsp+mmsize*1]
|
|
pmulld m3, [rsp+mmsize*2]
|
|
pmulld m4, [rsp+mmsize*3]
|
|
pmulld m5, [rsp+mmsize*4]
|
|
pmulld m0, [rsp+mmsize*5]
|
|
paddd m1, m2
|
|
paddd m3, m4
|
|
paddd m5, m0
|
|
paddd m1, m3
|
|
paddd m1, m5
|
|
|
|
movu m3, [lsrcq+cntq]
|
|
pshufb m7, m3, [ex_r]
|
|
pshufb m2, m3, [ex_g]
|
|
pshufb m3, [ex_b]
|
|
movu m0, [rsrcq+cntq]
|
|
pshufb m4, m0, [ex_r]
|
|
pshufb m5, m0, [ex_g]
|
|
pshufb m0, [ex_b]
|
|
pmulld m7, [rsp+mmsize*6]
|
|
pmulld m2, [rsp+mmsize*7]
|
|
pmulld m3, [rsp+mmsize*8]
|
|
pmulld m4, [rsp+mmsize*9]
|
|
pmulld m5, [rsp+mmsize*10]
|
|
pmulld m0, [rsp+mmsize*11]
|
|
paddd m7, m2
|
|
paddd m3, m4
|
|
paddd m5, m0
|
|
paddd m7, m3
|
|
paddd m7, m5
|
|
|
|
movu m4, [lsrcq+cntq]
|
|
pshufb m2, m4, [ex_r]
|
|
pshufb m3, m4, [ex_g]
|
|
pshufb m4, [ex_b]
|
|
movu m0, [rsrcq+cntq]
|
|
pshufb m5, m0, [ex_r]
|
|
pshufb m6, m0, [ex_g]
|
|
pshufb m0, [ex_b]
|
|
pmulld m2, m8
|
|
pmulld m3, m9
|
|
pmulld m4, m10
|
|
pmulld m5, m11
|
|
pmulld m6, m12
|
|
pmulld m0, m13
|
|
paddd m2, m3
|
|
paddd m4, m5
|
|
paddd m6, m0
|
|
paddd m2, m4
|
|
paddd m2, m6
|
|
|
|
psrld m1, 16
|
|
psrld m7, 16
|
|
psrld m2, 16
|
|
|
|
packusdw m1, m7
|
|
packusdw m2, m2
|
|
packuswb m1, m2
|
|
pshufb m1, [shuf]
|
|
|
|
movq [dstq+cntq+0], m1
|
|
psrldq m1, 8
|
|
movd [dstq+cntq+8], m1
|
|
add cntd, 12
|
|
sub od, 4
|
|
jg .loop
|
|
|
|
add dstq, dst_linesizeq
|
|
add lsrcq, l_linesizeq
|
|
add rsrcq, r_linesizeq
|
|
sub heightd, 1
|
|
jg .nextrow
|
|
RET
|