mirror of
				https://github.com/nyanmisaka/ffmpeg-rockchip.git
				synced 2025-10-31 20:42:49 +08:00 
			
		
		
		
	
		
			
				
	
	
		
			432 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			432 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
| ;******************************************************************************
 | |
| ;* VP9 MC SIMD optimizations
 | |
| ;*
 | |
| ;* Copyright (c) 2015 Ronald S. Bultje <rsbultje gmail com>
 | |
| ;*
 | |
| ;* This file is part of FFmpeg.
 | |
| ;*
 | |
| ;* FFmpeg is free software; you can redistribute it and/or
 | |
| ;* modify it under the terms of the GNU Lesser General Public
 | |
| ;* License as published by the Free Software Foundation; either
 | |
| ;* version 2.1 of the License, or (at your option) any later version.
 | |
| ;*
 | |
| ;* FFmpeg is distributed in the hope that it will be useful,
 | |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| ;* Lesser General Public License for more details.
 | |
| ;*
 | |
| ;* You should have received a copy of the GNU Lesser General Public
 | |
| ;* License along with FFmpeg; if not, write to the Free Software
 | |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
| ;******************************************************************************
 | |
| 
 | |
| %include "libavutil/x86/x86util.asm"
 | |
| 
 | |
| SECTION_RODATA 32
 | |
| 
 | |
| pd_64: times 8 dd 64
 | |
| 
 | |
| cextern pw_1023
 | |
| cextern pw_4095
 | |
| 
 | |
| SECTION .text
 | |
| 
 | |
| %macro filter_h4_fn 1-2 12
 | |
| cglobal vp9_%1_8tap_1d_h_4_10, 6, 6, %2, dst, dstride, src, sstride, h, filtery
 | |
|     mova        m5, [pw_1023]
 | |
| .body:
 | |
| %if notcpuflag(sse4) && ARCH_X86_64
 | |
|     pxor       m11, m11
 | |
| %endif
 | |
|     mova        m6, [pd_64]
 | |
|     mova        m7, [filteryq+ 0]
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     mova        m8, [filteryq+32]
 | |
|     mova        m9, [filteryq+64]
 | |
|     mova       m10, [filteryq+96]
 | |
| %endif
 | |
| .loop:
 | |
|     movh        m0, [srcq-6]
 | |
|     movh        m1, [srcq-4]
 | |
|     movh        m2, [srcq-2]
 | |
|     movh        m3, [srcq+0]
 | |
|     movh        m4, [srcq+2]
 | |
|     punpcklwd   m0, m1
 | |
|     punpcklwd   m2, m3
 | |
|     pmaddwd     m0, m7
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m2, m8
 | |
| %else
 | |
|     pmaddwd     m2, [filteryq+32]
 | |
| %endif
 | |
|     movu        m1, [srcq+4]
 | |
|     movu        m3, [srcq+6]
 | |
|     paddd       m0, m2
 | |
|     movu        m2, [srcq+8]
 | |
|     add       srcq, sstrideq
 | |
|     punpcklwd   m4, m1
 | |
|     punpcklwd   m3, m2
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m4, m9
 | |
|     pmaddwd     m3, m10
 | |
| %else
 | |
|     pmaddwd     m4, [filteryq+64]
 | |
|     pmaddwd     m3, [filteryq+96]
 | |
| %endif
 | |
|     paddd       m0, m4
 | |
|     paddd       m0, m3
 | |
|     paddd       m0, m6
 | |
|     psrad       m0, 7
 | |
| %if cpuflag(sse4)
 | |
|     packusdw    m0, m0
 | |
| %else
 | |
|     packssdw    m0, m0
 | |
| %endif
 | |
| %ifidn %1, avg
 | |
|     movh        m1, [dstq]
 | |
| %endif
 | |
|     pminsw      m0, m5
 | |
| %if notcpuflag(sse4)
 | |
| %if ARCH_X86_64
 | |
|     pmaxsw      m0, m11
 | |
| %else
 | |
|     pxor        m2, m2
 | |
|     pmaxsw      m0, m2
 | |
| %endif
 | |
| %endif
 | |
| %ifidn %1, avg
 | |
|     pavgw       m0, m1
 | |
| %endif
 | |
|     movh    [dstq], m0
 | |
|     add       dstq, dstrideq
 | |
|     dec         hd
 | |
|     jg .loop
 | |
|     RET
 | |
| 
 | |
| cglobal vp9_%1_8tap_1d_h_4_12, 6, 6, %2, dst, dstride, src, sstride, h, filtery
 | |
|     mova        m5, [pw_4095]
 | |
|     jmp mangle(private_prefix %+ _ %+ vp9_%1_8tap_1d_h_4_10 %+ SUFFIX).body
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| filter_h4_fn put
 | |
| filter_h4_fn avg
 | |
| 
 | |
| %macro filter_h_fn 1-2 12
 | |
| %assign %%px mmsize/2
 | |
| cglobal vp9_%1_8tap_1d_h_ %+ %%px %+ _10, 6, 6, %2, dst, dstride, src, sstride, h, filtery
 | |
|     mova        m5, [pw_1023]
 | |
| .body:
 | |
| %if notcpuflag(sse4) && ARCH_X86_64
 | |
|     pxor       m11, m11
 | |
| %endif
 | |
|     mova        m6, [pd_64]
 | |
|     mova        m7, [filteryq+ 0]
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     mova        m8, [filteryq+32]
 | |
|     mova        m9, [filteryq+64]
 | |
|     mova       m10, [filteryq+96]
 | |
| %endif
 | |
| .loop:
 | |
|     movu        m0, [srcq-6]
 | |
|     movu        m1, [srcq-4]
 | |
|     movu        m2, [srcq-2]
 | |
|     movu        m3, [srcq+0]
 | |
|     movu        m4, [srcq+2]
 | |
|     pmaddwd     m0, m7
 | |
|     pmaddwd     m1, m7
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m2, m8
 | |
|     pmaddwd     m3, m8
 | |
|     pmaddwd     m4, m9
 | |
| %else
 | |
|     pmaddwd     m2, [filteryq+32]
 | |
|     pmaddwd     m3, [filteryq+32]
 | |
|     pmaddwd     m4, [filteryq+64]
 | |
| %endif
 | |
|     paddd       m0, m2
 | |
|     paddd       m1, m3
 | |
|     paddd       m0, m4
 | |
|     movu        m2, [srcq+4]
 | |
|     movu        m3, [srcq+6]
 | |
|     movu        m4, [srcq+8]
 | |
|     add       srcq, sstrideq
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m2, m9
 | |
|     pmaddwd     m3, m10
 | |
|     pmaddwd     m4, m10
 | |
| %else
 | |
|     pmaddwd     m2, [filteryq+64]
 | |
|     pmaddwd     m3, [filteryq+96]
 | |
|     pmaddwd     m4, [filteryq+96]
 | |
| %endif
 | |
|     paddd       m1, m2
 | |
|     paddd       m0, m3
 | |
|     paddd       m1, m4
 | |
|     paddd       m0, m6
 | |
|     paddd       m1, m6
 | |
|     psrad       m0, 7
 | |
|     psrad       m1, 7
 | |
| %if cpuflag(sse4)
 | |
|     packusdw    m0, m0
 | |
|     packusdw    m1, m1
 | |
| %else
 | |
|     packssdw    m0, m0
 | |
|     packssdw    m1, m1
 | |
| %endif
 | |
|     punpcklwd   m0, m1
 | |
|     pminsw      m0, m5
 | |
| %if notcpuflag(sse4)
 | |
| %if ARCH_X86_64
 | |
|     pmaxsw      m0, m11
 | |
| %else
 | |
|     pxor        m2, m2
 | |
|     pmaxsw      m0, m2
 | |
| %endif
 | |
| %endif
 | |
| %ifidn %1, avg
 | |
|     pavgw       m0, [dstq]
 | |
| %endif
 | |
|     mova    [dstq], m0
 | |
|     add       dstq, dstrideq
 | |
|     dec         hd
 | |
|     jg .loop
 | |
|     RET
 | |
| 
 | |
| cglobal vp9_%1_8tap_1d_h_ %+ %%px %+ _12, 6, 6, %2, dst, dstride, src, sstride, h, filtery
 | |
|     mova        m5, [pw_4095]
 | |
|     jmp mangle(private_prefix %+ _ %+ vp9_%1_8tap_1d_h_ %+ %%px %+ _10 %+ SUFFIX).body
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| filter_h_fn put
 | |
| filter_h_fn avg
 | |
| %if HAVE_AVX2_EXTERNAL
 | |
| INIT_YMM avx2
 | |
| filter_h_fn put
 | |
| filter_h_fn avg
 | |
| %endif
 | |
| 
 | |
| %macro filter_v4_fn 1-2 12
 | |
| %if ARCH_X86_64
 | |
| cglobal vp9_%1_8tap_1d_v_4_10, 6, 8, %2, dst, dstride, src, sstride, h, filtery, src4, sstride3
 | |
| %else
 | |
| cglobal vp9_%1_8tap_1d_v_4_10, 4, 7, %2, dst, dstride, src, sstride, filtery, src4, sstride3
 | |
|     mov   filteryq, r5mp
 | |
| %define hd r4mp
 | |
| %endif
 | |
|     mova        m5, [pw_1023]
 | |
| .body:
 | |
| %if notcpuflag(sse4) && ARCH_X86_64
 | |
|     pxor       m11, m11
 | |
| %endif
 | |
|     mova        m6, [pd_64]
 | |
|     lea  sstride3q, [sstrideq*3]
 | |
|     lea      src4q, [srcq+sstrideq]
 | |
|     sub       srcq, sstride3q
 | |
|     mova        m7, [filteryq+  0]
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     mova        m8, [filteryq+ 32]
 | |
|     mova        m9, [filteryq+ 64]
 | |
|     mova       m10, [filteryq+ 96]
 | |
| %endif
 | |
| .loop:
 | |
|     ; FIXME maybe reuse loads from previous rows, or just
 | |
|     ; more generally unroll this to prevent multiple loads of
 | |
|     ; the same data?
 | |
|     movh        m0, [srcq]
 | |
|     movh        m1, [srcq+sstrideq]
 | |
|     movh        m2, [srcq+sstrideq*2]
 | |
|     movh        m3, [srcq+sstride3q]
 | |
|     add       srcq, sstrideq
 | |
|     movh        m4, [src4q]
 | |
|     punpcklwd   m0, m1
 | |
|     punpcklwd   m2, m3
 | |
|     pmaddwd     m0, m7
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m2, m8
 | |
| %else
 | |
|     pmaddwd     m2, [filteryq+ 32]
 | |
| %endif
 | |
|     movh        m1, [src4q+sstrideq]
 | |
|     movh        m3, [src4q+sstrideq*2]
 | |
|     paddd       m0, m2
 | |
|     movh        m2, [src4q+sstride3q]
 | |
|     add      src4q, sstrideq
 | |
|     punpcklwd   m4, m1
 | |
|     punpcklwd   m3, m2
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m4, m9
 | |
|     pmaddwd     m3, m10
 | |
| %else
 | |
|     pmaddwd     m4, [filteryq+ 64]
 | |
|     pmaddwd     m3, [filteryq+ 96]
 | |
| %endif
 | |
|     paddd       m0, m4
 | |
|     paddd       m0, m3
 | |
|     paddd       m0, m6
 | |
|     psrad       m0, 7
 | |
| %if cpuflag(sse4)
 | |
|     packusdw    m0, m0
 | |
| %else
 | |
|     packssdw    m0, m0
 | |
| %endif
 | |
| %ifidn %1, avg
 | |
|     movh        m1, [dstq]
 | |
| %endif
 | |
|     pminsw      m0, m5
 | |
| %if notcpuflag(sse4)
 | |
| %if ARCH_X86_64
 | |
|     pmaxsw      m0, m11
 | |
| %else
 | |
|     pxor        m2, m2
 | |
|     pmaxsw      m0, m2
 | |
| %endif
 | |
| %endif
 | |
| %ifidn %1, avg
 | |
|     pavgw       m0, m1
 | |
| %endif
 | |
|     movh    [dstq], m0
 | |
|     add       dstq, dstrideq
 | |
|     dec         hd
 | |
|     jg .loop
 | |
|     RET
 | |
| 
 | |
| %if ARCH_X86_64
 | |
| cglobal vp9_%1_8tap_1d_v_4_12, 6, 8, %2, dst, dstride, src, sstride, h, filtery, src4, sstride3
 | |
| %else
 | |
| cglobal vp9_%1_8tap_1d_v_4_12, 4, 7, %2, dst, dstride, src, sstride, filtery, src4, sstride3
 | |
|     mov   filteryq, r5mp
 | |
| %endif
 | |
|     mova        m5, [pw_4095]
 | |
|     jmp mangle(private_prefix %+ _ %+ vp9_%1_8tap_1d_v_4_10 %+ SUFFIX).body
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| filter_v4_fn put
 | |
| filter_v4_fn avg
 | |
| 
 | |
| %macro filter_v_fn 1-2 13
 | |
| %assign %%px mmsize/2
 | |
| %if ARCH_X86_64
 | |
| cglobal vp9_%1_8tap_1d_v_ %+ %%px %+ _10, 6, 8, %2, dst, dstride, src, sstride, h, filtery, src4, sstride3
 | |
| %else
 | |
| cglobal vp9_%1_8tap_1d_v_ %+ %%px %+ _10, 4, 7, %2, dst, dstride, src, sstride, filtery, src4, sstride3
 | |
|     mov   filteryq, r5mp
 | |
| %define hd r4mp
 | |
| %endif
 | |
|     mova        m5, [pw_1023]
 | |
| .body:
 | |
| %if notcpuflag(sse4) && ARCH_X86_64
 | |
|     pxor       m12, m12
 | |
| %endif
 | |
| %if ARCH_X86_64
 | |
|     mova       m11, [pd_64]
 | |
| %endif
 | |
|     lea  sstride3q, [sstrideq*3]
 | |
|     lea      src4q, [srcq+sstrideq]
 | |
|     sub       srcq, sstride3q
 | |
|     mova        m7, [filteryq+  0]
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     mova        m8, [filteryq+ 32]
 | |
|     mova        m9, [filteryq+ 64]
 | |
|     mova       m10, [filteryq+ 96]
 | |
| %endif
 | |
| .loop:
 | |
|     ; FIXME maybe reuse loads from previous rows, or just
 | |
|     ; more generally unroll this to prevent multiple loads of
 | |
|     ; the same data?
 | |
|     movu        m0, [srcq]
 | |
|     movu        m1, [srcq+sstrideq]
 | |
|     movu        m2, [srcq+sstrideq*2]
 | |
|     movu        m3, [srcq+sstride3q]
 | |
|     add       srcq, sstrideq
 | |
|     movu        m4, [src4q]
 | |
|     SBUTTERFLY  wd, 0, 1, 6
 | |
|     SBUTTERFLY  wd, 2, 3, 6
 | |
|     pmaddwd     m0, m7
 | |
|     pmaddwd     m1, m7
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m2, m8
 | |
|     pmaddwd     m3, m8
 | |
| %else
 | |
|     pmaddwd     m2, [filteryq+ 32]
 | |
|     pmaddwd     m3, [filteryq+ 32]
 | |
| %endif
 | |
|     paddd       m0, m2
 | |
|     paddd       m1, m3
 | |
|     movu        m2, [src4q+sstrideq]
 | |
|     movu        m3, [src4q+sstrideq*2]
 | |
|     SBUTTERFLY  wd, 4, 2, 6
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m4, m9
 | |
|     pmaddwd     m2, m9
 | |
| %else
 | |
|     pmaddwd     m4, [filteryq+ 64]
 | |
|     pmaddwd     m2, [filteryq+ 64]
 | |
| %endif
 | |
|     paddd       m0, m4
 | |
|     paddd       m1, m2
 | |
|     movu        m4, [src4q+sstride3q]
 | |
|     add      src4q, sstrideq
 | |
|     SBUTTERFLY  wd, 3, 4, 6
 | |
| %if ARCH_X86_64 && mmsize > 8
 | |
|     pmaddwd     m3, m10
 | |
|     pmaddwd     m4, m10
 | |
| %else
 | |
|     pmaddwd     m3, [filteryq+ 96]
 | |
|     pmaddwd     m4, [filteryq+ 96]
 | |
| %endif
 | |
|     paddd       m0, m3
 | |
|     paddd       m1, m4
 | |
| %if ARCH_X86_64
 | |
|     paddd       m0, m11
 | |
|     paddd       m1, m11
 | |
| %else
 | |
|     paddd       m0, [pd_64]
 | |
|     paddd       m1, [pd_64]
 | |
| %endif
 | |
|     psrad       m0, 7
 | |
|     psrad       m1, 7
 | |
| %if cpuflag(sse4)
 | |
|     packusdw    m0, m1
 | |
| %else
 | |
|     packssdw    m0, m1
 | |
| %endif
 | |
|     pminsw      m0, m5
 | |
| %if notcpuflag(sse4)
 | |
| %if ARCH_X86_64
 | |
|     pmaxsw      m0, m12
 | |
| %else
 | |
|     pxor        m2, m2
 | |
|     pmaxsw      m0, m2
 | |
| %endif
 | |
| %endif
 | |
| %ifidn %1, avg
 | |
|     pavgw       m0, [dstq]
 | |
| %endif
 | |
|     mova    [dstq], m0
 | |
|     add       dstq, dstrideq
 | |
|     dec         hd
 | |
|     jg .loop
 | |
|     RET
 | |
| 
 | |
| %if ARCH_X86_64
 | |
| cglobal vp9_%1_8tap_1d_v_ %+ %%px %+ _12, 6, 8, %2, dst, dstride, src, sstride, h, filtery, src4, sstride3
 | |
| %else
 | |
| cglobal vp9_%1_8tap_1d_v_ %+ %%px %+ _12, 4, 7, %2, dst, dstride, src, sstride, filtery, src4, sstride3
 | |
|     mov   filteryq, r5mp
 | |
| %endif
 | |
|     mova        m5, [pw_4095]
 | |
|     jmp mangle(private_prefix %+ _ %+ vp9_%1_8tap_1d_v_ %+ %%px %+ _10 %+ SUFFIX).body
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| filter_v_fn put
 | |
| filter_v_fn avg
 | |
| %if HAVE_AVX2_EXTERNAL
 | |
| INIT_YMM avx2
 | |
| filter_v_fn put
 | |
| filter_v_fn avg
 | |
| %endif
 | 
