mirror of
				https://github.com/nyanmisaka/ffmpeg-rockchip.git
				synced 2025-10-31 20:42:49 +08:00 
			
		
		
		
	
		
			
				
	
	
		
			488 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			488 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
| ;******************************************************************************
 | |
| ;* SIMD optimized MPEG-4 Parametric Stereo decoding functions
 | |
| ;*
 | |
| ;* Copyright (C) 2015 James Almer
 | |
| ;*
 | |
| ;* This file is part of FFmpeg.
 | |
| ;*
 | |
| ;* FFmpeg is free software; you can redistribute it and/or
 | |
| ;* modify it under the terms of the GNU Lesser General Public
 | |
| ;* License as published by the Free Software Foundation; either
 | |
| ;* version 2.1 of the License, or (at your option) any later version.
 | |
| ;*
 | |
| ;* FFmpeg is distributed in the hope that it will be useful,
 | |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| ;* Lesser General Public License for more details.
 | |
| ;*
 | |
| ;* You should have received a copy of the GNU Lesser General Public
 | |
| ;* License along with FFmpeg; if not, write to the Free Software
 | |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
| ;******************************************************************************
 | |
| 
 | |
| %include "libavutil/x86/x86util.asm"
 | |
| 
 | |
| SECTION_RODATA
 | |
| 
 | |
| ps_p1m1p1m1: dd 0, 0x80000000, 0, 0x80000000
 | |
| 
 | |
| SECTION .text
 | |
| 
 | |
| ;*************************************************************************
 | |
| ;void ff_ps_add_squares_<opt>(float *dst, const float (*src)[2], int n);
 | |
| ;*************************************************************************
 | |
| %macro PS_ADD_SQUARES 1
 | |
| cglobal ps_add_squares, 3, 3, %1, dst, src, n
 | |
|     shl    nd, 3
 | |
|     add  srcq, nq
 | |
|     neg    nq
 | |
| 
 | |
| align 16
 | |
| .loop:
 | |
|     movaps m0, [srcq+nq]
 | |
|     movaps m1, [srcq+nq+mmsize]
 | |
|     mulps  m0, m0
 | |
|     mulps  m1, m1
 | |
|     HADDPS m0, m1, m2
 | |
|     addps  m0, [dstq]
 | |
|     movaps [dstq], m0
 | |
|     add  dstq, mmsize
 | |
|     add    nq, mmsize*2
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse
 | |
| PS_ADD_SQUARES 2
 | |
| INIT_XMM sse3
 | |
| PS_ADD_SQUARES 3
 | |
| 
 | |
| ;*******************************************************************
 | |
| ;void ff_ps_mul_pair_single_sse(float (*dst)[2], float (*src0)[2],
 | |
| ;                                   float *src1, int n);
 | |
| ;*******************************************************************
 | |
| INIT_XMM sse
 | |
| cglobal ps_mul_pair_single, 4, 4, 4, dst, src1, src2, n
 | |
|     shl      nd, 3
 | |
|     add   src1q, nq
 | |
|     add    dstq, nq
 | |
|     neg      nq
 | |
| 
 | |
| align 16
 | |
| .loop:
 | |
|     movu     m0, [src1q+nq]
 | |
|     movu     m1, [src1q+nq+mmsize]
 | |
|     mova     m2, [src2q]
 | |
|     mova     m3, m2
 | |
|     unpcklps m2, m2
 | |
|     unpckhps m3, m3
 | |
|     mulps    m0, m2
 | |
|     mulps    m1, m3
 | |
|     mova [dstq+nq], m0
 | |
|     mova [dstq+nq+mmsize], m1
 | |
|     add   src2q, mmsize
 | |
|     add      nq, mmsize*2
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| 
 | |
| ;***********************************************************************
 | |
| ;void ff_ps_stereo_interpolate_sse3(float (*l)[2], float (*r)[2],
 | |
| ;                                   float h[2][4], float h_step[2][4],
 | |
| ;                                   int len);
 | |
| ;***********************************************************************
 | |
| INIT_XMM sse3
 | |
| cglobal ps_stereo_interpolate, 5, 5, 6, l, r, h, h_step, n
 | |
|     movaps   m0, [hq]
 | |
|     movaps   m1, [h_stepq]
 | |
|     unpcklps m4, m0, m0
 | |
|     unpckhps m0, m0
 | |
|     unpcklps m5, m1, m1
 | |
|     unpckhps m1, m1
 | |
|     shl      nd, 3
 | |
|     add      lq, nq
 | |
|     add      rq, nq
 | |
|     neg      nq
 | |
| 
 | |
| align 16
 | |
| .loop:
 | |
|     addps    m4, m5
 | |
|     addps    m0, m1
 | |
|     movddup  m2, [lq+nq]
 | |
|     movddup  m3, [rq+nq]
 | |
|     mulps    m2, m4
 | |
|     mulps    m3, m0
 | |
|     addps    m2, m3
 | |
|     movsd  [lq+nq], m2
 | |
|     movhps [rq+nq], m2
 | |
|     add      nq, 8
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| 
 | |
| ;***************************************************************************
 | |
| ;void ps_stereo_interpolate_ipdopd_sse3(float (*l)[2], float (*r)[2],
 | |
| ;                                       float h[2][4], float h_step[2][4],
 | |
| ;                                       int len);
 | |
| ;***************************************************************************
 | |
| INIT_XMM sse3
 | |
| cglobal ps_stereo_interpolate_ipdopd, 5, 5, 10, l, r, h, h_step, n
 | |
|     movaps   m0, [hq]
 | |
|     movaps   m1, [hq+mmsize]
 | |
| %if ARCH_X86_64
 | |
|     movaps   m8, [h_stepq]
 | |
|     movaps   m9, [h_stepq+mmsize]
 | |
|     %define  H_STEP0 m8
 | |
|     %define  H_STEP1 m9
 | |
| %else
 | |
|     %define  H_STEP0 [h_stepq]
 | |
|     %define  H_STEP1 [h_stepq+mmsize]
 | |
| %endif
 | |
|     shl      nd, 3
 | |
|     add      lq, nq
 | |
|     add      rq, nq
 | |
|     neg      nq
 | |
| 
 | |
| align 16
 | |
| .loop:
 | |
|     addps    m0, H_STEP0
 | |
|     addps    m1, H_STEP1
 | |
|     movddup  m2, [lq+nq]
 | |
|     movddup  m3, [rq+nq]
 | |
|     shufps   m4, m2, m2, q2301
 | |
|     shufps   m5, m3, m3, q2301
 | |
|     unpcklps m6, m0, m0
 | |
|     unpckhps m7, m0, m0
 | |
|     mulps    m2, m6
 | |
|     mulps    m3, m7
 | |
|     unpcklps m6, m1, m1
 | |
|     unpckhps m7, m1, m1
 | |
|     mulps    m4, m6
 | |
|     mulps    m5, m7
 | |
|     addps    m2, m3
 | |
|     addsubps m2, m4
 | |
|     addsubps m2, m5
 | |
|     movsd  [lq+nq], m2
 | |
|     movhps [rq+nq], m2
 | |
|     add      nq, 8
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| 
 | |
| ;**********************************************************
 | |
| ;void ps_hybrid_analysis_ileave_sse(float out[2][38][64],
 | |
| ;                                   float (*in)[32][2],
 | |
| ;                                   int i, int len)
 | |
| ;**********************************************************
 | |
| INIT_XMM sse
 | |
| cglobal ps_hybrid_analysis_ileave, 3, 7, 5, out, in, i, len, in0, in1, tmp
 | |
|     movsxdifnidn        iq, id
 | |
|     mov               lend, 32 << 3
 | |
|     lea                inq, [inq+iq*4]
 | |
|     mov               tmpd, id
 | |
|     shl               tmpd, 8
 | |
|     add               outq, tmpq
 | |
|     mov               tmpd, 64
 | |
|     sub               tmpd, id
 | |
|     mov                 id, tmpd
 | |
| 
 | |
|     test                id, 1
 | |
|     jne .loop4
 | |
|     test                id, 2
 | |
|     jne .loop8
 | |
| 
 | |
| align 16
 | |
| .loop16:
 | |
|     mov               in0q, inq
 | |
|     mov               in1q, 38*64*4
 | |
|     add               in1q, in0q
 | |
|     mov               tmpd, lend
 | |
| 
 | |
| .inner_loop16:
 | |
|     movaps              m0, [in0q]
 | |
|     movaps              m1, [in1q]
 | |
|     movaps              m2, [in0q+lenq]
 | |
|     movaps              m3, [in1q+lenq]
 | |
|     TRANSPOSE4x4PS 0, 1, 2, 3, 4
 | |
|     movaps          [outq], m0
 | |
|     movaps     [outq+lenq], m1
 | |
|     movaps   [outq+lenq*2], m2
 | |
|     movaps [outq+3*32*2*4], m3
 | |
|     lea               in0q, [in0q+lenq*2]
 | |
|     lea               in1q, [in1q+lenq*2]
 | |
|     add               outq, mmsize
 | |
|     sub               tmpd, mmsize
 | |
|     jg .inner_loop16
 | |
|     add                inq, 16
 | |
|     add               outq, 3*32*2*4
 | |
|     sub                 id, 4
 | |
|     jg .loop16
 | |
|     RET
 | |
| 
 | |
| align 16
 | |
| .loop8:
 | |
|     mov               in0q, inq
 | |
|     mov               in1q, 38*64*4
 | |
|     add               in1q, in0q
 | |
|     mov               tmpd, lend
 | |
| 
 | |
| .inner_loop8:
 | |
|     movlps              m0, [in0q]
 | |
|     movlps              m1, [in1q]
 | |
|     movhps              m0, [in0q+lenq]
 | |
|     movhps              m1, [in1q+lenq]
 | |
|     SBUTTERFLYPS 0, 1, 2
 | |
|     SBUTTERFLYPD 0, 1, 2
 | |
|     movaps          [outq], m0
 | |
|     movaps     [outq+lenq], m1
 | |
|     lea               in0q, [in0q+lenq*2]
 | |
|     lea               in1q, [in1q+lenq*2]
 | |
|     add               outq, mmsize
 | |
|     sub               tmpd, mmsize
 | |
|     jg .inner_loop8
 | |
|     add                inq, 8
 | |
|     add               outq, lenq
 | |
|     sub                 id, 2
 | |
|     jg .loop16
 | |
|     RET
 | |
| 
 | |
| align 16
 | |
| .loop4:
 | |
|     mov               in0q, inq
 | |
|     mov               in1q, 38*64*4
 | |
|     add               in1q, in0q
 | |
|     mov               tmpd, lend
 | |
| 
 | |
| .inner_loop4:
 | |
|     movss               m0, [in0q]
 | |
|     movss               m1, [in1q]
 | |
|     movss               m2, [in0q+lenq]
 | |
|     movss               m3, [in1q+lenq]
 | |
|     movlhps             m0, m1
 | |
|     movlhps             m2, m3
 | |
|     shufps              m0, m2, q2020
 | |
|     movaps          [outq], m0
 | |
|     lea               in0q, [in0q+lenq*2]
 | |
|     lea               in1q, [in1q+lenq*2]
 | |
|     add               outq, mmsize
 | |
|     sub               tmpd, mmsize
 | |
|     jg .inner_loop4
 | |
|     add                inq, 4
 | |
|     sub                 id, 1
 | |
|     test                id, 2
 | |
|     jne .loop8
 | |
|     cmp                 id, 4
 | |
|     jge .loop16
 | |
|     RET
 | |
| 
 | |
| ;***********************************************************
 | |
| ;void ps_hybrid_synthesis_deint_sse4(float out[2][38][64],
 | |
| ;                                    float (*in)[32][2],
 | |
| ;                                    int i, int len)
 | |
| ;***********************************************************
 | |
| %macro HYBRID_SYNTHESIS_DEINT 0
 | |
| cglobal ps_hybrid_synthesis_deint, 3, 7, 5, out, in, i, len, out0, out1, tmp
 | |
| %if cpuflag(sse4)
 | |
| %define MOVH movsd
 | |
| %else
 | |
| %define MOVH movlps
 | |
| %endif
 | |
|     movsxdifnidn        iq, id
 | |
|     mov               lend, 32 << 3
 | |
|     lea               outq, [outq+iq*4]
 | |
|     mov               tmpd, id
 | |
|     shl               tmpd, 8
 | |
|     add                inq, tmpq
 | |
|     mov               tmpd, 64
 | |
|     sub               tmpd, id
 | |
|     mov                 id, tmpd
 | |
| 
 | |
|     test                id, 1
 | |
|     jne .loop4
 | |
|     test                id, 2
 | |
|     jne .loop8
 | |
| 
 | |
| align 16
 | |
| .loop16:
 | |
|     mov              out0q, outq
 | |
|     mov              out1q, 38*64*4
 | |
|     add              out1q, out0q
 | |
|     mov               tmpd, lend
 | |
| 
 | |
| .inner_loop16:
 | |
|     movaps              m0, [inq]
 | |
|     movaps              m1, [inq+lenq]
 | |
|     movaps              m2, [inq+lenq*2]
 | |
|     movaps              m3, [inq+3*32*2*4]
 | |
|     TRANSPOSE4x4PS 0, 1, 2, 3, 4
 | |
|     movaps         [out0q], m0
 | |
|     movaps         [out1q], m1
 | |
|     movaps    [out0q+lenq], m2
 | |
|     movaps    [out1q+lenq], m3
 | |
|     lea              out0q, [out0q+lenq*2]
 | |
|     lea              out1q, [out1q+lenq*2]
 | |
|     add                inq, mmsize
 | |
|     sub               tmpd, mmsize
 | |
|     jg .inner_loop16
 | |
|     add               outq, 16
 | |
|     add                inq, 3*32*2*4
 | |
|     sub                 id, 4
 | |
|     jg .loop16
 | |
|     RET
 | |
| 
 | |
| align 16
 | |
| .loop8:
 | |
|     mov              out0q, outq
 | |
|     mov              out1q, 38*64*4
 | |
|     add              out1q, out0q
 | |
|     mov               tmpd, lend
 | |
| 
 | |
| .inner_loop8:
 | |
|     movaps              m0, [inq]
 | |
|     movaps              m1, [inq+lenq]
 | |
|     SBUTTERFLYPS 0, 1, 2
 | |
|     SBUTTERFLYPD 0, 1, 2
 | |
|     MOVH           [out0q], m0
 | |
|     MOVH           [out1q], m1
 | |
|     movhps    [out0q+lenq], m0
 | |
|     movhps    [out1q+lenq], m1
 | |
|     lea              out0q, [out0q+lenq*2]
 | |
|     lea              out1q, [out1q+lenq*2]
 | |
|     add                inq, mmsize
 | |
|     sub               tmpd, mmsize
 | |
|     jg .inner_loop8
 | |
|     add               outq, 8
 | |
|     add                inq, lenq
 | |
|     sub                 id, 2
 | |
|     jg .loop16
 | |
|     RET
 | |
| 
 | |
| align 16
 | |
| .loop4:
 | |
|     mov              out0q, outq
 | |
|     mov              out1q, 38*64*4
 | |
|     add              out1q, out0q
 | |
|     mov               tmpd, lend
 | |
| 
 | |
| .inner_loop4:
 | |
|     movaps              m0, [inq]
 | |
|     movss          [out0q], m0
 | |
| %if cpuflag(sse4)
 | |
|     extractps      [out1q], m0, 1
 | |
|     extractps [out0q+lenq], m0, 2
 | |
|     extractps [out1q+lenq], m0, 3
 | |
| %else
 | |
|     movhlps             m1, m0
 | |
|     movss     [out0q+lenq], m1
 | |
|     shufps              m0, m0, 0xb1
 | |
|     movss          [out1q], m0
 | |
|     movhlps             m1, m0
 | |
|     movss     [out1q+lenq], m1
 | |
| %endif
 | |
|     lea              out0q, [out0q+lenq*2]
 | |
|     lea              out1q, [out1q+lenq*2]
 | |
|     add                inq, mmsize
 | |
|     sub               tmpd, mmsize
 | |
|     jg .inner_loop4
 | |
|     add               outq, 4
 | |
|     sub                 id, 1
 | |
|     test                id, 2
 | |
|     jne .loop8
 | |
|     cmp                 id, 4
 | |
|     jge .loop16
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse
 | |
| HYBRID_SYNTHESIS_DEINT
 | |
| INIT_XMM sse4
 | |
| HYBRID_SYNTHESIS_DEINT
 | |
| 
 | |
| ;*******************************************************************
 | |
| ;void ff_ps_hybrid_analysis_<opt>(float (*out)[2], float (*in)[2],
 | |
| ;                                 const float (*filter)[8][2],
 | |
| ;                                 ptrdiff_t stride, int n);
 | |
| ;*******************************************************************
 | |
| %macro PS_HYBRID_ANALYSIS_LOOP 3
 | |
|     movu     %1, [inq+mmsize*%3]
 | |
|     movu     m1, [inq+mmsize*(5-%3)+8]
 | |
| %if cpuflag(sse3)
 | |
|     pshufd   %2, %1, q2301
 | |
|     pshufd   m4, m1, q0123
 | |
|     pshufd   m1, m1, q1032
 | |
|     pshufd   m2, [filterq+nq+mmsize*%3], q2301
 | |
|     addsubps %2, m4
 | |
|     addsubps %1, m1
 | |
| %else
 | |
|     mova     m2, [filterq+nq+mmsize*%3]
 | |
|     mova     %2, %1
 | |
|     mova     m4, m1
 | |
|     shufps   %2, %2, q2301
 | |
|     shufps   m4, m4, q0123
 | |
|     shufps   m1, m1, q1032
 | |
|     shufps   m2, m2, q2301
 | |
|     xorps    m4, m7
 | |
|     xorps    m1, m7
 | |
|     subps    %2, m4
 | |
|     subps    %1, m1
 | |
| %endif
 | |
|     mulps    %2, m2
 | |
|     mulps    %1, m2
 | |
| %if %3
 | |
|     addps    m3, %2
 | |
|     addps    m0, %1
 | |
| %endif
 | |
| %endmacro
 | |
| 
 | |
| %macro PS_HYBRID_ANALYSIS 0
 | |
| cglobal ps_hybrid_analysis, 5, 5, 8, out, in, filter, stride, n
 | |
| %if cpuflag(sse3)
 | |
| %define MOVH movsd
 | |
| %else
 | |
| %define MOVH movlps
 | |
| %endif
 | |
|     shl strideq, 3
 | |
|     shl nd, 6
 | |
|     add filterq, nq
 | |
|     neg nq
 | |
|     mova m7, [ps_p1m1p1m1]
 | |
| 
 | |
| align 16
 | |
| .loop:
 | |
|     PS_HYBRID_ANALYSIS_LOOP m0, m3, 0
 | |
|     PS_HYBRID_ANALYSIS_LOOP m5, m6, 1
 | |
|     PS_HYBRID_ANALYSIS_LOOP m5, m6, 2
 | |
| 
 | |
| %if cpuflag(sse3)
 | |
|     pshufd   m3, m3, q2301
 | |
|     xorps    m0, m7
 | |
|     hsubps   m3, m0
 | |
|     pshufd   m1, m3, q0020
 | |
|     pshufd   m3, m3, q0031
 | |
|     addps    m1, m3
 | |
|     movsd    m2, [inq+6*8]
 | |
| %else
 | |
|     mova     m1, m3
 | |
|     mova     m2, m0
 | |
|     shufps   m1, m1, q2301
 | |
|     shufps   m2, m2, q2301
 | |
|     subps    m1, m3
 | |
|     addps    m2, m0
 | |
|     unpcklps m3, m1, m2
 | |
|     unpckhps m1, m2
 | |
|     addps    m1, m3
 | |
|     movu     m2, [inq+6*8] ; faster than movlps and no risk of overread
 | |
| %endif
 | |
|     movss    m3, [filterq+nq+8*6]
 | |
|     SPLATD   m3
 | |
|     mulps    m2, m3
 | |
|     addps    m1, m2
 | |
|     MOVH [outq], m1
 | |
|     add    outq, strideq
 | |
|     add      nq, 64
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse
 | |
| PS_HYBRID_ANALYSIS
 | |
| INIT_XMM sse3
 | |
| PS_HYBRID_ANALYSIS
 | 
