1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525 |
- /*****************************************************************************
- * pixel.c: pixel metrics
- *****************************************************************************
- * Copyright (C) 2003-2018 x264 project
- *
- * Authors: Loren Merritt <lorenm@u.washington.edu>
- * Laurent Aimar <fenrir@via.ecp.fr>
- * Fiona Glaser <fiona@x264.com>
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License as published by
- * the Free Software Foundation; either version 2 of the License, or
- * (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License
- * along with this program; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA.
- *
- * This program is also available under a commercial proprietary license.
- * For more information, contact us at licensing@x264.com.
- *****************************************************************************/
- #include "common.h"
- #if HAVE_MMX
- # include "x86/pixel.h"
- # include "x86/predict.h"
- #endif
- #if ARCH_PPC
- # include "ppc/pixel.h"
- #endif
- #if ARCH_ARM
- # include "arm/pixel.h"
- # include "arm/predict.h"
- #endif
- #if ARCH_AARCH64
- # include "aarch64/pixel.h"
- # include "aarch64/predict.h"
- #endif
- #if ARCH_MIPS
- # include "mips/pixel.h"
- #endif
- /****************************************************************************
- * pixel_sad_WxH
- ****************************************************************************/
- #define PIXEL_SAD_C( name, lx, ly ) \
- static int name( pixel *pix1, intptr_t i_stride_pix1, \
- pixel *pix2, intptr_t i_stride_pix2 ) \
- { \
- int i_sum = 0; \
- for( int y = 0; y < ly; y++ ) \
- { \
- for( int x = 0; x < lx; x++ ) \
- { \
- i_sum += abs( pix1[x] - pix2[x] ); \
- } \
- pix1 += i_stride_pix1; \
- pix2 += i_stride_pix2; \
- } \
- return i_sum; \
- }
- PIXEL_SAD_C( x264_pixel_sad_16x16, 16, 16 )
- PIXEL_SAD_C( x264_pixel_sad_16x8, 16, 8 )
- PIXEL_SAD_C( x264_pixel_sad_8x16, 8, 16 )
- PIXEL_SAD_C( x264_pixel_sad_8x8, 8, 8 )
- PIXEL_SAD_C( x264_pixel_sad_8x4, 8, 4 )
- PIXEL_SAD_C( x264_pixel_sad_4x16, 4, 16 )
- PIXEL_SAD_C( x264_pixel_sad_4x8, 4, 8 )
- PIXEL_SAD_C( x264_pixel_sad_4x4, 4, 4 )
- /****************************************************************************
- * pixel_ssd_WxH
- ****************************************************************************/
- #define PIXEL_SSD_C( name, lx, ly ) \
- static int name( pixel *pix1, intptr_t i_stride_pix1, \
- pixel *pix2, intptr_t i_stride_pix2 ) \
- { \
- int i_sum = 0; \
- for( int y = 0; y < ly; y++ ) \
- { \
- for( int x = 0; x < lx; x++ ) \
- { \
- int d = pix1[x] - pix2[x]; \
- i_sum += d*d; \
- } \
- pix1 += i_stride_pix1; \
- pix2 += i_stride_pix2; \
- } \
- return i_sum; \
- }
- PIXEL_SSD_C( x264_pixel_ssd_16x16, 16, 16 )
- PIXEL_SSD_C( x264_pixel_ssd_16x8, 16, 8 )
- PIXEL_SSD_C( x264_pixel_ssd_8x16, 8, 16 )
- PIXEL_SSD_C( x264_pixel_ssd_8x8, 8, 8 )
- PIXEL_SSD_C( x264_pixel_ssd_8x4, 8, 4 )
- PIXEL_SSD_C( x264_pixel_ssd_4x16, 4, 16 )
- PIXEL_SSD_C( x264_pixel_ssd_4x8, 4, 8 )
- PIXEL_SSD_C( x264_pixel_ssd_4x4, 4, 4 )
- uint64_t x264_pixel_ssd_wxh( x264_pixel_function_t *pf, pixel *pix1, intptr_t i_pix1,
- pixel *pix2, intptr_t i_pix2, int i_width, int i_height )
- {
- uint64_t i_ssd = 0;
- int y;
- int align = !(((intptr_t)pix1 | (intptr_t)pix2 | i_pix1 | i_pix2) & 15);
- #define SSD(size) i_ssd += pf->ssd[size]( pix1 + y*i_pix1 + x, i_pix1, \
- pix2 + y*i_pix2 + x, i_pix2 );
- for( y = 0; y < i_height-15; y += 16 )
- {
- int x = 0;
- if( align )
- for( ; x < i_width-15; x += 16 )
- SSD(PIXEL_16x16);
- for( ; x < i_width-7; x += 8 )
- SSD(PIXEL_8x16);
- }
- if( y < i_height-7 )
- for( int x = 0; x < i_width-7; x += 8 )
- SSD(PIXEL_8x8);
- #undef SSD
- #define SSD1 { int d = pix1[y*i_pix1+x] - pix2[y*i_pix2+x]; i_ssd += d*d; }
- if( i_width & 7 )
- {
- for( y = 0; y < (i_height & ~7); y++ )
- for( int x = i_width & ~7; x < i_width; x++ )
- SSD1;
- }
- if( i_height & 7 )
- {
- for( y = i_height & ~7; y < i_height; y++ )
- for( int x = 0; x < i_width; x++ )
- SSD1;
- }
- #undef SSD1
- return i_ssd;
- }
- static void pixel_ssd_nv12_core( pixel *pixuv1, intptr_t stride1, pixel *pixuv2, intptr_t stride2,
- int width, int height, uint64_t *ssd_u, uint64_t *ssd_v )
- {
- *ssd_u = 0, *ssd_v = 0;
- for( int y = 0; y < height; y++, pixuv1+=stride1, pixuv2+=stride2 )
- for( int x = 0; x < width; x++ )
- {
- int du = pixuv1[2*x] - pixuv2[2*x];
- int dv = pixuv1[2*x+1] - pixuv2[2*x+1];
- *ssd_u += du*du;
- *ssd_v += dv*dv;
- }
- }
- void x264_pixel_ssd_nv12( x264_pixel_function_t *pf, pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2,
- int i_width, int i_height, uint64_t *ssd_u, uint64_t *ssd_v )
- {
- pf->ssd_nv12_core( pix1, i_pix1, pix2, i_pix2, i_width&~7, i_height, ssd_u, ssd_v );
- if( i_width&7 )
- {
- uint64_t tmp[2];
- pixel_ssd_nv12_core( pix1+(i_width&~7), i_pix1, pix2+(i_width&~7), i_pix2, i_width&7, i_height, &tmp[0], &tmp[1] );
- *ssd_u += tmp[0];
- *ssd_v += tmp[1];
- }
- }
- /****************************************************************************
- * pixel_var_wxh
- ****************************************************************************/
- #define PIXEL_VAR_C( name, w, h ) \
- static uint64_t name( pixel *pix, intptr_t i_stride ) \
- { \
- uint32_t sum = 0, sqr = 0; \
- for( int y = 0; y < h; y++ ) \
- { \
- for( int x = 0; x < w; x++ ) \
- { \
- sum += pix[x]; \
- sqr += pix[x] * pix[x]; \
- } \
- pix += i_stride; \
- } \
- return sum + ((uint64_t)sqr << 32); \
- }
- PIXEL_VAR_C( pixel_var_16x16, 16, 16 )
- PIXEL_VAR_C( pixel_var_8x16, 8, 16 )
- PIXEL_VAR_C( pixel_var_8x8, 8, 8 )
- /****************************************************************************
- * pixel_var2_wxh
- ****************************************************************************/
- #define PIXEL_VAR2_C( name, h, shift ) \
- static int name( pixel *fenc, pixel *fdec, int ssd[2] ) \
- { \
- int sum_u = 0, sum_v = 0, sqr_u = 0, sqr_v = 0; \
- for( int y = 0; y < h; y++ ) \
- { \
- for( int x = 0; x < 8; x++ ) \
- { \
- int diff_u = fenc[x] - fdec[x]; \
- int diff_v = fenc[x+FENC_STRIDE/2] - fdec[x+FDEC_STRIDE/2]; \
- sum_u += diff_u; \
- sum_v += diff_v; \
- sqr_u += diff_u * diff_u; \
- sqr_v += diff_v * diff_v; \
- } \
- fenc += FENC_STRIDE; \
- fdec += FDEC_STRIDE; \
- } \
- ssd[0] = sqr_u; \
- ssd[1] = sqr_v; \
- return sqr_u - ((int64_t)sum_u * sum_u >> shift) + \
- sqr_v - ((int64_t)sum_v * sum_v >> shift); \
- }
- PIXEL_VAR2_C( pixel_var2_8x16, 16, 7 )
- PIXEL_VAR2_C( pixel_var2_8x8, 8, 6 )
- #if BIT_DEPTH > 8
- typedef uint32_t sum_t;
- typedef uint64_t sum2_t;
- #else
- typedef uint16_t sum_t;
- typedef uint32_t sum2_t;
- #endif
- #define BITS_PER_SUM (8 * sizeof(sum_t))
- #define HADAMARD4(d0, d1, d2, d3, s0, s1, s2, s3) {\
- sum2_t t0 = s0 + s1;\
- sum2_t t1 = s0 - s1;\
- sum2_t t2 = s2 + s3;\
- sum2_t t3 = s2 - s3;\
- d0 = t0 + t2;\
- d2 = t0 - t2;\
- d1 = t1 + t3;\
- d3 = t1 - t3;\
- }
- // in: a pseudo-simd number of the form x+(y<<16)
- // return: abs(x)+(abs(y)<<16)
- static ALWAYS_INLINE sum2_t abs2( sum2_t a )
- {
- sum2_t s = ((a>>(BITS_PER_SUM-1))&(((sum2_t)1<<BITS_PER_SUM)+1))*((sum_t)-1);
- return (a+s)^s;
- }
- /****************************************************************************
- * pixel_satd_WxH: sum of 4x4 Hadamard transformed differences
- ****************************************************************************/
- static NOINLINE int x264_pixel_satd_4x4( pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2 )
- {
- sum2_t tmp[4][2];
- sum2_t a0, a1, a2, a3, b0, b1;
- sum2_t sum = 0;
- for( int i = 0; i < 4; i++, pix1 += i_pix1, pix2 += i_pix2 )
- {
- a0 = pix1[0] - pix2[0];
- a1 = pix1[1] - pix2[1];
- b0 = (a0+a1) + ((a0-a1)<<BITS_PER_SUM);
- a2 = pix1[2] - pix2[2];
- a3 = pix1[3] - pix2[3];
- b1 = (a2+a3) + ((a2-a3)<<BITS_PER_SUM);
- tmp[i][0] = b0 + b1;
- tmp[i][1] = b0 - b1;
- }
- for( int i = 0; i < 2; i++ )
- {
- HADAMARD4( a0, a1, a2, a3, tmp[0][i], tmp[1][i], tmp[2][i], tmp[3][i] );
- a0 = abs2(a0) + abs2(a1) + abs2(a2) + abs2(a3);
- sum += ((sum_t)a0) + (a0>>BITS_PER_SUM);
- }
- return sum >> 1;
- }
- static NOINLINE int x264_pixel_satd_8x4( pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2 )
- {
- sum2_t tmp[4][4];
- sum2_t a0, a1, a2, a3;
- sum2_t sum = 0;
- for( int i = 0; i < 4; i++, pix1 += i_pix1, pix2 += i_pix2 )
- {
- a0 = (pix1[0] - pix2[0]) + ((sum2_t)(pix1[4] - pix2[4]) << BITS_PER_SUM);
- a1 = (pix1[1] - pix2[1]) + ((sum2_t)(pix1[5] - pix2[5]) << BITS_PER_SUM);
- a2 = (pix1[2] - pix2[2]) + ((sum2_t)(pix1[6] - pix2[6]) << BITS_PER_SUM);
- a3 = (pix1[3] - pix2[3]) + ((sum2_t)(pix1[7] - pix2[7]) << BITS_PER_SUM);
- HADAMARD4( tmp[i][0], tmp[i][1], tmp[i][2], tmp[i][3], a0,a1,a2,a3 );
- }
- for( int i = 0; i < 4; i++ )
- {
- HADAMARD4( a0, a1, a2, a3, tmp[0][i], tmp[1][i], tmp[2][i], tmp[3][i] );
- sum += abs2(a0) + abs2(a1) + abs2(a2) + abs2(a3);
- }
- return (((sum_t)sum) + (sum>>BITS_PER_SUM)) >> 1;
- }
- #define PIXEL_SATD_C( w, h, sub )\
- static int x264_pixel_satd_##w##x##h( pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2 )\
- {\
- int sum = sub( pix1, i_pix1, pix2, i_pix2 )\
- + sub( pix1+4*i_pix1, i_pix1, pix2+4*i_pix2, i_pix2 );\
- if( w==16 )\
- sum+= sub( pix1+8, i_pix1, pix2+8, i_pix2 )\
- + sub( pix1+8+4*i_pix1, i_pix1, pix2+8+4*i_pix2, i_pix2 );\
- if( h==16 )\
- sum+= sub( pix1+8*i_pix1, i_pix1, pix2+8*i_pix2, i_pix2 )\
- + sub( pix1+12*i_pix1, i_pix1, pix2+12*i_pix2, i_pix2 );\
- if( w==16 && h==16 )\
- sum+= sub( pix1+8+8*i_pix1, i_pix1, pix2+8+8*i_pix2, i_pix2 )\
- + sub( pix1+8+12*i_pix1, i_pix1, pix2+8+12*i_pix2, i_pix2 );\
- return sum;\
- }
- PIXEL_SATD_C( 16, 16, x264_pixel_satd_8x4 )
- PIXEL_SATD_C( 16, 8, x264_pixel_satd_8x4 )
- PIXEL_SATD_C( 8, 16, x264_pixel_satd_8x4 )
- PIXEL_SATD_C( 8, 8, x264_pixel_satd_8x4 )
- PIXEL_SATD_C( 4, 16, x264_pixel_satd_4x4 )
- PIXEL_SATD_C( 4, 8, x264_pixel_satd_4x4 )
- static NOINLINE int sa8d_8x8( pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2 )
- {
- sum2_t tmp[8][4];
- sum2_t a0, a1, a2, a3, a4, a5, a6, a7, b0, b1, b2, b3;
- sum2_t sum = 0;
- for( int i = 0; i < 8; i++, pix1 += i_pix1, pix2 += i_pix2 )
- {
- a0 = pix1[0] - pix2[0];
- a1 = pix1[1] - pix2[1];
- b0 = (a0+a1) + ((a0-a1)<<BITS_PER_SUM);
- a2 = pix1[2] - pix2[2];
- a3 = pix1[3] - pix2[3];
- b1 = (a2+a3) + ((a2-a3)<<BITS_PER_SUM);
- a4 = pix1[4] - pix2[4];
- a5 = pix1[5] - pix2[5];
- b2 = (a4+a5) + ((a4-a5)<<BITS_PER_SUM);
- a6 = pix1[6] - pix2[6];
- a7 = pix1[7] - pix2[7];
- b3 = (a6+a7) + ((a6-a7)<<BITS_PER_SUM);
- HADAMARD4( tmp[i][0], tmp[i][1], tmp[i][2], tmp[i][3], b0,b1,b2,b3 );
- }
- for( int i = 0; i < 4; i++ )
- {
- HADAMARD4( a0, a1, a2, a3, tmp[0][i], tmp[1][i], tmp[2][i], tmp[3][i] );
- HADAMARD4( a4, a5, a6, a7, tmp[4][i], tmp[5][i], tmp[6][i], tmp[7][i] );
- b0 = abs2(a0+a4) + abs2(a0-a4);
- b0 += abs2(a1+a5) + abs2(a1-a5);
- b0 += abs2(a2+a6) + abs2(a2-a6);
- b0 += abs2(a3+a7) + abs2(a3-a7);
- sum += (sum_t)b0 + (b0>>BITS_PER_SUM);
- }
- return sum;
- }
- static int x264_pixel_sa8d_8x8( pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2 )
- {
- int sum = sa8d_8x8( pix1, i_pix1, pix2, i_pix2 );
- return (sum+2)>>2;
- }
- static int x264_pixel_sa8d_16x16( pixel *pix1, intptr_t i_pix1, pixel *pix2, intptr_t i_pix2 )
- {
- int sum = sa8d_8x8( pix1, i_pix1, pix2, i_pix2 )
- + sa8d_8x8( pix1+8, i_pix1, pix2+8, i_pix2 )
- + sa8d_8x8( pix1+8*i_pix1, i_pix1, pix2+8*i_pix2, i_pix2 )
- + sa8d_8x8( pix1+8+8*i_pix1, i_pix1, pix2+8+8*i_pix2, i_pix2 );
- return (sum+2)>>2;
- }
- static NOINLINE uint64_t pixel_hadamard_ac( pixel *pix, intptr_t stride )
- {
- sum2_t tmp[32];
- sum2_t a0, a1, a2, a3, dc;
- sum2_t sum4 = 0, sum8 = 0;
- for( int i = 0; i < 8; i++, pix+=stride )
- {
- sum2_t *t = tmp + (i&3) + (i&4)*4;
- a0 = (pix[0]+pix[1]) + ((sum2_t)(pix[0]-pix[1])<<BITS_PER_SUM);
- a1 = (pix[2]+pix[3]) + ((sum2_t)(pix[2]-pix[3])<<BITS_PER_SUM);
- t[0] = a0 + a1;
- t[4] = a0 - a1;
- a2 = (pix[4]+pix[5]) + ((sum2_t)(pix[4]-pix[5])<<BITS_PER_SUM);
- a3 = (pix[6]+pix[7]) + ((sum2_t)(pix[6]-pix[7])<<BITS_PER_SUM);
- t[8] = a2 + a3;
- t[12] = a2 - a3;
- }
- for( int i = 0; i < 8; i++ )
- {
- HADAMARD4( a0, a1, a2, a3, tmp[i*4+0], tmp[i*4+1], tmp[i*4+2], tmp[i*4+3] );
- tmp[i*4+0] = a0;
- tmp[i*4+1] = a1;
- tmp[i*4+2] = a2;
- tmp[i*4+3] = a3;
- sum4 += abs2(a0) + abs2(a1) + abs2(a2) + abs2(a3);
- }
- for( int i = 0; i < 8; i++ )
- {
- HADAMARD4( a0,a1,a2,a3, tmp[i], tmp[8+i], tmp[16+i], tmp[24+i] );
- sum8 += abs2(a0) + abs2(a1) + abs2(a2) + abs2(a3);
- }
- dc = (sum_t)(tmp[0] + tmp[8] + tmp[16] + tmp[24]);
- sum4 = (sum_t)sum4 + (sum4>>BITS_PER_SUM) - dc;
- sum8 = (sum_t)sum8 + (sum8>>BITS_PER_SUM) - dc;
- return ((uint64_t)sum8<<32) + sum4;
- }
- #define HADAMARD_AC(w,h) \
- static uint64_t x264_pixel_hadamard_ac_##w##x##h( pixel *pix, intptr_t stride )\
- {\
- uint64_t sum = pixel_hadamard_ac( pix, stride );\
- if( w==16 )\
- sum += pixel_hadamard_ac( pix+8, stride );\
- if( h==16 )\
- sum += pixel_hadamard_ac( pix+8*stride, stride );\
- if( w==16 && h==16 )\
- sum += pixel_hadamard_ac( pix+8*stride+8, stride );\
- return ((sum>>34)<<32) + ((uint32_t)sum>>1);\
- }
- HADAMARD_AC( 16, 16 )
- HADAMARD_AC( 16, 8 )
- HADAMARD_AC( 8, 16 )
- HADAMARD_AC( 8, 8 )
- /****************************************************************************
- * pixel_sad_x4
- ****************************************************************************/
- #define SAD_X( size ) \
- static void x264_pixel_sad_x3_##size( pixel *fenc, pixel *pix0, pixel *pix1, pixel *pix2,\
- intptr_t i_stride, int scores[3] )\
- {\
- scores[0] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix0, i_stride );\
- scores[1] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix1, i_stride );\
- scores[2] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix2, i_stride );\
- }\
- static void x264_pixel_sad_x4_##size( pixel *fenc, pixel *pix0, pixel *pix1,pixel *pix2, pixel *pix3,\
- intptr_t i_stride, int scores[4] )\
- {\
- scores[0] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix0, i_stride );\
- scores[1] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix1, i_stride );\
- scores[2] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix2, i_stride );\
- scores[3] = x264_pixel_sad_##size( fenc, FENC_STRIDE, pix3, i_stride );\
- }
- SAD_X( 16x16 )
- SAD_X( 16x8 )
- SAD_X( 8x16 )
- SAD_X( 8x8 )
- SAD_X( 8x4 )
- SAD_X( 4x8 )
- SAD_X( 4x4 )
- /****************************************************************************
- * pixel_satd_x4
- * no faster than single satd, but needed for satd to be a drop-in replacement for sad
- ****************************************************************************/
- #define SATD_X( size, cpu ) \
- static void x264_pixel_satd_x3_##size##cpu( pixel *fenc, pixel *pix0, pixel *pix1, pixel *pix2,\
- intptr_t i_stride, int scores[3] )\
- {\
- scores[0] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix0, i_stride );\
- scores[1] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix1, i_stride );\
- scores[2] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix2, i_stride );\
- }\
- static void x264_pixel_satd_x4_##size##cpu( pixel *fenc, pixel *pix0, pixel *pix1, pixel *pix2, pixel *pix3,\
- intptr_t i_stride, int scores[4] )\
- {\
- scores[0] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix0, i_stride );\
- scores[1] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix1, i_stride );\
- scores[2] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix2, i_stride );\
- scores[3] = x264_pixel_satd_##size##cpu( fenc, FENC_STRIDE, pix3, i_stride );\
- }
- #define SATD_X_DECL6( cpu )\
- SATD_X( 16x16, cpu )\
- SATD_X( 16x8, cpu )\
- SATD_X( 8x16, cpu )\
- SATD_X( 8x8, cpu )\
- SATD_X( 8x4, cpu )\
- SATD_X( 4x8, cpu )
- #define SATD_X_DECL7( cpu )\
- SATD_X_DECL6( cpu )\
- SATD_X( 4x4, cpu )
- SATD_X_DECL7()
- #if HAVE_MMX
- SATD_X_DECL7( _mmx2 )
- #if !HIGH_BIT_DEPTH
- SATD_X_DECL6( _sse2 )
- SATD_X_DECL7( _ssse3 )
- SATD_X_DECL6( _ssse3_atom )
- SATD_X_DECL7( _sse4 )
- SATD_X_DECL7( _avx )
- SATD_X_DECL7( _xop )
- SATD_X_DECL7( _avx512 )
- #endif // !HIGH_BIT_DEPTH
- #endif
- #if !HIGH_BIT_DEPTH
- #if HAVE_ARMV6 || ARCH_AARCH64
- SATD_X_DECL7( _neon )
- #endif
- #endif // !HIGH_BIT_DEPTH
- #define INTRA_MBCMP_8x8( mbcmp, cpu, cpu2 )\
- static void intra_##mbcmp##_x3_8x8##cpu( pixel *fenc, pixel edge[36], int res[3] )\
- {\
- ALIGNED_ARRAY_16( pixel, pix, [8*FDEC_STRIDE] );\
- x264_predict_8x8_v##cpu2( pix, edge );\
- res[0] = x264_pixel_##mbcmp##_8x8##cpu( pix, FDEC_STRIDE, fenc, FENC_STRIDE );\
- x264_predict_8x8_h##cpu2( pix, edge );\
- res[1] = x264_pixel_##mbcmp##_8x8##cpu( pix, FDEC_STRIDE, fenc, FENC_STRIDE );\
- x264_predict_8x8_dc##cpu2( pix, edge );\
- res[2] = x264_pixel_##mbcmp##_8x8##cpu( pix, FDEC_STRIDE, fenc, FENC_STRIDE );\
- }
- INTRA_MBCMP_8x8( sad,, _c )
- INTRA_MBCMP_8x8(sa8d,, _c )
- #if HIGH_BIT_DEPTH && HAVE_MMX
- #define x264_predict_8x8_v_sse2 x264_predict_8x8_v_sse
- INTRA_MBCMP_8x8( sad, _mmx2, _c )
- INTRA_MBCMP_8x8(sa8d, _sse2, _sse2 )
- #endif
- #if !HIGH_BIT_DEPTH && (HAVE_ARMV6 || ARCH_AARCH64)
- INTRA_MBCMP_8x8( sad, _neon, _neon )
- INTRA_MBCMP_8x8(sa8d, _neon, _neon )
- #endif
- #define INTRA_MBCMP( mbcmp, size, pred1, pred2, pred3, chroma, cpu, cpu2 )\
- static void intra_##mbcmp##_x3_##size##chroma##cpu( pixel *fenc, pixel *fdec, int res[3] )\
- {\
- x264_predict_##size##chroma##_##pred1##cpu2( fdec );\
- res[0] = x264_pixel_##mbcmp##_##size##cpu( fdec, FDEC_STRIDE, fenc, FENC_STRIDE );\
- x264_predict_##size##chroma##_##pred2##cpu2( fdec );\
- res[1] = x264_pixel_##mbcmp##_##size##cpu( fdec, FDEC_STRIDE, fenc, FENC_STRIDE );\
- x264_predict_##size##chroma##_##pred3##cpu2( fdec );\
- res[2] = x264_pixel_##mbcmp##_##size##cpu( fdec, FDEC_STRIDE, fenc, FENC_STRIDE );\
- }
- INTRA_MBCMP( sad, 4x4, v, h, dc, ,, _c )
- INTRA_MBCMP(satd, 4x4, v, h, dc, ,, _c )
- INTRA_MBCMP( sad, 8x8, dc, h, v, c,, _c )
- INTRA_MBCMP(satd, 8x8, dc, h, v, c,, _c )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c,, _c )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c,, _c )
- INTRA_MBCMP( sad, 16x16, v, h, dc, ,, _c )
- INTRA_MBCMP(satd, 16x16, v, h, dc, ,, _c )
- #if HAVE_MMX
- #if HIGH_BIT_DEPTH
- #define x264_predict_8x8c_v_mmx2 x264_predict_8x8c_v_mmx
- #define x264_predict_8x16c_v_mmx2 x264_predict_8x16c_v_c
- #define x264_predict_16x16_dc_mmx2 x264_predict_16x16_dc_c
- #define x264_predict_8x8c_v_sse2 x264_predict_8x8c_v_sse
- #define x264_predict_8x16c_v_sse2 x264_predict_8x16c_v_sse
- #define x264_predict_16x16_v_sse2 x264_predict_16x16_v_sse
- INTRA_MBCMP( sad, 4x4, v, h, dc, , _mmx2, _c )
- INTRA_MBCMP( sad, 8x8, dc, h, v, c, _mmx2, _mmx2 )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _mmx2, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _mmx2, _mmx2 )
- INTRA_MBCMP( sad, 16x16, v, h, dc, , _mmx2, _mmx2 )
- INTRA_MBCMP( sad, 8x8, dc, h, v, c, _sse2, _sse2 )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _sse2, _sse2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _sse2, _sse2 )
- INTRA_MBCMP( sad, 16x16, v, h, dc, , _sse2, _sse2 )
- INTRA_MBCMP( sad, 8x8, dc, h, v, c, _ssse3, _sse2 )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _ssse3, _sse2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _ssse3, _sse2 )
- INTRA_MBCMP( sad, 16x16, v, h, dc, , _ssse3, _sse2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _sse4, _sse2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _avx, _sse2 )
- #else
- #define x264_predict_8x16c_v_mmx2 x264_predict_8x16c_v_mmx
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _mmx2, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _mmx2, _mmx2 )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _sse2, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _sse2, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _ssse3, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _sse4, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _avx, _mmx2 )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _xop, _mmx2 )
- #endif
- #endif
- #if !HIGH_BIT_DEPTH && HAVE_ARMV6
- INTRA_MBCMP( sad, 4x4, v, h, dc, , _neon, _armv6 )
- INTRA_MBCMP(satd, 4x4, v, h, dc, , _neon, _armv6 )
- INTRA_MBCMP( sad, 8x8, dc, h, v, c, _neon, _neon )
- INTRA_MBCMP(satd, 8x8, dc, h, v, c, _neon, _neon )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _neon, _c )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _neon, _c )
- INTRA_MBCMP( sad, 16x16, v, h, dc, , _neon, _neon )
- INTRA_MBCMP(satd, 16x16, v, h, dc, , _neon, _neon )
- #endif
- #if !HIGH_BIT_DEPTH && ARCH_AARCH64
- INTRA_MBCMP( sad, 4x4, v, h, dc, , _neon, _neon )
- INTRA_MBCMP(satd, 4x4, v, h, dc, , _neon, _neon )
- INTRA_MBCMP( sad, 8x8, dc, h, v, c, _neon, _neon )
- INTRA_MBCMP(satd, 8x8, dc, h, v, c, _neon, _neon )
- INTRA_MBCMP( sad, 8x16, dc, h, v, c, _neon, _neon )
- INTRA_MBCMP(satd, 8x16, dc, h, v, c, _neon, _neon )
- INTRA_MBCMP( sad, 16x16, v, h, dc, , _neon, _neon )
- INTRA_MBCMP(satd, 16x16, v, h, dc, , _neon, _neon )
- #endif
- // No C implementation of intra_satd_x9. See checkasm for its behavior,
- // or see mb_analyse_intra for the entirely different algorithm we
- // use when lacking an asm implementation of it.
- /****************************************************************************
- * structural similarity metric
- ****************************************************************************/
- static void ssim_4x4x2_core( const pixel *pix1, intptr_t stride1,
- const pixel *pix2, intptr_t stride2,
- int sums[2][4] )
- {
- for( int z = 0; z < 2; z++ )
- {
- uint32_t s1 = 0, s2 = 0, ss = 0, s12 = 0;
- for( int y = 0; y < 4; y++ )
- for( int x = 0; x < 4; x++ )
- {
- int a = pix1[x+y*stride1];
- int b = pix2[x+y*stride2];
- s1 += a;
- s2 += b;
- ss += a*a;
- ss += b*b;
- s12 += a*b;
- }
- sums[z][0] = s1;
- sums[z][1] = s2;
- sums[z][2] = ss;
- sums[z][3] = s12;
- pix1 += 4;
- pix2 += 4;
- }
- }
- static float ssim_end1( int s1, int s2, int ss, int s12 )
- {
- /* Maximum value for 10-bit is: ss*64 = (2^10-1)^2*16*4*64 = 4286582784, which will overflow in some cases.
- * s1*s1, s2*s2, and s1*s2 also obtain this value for edge cases: ((2^10-1)*16*4)^2 = 4286582784.
- * Maximum value for 9-bit is: ss*64 = (2^9-1)^2*16*4*64 = 1069551616, which will not overflow. */
- #if BIT_DEPTH > 9
- #define type float
- static const float ssim_c1 = .01*.01*PIXEL_MAX*PIXEL_MAX*64;
- static const float ssim_c2 = .03*.03*PIXEL_MAX*PIXEL_MAX*64*63;
- #else
- #define type int
- static const int ssim_c1 = (int)(.01*.01*PIXEL_MAX*PIXEL_MAX*64 + .5);
- static const int ssim_c2 = (int)(.03*.03*PIXEL_MAX*PIXEL_MAX*64*63 + .5);
- #endif
- type fs1 = s1;
- type fs2 = s2;
- type fss = ss;
- type fs12 = s12;
- type vars = fss*64 - fs1*fs1 - fs2*fs2;
- type covar = fs12*64 - fs1*fs2;
- return (float)(2*fs1*fs2 + ssim_c1) * (float)(2*covar + ssim_c2)
- / ((float)(fs1*fs1 + fs2*fs2 + ssim_c1) * (float)(vars + ssim_c2));
- #undef type
- }
- static float ssim_end4( int sum0[5][4], int sum1[5][4], int width )
- {
- float ssim = 0.0;
- for( int i = 0; i < width; i++ )
- ssim += ssim_end1( sum0[i][0] + sum0[i+1][0] + sum1[i][0] + sum1[i+1][0],
- sum0[i][1] + sum0[i+1][1] + sum1[i][1] + sum1[i+1][1],
- sum0[i][2] + sum0[i+1][2] + sum1[i][2] + sum1[i+1][2],
- sum0[i][3] + sum0[i+1][3] + sum1[i][3] + sum1[i+1][3] );
- return ssim;
- }
- float x264_pixel_ssim_wxh( x264_pixel_function_t *pf,
- pixel *pix1, intptr_t stride1,
- pixel *pix2, intptr_t stride2,
- int width, int height, void *buf, int *cnt )
- {
- int z = 0;
- float ssim = 0.0;
- int (*sum0)[4] = buf;
- int (*sum1)[4] = sum0 + (width >> 2) + 3;
- width >>= 2;
- height >>= 2;
- for( int y = 1; y < height; y++ )
- {
- for( ; z <= y; z++ )
- {
- XCHG( void*, sum0, sum1 );
- for( int x = 0; x < width; x+=2 )
- pf->ssim_4x4x2_core( &pix1[4*(x+z*stride1)], stride1, &pix2[4*(x+z*stride2)], stride2, &sum0[x] );
- }
- for( int x = 0; x < width-1; x += 4 )
- ssim += pf->ssim_end4( sum0+x, sum1+x, X264_MIN(4,width-x-1) );
- }
- *cnt = (height-1) * (width-1);
- return ssim;
- }
- static int pixel_vsad( pixel *src, intptr_t stride, int height )
- {
- int score = 0;
- for( int i = 1; i < height; i++, src += stride )
- for( int j = 0; j < 16; j++ )
- score += abs(src[j] - src[j+stride]);
- return score;
- }
- int x264_field_vsad( x264_t *h, int mb_x, int mb_y )
- {
- int score_field, score_frame;
- int stride = h->fenc->i_stride[0];
- int mb_stride = h->mb.i_mb_stride;
- pixel *fenc = h->fenc->plane[0] + 16 * (mb_x + mb_y * stride);
- int mb_xy = mb_x + mb_y*mb_stride;
- /* We don't want to analyze pixels outside the frame, as it gives inaccurate results. */
- int mbpair_height = X264_MIN( h->param.i_height - mb_y * 16, 32 );
- score_frame = h->pixf.vsad( fenc, stride, mbpair_height );
- score_field = h->pixf.vsad( fenc, stride*2, mbpair_height >> 1 );
- score_field += h->pixf.vsad( fenc+stride, stride*2, mbpair_height >> 1 );
- if( mb_x > 0 )
- score_field += 512 - h->mb.field[mb_xy -1]*1024;
- if( mb_y > 0 )
- score_field += 512 - h->mb.field[mb_xy-mb_stride]*1024;
- return (score_field < score_frame);
- }
- static int pixel_asd8( pixel *pix1, intptr_t stride1, pixel *pix2, intptr_t stride2, int height )
- {
- int sum = 0;
- for( int y = 0; y < height; y++, pix1 += stride1, pix2 += stride2 )
- for( int x = 0; x < 8; x++ )
- sum += pix1[x] - pix2[x];
- return abs( sum );
- }
- /****************************************************************************
- * successive elimination
- ****************************************************************************/
- static int x264_pixel_ads4( int enc_dc[4], uint16_t *sums, int delta,
- uint16_t *cost_mvx, int16_t *mvs, int width, int thresh )
- {
- int nmv = 0;
- for( int i = 0; i < width; i++, sums++ )
- {
- int ads = abs( enc_dc[0] - sums[0] )
- + abs( enc_dc[1] - sums[8] )
- + abs( enc_dc[2] - sums[delta] )
- + abs( enc_dc[3] - sums[delta+8] )
- + cost_mvx[i];
- if( ads < thresh )
- mvs[nmv++] = i;
- }
- return nmv;
- }
- static int x264_pixel_ads2( int enc_dc[2], uint16_t *sums, int delta,
- uint16_t *cost_mvx, int16_t *mvs, int width, int thresh )
- {
- int nmv = 0;
- for( int i = 0; i < width; i++, sums++ )
- {
- int ads = abs( enc_dc[0] - sums[0] )
- + abs( enc_dc[1] - sums[delta] )
- + cost_mvx[i];
- if( ads < thresh )
- mvs[nmv++] = i;
- }
- return nmv;
- }
- static int x264_pixel_ads1( int enc_dc[1], uint16_t *sums, int delta,
- uint16_t *cost_mvx, int16_t *mvs, int width, int thresh )
- {
- int nmv = 0;
- for( int i = 0; i<width; i++, sums++ )
- {
- int ads = abs( enc_dc[0] - sums[0] )
- + cost_mvx[i];
- if( ads < thresh )
- mvs[nmv++] = i;
- }
- return nmv;
- }
- /****************************************************************************
- * x264_pixel_init:
- ****************************************************************************/
- void x264_pixel_init( int cpu, x264_pixel_function_t *pixf )
- {
- memset( pixf, 0, sizeof(*pixf) );
- #define INIT2_NAME( name1, name2, cpu ) \
- pixf->name1[PIXEL_16x16] = x264_pixel_##name2##_16x16##cpu;\
- pixf->name1[PIXEL_16x8] = x264_pixel_##name2##_16x8##cpu;
- #define INIT4_NAME( name1, name2, cpu ) \
- INIT2_NAME( name1, name2, cpu ) \
- pixf->name1[PIXEL_8x16] = x264_pixel_##name2##_8x16##cpu;\
- pixf->name1[PIXEL_8x8] = x264_pixel_##name2##_8x8##cpu;
- #define INIT5_NAME( name1, name2, cpu ) \
- INIT4_NAME( name1, name2, cpu ) \
- pixf->name1[PIXEL_8x4] = x264_pixel_##name2##_8x4##cpu;
- #define INIT6_NAME( name1, name2, cpu ) \
- INIT5_NAME( name1, name2, cpu ) \
- pixf->name1[PIXEL_4x8] = x264_pixel_##name2##_4x8##cpu;
- #define INIT7_NAME( name1, name2, cpu ) \
- INIT6_NAME( name1, name2, cpu ) \
- pixf->name1[PIXEL_4x4] = x264_pixel_##name2##_4x4##cpu;
- #define INIT8_NAME( name1, name2, cpu ) \
- INIT7_NAME( name1, name2, cpu ) \
- pixf->name1[PIXEL_4x16] = x264_pixel_##name2##_4x16##cpu;
- #define INIT2( name, cpu ) INIT2_NAME( name, name, cpu )
- #define INIT4( name, cpu ) INIT4_NAME( name, name, cpu )
- #define INIT5( name, cpu ) INIT5_NAME( name, name, cpu )
- #define INIT6( name, cpu ) INIT6_NAME( name, name, cpu )
- #define INIT7( name, cpu ) INIT7_NAME( name, name, cpu )
- #define INIT8( name, cpu ) INIT8_NAME( name, name, cpu )
- #define INIT_ADS( cpu ) \
- pixf->ads[PIXEL_16x16] = x264_pixel_ads4##cpu;\
- pixf->ads[PIXEL_16x8] = x264_pixel_ads2##cpu;\
- pixf->ads[PIXEL_8x8] = x264_pixel_ads1##cpu;
- INIT8( sad, );
- INIT8_NAME( sad_aligned, sad, );
- INIT7( sad_x3, );
- INIT7( sad_x4, );
- INIT8( ssd, );
- INIT8( satd, );
- INIT7( satd_x3, );
- INIT7( satd_x4, );
- INIT4( hadamard_ac, );
- INIT_ADS( );
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8;
- pixf->var[PIXEL_16x16] = pixel_var_16x16;
- pixf->var[PIXEL_8x16] = pixel_var_8x16;
- pixf->var[PIXEL_8x8] = pixel_var_8x8;
- pixf->var2[PIXEL_8x16] = pixel_var2_8x16;
- pixf->var2[PIXEL_8x8] = pixel_var2_8x8;
- pixf->ssd_nv12_core = pixel_ssd_nv12_core;
- pixf->ssim_4x4x2_core = ssim_4x4x2_core;
- pixf->ssim_end4 = ssim_end4;
- pixf->vsad = pixel_vsad;
- pixf->asd8 = pixel_asd8;
- pixf->intra_sad_x3_4x4 = intra_sad_x3_4x4;
- pixf->intra_satd_x3_4x4 = intra_satd_x3_4x4;
- pixf->intra_sad_x3_8x8 = intra_sad_x3_8x8;
- pixf->intra_sa8d_x3_8x8 = intra_sa8d_x3_8x8;
- pixf->intra_sad_x3_8x8c = intra_sad_x3_8x8c;
- pixf->intra_satd_x3_8x8c = intra_satd_x3_8x8c;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c;
- pixf->intra_sad_x3_16x16 = intra_sad_x3_16x16;
- pixf->intra_satd_x3_16x16 = intra_satd_x3_16x16;
- #if HIGH_BIT_DEPTH
- #if HAVE_MMX
- if( cpu&X264_CPU_MMX2 )
- {
- INIT7( sad, _mmx2 );
- INIT7_NAME( sad_aligned, sad, _mmx2 );
- INIT7( sad_x3, _mmx2 );
- INIT7( sad_x4, _mmx2 );
- INIT8( satd, _mmx2 );
- INIT7( satd_x3, _mmx2 );
- INIT7( satd_x4, _mmx2 );
- INIT4( hadamard_ac, _mmx2 );
- INIT8( ssd, _mmx2 );
- INIT_ADS( _mmx2 );
- pixf->intra_sad_x3_4x4 = intra_sad_x3_4x4_mmx2;
- pixf->intra_satd_x3_4x4 = x264_intra_satd_x3_4x4_mmx2;
- pixf->intra_sad_x3_8x8 = intra_sad_x3_8x8_mmx2;
- pixf->intra_sad_x3_8x8c = intra_sad_x3_8x8c_mmx2;
- pixf->intra_satd_x3_8x8c = x264_intra_satd_x3_8x8c_mmx2;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_mmx2;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_mmx2;
- pixf->intra_sad_x3_16x16 = intra_sad_x3_16x16_mmx2;
- pixf->intra_satd_x3_16x16 = x264_intra_satd_x3_16x16_mmx2;
- }
- if( cpu&X264_CPU_SSE2 )
- {
- INIT4_NAME( sad_aligned, sad, _sse2_aligned );
- INIT5( ssd, _sse2 );
- INIT6( satd, _sse2 );
- pixf->satd[PIXEL_4x16] = x264_pixel_satd_4x16_sse2;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_sse2;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_sse2;
- pixf->ssim_end4 = x264_pixel_ssim_end4_sse2;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_sse2;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_sse2;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_sse2;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_sse2;
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16_sse2;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_sse2;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_sse2;
- #endif
- pixf->intra_sad_x3_4x4 = x264_intra_sad_x3_4x4_sse2;
- pixf->intra_sad_x3_8x8 = x264_intra_sad_x3_8x8_sse2;
- pixf->intra_sa8d_x3_8x8 = intra_sa8d_x3_8x8_sse2;
- }
- if( (cpu&X264_CPU_SSE2) && !(cpu&X264_CPU_SSE2_IS_SLOW) )
- {
- INIT5( sad, _sse2 );
- INIT2( sad_x3, _sse2 );
- INIT2( sad_x4, _sse2 );
- INIT_ADS( _sse2 );
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- INIT4( hadamard_ac, _sse2 );
- }
- pixf->vsad = x264_pixel_vsad_sse2;
- pixf->asd8 = x264_pixel_asd8_sse2;
- pixf->intra_sad_x3_8x8 = x264_intra_sad_x3_8x8_sse2;
- pixf->intra_sad_x3_8x8c = intra_sad_x3_8x8c_sse2;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_sse2;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_sse2;
- pixf->intra_sad_x3_16x16 = intra_sad_x3_16x16_sse2;
- }
- if( cpu&X264_CPU_SSE2_IS_FAST )
- {
- pixf->sad[PIXEL_8x16] = x264_pixel_sad_8x16_sse2;
- pixf->sad_x3[PIXEL_8x16] = x264_pixel_sad_x3_8x16_sse2;
- pixf->sad_x3[PIXEL_8x8] = x264_pixel_sad_x3_8x8_sse2;
- pixf->sad_x3[PIXEL_8x4] = x264_pixel_sad_x3_8x4_sse2;
- pixf->sad_x4[PIXEL_8x16] = x264_pixel_sad_x4_8x16_sse2;
- pixf->sad_x4[PIXEL_8x8] = x264_pixel_sad_x4_8x8_sse2;
- pixf->sad_x4[PIXEL_8x4] = x264_pixel_sad_x4_8x4_sse2;
- }
- if( cpu&X264_CPU_SSSE3 )
- {
- INIT4_NAME( sad_aligned, sad, _ssse3_aligned );
- pixf->sad_aligned[PIXEL_4x4] = x264_pixel_sad_4x4_ssse3;
- pixf->sad_aligned[PIXEL_4x8] = x264_pixel_sad_4x8_ssse3;
- INIT7( sad, _ssse3 );
- INIT7( sad_x3, _ssse3 );
- INIT7( sad_x4, _ssse3 );
- #if ARCH_X86 || !defined( __MACH__ )
- INIT_ADS( _ssse3 );
- #endif
- INIT6( satd, _ssse3 );
- pixf->satd[PIXEL_4x16] = x264_pixel_satd_4x16_ssse3;
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- INIT4( hadamard_ac, _ssse3 );
- }
- pixf->vsad = x264_pixel_vsad_ssse3;
- pixf->asd8 = x264_pixel_asd8_ssse3;
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_ssse3;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_ssse3;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_ssse3;
- #endif
- pixf->intra_sad_x3_4x4 = x264_intra_sad_x3_4x4_ssse3;
- pixf->intra_sad_x3_8x8 = x264_intra_sad_x3_8x8_ssse3;
- pixf->intra_sad_x3_8x8c = intra_sad_x3_8x8c_ssse3;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_ssse3;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_ssse3;
- pixf->intra_sad_x3_16x16 = intra_sad_x3_16x16_ssse3;
- }
- if( cpu&X264_CPU_SSE4 )
- {
- INIT6( satd, _sse4 );
- pixf->satd[PIXEL_4x16] = x264_pixel_satd_4x16_sse4;
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- INIT4( hadamard_ac, _sse4 );
- }
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_sse4;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_sse4;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_sse4;
- #endif
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_sse4;
- }
- if( cpu&X264_CPU_AVX )
- {
- INIT5_NAME( sad_aligned, sad, _ssse3 ); /* AVX-capable CPUs doesn't benefit from an aligned version */
- #if ARCH_X86 || !defined( __MACH__ )
- INIT_ADS( _avx );
- #endif
- INIT6( satd, _avx );
- pixf->satd[PIXEL_4x16] = x264_pixel_satd_4x16_avx;
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- INIT4( hadamard_ac, _avx );
- }
- pixf->intra_sad_x3_4x4 = x264_intra_sad_x3_4x4_avx;
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_avx;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_avx;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_avx;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_avx;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_avx;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_avx;
- pixf->ssim_end4 = x264_pixel_ssim_end4_avx;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_avx;
- #endif
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_avx;
- }
- if( cpu&X264_CPU_XOP )
- {
- INIT5( sad_x3, _xop );
- INIT5( sad_x4, _xop );
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_xop;
- pixf->vsad = x264_pixel_vsad_xop;
- pixf->asd8 = x264_pixel_asd8_xop;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_xop;
- #endif
- }
- if( cpu&X264_CPU_AVX2 )
- {
- INIT2( ssd, _avx2 );
- INIT2( sad, _avx2 );
- INIT2_NAME( sad_aligned, sad, _avx2 );
- INIT2( sad_x3, _avx2 );
- INIT2( sad_x4, _avx2 );
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_avx2;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_avx2;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_avx2;
- pixf->vsad = x264_pixel_vsad_avx2;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_avx2;
- pixf->intra_sad_x3_8x8 = x264_intra_sad_x3_8x8_avx2;
- }
- if( cpu&X264_CPU_AVX512 )
- {
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_avx512;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_avx512;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_avx512;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_avx512;
- }
- #endif // HAVE_MMX
- #else // !HIGH_BIT_DEPTH
- #if HAVE_MMX
- if( cpu&X264_CPU_MMX )
- {
- INIT8( ssd, _mmx );
- }
- if( cpu&X264_CPU_MMX2 )
- {
- INIT8( sad, _mmx2 );
- INIT8_NAME( sad_aligned, sad, _mmx2 );
- INIT7( sad_x3, _mmx2 );
- INIT7( sad_x4, _mmx2 );
- INIT8( satd, _mmx2 );
- INIT7( satd_x3, _mmx2 );
- INIT7( satd_x4, _mmx2 );
- INIT4( hadamard_ac, _mmx2 );
- INIT_ADS( _mmx2 );
- #if ARCH_X86
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16_mmx2;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_mmx2;
- pixf->intra_sa8d_x3_8x8 = x264_intra_sa8d_x3_8x8_mmx2;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_mmx2;
- pixf->vsad = x264_pixel_vsad_mmx2;
- if( cpu&X264_CPU_CACHELINE_32 )
- {
- INIT5( sad, _cache32_mmx2 );
- INIT4( sad_x3, _cache32_mmx2 );
- INIT4( sad_x4, _cache32_mmx2 );
- }
- else if( cpu&X264_CPU_CACHELINE_64 && !(cpu&X264_CPU_SLOW_ATOM) )
- {
- INIT5( sad, _cache64_mmx2 );
- INIT4( sad_x3, _cache64_mmx2 );
- INIT4( sad_x4, _cache64_mmx2 );
- }
- #else
- if( cpu&X264_CPU_CACHELINE_64 && !(cpu&X264_CPU_SLOW_ATOM) )
- {
- pixf->sad[PIXEL_8x16] = x264_pixel_sad_8x16_cache64_mmx2;
- pixf->sad[PIXEL_8x8] = x264_pixel_sad_8x8_cache64_mmx2;
- pixf->sad[PIXEL_8x4] = x264_pixel_sad_8x4_cache64_mmx2;
- pixf->sad_x3[PIXEL_8x16] = x264_pixel_sad_x3_8x16_cache64_mmx2;
- pixf->sad_x3[PIXEL_8x8] = x264_pixel_sad_x3_8x8_cache64_mmx2;
- pixf->sad_x4[PIXEL_8x16] = x264_pixel_sad_x4_8x16_cache64_mmx2;
- pixf->sad_x4[PIXEL_8x8] = x264_pixel_sad_x4_8x8_cache64_mmx2;
- }
- #endif
- pixf->intra_satd_x3_16x16 = x264_intra_satd_x3_16x16_mmx2;
- pixf->intra_sad_x3_16x16 = x264_intra_sad_x3_16x16_mmx2;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_mmx2;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_mmx2;
- pixf->intra_satd_x3_8x8c = x264_intra_satd_x3_8x8c_mmx2;
- pixf->intra_sad_x3_8x8c = x264_intra_sad_x3_8x8c_mmx2;
- pixf->intra_sad_x3_8x8 = x264_intra_sad_x3_8x8_mmx2;
- pixf->intra_satd_x3_4x4 = x264_intra_satd_x3_4x4_mmx2;
- pixf->intra_sad_x3_4x4 = x264_intra_sad_x3_4x4_mmx2;
- }
- if( cpu&X264_CPU_SSE2 )
- {
- INIT5( ssd, _sse2slow );
- INIT2_NAME( sad_aligned, sad, _sse2_aligned );
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_sse2;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_sse2;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_sse2;
- pixf->ssim_end4 = x264_pixel_ssim_end4_sse2;
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16_sse2;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_sse2;
- #if ARCH_X86_64
- pixf->intra_sa8d_x3_8x8 = x264_intra_sa8d_x3_8x8_sse2;
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_sse2;
- #endif
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_sse2;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_sse2;
- pixf->vsad = x264_pixel_vsad_sse2;
- pixf->asd8 = x264_pixel_asd8_sse2;
- }
- if( (cpu&X264_CPU_SSE2) && !(cpu&X264_CPU_SSE2_IS_SLOW) )
- {
- INIT2( sad, _sse2 );
- INIT2( sad_x3, _sse2 );
- INIT2( sad_x4, _sse2 );
- INIT6( satd, _sse2 );
- pixf->satd[PIXEL_4x16] = x264_pixel_satd_4x16_sse2;
- INIT6( satd_x3, _sse2 );
- INIT6( satd_x4, _sse2 );
- INIT4( hadamard_ac, _sse2 );
- INIT_ADS( _sse2 );
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_sse2;
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_sse2;
- pixf->intra_sad_x3_16x16 = x264_intra_sad_x3_16x16_sse2;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_sse2;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_sse2;
- if( cpu&X264_CPU_CACHELINE_64 )
- {
- INIT2( ssd, _sse2); /* faster for width 16 on p4 */
- #if ARCH_X86
- INIT2( sad, _cache64_sse2 );
- INIT2( sad_x3, _cache64_sse2 );
- INIT2( sad_x4, _cache64_sse2 );
- #endif
- if( cpu&X264_CPU_SSE2_IS_FAST )
- {
- pixf->sad_x3[PIXEL_8x16] = x264_pixel_sad_x3_8x16_cache64_sse2;
- pixf->sad_x4[PIXEL_8x16] = x264_pixel_sad_x4_8x16_cache64_sse2;
- }
- }
- }
- if( cpu&X264_CPU_SSE2_IS_FAST && !(cpu&X264_CPU_CACHELINE_64) )
- {
- pixf->sad_aligned[PIXEL_8x16] = x264_pixel_sad_8x16_sse2;
- pixf->sad[PIXEL_8x16] = x264_pixel_sad_8x16_sse2;
- pixf->sad_x3[PIXEL_8x16] = x264_pixel_sad_x3_8x16_sse2;
- pixf->sad_x3[PIXEL_8x8] = x264_pixel_sad_x3_8x8_sse2;
- pixf->sad_x3[PIXEL_8x4] = x264_pixel_sad_x3_8x4_sse2;
- pixf->sad_x4[PIXEL_8x16] = x264_pixel_sad_x4_8x16_sse2;
- pixf->sad_x4[PIXEL_8x8] = x264_pixel_sad_x4_8x8_sse2;
- pixf->sad_x4[PIXEL_8x4] = x264_pixel_sad_x4_8x4_sse2;
- }
- if( (cpu&X264_CPU_SSE3) && (cpu&X264_CPU_CACHELINE_64) )
- {
- INIT2( sad, _sse3 );
- INIT2( sad_x3, _sse3 );
- INIT2( sad_x4, _sse3 );
- }
- if( cpu&X264_CPU_SSSE3 )
- {
- INIT4( hadamard_ac, _ssse3 );
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- pixf->intra_sad_x9_4x4 = x264_intra_sad_x9_4x4_ssse3;
- pixf->intra_satd_x9_4x4 = x264_intra_satd_x9_4x4_ssse3;
- pixf->intra_sad_x9_8x8 = x264_intra_sad_x9_8x8_ssse3;
- #if ARCH_X86_64
- pixf->intra_sa8d_x9_8x8 = x264_intra_sa8d_x9_8x8_ssse3;
- #endif
- }
- #if ARCH_X86 || !defined( __MACH__ )
- INIT_ADS( _ssse3 );
- #endif
- if( cpu&X264_CPU_SLOW_ATOM )
- {
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_ssse3_atom;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_ssse3_atom;
- INIT6( satd, _ssse3_atom );
- pixf->satd[PIXEL_4x16] = x264_pixel_satd_4x16_ssse3_atom;
- INIT6( satd_x3, _ssse3_atom );
- INIT6( satd_x4, _ssse3_atom );
- INIT4( hadamard_ac, _ssse3_atom );
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_ssse3_atom;
- #endif
- }
- else
- {
- INIT8( ssd, _ssse3 );
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_ssse3;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_ssse3;
- INIT8( satd, _ssse3 );
- INIT7( satd_x3, _ssse3 );
- INIT7( satd_x4, _ssse3 );
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_ssse3;
- #endif
- }
- pixf->intra_satd_x3_16x16 = x264_intra_satd_x3_16x16_ssse3;
- if( !(cpu&X264_CPU_SLOW_PSHUFB) )
- pixf->intra_sad_x3_16x16 = x264_intra_sad_x3_16x16_ssse3;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_ssse3;
- pixf->intra_satd_x3_8x8c = x264_intra_satd_x3_8x8c_ssse3;
- pixf->intra_sad_x3_8x8c = x264_intra_sad_x3_8x8c_ssse3;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_ssse3;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_ssse3;
- pixf->asd8 = x264_pixel_asd8_ssse3;
- if( cpu&X264_CPU_CACHELINE_64 )
- {
- INIT2( sad, _cache64_ssse3 );
- INIT2( sad_x3, _cache64_ssse3 );
- INIT2( sad_x4, _cache64_ssse3 );
- }
- else
- {
- INIT2( sad_x3, _ssse3 );
- INIT5( sad_x4, _ssse3 );
- }
- if( (cpu&X264_CPU_SLOW_ATOM) || (cpu&X264_CPU_SLOW_SHUFFLE) )
- {
- INIT5( ssd, _sse2 ); /* on conroe, sse2 is faster for width8/16 */
- }
- }
- if( cpu&X264_CPU_SSE4 )
- {
- INIT8( satd, _sse4 );
- INIT7( satd_x3, _sse4 );
- INIT7( satd_x4, _sse4 );
- INIT4( hadamard_ac, _sse4 );
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- pixf->intra_sad_x9_4x4 = x264_intra_sad_x9_4x4_sse4;
- pixf->intra_satd_x9_4x4 = x264_intra_satd_x9_4x4_sse4;
- pixf->intra_sad_x9_8x8 = x264_intra_sad_x9_8x8_sse4;
- #if ARCH_X86_64
- pixf->intra_sa8d_x9_8x8 = x264_intra_sa8d_x9_8x8_sse4;
- #endif
- }
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_sse4;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_sse4;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_sse4;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_sse4;
- #endif
- }
- if( cpu&X264_CPU_AVX )
- {
- INIT2_NAME( sad_aligned, sad, _sse2 ); /* AVX-capable CPUs doesn't benefit from an aligned version */
- INIT2( sad_x3, _avx );
- INIT2( sad_x4, _avx );
- INIT8( satd, _avx );
- INIT7( satd_x3, _avx );
- INIT7( satd_x4, _avx );
- #if ARCH_X86 || !defined( __MACH__ )
- INIT_ADS( _avx );
- #endif
- INIT4( hadamard_ac, _avx );
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- pixf->intra_sad_x9_4x4 = x264_intra_sad_x9_4x4_avx;
- pixf->intra_satd_x9_4x4 = x264_intra_satd_x9_4x4_avx;
- pixf->intra_sad_x9_8x8 = x264_intra_sad_x9_8x8_avx;
- #if ARCH_X86_64
- pixf->intra_sa8d_x9_8x8 = x264_intra_sa8d_x9_8x8_avx;
- #endif
- }
- INIT5( ssd, _avx );
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_avx;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_avx;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_avx;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_avx;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_avx;
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_avx;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_avx;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_avx;
- pixf->ssim_end4 = x264_pixel_ssim_end4_avx;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_avx;
- #endif
- }
- if( cpu&X264_CPU_XOP )
- {
- INIT7( satd, _xop );
- INIT7( satd_x3, _xop );
- INIT7( satd_x4, _xop );
- INIT4( hadamard_ac, _xop );
- if( !(cpu&X264_CPU_STACK_MOD4) )
- {
- pixf->intra_satd_x9_4x4 = x264_intra_satd_x9_4x4_xop;
- }
- INIT5( ssd, _xop );
- pixf->sa8d[PIXEL_16x16]= x264_pixel_sa8d_16x16_xop;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_xop;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_xop;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_xop;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_xop;
- #endif
- }
- if( cpu&X264_CPU_AVX2 )
- {
- INIT2( ssd, _avx2 );
- INIT2( sad_x3, _avx2 );
- INIT2( sad_x4, _avx2 );
- INIT4( satd, _avx2 );
- INIT2( hadamard_ac, _avx2 );
- #if ARCH_X86 || !defined( __MACH__ )
- INIT_ADS( _avx2 );
- #endif
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_avx2;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_avx2;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_avx2;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_avx2;
- pixf->intra_sad_x3_16x16 = x264_intra_sad_x3_16x16_avx2;
- pixf->intra_sad_x9_8x8 = x264_intra_sad_x9_8x8_avx2;
- pixf->intra_sad_x3_8x8c = x264_intra_sad_x3_8x8c_avx2;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_avx2;
- #if ARCH_X86_64
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_avx2;
- #endif
- }
- if( cpu&X264_CPU_AVX512 )
- {
- INIT8( sad, _avx512 );
- INIT8_NAME( sad_aligned, sad, _avx512 );
- INIT7( sad_x3, _avx512 );
- INIT7( sad_x4, _avx512 );
- INIT8( satd, _avx512 );
- INIT7( satd_x3, _avx512 );
- INIT7( satd_x4, _avx512 );
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_avx512;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_avx512;
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_avx512;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_avx512;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_avx512;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_avx512;
- }
- #endif //HAVE_MMX
- #if HAVE_ARMV6
- if( cpu&X264_CPU_ARMV6 )
- {
- pixf->sad[PIXEL_4x8] = x264_pixel_sad_4x8_armv6;
- pixf->sad[PIXEL_4x4] = x264_pixel_sad_4x4_armv6;
- pixf->sad_aligned[PIXEL_4x8] = x264_pixel_sad_4x8_armv6;
- pixf->sad_aligned[PIXEL_4x4] = x264_pixel_sad_4x4_armv6;
- }
- if( cpu&X264_CPU_NEON )
- {
- INIT5( sad, _neon );
- INIT5( sad_aligned, _neon );
- INIT7( sad_x3, _neon );
- INIT7( sad_x4, _neon );
- INIT7( ssd, _neon );
- INIT7( satd, _neon );
- INIT7( satd_x3, _neon );
- INIT7( satd_x4, _neon );
- INIT4( hadamard_ac, _neon );
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_neon;
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16_neon;
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_neon;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_neon;
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_neon;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_neon;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_neon;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_neon;
- pixf->vsad = x264_pixel_vsad_neon;
- pixf->asd8 = x264_pixel_asd8_neon;
- pixf->intra_sad_x3_4x4 = intra_sad_x3_4x4_neon;
- pixf->intra_satd_x3_4x4 = intra_satd_x3_4x4_neon;
- pixf->intra_sad_x3_8x8 = intra_sad_x3_8x8_neon;
- pixf->intra_sa8d_x3_8x8 = intra_sa8d_x3_8x8_neon;
- pixf->intra_sad_x3_8x8c = intra_sad_x3_8x8c_neon;
- pixf->intra_satd_x3_8x8c = intra_satd_x3_8x8c_neon;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_neon;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_neon;
- pixf->intra_sad_x3_16x16 = intra_sad_x3_16x16_neon;
- pixf->intra_satd_x3_16x16 = intra_satd_x3_16x16_neon;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_neon;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_neon;
- pixf->ssim_end4 = x264_pixel_ssim_end4_neon;
- if( cpu&X264_CPU_FAST_NEON_MRC )
- {
- pixf->sad[PIXEL_4x8] = x264_pixel_sad_4x8_neon;
- pixf->sad[PIXEL_4x4] = x264_pixel_sad_4x4_neon;
- pixf->sad_aligned[PIXEL_4x8] = x264_pixel_sad_aligned_4x8_neon;
- pixf->sad_aligned[PIXEL_4x4] = x264_pixel_sad_aligned_4x4_neon;
- }
- else // really just scheduled for dual issue / A8
- {
- INIT5( sad_aligned, _neon_dual );
- }
- }
- #endif
- #if ARCH_AARCH64
- if( cpu&X264_CPU_NEON )
- {
- INIT8( sad, _neon );
- // AArch64 has no distinct instructions for aligned load/store
- INIT8_NAME( sad_aligned, sad, _neon );
- INIT7( sad_x3, _neon );
- INIT7( sad_x4, _neon );
- INIT8( ssd, _neon );
- INIT8( satd, _neon );
- INIT7( satd_x3, _neon );
- INIT7( satd_x4, _neon );
- INIT4( hadamard_ac, _neon );
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_neon;
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16_neon;
- pixf->sa8d_satd[PIXEL_16x16] = x264_pixel_sa8d_satd_16x16_neon;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_neon;
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_neon;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_neon;
- pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_neon;
- pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_neon;
- pixf->vsad = x264_pixel_vsad_neon;
- pixf->asd8 = x264_pixel_asd8_neon;
- pixf->intra_sad_x3_4x4 = intra_sad_x3_4x4_neon;
- pixf->intra_satd_x3_4x4 = intra_satd_x3_4x4_neon;
- pixf->intra_sad_x3_8x8 = intra_sad_x3_8x8_neon;
- pixf->intra_sa8d_x3_8x8 = intra_sa8d_x3_8x8_neon;
- pixf->intra_sad_x3_8x8c = intra_sad_x3_8x8c_neon;
- pixf->intra_satd_x3_8x8c = intra_satd_x3_8x8c_neon;
- pixf->intra_sad_x3_8x16c = intra_sad_x3_8x16c_neon;
- pixf->intra_satd_x3_8x16c = intra_satd_x3_8x16c_neon;
- pixf->intra_sad_x3_16x16 = intra_sad_x3_16x16_neon;
- pixf->intra_satd_x3_16x16 = intra_satd_x3_16x16_neon;
- pixf->ssd_nv12_core = x264_pixel_ssd_nv12_core_neon;
- pixf->ssim_4x4x2_core = x264_pixel_ssim_4x4x2_core_neon;
- pixf->ssim_end4 = x264_pixel_ssim_end4_neon;
- }
- #endif // ARCH_AARCH64
- #if HAVE_MSA
- if( cpu&X264_CPU_MSA )
- {
- INIT8( sad, _msa );
- INIT8_NAME( sad_aligned, sad, _msa );
- INIT8( ssd, _msa );
- INIT7( sad_x3, _msa );
- INIT7( sad_x4, _msa );
- INIT8( satd, _msa );
- INIT4( hadamard_ac, _msa );
- pixf->intra_sad_x3_4x4 = x264_intra_sad_x3_4x4_msa;
- pixf->intra_sad_x3_8x8 = x264_intra_sad_x3_8x8_msa;
- pixf->intra_sad_x3_8x8c = x264_intra_sad_x3_8x8c_msa;
- pixf->intra_sad_x3_16x16 = x264_intra_sad_x3_16x16_msa;
- pixf->intra_satd_x3_4x4 = x264_intra_satd_x3_4x4_msa;
- pixf->intra_satd_x3_16x16 = x264_intra_satd_x3_16x16_msa;
- pixf->intra_satd_x3_8x8c = x264_intra_satd_x3_8x8c_msa;
- pixf->intra_sa8d_x3_8x8 = x264_intra_sa8d_x3_8x8_msa;
- pixf->ssim_4x4x2_core = x264_ssim_4x4x2_core_msa;
- pixf->var[PIXEL_16x16] = x264_pixel_var_16x16_msa;
- pixf->var[PIXEL_8x16] = x264_pixel_var_8x16_msa;
- pixf->var[PIXEL_8x8] = x264_pixel_var_8x8_msa;
- //pixf->var2[PIXEL_8x16] = x264_pixel_var2_8x16_msa;
- //pixf->var2[PIXEL_8x8] = x264_pixel_var2_8x8_msa;
- pixf->sa8d[PIXEL_16x16] = x264_pixel_sa8d_16x16_msa;
- pixf->sa8d[PIXEL_8x8] = x264_pixel_sa8d_8x8_msa;
- }
- #endif // HAVE_MSA
- #endif // HIGH_BIT_DEPTH
- #if HAVE_ALTIVEC
- if( cpu&X264_CPU_ALTIVEC )
- {
- x264_pixel_init_altivec( pixf );
- }
- #endif
- pixf->ads[PIXEL_8x16] =
- pixf->ads[PIXEL_8x4] =
- pixf->ads[PIXEL_4x8] = pixf->ads[PIXEL_16x8];
- pixf->ads[PIXEL_4x4] = pixf->ads[PIXEL_8x8];
- }
|