package/x265: bump version to 3.5
Rebased patch 0001 based on Gentoo patch
https://gitweb.gentoo.org/repo/gentoo.git/tree/media-libs/x265/files/arm-r1.patch
Removed patch 0002, a different fix was applied upstream:
ec7396adaa
Update project URL and _SITE.
Signed-off-by: Bernd Kuhls <bernd.kuhls@t-online.de>
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@bootlin.com>
This commit is contained in:
parent
bbc52f3ff3
commit
724bc9b83d
@ -17,36 +17,72 @@ Backported from upstream pull request:
|
|||||||
https://bitbucket.org/multicoreware/x265/pull-requests/21
|
https://bitbucket.org/multicoreware/x265/pull-requests/21
|
||||||
|
|
||||||
Signed-off-by: Bernd Kuhls <bernd.kuhls@t-online.de>
|
Signed-off-by: Bernd Kuhls <bernd.kuhls@t-online.de>
|
||||||
|
[Bernd: rebased for 3.5 based on
|
||||||
|
https://gitweb.gentoo.org/repo/gentoo.git/tree/media-libs/x265/files/arm-r1.patch]
|
||||||
|
|
||||||
diff -r 58dddcf01b7d source/CMakeLists.txt
|
diff -r 58dddcf01b7d source/CMakeLists.txt
|
||||||
--- a/source/CMakeLists.txt Mon Jan 23 09:50:33 2017 +0530
|
--- a/source/CMakeLists.txt Mon Jan 23 09:50:33 2017 +0530
|
||||||
+++ b/source/CMakeLists.txt Mon Jan 23 21:03:10 2017 +0100
|
+++ b/source/CMakeLists.txt Mon Jan 23 21:03:10 2017 +0100
|
||||||
@@ -194,28 +194,6 @@
|
@@ -40,7 +40,7 @@ SET(CMAKE_MODULE_PATH "${PROJECT_SOURCE_
|
||||||
if(NATIVE_BUILD)
|
# System architecture detection
|
||||||
if(INTEL_CXX)
|
string(TOLOWER "${CMAKE_SYSTEM_PROCESSOR}" SYSPROC)
|
||||||
add_definitions(-xhost)
|
set(X86_ALIASES x86 i386 i686 x86_64 amd64)
|
||||||
|
-set(ARM_ALIASES armv6l armv7l aarch64)
|
||||||
|
+set(ARM_ALIASES armv6l armv6l armv7l armv7a aarch64)
|
||||||
|
list(FIND X86_ALIASES "${SYSPROC}" X86MATCH)
|
||||||
|
list(FIND ARM_ALIASES "${SYSPROC}" ARMMATCH)
|
||||||
|
set(POWER_ALIASES ppc64 ppc64le)
|
||||||
|
@@ -239,11 +239,7 @@ if(GCC)
|
||||||
|
endif()
|
||||||
|
endif()
|
||||||
|
if(ARM AND CROSS_COMPILE_ARM)
|
||||||
|
- if(ARM64)
|
||||||
|
- set(ARM_ARGS -fPIC)
|
||||||
- else()
|
- else()
|
||||||
- add_definitions(-march=native)
|
- set(ARM_ARGS -march=armv6 -mfloat-abi=soft -mfpu=vfp -marm -fPIC)
|
||||||
- endif()
|
- endif()
|
||||||
- elseif(X86 AND NOT X64)
|
+ set(ARM_ARGS -fPIC)
|
||||||
- string(FIND "${CMAKE_CXX_FLAGS}" "-march" marchPos)
|
message(STATUS "cross compile arm")
|
||||||
- if(marchPos LESS "0")
|
elseif(ARM)
|
||||||
- add_definitions(-march=i686)
|
if(ARM64)
|
||||||
- if(WIN32 AND NOT INTEL_CXX AND NOT CLANG AND
|
@@ -252,10 +248,10 @@ if(GCC)
|
||||||
- CMAKE_CXX_COMPILER_VERSION VERSION_GREATER 6.0 AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 7.0)
|
else()
|
||||||
- add_definitions(-mpreferred-stack-boundary=2)
|
find_package(Neon)
|
||||||
- endif()
|
if(CPU_HAS_NEON)
|
||||||
- endif()
|
- set(ARM_ARGS -mcpu=native -mfloat-abi=hard -mfpu=neon -marm -fPIC)
|
||||||
- endif()
|
+ set(ARM_ARGS -mfpu=neon -fPIC)
|
||||||
- if(ARM AND CROSS_COMPILE_ARM)
|
add_definitions(-DHAVE_NEON)
|
||||||
|
else()
|
||||||
|
- set(ARM_ARGS -mcpu=native -mfloat-abi=hard -mfpu=vfp -marm)
|
||||||
|
+ set(ARM_ARGS -fPIC)
|
||||||
|
endif()
|
||||||
|
endif()
|
||||||
|
endif()
|
||||||
|
--- a/source/dynamicHDR10/CMakeLists.txt
|
||||||
|
+++ b/source/dynamicHDR10/CMakeLists.txt
|
||||||
|
@@ -43,14 +43,14 @@ if(GCC)
|
||||||
|
endif()
|
||||||
|
endif()
|
||||||
|
if(ARM AND CROSS_COMPILE_ARM)
|
||||||
- set(ARM_ARGS -march=armv6 -mfloat-abi=soft -mfpu=vfp -marm -fPIC)
|
- set(ARM_ARGS -march=armv6 -mfloat-abi=soft -mfpu=vfp -marm -fPIC)
|
||||||
- elseif(ARM)
|
+ set(ARM_ARGS -fPIC)
|
||||||
- find_package(Neon)
|
elseif(ARM)
|
||||||
- if(CPU_HAS_NEON)
|
find_package(Neon)
|
||||||
|
if(CPU_HAS_NEON)
|
||||||
- set(ARM_ARGS -mcpu=native -mfloat-abi=hard -mfpu=neon -marm -fPIC)
|
- set(ARM_ARGS -mcpu=native -mfloat-abi=hard -mfpu=neon -marm -fPIC)
|
||||||
- add_definitions(-DHAVE_NEON)
|
+ set(ARM_ARGS -fPIC)
|
||||||
- else()
|
add_definitions(-DHAVE_NEON)
|
||||||
|
else()
|
||||||
- set(ARM_ARGS -mcpu=native -mfloat-abi=hard -mfpu=vfp -marm)
|
- set(ARM_ARGS -mcpu=native -mfloat-abi=hard -mfpu=vfp -marm)
|
||||||
|
+ set(ARM_ARGS -fPIC)
|
||||||
endif()
|
endif()
|
||||||
endif()
|
endif()
|
||||||
add_definitions(${ARM_ARGS})
|
add_definitions(${ARM_ARGS})
|
||||||
|
@@ -150,4 +150,4 @@ set(BIN_INSTALL_DIR bin CACHE STRING "Install location of executables")
|
||||||
|
option(ENABLE_SHARED "Build shared library" OFF)
|
||||||
|
|
||||||
|
install(FILES hdr10plus.h DESTINATION include)
|
||||||
|
-endif()
|
||||||
|
\ No newline at end of file
|
||||||
|
+endif()
|
||||||
|
|
||||||
|
@ -1,354 +0,0 @@
|
|||||||
Fixes "arm assembly fail to compile on 1.8"
|
|
||||||
|
|
||||||
Downloaded from upstream bug report:
|
|
||||||
https://bitbucket.org/multicoreware/x265/issues/406
|
|
||||||
|
|
||||||
Signed-off-by: Bernd Kuhls <bernd.kuhls@t-online.de>
|
|
||||||
|
|
||||||
--- ./source/common/arm/asm-primitives.cpp.orig 2018-05-21 02:33:10.000000000 -0600
|
|
||||||
+++ ./source/common/arm/asm-primitives.cpp 2018-05-28 20:38:37.302378303 -0600
|
|
||||||
@@ -48,77 +48,77 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.ssim_4x4x2_core = PFX(ssim_4x4x2_core_neon);
|
|
||||||
|
|
||||||
// addAvg
|
|
||||||
- p.pu[LUMA_4x4].addAvg = PFX(addAvg_4x4_neon);
|
|
||||||
- p.pu[LUMA_4x8].addAvg = PFX(addAvg_4x8_neon);
|
|
||||||
- p.pu[LUMA_4x16].addAvg = PFX(addAvg_4x16_neon);
|
|
||||||
- p.pu[LUMA_8x4].addAvg = PFX(addAvg_8x4_neon);
|
|
||||||
- p.pu[LUMA_8x8].addAvg = PFX(addAvg_8x8_neon);
|
|
||||||
- p.pu[LUMA_8x16].addAvg = PFX(addAvg_8x16_neon);
|
|
||||||
- p.pu[LUMA_8x32].addAvg = PFX(addAvg_8x32_neon);
|
|
||||||
- p.pu[LUMA_12x16].addAvg = PFX(addAvg_12x16_neon);
|
|
||||||
- p.pu[LUMA_16x4].addAvg = PFX(addAvg_16x4_neon);
|
|
||||||
- p.pu[LUMA_16x8].addAvg = PFX(addAvg_16x8_neon);
|
|
||||||
- p.pu[LUMA_16x12].addAvg = PFX(addAvg_16x12_neon);
|
|
||||||
- p.pu[LUMA_16x16].addAvg = PFX(addAvg_16x16_neon);
|
|
||||||
- p.pu[LUMA_16x32].addAvg = PFX(addAvg_16x32_neon);
|
|
||||||
- p.pu[LUMA_16x64].addAvg = PFX(addAvg_16x64_neon);
|
|
||||||
- p.pu[LUMA_24x32].addAvg = PFX(addAvg_24x32_neon);
|
|
||||||
- p.pu[LUMA_32x8].addAvg = PFX(addAvg_32x8_neon);
|
|
||||||
- p.pu[LUMA_32x16].addAvg = PFX(addAvg_32x16_neon);
|
|
||||||
- p.pu[LUMA_32x24].addAvg = PFX(addAvg_32x24_neon);
|
|
||||||
- p.pu[LUMA_32x32].addAvg = PFX(addAvg_32x32_neon);
|
|
||||||
- p.pu[LUMA_32x64].addAvg = PFX(addAvg_32x64_neon);
|
|
||||||
- p.pu[LUMA_48x64].addAvg = PFX(addAvg_48x64_neon);
|
|
||||||
- p.pu[LUMA_64x16].addAvg = PFX(addAvg_64x16_neon);
|
|
||||||
- p.pu[LUMA_64x32].addAvg = PFX(addAvg_64x32_neon);
|
|
||||||
- p.pu[LUMA_64x48].addAvg = PFX(addAvg_64x48_neon);
|
|
||||||
- p.pu[LUMA_64x64].addAvg = PFX(addAvg_64x64_neon);
|
|
||||||
+ p.pu[LUMA_4x4].addAvg[ALIGNED] = PFX(addAvg_4x4_neon);
|
|
||||||
+ p.pu[LUMA_4x8].addAvg[ALIGNED] = PFX(addAvg_4x8_neon);
|
|
||||||
+ p.pu[LUMA_4x16].addAvg[ALIGNED] = PFX(addAvg_4x16_neon);
|
|
||||||
+ p.pu[LUMA_8x4].addAvg[ALIGNED] = PFX(addAvg_8x4_neon);
|
|
||||||
+ p.pu[LUMA_8x8].addAvg[ALIGNED] = PFX(addAvg_8x8_neon);
|
|
||||||
+ p.pu[LUMA_8x16].addAvg[ALIGNED] = PFX(addAvg_8x16_neon);
|
|
||||||
+ p.pu[LUMA_8x32].addAvg[ALIGNED] = PFX(addAvg_8x32_neon);
|
|
||||||
+ p.pu[LUMA_12x16].addAvg[ALIGNED] = PFX(addAvg_12x16_neon);
|
|
||||||
+ p.pu[LUMA_16x4].addAvg[ALIGNED] = PFX(addAvg_16x4_neon);
|
|
||||||
+ p.pu[LUMA_16x8].addAvg[ALIGNED] = PFX(addAvg_16x8_neon);
|
|
||||||
+ p.pu[LUMA_16x12].addAvg[ALIGNED] = PFX(addAvg_16x12_neon);
|
|
||||||
+ p.pu[LUMA_16x16].addAvg[ALIGNED] = PFX(addAvg_16x16_neon);
|
|
||||||
+ p.pu[LUMA_16x32].addAvg[ALIGNED] = PFX(addAvg_16x32_neon);
|
|
||||||
+ p.pu[LUMA_16x64].addAvg[ALIGNED] = PFX(addAvg_16x64_neon);
|
|
||||||
+ p.pu[LUMA_24x32].addAvg[ALIGNED] = PFX(addAvg_24x32_neon);
|
|
||||||
+ p.pu[LUMA_32x8].addAvg[ALIGNED] = PFX(addAvg_32x8_neon);
|
|
||||||
+ p.pu[LUMA_32x16].addAvg[ALIGNED] = PFX(addAvg_32x16_neon);
|
|
||||||
+ p.pu[LUMA_32x24].addAvg[ALIGNED] = PFX(addAvg_32x24_neon);
|
|
||||||
+ p.pu[LUMA_32x32].addAvg[ALIGNED] = PFX(addAvg_32x32_neon);
|
|
||||||
+ p.pu[LUMA_32x64].addAvg[ALIGNED] = PFX(addAvg_32x64_neon);
|
|
||||||
+ p.pu[LUMA_48x64].addAvg[ALIGNED] = PFX(addAvg_48x64_neon);
|
|
||||||
+ p.pu[LUMA_64x16].addAvg[ALIGNED] = PFX(addAvg_64x16_neon);
|
|
||||||
+ p.pu[LUMA_64x32].addAvg[ALIGNED] = PFX(addAvg_64x32_neon);
|
|
||||||
+ p.pu[LUMA_64x48].addAvg[ALIGNED] = PFX(addAvg_64x48_neon);
|
|
||||||
+ p.pu[LUMA_64x64].addAvg[ALIGNED] = PFX(addAvg_64x64_neon);
|
|
||||||
|
|
||||||
// chroma addAvg
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_4x2].addAvg = PFX(addAvg_4x2_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_4x4].addAvg = PFX(addAvg_4x4_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_4x8].addAvg = PFX(addAvg_4x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_4x16].addAvg = PFX(addAvg_4x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_6x8].addAvg = PFX(addAvg_6x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_8x2].addAvg = PFX(addAvg_8x2_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_8x4].addAvg = PFX(addAvg_8x4_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_8x6].addAvg = PFX(addAvg_8x6_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_8x8].addAvg = PFX(addAvg_8x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_8x16].addAvg = PFX(addAvg_8x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_8x32].addAvg = PFX(addAvg_8x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_12x16].addAvg = PFX(addAvg_12x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_16x4].addAvg = PFX(addAvg_16x4_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_16x8].addAvg = PFX(addAvg_16x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_16x12].addAvg = PFX(addAvg_16x12_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_16x16].addAvg = PFX(addAvg_16x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_16x32].addAvg = PFX(addAvg_16x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_24x32].addAvg = PFX(addAvg_24x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_32x8].addAvg = PFX(addAvg_32x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_32x16].addAvg = PFX(addAvg_32x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_32x24].addAvg = PFX(addAvg_32x24_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].pu[CHROMA_420_32x32].addAvg = PFX(addAvg_32x32_neon);
|
|
||||||
-
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_4x8].addAvg = PFX(addAvg_4x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_4x16].addAvg = PFX(addAvg_4x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_4x32].addAvg = PFX(addAvg_4x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_6x16].addAvg = PFX(addAvg_6x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_8x4].addAvg = PFX(addAvg_8x4_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_8x8].addAvg = PFX(addAvg_8x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_8x12].addAvg = PFX(addAvg_8x12_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_8x16].addAvg = PFX(addAvg_8x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_8x32].addAvg = PFX(addAvg_8x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_8x64].addAvg = PFX(addAvg_8x64_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_12x32].addAvg = PFX(addAvg_12x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_16x8].addAvg = PFX(addAvg_16x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_16x16].addAvg = PFX(addAvg_16x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_16x24].addAvg = PFX(addAvg_16x24_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_16x32].addAvg = PFX(addAvg_16x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_16x64].addAvg = PFX(addAvg_16x64_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_24x64].addAvg = PFX(addAvg_24x64_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_32x16].addAvg = PFX(addAvg_32x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_32x32].addAvg = PFX(addAvg_32x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_32x48].addAvg = PFX(addAvg_32x48_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].pu[CHROMA_422_32x64].addAvg = PFX(addAvg_32x64_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_4x2].addAvg[ALIGNED] = PFX(addAvg_4x2_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_4x4].addAvg[ALIGNED] = PFX(addAvg_4x4_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_4x8].addAvg[ALIGNED] = PFX(addAvg_4x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_4x16].addAvg[ALIGNED] = PFX(addAvg_4x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_6x8].addAvg[ALIGNED] = PFX(addAvg_6x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_8x2].addAvg[ALIGNED] = PFX(addAvg_8x2_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_8x4].addAvg[ALIGNED] = PFX(addAvg_8x4_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_8x6].addAvg[ALIGNED] = PFX(addAvg_8x6_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_8x8].addAvg[ALIGNED] = PFX(addAvg_8x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_8x16].addAvg[ALIGNED] = PFX(addAvg_8x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_8x32].addAvg[ALIGNED] = PFX(addAvg_8x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_12x16].addAvg[ALIGNED] = PFX(addAvg_12x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_16x4].addAvg[ALIGNED] = PFX(addAvg_16x4_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_16x8].addAvg[ALIGNED] = PFX(addAvg_16x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_16x12].addAvg[ALIGNED] = PFX(addAvg_16x12_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_16x16].addAvg[ALIGNED] = PFX(addAvg_16x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_16x32].addAvg[ALIGNED] = PFX(addAvg_16x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_24x32].addAvg[ALIGNED] = PFX(addAvg_24x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_32x8].addAvg[ALIGNED] = PFX(addAvg_32x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_32x16].addAvg[ALIGNED] = PFX(addAvg_32x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_32x24].addAvg[ALIGNED] = PFX(addAvg_32x24_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].pu[CHROMA_420_32x32].addAvg[ALIGNED] = PFX(addAvg_32x32_neon);
|
|
||||||
+
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_4x8].addAvg[ALIGNED] = PFX(addAvg_4x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_4x16].addAvg[ALIGNED] = PFX(addAvg_4x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_4x32].addAvg[ALIGNED] = PFX(addAvg_4x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_6x16].addAvg[ALIGNED] = PFX(addAvg_6x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_8x4].addAvg[ALIGNED] = PFX(addAvg_8x4_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_8x8].addAvg[ALIGNED] = PFX(addAvg_8x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_8x12].addAvg[ALIGNED] = PFX(addAvg_8x12_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_8x16].addAvg[ALIGNED] = PFX(addAvg_8x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_8x32].addAvg[ALIGNED] = PFX(addAvg_8x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_8x64].addAvg[ALIGNED] = PFX(addAvg_8x64_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_12x32].addAvg[ALIGNED] = PFX(addAvg_12x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_16x8].addAvg[ALIGNED] = PFX(addAvg_16x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_16x16].addAvg[ALIGNED] = PFX(addAvg_16x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_16x24].addAvg[ALIGNED] = PFX(addAvg_16x24_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_16x32].addAvg[ALIGNED] = PFX(addAvg_16x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_16x64].addAvg[ALIGNED] = PFX(addAvg_16x64_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_24x64].addAvg[ALIGNED] = PFX(addAvg_24x64_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_32x16].addAvg[ALIGNED] = PFX(addAvg_32x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_32x32].addAvg[ALIGNED] = PFX(addAvg_32x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_32x48].addAvg[ALIGNED] = PFX(addAvg_32x48_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].pu[CHROMA_422_32x64].addAvg[ALIGNED] = PFX(addAvg_32x64_neon);
|
|
||||||
|
|
||||||
// quant
|
|
||||||
p.quant = PFX(quant_neon);
|
|
||||||
@@ -402,7 +402,7 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.scale2D_64to32 = PFX(scale2D_64to32_neon);
|
|
||||||
|
|
||||||
// scale1D_128to64
|
|
||||||
- p.scale1D_128to64 = PFX(scale1D_128to64_neon);
|
|
||||||
+ p.scale1D_128to64[ALIGNED] = PFX(scale1D_128to64_neon);
|
|
||||||
|
|
||||||
// copy_count
|
|
||||||
p.cu[BLOCK_4x4].copy_cnt = PFX(copy_cnt_4_neon);
|
|
||||||
@@ -411,37 +411,37 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.cu[BLOCK_32x32].copy_cnt = PFX(copy_cnt_32_neon);
|
|
||||||
|
|
||||||
// filterPixelToShort
|
|
||||||
- p.pu[LUMA_4x4].convert_p2s = PFX(filterPixelToShort_4x4_neon);
|
|
||||||
- p.pu[LUMA_4x8].convert_p2s = PFX(filterPixelToShort_4x8_neon);
|
|
||||||
- p.pu[LUMA_4x16].convert_p2s = PFX(filterPixelToShort_4x16_neon);
|
|
||||||
- p.pu[LUMA_8x4].convert_p2s = PFX(filterPixelToShort_8x4_neon);
|
|
||||||
- p.pu[LUMA_8x8].convert_p2s = PFX(filterPixelToShort_8x8_neon);
|
|
||||||
- p.pu[LUMA_8x16].convert_p2s = PFX(filterPixelToShort_8x16_neon);
|
|
||||||
- p.pu[LUMA_8x32].convert_p2s = PFX(filterPixelToShort_8x32_neon);
|
|
||||||
- p.pu[LUMA_12x16].convert_p2s = PFX(filterPixelToShort_12x16_neon);
|
|
||||||
- p.pu[LUMA_16x4].convert_p2s = PFX(filterPixelToShort_16x4_neon);
|
|
||||||
- p.pu[LUMA_16x8].convert_p2s = PFX(filterPixelToShort_16x8_neon);
|
|
||||||
- p.pu[LUMA_16x12].convert_p2s = PFX(filterPixelToShort_16x12_neon);
|
|
||||||
- p.pu[LUMA_16x16].convert_p2s = PFX(filterPixelToShort_16x16_neon);
|
|
||||||
- p.pu[LUMA_16x32].convert_p2s = PFX(filterPixelToShort_16x32_neon);
|
|
||||||
- p.pu[LUMA_16x64].convert_p2s = PFX(filterPixelToShort_16x64_neon);
|
|
||||||
- p.pu[LUMA_24x32].convert_p2s = PFX(filterPixelToShort_24x32_neon);
|
|
||||||
- p.pu[LUMA_32x8].convert_p2s = PFX(filterPixelToShort_32x8_neon);
|
|
||||||
- p.pu[LUMA_32x16].convert_p2s = PFX(filterPixelToShort_32x16_neon);
|
|
||||||
- p.pu[LUMA_32x24].convert_p2s = PFX(filterPixelToShort_32x24_neon);
|
|
||||||
- p.pu[LUMA_32x32].convert_p2s = PFX(filterPixelToShort_32x32_neon);
|
|
||||||
- p.pu[LUMA_32x64].convert_p2s = PFX(filterPixelToShort_32x64_neon);
|
|
||||||
- p.pu[LUMA_48x64].convert_p2s = PFX(filterPixelToShort_48x64_neon);
|
|
||||||
- p.pu[LUMA_64x16].convert_p2s = PFX(filterPixelToShort_64x16_neon);
|
|
||||||
- p.pu[LUMA_64x32].convert_p2s = PFX(filterPixelToShort_64x32_neon);
|
|
||||||
- p.pu[LUMA_64x48].convert_p2s = PFX(filterPixelToShort_64x48_neon);
|
|
||||||
- p.pu[LUMA_64x64].convert_p2s = PFX(filterPixelToShort_64x64_neon);
|
|
||||||
+ p.pu[LUMA_4x4].convert_p2s[ALIGNED] = PFX(filterPixelToShort_4x4_neon);
|
|
||||||
+ p.pu[LUMA_4x8].convert_p2s[ALIGNED] = PFX(filterPixelToShort_4x8_neon);
|
|
||||||
+ p.pu[LUMA_4x16].convert_p2s[ALIGNED] = PFX(filterPixelToShort_4x16_neon);
|
|
||||||
+ p.pu[LUMA_8x4].convert_p2s[ALIGNED] = PFX(filterPixelToShort_8x4_neon);
|
|
||||||
+ p.pu[LUMA_8x8].convert_p2s[ALIGNED] = PFX(filterPixelToShort_8x8_neon);
|
|
||||||
+ p.pu[LUMA_8x16].convert_p2s[ALIGNED] = PFX(filterPixelToShort_8x16_neon);
|
|
||||||
+ p.pu[LUMA_8x32].convert_p2s[ALIGNED] = PFX(filterPixelToShort_8x32_neon);
|
|
||||||
+ p.pu[LUMA_12x16].convert_p2s[ALIGNED] = PFX(filterPixelToShort_12x16_neon);
|
|
||||||
+ p.pu[LUMA_16x4].convert_p2s[ALIGNED] = PFX(filterPixelToShort_16x4_neon);
|
|
||||||
+ p.pu[LUMA_16x8].convert_p2s[ALIGNED] = PFX(filterPixelToShort_16x8_neon);
|
|
||||||
+ p.pu[LUMA_16x12].convert_p2s[ALIGNED] = PFX(filterPixelToShort_16x12_neon);
|
|
||||||
+ p.pu[LUMA_16x16].convert_p2s[ALIGNED] = PFX(filterPixelToShort_16x16_neon);
|
|
||||||
+ p.pu[LUMA_16x32].convert_p2s[ALIGNED] = PFX(filterPixelToShort_16x32_neon);
|
|
||||||
+ p.pu[LUMA_16x64].convert_p2s[ALIGNED] = PFX(filterPixelToShort_16x64_neon);
|
|
||||||
+ p.pu[LUMA_24x32].convert_p2s[ALIGNED] = PFX(filterPixelToShort_24x32_neon);
|
|
||||||
+ p.pu[LUMA_32x8].convert_p2s[ALIGNED] = PFX(filterPixelToShort_32x8_neon);
|
|
||||||
+ p.pu[LUMA_32x16].convert_p2s[ALIGNED] = PFX(filterPixelToShort_32x16_neon);
|
|
||||||
+ p.pu[LUMA_32x24].convert_p2s[ALIGNED] = PFX(filterPixelToShort_32x24_neon);
|
|
||||||
+ p.pu[LUMA_32x32].convert_p2s[ALIGNED] = PFX(filterPixelToShort_32x32_neon);
|
|
||||||
+ p.pu[LUMA_32x64].convert_p2s[ALIGNED] = PFX(filterPixelToShort_32x64_neon);
|
|
||||||
+ p.pu[LUMA_48x64].convert_p2s[ALIGNED] = PFX(filterPixelToShort_48x64_neon);
|
|
||||||
+ p.pu[LUMA_64x16].convert_p2s[ALIGNED] = PFX(filterPixelToShort_64x16_neon);
|
|
||||||
+ p.pu[LUMA_64x32].convert_p2s[ALIGNED] = PFX(filterPixelToShort_64x32_neon);
|
|
||||||
+ p.pu[LUMA_64x48].convert_p2s[ALIGNED] = PFX(filterPixelToShort_64x48_neon);
|
|
||||||
+ p.pu[LUMA_64x64].convert_p2s[ALIGNED] = PFX(filterPixelToShort_64x64_neon);
|
|
||||||
|
|
||||||
// Block_fill
|
|
||||||
- p.cu[BLOCK_4x4].blockfill_s = PFX(blockfill_s_4x4_neon);
|
|
||||||
- p.cu[BLOCK_8x8].blockfill_s = PFX(blockfill_s_8x8_neon);
|
|
||||||
- p.cu[BLOCK_16x16].blockfill_s = PFX(blockfill_s_16x16_neon);
|
|
||||||
- p.cu[BLOCK_32x32].blockfill_s = PFX(blockfill_s_32x32_neon);
|
|
||||||
+ p.cu[BLOCK_4x4].blockfill_s[ALIGNED] = PFX(blockfill_s_4x4_neon);
|
|
||||||
+ p.cu[BLOCK_8x8].blockfill_s[ALIGNED] = PFX(blockfill_s_8x8_neon);
|
|
||||||
+ p.cu[BLOCK_16x16].blockfill_s[ALIGNED] = PFX(blockfill_s_16x16_neon);
|
|
||||||
+ p.cu[BLOCK_32x32].blockfill_s[ALIGNED] = PFX(blockfill_s_32x32_neon);
|
|
||||||
|
|
||||||
// Blockcopy_ss
|
|
||||||
p.cu[BLOCK_4x4].copy_ss = PFX(blockcopy_ss_4x4_neon);
|
|
||||||
@@ -495,21 +495,21 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.chroma[X265_CSP_I422].cu[BLOCK_422_32x64].copy_sp = PFX(blockcopy_sp_32x64_neon);
|
|
||||||
|
|
||||||
// pixel_add_ps
|
|
||||||
- p.cu[BLOCK_4x4].add_ps = PFX(pixel_add_ps_4x4_neon);
|
|
||||||
- p.cu[BLOCK_8x8].add_ps = PFX(pixel_add_ps_8x8_neon);
|
|
||||||
- p.cu[BLOCK_16x16].add_ps = PFX(pixel_add_ps_16x16_neon);
|
|
||||||
- p.cu[BLOCK_32x32].add_ps = PFX(pixel_add_ps_32x32_neon);
|
|
||||||
- p.cu[BLOCK_64x64].add_ps = PFX(pixel_add_ps_64x64_neon);
|
|
||||||
+ p.cu[BLOCK_4x4].add_ps[ALIGNED] = PFX(pixel_add_ps_4x4_neon);
|
|
||||||
+ p.cu[BLOCK_8x8].add_ps[ALIGNED] = PFX(pixel_add_ps_8x8_neon);
|
|
||||||
+ p.cu[BLOCK_16x16].add_ps[ALIGNED] = PFX(pixel_add_ps_16x16_neon);
|
|
||||||
+ p.cu[BLOCK_32x32].add_ps[ALIGNED] = PFX(pixel_add_ps_32x32_neon);
|
|
||||||
+ p.cu[BLOCK_64x64].add_ps[ALIGNED] = PFX(pixel_add_ps_64x64_neon);
|
|
||||||
|
|
||||||
// chroma add_ps
|
|
||||||
- p.chroma[X265_CSP_I420].cu[BLOCK_420_4x4].add_ps = PFX(pixel_add_ps_4x4_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].cu[BLOCK_420_8x8].add_ps = PFX(pixel_add_ps_8x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].cu[BLOCK_420_16x16].add_ps = PFX(pixel_add_ps_16x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I420].cu[BLOCK_420_32x32].add_ps = PFX(pixel_add_ps_32x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].cu[BLOCK_422_4x8].add_ps = PFX(pixel_add_ps_4x8_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].cu[BLOCK_422_8x16].add_ps = PFX(pixel_add_ps_8x16_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].cu[BLOCK_422_16x32].add_ps = PFX(pixel_add_ps_16x32_neon);
|
|
||||||
- p.chroma[X265_CSP_I422].cu[BLOCK_422_32x64].add_ps = PFX(pixel_add_ps_32x64_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].cu[BLOCK_420_4x4].add_ps[ALIGNED] = PFX(pixel_add_ps_4x4_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].cu[BLOCK_420_8x8].add_ps[ALIGNED] = PFX(pixel_add_ps_8x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].cu[BLOCK_420_16x16].add_ps[ALIGNED] = PFX(pixel_add_ps_16x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I420].cu[BLOCK_420_32x32].add_ps[ALIGNED] = PFX(pixel_add_ps_32x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].cu[BLOCK_422_4x8].add_ps[ALIGNED] = PFX(pixel_add_ps_4x8_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].cu[BLOCK_422_8x16].add_ps[ALIGNED] = PFX(pixel_add_ps_8x16_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].cu[BLOCK_422_16x32].add_ps[ALIGNED] = PFX(pixel_add_ps_16x32_neon);
|
|
||||||
+ p.chroma[X265_CSP_I422].cu[BLOCK_422_32x64].add_ps[ALIGNED] = PFX(pixel_add_ps_32x64_neon);
|
|
||||||
|
|
||||||
// cpy2Dto1D_shr
|
|
||||||
p.cu[BLOCK_4x4].cpy2Dto1D_shr = PFX(cpy2Dto1D_shr_4x4_neon);
|
|
||||||
@@ -518,10 +518,10 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.cu[BLOCK_32x32].cpy2Dto1D_shr = PFX(cpy2Dto1D_shr_32x32_neon);
|
|
||||||
|
|
||||||
// ssd_s
|
|
||||||
- p.cu[BLOCK_4x4].ssd_s = PFX(pixel_ssd_s_4x4_neon);
|
|
||||||
- p.cu[BLOCK_8x8].ssd_s = PFX(pixel_ssd_s_8x8_neon);
|
|
||||||
- p.cu[BLOCK_16x16].ssd_s = PFX(pixel_ssd_s_16x16_neon);
|
|
||||||
- p.cu[BLOCK_32x32].ssd_s = PFX(pixel_ssd_s_32x32_neon);
|
|
||||||
+ p.cu[BLOCK_4x4].ssd_s[ALIGNED] = PFX(pixel_ssd_s_4x4_neon);
|
|
||||||
+ p.cu[BLOCK_8x8].ssd_s[ALIGNED] = PFX(pixel_ssd_s_8x8_neon);
|
|
||||||
+ p.cu[BLOCK_16x16].ssd_s[ALIGNED] = PFX(pixel_ssd_s_16x16_neon);
|
|
||||||
+ p.cu[BLOCK_32x32].ssd_s[ALIGNED] = PFX(pixel_ssd_s_32x32_neon);
|
|
||||||
|
|
||||||
// sse_ss
|
|
||||||
p.cu[BLOCK_4x4].sse_ss = PFX(pixel_sse_ss_4x4_neon);
|
|
||||||
@@ -548,10 +548,10 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.chroma[X265_CSP_I422].cu[BLOCK_422_32x64].sub_ps = PFX(pixel_sub_ps_32x64_neon);
|
|
||||||
|
|
||||||
// calc_Residual
|
|
||||||
- p.cu[BLOCK_4x4].calcresidual = PFX(getResidual4_neon);
|
|
||||||
- p.cu[BLOCK_8x8].calcresidual = PFX(getResidual8_neon);
|
|
||||||
- p.cu[BLOCK_16x16].calcresidual = PFX(getResidual16_neon);
|
|
||||||
- p.cu[BLOCK_32x32].calcresidual = PFX(getResidual32_neon);
|
|
||||||
+ p.cu[BLOCK_4x4].calcresidual[ALIGNED] = PFX(getResidual4_neon);
|
|
||||||
+ p.cu[BLOCK_8x8].calcresidual[ALIGNED] = PFX(getResidual8_neon);
|
|
||||||
+ p.cu[BLOCK_16x16].calcresidual[ALIGNED] = PFX(getResidual16_neon);
|
|
||||||
+ p.cu[BLOCK_32x32].calcresidual[ALIGNED] = PFX(getResidual32_neon);
|
|
||||||
|
|
||||||
// sse_pp
|
|
||||||
p.cu[BLOCK_4x4].sse_pp = PFX(pixel_sse_pp_4x4_neon);
|
|
||||||
@@ -722,31 +722,31 @@ void setupAssemblyPrimitives(EncoderPrim
|
|
||||||
p.pu[LUMA_64x64].sad_x4 = PFX(sad_x4_64x64_neon);
|
|
||||||
|
|
||||||
// pixel_avg_pp
|
|
||||||
- p.pu[LUMA_4x4].pixelavg_pp = PFX(pixel_avg_pp_4x4_neon);
|
|
||||||
- p.pu[LUMA_4x8].pixelavg_pp = PFX(pixel_avg_pp_4x8_neon);
|
|
||||||
- p.pu[LUMA_4x16].pixelavg_pp = PFX(pixel_avg_pp_4x16_neon);
|
|
||||||
- p.pu[LUMA_8x4].pixelavg_pp = PFX(pixel_avg_pp_8x4_neon);
|
|
||||||
- p.pu[LUMA_8x8].pixelavg_pp = PFX(pixel_avg_pp_8x8_neon);
|
|
||||||
- p.pu[LUMA_8x16].pixelavg_pp = PFX(pixel_avg_pp_8x16_neon);
|
|
||||||
- p.pu[LUMA_8x32].pixelavg_pp = PFX(pixel_avg_pp_8x32_neon);
|
|
||||||
- p.pu[LUMA_12x16].pixelavg_pp = PFX(pixel_avg_pp_12x16_neon);
|
|
||||||
- p.pu[LUMA_16x4].pixelavg_pp = PFX(pixel_avg_pp_16x4_neon);
|
|
||||||
- p.pu[LUMA_16x8].pixelavg_pp = PFX(pixel_avg_pp_16x8_neon);
|
|
||||||
- p.pu[LUMA_16x12].pixelavg_pp = PFX(pixel_avg_pp_16x12_neon);
|
|
||||||
- p.pu[LUMA_16x16].pixelavg_pp = PFX(pixel_avg_pp_16x16_neon);
|
|
||||||
- p.pu[LUMA_16x32].pixelavg_pp = PFX(pixel_avg_pp_16x32_neon);
|
|
||||||
- p.pu[LUMA_16x64].pixelavg_pp = PFX(pixel_avg_pp_16x64_neon);
|
|
||||||
- p.pu[LUMA_24x32].pixelavg_pp = PFX(pixel_avg_pp_24x32_neon);
|
|
||||||
- p.pu[LUMA_32x8].pixelavg_pp = PFX(pixel_avg_pp_32x8_neon);
|
|
||||||
- p.pu[LUMA_32x16].pixelavg_pp = PFX(pixel_avg_pp_32x16_neon);
|
|
||||||
- p.pu[LUMA_32x24].pixelavg_pp = PFX(pixel_avg_pp_32x24_neon);
|
|
||||||
- p.pu[LUMA_32x32].pixelavg_pp = PFX(pixel_avg_pp_32x32_neon);
|
|
||||||
- p.pu[LUMA_32x64].pixelavg_pp = PFX(pixel_avg_pp_32x64_neon);
|
|
||||||
- p.pu[LUMA_48x64].pixelavg_pp = PFX(pixel_avg_pp_48x64_neon);
|
|
||||||
- p.pu[LUMA_64x16].pixelavg_pp = PFX(pixel_avg_pp_64x16_neon);
|
|
||||||
- p.pu[LUMA_64x32].pixelavg_pp = PFX(pixel_avg_pp_64x32_neon);
|
|
||||||
- p.pu[LUMA_64x48].pixelavg_pp = PFX(pixel_avg_pp_64x48_neon);
|
|
||||||
- p.pu[LUMA_64x64].pixelavg_pp = PFX(pixel_avg_pp_64x64_neon);
|
|
||||||
+ p.pu[LUMA_4x4].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_4x4_neon);
|
|
||||||
+ p.pu[LUMA_4x8].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_4x8_neon);
|
|
||||||
+ p.pu[LUMA_4x16].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_4x16_neon);
|
|
||||||
+ p.pu[LUMA_8x4].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_8x4_neon);
|
|
||||||
+ p.pu[LUMA_8x8].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_8x8_neon);
|
|
||||||
+ p.pu[LUMA_8x16].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_8x16_neon);
|
|
||||||
+ p.pu[LUMA_8x32].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_8x32_neon);
|
|
||||||
+ p.pu[LUMA_12x16].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_12x16_neon);
|
|
||||||
+ p.pu[LUMA_16x4].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_16x4_neon);
|
|
||||||
+ p.pu[LUMA_16x8].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_16x8_neon);
|
|
||||||
+ p.pu[LUMA_16x12].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_16x12_neon);
|
|
||||||
+ p.pu[LUMA_16x16].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_16x16_neon);
|
|
||||||
+ p.pu[LUMA_16x32].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_16x32_neon);
|
|
||||||
+ p.pu[LUMA_16x64].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_16x64_neon);
|
|
||||||
+ p.pu[LUMA_24x32].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_24x32_neon);
|
|
||||||
+ p.pu[LUMA_32x8].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_32x8_neon);
|
|
||||||
+ p.pu[LUMA_32x16].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_32x16_neon);
|
|
||||||
+ p.pu[LUMA_32x24].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_32x24_neon);
|
|
||||||
+ p.pu[LUMA_32x32].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_32x32_neon);
|
|
||||||
+ p.pu[LUMA_32x64].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_32x64_neon);
|
|
||||||
+ p.pu[LUMA_48x64].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_48x64_neon);
|
|
||||||
+ p.pu[LUMA_64x16].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_64x16_neon);
|
|
||||||
+ p.pu[LUMA_64x32].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_64x32_neon);
|
|
||||||
+ p.pu[LUMA_64x48].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_64x48_neon);
|
|
||||||
+ p.pu[LUMA_64x64].pixelavg_pp[ALIGNED] = PFX(pixel_avg_pp_64x64_neon);
|
|
||||||
|
|
||||||
// planecopy
|
|
||||||
p.planecopy_cp = PFX(pixel_planecopy_cp_neon);
|
|
@ -11,7 +11,7 @@ config BR2_PACKAGE_X265
|
|||||||
General Public License (GPL) 2 license or a commercial
|
General Public License (GPL) 2 license or a commercial
|
||||||
license, similar to the x264 project.
|
license, similar to the x264 project.
|
||||||
|
|
||||||
https://bitbucket.org/multicoreware/x265/wiki/Home
|
https://bitbucket.org/multicoreware/x265_git/wiki/Home
|
||||||
|
|
||||||
if BR2_PACKAGE_X265
|
if BR2_PACKAGE_X265
|
||||||
|
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
# From https://mailman.videolan.org/pipermail/x265-devel/2020-February/012880.html
|
# From https://mailman.videolan.org/pipermail/x265-devel/2021-March/013257.html
|
||||||
md5 0c8c747b59b5411dea8cf557554636c1 x265_3.3.tar.gz
|
md5 deb5df5cb2ec17bdbae6ac6bbc3b1eef x265_3.5.tar.gz
|
||||||
# Locally generated
|
# Locally generated
|
||||||
sha256 f26e148ed1f4dfb33fd1eb3ff5e60e08078d1b2017e88bcbb045b3fb58300b9c x265_3.3.tar.gz
|
sha256 e70a3335cacacbba0b3a20ec6fecd6783932288ebc8163ad74bcc9606477cae8 x265_3.5.tar.gz
|
||||||
sha256 d8afb1bcc7a2cfc603683b168d6987ef0a48e59e0da3693bf55c5d33b67e2b49 COPYING
|
sha256 d8afb1bcc7a2cfc603683b168d6987ef0a48e59e0da3693bf55c5d33b67e2b49 COPYING
|
||||||
|
@ -4,9 +4,9 @@
|
|||||||
#
|
#
|
||||||
################################################################################
|
################################################################################
|
||||||
|
|
||||||
X265_VERSION = 3.3
|
X265_VERSION = 3.5
|
||||||
X265_SOURCE = x265_$(X265_VERSION).tar.gz
|
X265_SOURCE = x265_$(X265_VERSION).tar.gz
|
||||||
X265_SITE = https://bitbucket.org/multicoreware/x265/downloads
|
X265_SITE = https://bitbucket.org/multicoreware/x265_git/downloads
|
||||||
X265_LICENSE = GPL-2.0+
|
X265_LICENSE = GPL-2.0+
|
||||||
X265_LICENSE_FILES = COPYING
|
X265_LICENSE_FILES = COPYING
|
||||||
X265_CPE_ID_VENDOR = multicorewareinc
|
X265_CPE_ID_VENDOR = multicorewareinc
|
||||||
|
Loading…
Reference in New Issue
Block a user