summaryrefslogtreecommitdiffstats
path: root/libavcodec
Commit message (Collapse)AuthorAgeFilesLines
* lavc/copy_block: style fixClément Bœsch2017-03-201-4/+3
|
* Merge commit '73f5e17a203713c4ac4e5a821809823b383b195f'Clément Bœsch2017-03-201-6/+7
|\ | | | | | | | | | | | | * commit '73f5e17a203713c4ac4e5a821809823b383b195f': copy_block: Change type of array stride parameters to ptrdiff_t Merged-by: Clément Bœsch <u@pkh.me>
| * copy_block: Change type of array stride parameters to ptrdiff_tDiego Biurrun2016-09-041-5/+6
| | | | | | | | ptrdiff_t is the correct type for array strides and similar.
* | Merge commit '21e500ba647aec233d5930d3d1081489d0d53ceb'Clément Bœsch2017-03-201-6/+6
|\ \ | |/ | | | | | | | | | | * commit '21e500ba647aec233d5930d3d1081489d0d53ceb': svq1dec: Change type of array pitch parameters to ptrdiff_t Merged-by: Clément Bœsch <u@pkh.me>
| * svq1dec: Change type of array pitch parameters to ptrdiff_tDiego Biurrun2016-09-041-6/+6
| | | | | | | | ptrdiff_t is the correct type for array pitches and similar.
* | Merge commit '746c56b7730ce09397d3a8354acc131285e9d829'Clément Bœsch2017-03-205-68/+69
|\ \ | |/ | | | | | | | | | | * commit '746c56b7730ce09397d3a8354acc131285e9d829': indeo: Change type of array pitch parameters to ptrdiff_t Merged-by: Clément Bœsch <u@pkh.me>
| * indeo: Change type of array pitch parameters to ptrdiff_tDiego Biurrun2016-09-045-68/+69
| | | | | | | | ptrdiff_t is the correct type for array pitches and similar.
| * lavc: Remove old vaapi decode infrastructureMark Thompson2016-08-307-329/+69
| | | | | | | | | | Deprecates struct vaapi_context and the installed header vaapi.h, to be removed at the next version bump.
| * vaapi_mpeg4: Convert to use the new VAAPI hwaccel codeMark Thompson2016-08-301-82/+133
| |
| * vaapi_vc1: Convert to use the new VAAPI hwaccel codeMark Thompson2016-08-301-140/+206
| |
| * vaapi_mpeg2: Convert to use the new VAAPI hwaccel codeMark Thompson2016-08-301-61/+106
| |
| * vaapi_h264: Convert to use the new VAAPI hwaccel codeMark Thompson2016-08-301-98/+137
| |
| * lavc: Rewrite VAAPI decode infrastructureMark Thompson2016-08-303-2/+627
| | | | | | | | | | Moves much of the setup logic for VAAPI decoding into lavc; the user now need only provide the hw_frames_ctx.
| * wmavoice: Simplify GetBitContext initializationDiego Biurrun2016-08-301-3/+1
| |
| * libvpx: Cast a pointer to const to squelch a warningDiego Biurrun2016-08-261-1/+1
| | | | | | | | | | | | | | libavcodec/libvpxdec.c:100:57: warning: passing argument 3 of 'av_image_copy' from incompatible pointer type av_image_copy(picture->data, picture->linesize, img->planes, libavutil/imgutils.h:116:6: note: expected 'const uint8_t **' but argument is of type 'unsigned char **' void av_image_copy(uint8_t *dst_data[4], int dst_linesizes[4],
* | avcodec/mjpegdec: quant_matrixes can be up to 65535, use uint16_tMichael Niedermayer2017-03-202-6/+6
| | | | | | | | | | | | | | | | Fixes invalid shift Fixes: 870/clusterfuzz-testcase-5649105424482304 Found-by: continuous fuzzing process https://github.com/google/oss-fuzz/tree/master/targets/ffmpeg Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
* | avcodec/mjpegdec: Check quant_matrixes values for being non zeroMichael Niedermayer2017-03-201-0/+4
| | | | | | | | Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
* | avcodec/vp56: Check avctx->error_concealment before enabling ECMichael Niedermayer2017-03-201-1/+1
| | | | | | | | | | | | | | | | Fixes timeout with 847/clusterfuzz-testcase-5291877358108672 Fixes timeout with 850/clusterfuzz-testcase-5721296509861888 Found-by: continuous fuzzing process https://github.com/google/oss-fuzz/tree/master/targets/ffmpeg Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
* | avcodec/h264_direct: Fix runtime error: signed integer overflow: -9 - ↵Michael Niedermayer2017-03-201-1/+6
| | | | | | | | | | | | | | | | | | | | | | | | 2147483647 cannot be represented in type 'int' Fixes: 864/clusterfuzz-testcase-4774385942528000 See: [FFmpeg-devel] [PATCH 1/2] avcodec/h264_direct: Fix runtime error: signed integer overflow: 2147483647 - -14133 cannot be represented in type 'int' See: [FFmpeg-devel] [PATCH 2/2] avcodec/h264_direct: Fix runtime error: signed integer overflow: -9 - 2147483647 cannot be represented in type 'int' Found-by: continuous fuzzing process https://github.com/google/oss-fuzz/tree/master/targets/ffmpeg Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
* | avcodec/tiff: Check stripsize strippos for overflowMichael Niedermayer2017-03-201-0/+10
| | | | | | | | | | | | | | Fixes: 861/clusterfuzz-testcase-5688284384591872 Found-by: continuous fuzzing process https://github.com/google/oss-fuzz/tree/master/targets/ffmpeg Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
* | aarch64: vp9itxfm16: Do a simpler half/quarter idct16/idct32 when possibleMartin Storsjö2017-03-191-58/+547
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This work is sponsored by, and copyright, Google. This avoids loading and calculating coefficients that we know will be zero, and avoids filling the temp buffer with zeros in places where we know the second pass won't read. This gives a pretty substantial speedup for the smaller subpartitions. The code size increases from 21512 bytes to 31400 bytes. The idct16/32_end macros are moved above the individual functions; the instructions themselves are unchanged, but since new functions are added at the same place where the code is moved from, the diff looks rather messy. Before: vp9_inv_dct_dct_16x16_sub1_add_10_neon: 284.6 vp9_inv_dct_dct_16x16_sub2_add_10_neon: 1902.7 vp9_inv_dct_dct_16x16_sub4_add_10_neon: 1903.0 vp9_inv_dct_dct_16x16_sub8_add_10_neon: 2201.1 vp9_inv_dct_dct_16x16_sub12_add_10_neon: 2510.0 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 2821.3 vp9_inv_dct_dct_32x32_sub1_add_10_neon: 1011.6 vp9_inv_dct_dct_32x32_sub2_add_10_neon: 9716.5 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 9704.9 vp9_inv_dct_dct_32x32_sub8_add_10_neon: 10641.7 vp9_inv_dct_dct_32x32_sub12_add_10_neon: 11555.7 vp9_inv_dct_dct_32x32_sub16_add_10_neon: 12499.8 vp9_inv_dct_dct_32x32_sub20_add_10_neon: 13403.7 vp9_inv_dct_dct_32x32_sub24_add_10_neon: 14335.8 vp9_inv_dct_dct_32x32_sub28_add_10_neon: 15253.6 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 16179.5 After: vp9_inv_dct_dct_16x16_sub1_add_10_neon: 282.8 vp9_inv_dct_dct_16x16_sub2_add_10_neon: 1142.4 vp9_inv_dct_dct_16x16_sub4_add_10_neon: 1139.0 vp9_inv_dct_dct_16x16_sub8_add_10_neon: 1772.9 vp9_inv_dct_dct_16x16_sub12_add_10_neon: 2515.2 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 2823.5 vp9_inv_dct_dct_32x32_sub1_add_10_neon: 1012.7 vp9_inv_dct_dct_32x32_sub2_add_10_neon: 6944.4 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 6944.2 vp9_inv_dct_dct_32x32_sub8_add_10_neon: 7609.8 vp9_inv_dct_dct_32x32_sub12_add_10_neon: 9953.4 vp9_inv_dct_dct_32x32_sub16_add_10_neon: 10770.1 vp9_inv_dct_dct_32x32_sub20_add_10_neon: 13418.8 vp9_inv_dct_dct_32x32_sub24_add_10_neon: 14330.7 vp9_inv_dct_dct_32x32_sub28_add_10_neon: 15257.1 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 16190.6 Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm: vp9itxfm16: Do a simpler half/quarter idct16/idct32 when possibleMartin Storsjö2017-03-191-60/+469
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This work is sponsored by, and copyright, Google. This avoids loading and calculating coefficients that we know will be zero, and avoids filling the temp buffer with zeros in places where we know the second pass won't read. This gives a pretty substantial speedup for the smaller subpartitions. The code size increases from 14516 bytes to 22484 bytes. The idct16/32_end macros are moved above the individual functions; the instructions themselves are unchanged, but since new functions are added at the same place where the code is moved from, the diff looks rather messy. Before: Cortex A7 A8 A9 A53 vp9_inv_dct_dct_16x16_sub1_add_10_neon: 454.0 270.7 418.5 295.4 vp9_inv_dct_dct_16x16_sub2_add_10_neon: 3840.2 3244.8 3700.1 2337.9 vp9_inv_dct_dct_16x16_sub4_add_10_neon: 4212.5 3575.4 3996.9 2571.6 vp9_inv_dct_dct_16x16_sub8_add_10_neon: 5174.4 4270.5 4615.5 3031.9 vp9_inv_dct_dct_16x16_sub12_add_10_neon: 5676.0 4908.5 5226.5 3491.3 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 6403.9 5589.0 5839.8 3948.5 vp9_inv_dct_dct_32x32_sub1_add_10_neon: 1710.7 944.7 1582.1 1045.4 vp9_inv_dct_dct_32x32_sub2_add_10_neon: 21040.7 16706.1 18687.7 13193.1 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 22197.7 18282.7 19577.5 13918.6 vp9_inv_dct_dct_32x32_sub8_add_10_neon: 24511.5 20911.5 21472.5 15367.5 vp9_inv_dct_dct_32x32_sub12_add_10_neon: 26939.5 24264.3 23239.1 16830.3 vp9_inv_dct_dct_32x32_sub16_add_10_neon: 29419.5 26845.1 25020.6 18259.9 vp9_inv_dct_dct_32x32_sub20_add_10_neon: 31146.4 29633.5 26803.3 19721.7 vp9_inv_dct_dct_32x32_sub24_add_10_neon: 33376.3 32507.8 28642.4 21174.2 vp9_inv_dct_dct_32x32_sub28_add_10_neon: 35629.4 35439.6 30416.5 22625.7 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 37269.9 37914.9 32271.9 24078.9 After: vp9_inv_dct_dct_16x16_sub1_add_10_neon: 454.0 276.0 418.5 295.1 vp9_inv_dct_dct_16x16_sub2_add_10_neon: 2336.2 1886.0 2251.0 1458.6 vp9_inv_dct_dct_16x16_sub4_add_10_neon: 2531.0 2054.7 2402.8 1591.1 vp9_inv_dct_dct_16x16_sub8_add_10_neon: 3848.6 3491.1 3845.7 2554.8 vp9_inv_dct_dct_16x16_sub12_add_10_neon: 5703.8 4831.6 5230.8 3493.4 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 6399.5 5567.0 5832.4 3951.5 vp9_inv_dct_dct_32x32_sub1_add_10_neon: 1722.1 938.5 1577.3 1044.5 vp9_inv_dct_dct_32x32_sub2_add_10_neon: 15003.5 11576.8 13105.8 9602.2 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 15768.5 12677.2 13726.0 10138.1 vp9_inv_dct_dct_32x32_sub8_add_10_neon: 17278.8 14825.4 14907.5 11185.7 vp9_inv_dct_dct_32x32_sub12_add_10_neon: 22335.7 21544.5 20379.5 15019.8 vp9_inv_dct_dct_32x32_sub16_add_10_neon: 24165.6 23881.7 21938.6 16308.2 vp9_inv_dct_dct_32x32_sub20_add_10_neon: 31082.2 30860.9 26835.3 19711.3 vp9_inv_dct_dct_32x32_sub24_add_10_neon: 33102.6 31922.8 28638.3 21161.0 vp9_inv_dct_dct_32x32_sub28_add_10_neon: 35104.9 34867.5 30411.7 22621.2 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 37438.1 39103.4 32217.8 24067.6 Signed-off-by: Martin Storsjö <martin@martin.st>
* | aarch64: vp9itxfm16: Move the load_add_store macro out from the itxfm16 ↵Martin Storsjö2017-03-191-49/+49
| | | | | | | | | | | | | | | | | | pass2 function This allows reusing the macro for a separate implementation of the pass2 function. Signed-off-by: Martin Storsjö <martin@martin.st>
* | aarch64: vp9itxfm16: Make the larger core transforms standalone functionsMartin Storsjö2017-03-191-17/+28
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This work is sponsored by, and copyright, Google. This reduces the code size of libavcodec/aarch64/vp9itxfm_16bpp_neon.o from 26288 to 21512 bytes. This gives a small slowdown of a couple of tens of cycles, but makes it more feasible to add more optimized versions of these transforms. Before: vp9_inv_dct_dct_16x16_sub4_add_10_neon: 1887.4 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 2801.5 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 9691.4 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 16154.9 After: vp9_inv_dct_dct_16x16_sub4_add_10_neon: 1899.5 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 2827.2 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 9714.7 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 16175.9 Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm: vp9itxfm16: Make the larger core transforms standalone functionsMartin Storsjö2017-03-191-16/+27
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This work is sponsored by, and copyright, Google. This reduces the code size of libavcodec/arm/vp9itxfm_16bpp_neon.o from 17500 to 14516 bytes. This gives a small slowdown of a couple tens of cycles, up to around 150 cycles for the full case of the largest transform, but makes it more feasible to add more optimized versions of these transforms. Before: Cortex A7 A8 A9 A53 vp9_inv_dct_dct_16x16_sub4_add_10_neon: 4237.4 3561.5 3971.8 2525.3 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 6371.9 5452.0 5779.3 3910.5 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 22068.8 17867.5 19555.2 13871.6 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 37268.9 38684.2 32314.2 23969.0 After: vp9_inv_dct_dct_16x16_sub4_add_10_neon: 4375.1 3571.9 4283.8 2567.2 vp9_inv_dct_dct_16x16_sub16_add_10_neon: 6415.6 5578.9 5844.6 3948.3 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 22653.7 18079.7 19603.7 13905.3 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 37593.2 38862.2 32235.8 24070.9 Signed-off-by: Martin Storsjö <martin@martin.st>
* | aarch64: vp9itxfm16: Restructure the idct32 store macrosMartin Storsjö2017-03-191-45/+45
| | | | | | | | | | | | | | This avoids concatenation, which can't be used if the whole macro is wrapped within another macro. Signed-off-by: Martin Storsjö <martin@martin.st>
* | aarch64: vp9itxfm16: Avoid .irp when it doesn't save any linesMartin Storsjö2017-03-191-12/+12
| | | | | | | | | | | | This makes the code a bit more readable. Signed-off-by: Martin Storsjö <martin@martin.st>
* | aarch64: vp9itxfm16: Fix a typo in a commentMartin Storsjö2017-03-191-1/+1
| | | | | | | | Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm: vp9itxfm16: Avoid reloading the idct32 coefficientsMartin Storsjö2017-03-191-59/+69
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Keep the idct32 coefficients in narrow form in q6-q7, and idct16 coefficients in lengthened 32 bit form in q0-q3. Avoid clobbering q0-q3 in the pass1 function, and squeeze the idct16 coefficients into q0-q1 in the pass2 function to avoid reloading them. The idct16 coefficients are clobbered and reloaded within idct32_odd though, since that turns out to be faster than narrowing them and swapping them into q6-q7. Before: Cortex A7 A8 A9 A53 vp9_inv_dct_dct_32x32_sub4_add_10_neon: 22653.8 18268.4 19598.0 14079.0 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 37699.0 38665.2 32542.3 24472.2 After: vp9_inv_dct_dct_32x32_sub4_add_10_neon: 22270.8 18159.3 19531.0 13865.0 vp9_inv_dct_dct_32x32_sub32_add_10_neon: 37523.3 37731.6 32181.7 24071.2 Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm: vp9itxfm16: Fix vertical alignmentMartin Storsjö2017-03-191-10/+10
| | | | | | | | Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm: vp9itxfm16: Use the right lane sizeMartin Storsjö2017-03-191-2/+2
| | | | | | | | | | | | | | This makes the code slightly clearer, but doesn't make any functional difference. Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm/aarch64: vp9: Fix vertical alignmentMartin Storsjö2017-03-193-26/+26
| | | | | | | | | | | | | | | | | | | | | | | | | | Align the second/third operands as they usually are. Due to the wildly varying sizes of the written out operands in aarch64 assembly, the column alignment is usually not as clear as in arm assembly. This is cherrypicked from libav commit 7995ebfad12002033c73feed422a1cfc62081e8f. Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm/aarch64: vp9itxfm: Skip loading the min_eob pointer when it won't be usedMartin Storsjö2017-03-192-3/+4
| | | | | | | | | | | | | | | | | | | | In the half/quarter cases where we don't use the min_eob array, defer loading the pointer until we know it will be needed. This is cherrypicked from libav commit 3a0d5e206d24d41d87a25ba16a79b2ea04c39d4c. Signed-off-by: Martin Storsjö <martin@martin.st>
* | arm: vp9itxfm: Template the quarter/half idct32 functionMartin Storsjö2017-03-191-37/+20
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This reduces the number of lines and reduces the duplication. Also simplify the eob check for the half case. If we are in the half case, we know we at least will need to do the first three slices, we only need to check eob for the fourth one, so we can hardcode the value to check against instead of loading from the min_eob array. Since at most one slice can be skipped in the first pass, we can unroll the loop for filling zeros completely, as it was done for the quarter case before. This allows skipping loading the min_eob pointer when using the quarter/half cases. This is cherrypicked from libav commit 98ee855ae0cc118bd1d20921d6bdb14731832462. Signed-off-by: Martin Storsjö <martin@martin.st>
* | Merge commit '721d57e608dc4fd6c86f27c5ae76ef559d646220'James Almer2017-03-198-43/+50
|\ \ | |/ | | | | | | | | | | * commit '721d57e608dc4fd6c86f27c5ae76ef559d646220': vp56: Separate VP5 and VP6 dsp initialization Merged-by: James Almer <jamrial@gmail.com>
| * vp56: Separate VP5 and VP6 dsp initializationDiego Biurrun2016-08-268-43/+49
| | | | | | | | | | VP5 has no arch-specific optimizations (nor will it get some in the future), so it makes no sense to try to share dsp init code with VP6.
* | Merge commit '3fd22538bc0e0de84b31335266b4b1577d3d609e'James Almer2017-03-195-16/+18
|\ \ | |/ | | | | | | | | | | * commit '3fd22538bc0e0de84b31335266b4b1577d3d609e': prores: Change type of stride parameters to ptrdiff_t Merged-by: James Almer <jamrial@gmail.com>
| * prores: Change type of stride parameters to ptrdiff_tDiego Biurrun2016-08-265-18/+19
| | | | | | | | | | | | | | This avoids SIMD-optimized functions having to sign-extend their line size argument manually to be able to do pointer arithmetic. Also adjust parameter names to be "linesize" everywhere.
* | Merge commit 'f81be06cf614919d71ded29b8f595bef40123ad8'James Almer2017-03-196-49/+75
|\ \ | |/ | | | | | | | | | | * commit 'f81be06cf614919d71ded29b8f595bef40123ad8': cavs: Change type of stride parameters to ptrdiff_t Merged-by: James Almer <jamrial@gmail.com>
| * cavs: Change type of stride parameters to ptrdiff_tDiego Biurrun2016-08-266-49/+75
| | | | | | | | ptrdiff_t is the correct type for array strides and similar.
* | Merge commit '802727b538b484e3f9d1345bfcc4ab24cfea8898'James Almer2017-03-193-21/+21
|\ \ | |/ | | | | | | | | | | * commit '802727b538b484e3f9d1345bfcc4ab24cfea8898': vp8: Update some assembly comments left unchanged in bd66f073fe7286bd3c Merged-by: James Almer <jamrial@gmail.com>
| * vp8: Update some assembly comments left unchanged in bd66f073fe7286bd3cDiego Biurrun2016-08-263-21/+21
| |
* | Merge commit '87c6c78604e4dd16f1f45862b27ca006da010527'James Almer2017-03-192-11/+12
|\ \ | |/ | | | | | | | | | | * commit '87c6c78604e4dd16f1f45862b27ca006da010527': vp8: Change type of stride parameters to ptrdiff_t Merged-by: James Almer <jamrial@gmail.com>
| * vp8: Change type of stride parameters to ptrdiff_tDiego Biurrun2016-08-262-10/+11
| | | | | | | | ptrdiff_t is the correct type for array strides and similar.
* | Merge commit 'd9d26a3674f31f482f54e936fcb382160830877a'James Almer2017-03-198-20/+20
|\ \ | |/ | | | | | | | | | | * commit 'd9d26a3674f31f482f54e936fcb382160830877a': vp56: Change type of stride parameters to ptrdiff_t Merged-by: James Almer <jamrial@gmail.com>
| * vp56: Change type of stride parameters to ptrdiff_tDiego Biurrun2016-08-268-20/+20
| | | | | | | | | | This avoids SIMD-optimized functions having to sign-extend their line size argument manually to be able to do pointer arithmetic.
* | Merge commit '6892df9294d93322d43255ada299507465bc93c8'Clément Bœsch2017-03-196-39/+27
|\ \ | |/ | | | | | | | | | | * commit '6892df9294d93322d43255ada299507465bc93c8': vp3: Change type of stride parameters to ptrdiff_t Merged-by: Clément Bœsch <u@pkh.me>
| * vp3: Change type of stride parameters to ptrdiff_tDiego Biurrun2016-08-266-39/+27
| | | | | | | | | | | | | | This avoids SIMD-optimized functions having to sign-extend their stride argument manually to be able to do pointer arithmetic. Also adjust parameter names to be "stride" everywhere.
* | Merge commit 'd42809f9835a4e9e5c7c63210abb09ad0ef19cfb'Clément Bœsch2017-03-192-0/+8
|\ \ | |/ | | | | | | | | | | * commit 'd42809f9835a4e9e5c7c63210abb09ad0ef19cfb': av1: Add codec_id and basic demuxing support Merged-by: Clément Bœsch <u@pkh.me>
| * av1: Add codec_id and basic demuxing supportLuca Barbato2016-08-242-0/+8
| | | | | | | | Signed-off-by: Diego Biurrun <diego@biurrun.de>
OpenPOWER on IntegriCloud