summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--libavcodec/ppc/pixblockdsp.c43
1 files changed, 20 insertions, 23 deletions
diff --git a/libavcodec/ppc/pixblockdsp.c b/libavcodec/ppc/pixblockdsp.c
index 698d655..a500a70 100644
--- a/libavcodec/ppc/pixblockdsp.c
+++ b/libavcodec/ppc/pixblockdsp.c
@@ -39,24 +39,22 @@ static void get_pixels_altivec(int16_t *restrict block, const uint8_t *pixels,
int line_size)
{
int i;
- vector unsigned char perm = vec_lvsl(0, pixels);
- const vector unsigned char zero =
- (const vector unsigned char) vec_splat_u8(0);
+ vec_u8 perm = vec_lvsl(0, pixels);
+ const vec_u8 zero = (const vec_u8)vec_splat_u8(0);
for (i = 0; i < 8; i++) {
/* Read potentially unaligned pixels.
* We're reading 16 pixels, and actually only want 8,
* but we simply ignore the extras. */
- vector unsigned char pixl = vec_ld(0, pixels);
- vector unsigned char pixr = vec_ld(7, pixels);
- vector unsigned char bytes = vec_perm(pixl, pixr, perm);
+ vec_u8 pixl = vec_ld(0, pixels);
+ vec_u8 pixr = vec_ld(7, pixels);
+ vec_u8 bytes = vec_perm(pixl, pixr, perm);
// Convert the bytes into shorts.
- vector signed short shorts = (vector signed short) vec_mergeh(zero,
- bytes);
+ vec_s16 shorts = (vec_s16)vec_mergeh(zero, bytes);
// Save the data to the block, we assume the block is 16-byte aligned.
- vec_st(shorts, i * 16, (vector signed short *) block);
+ vec_st(shorts, i * 16, (vec_s16 *)block);
pixels += line_size;
}
@@ -66,22 +64,21 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1,
const uint8_t *s2, int stride)
{
int i;
- vector unsigned char perm1 = vec_lvsl(0, s1);
- vector unsigned char perm2 = vec_lvsl(0, s2);
- const vector unsigned char zero =
- (const vector unsigned char) vec_splat_u8(0);
- vector signed short shorts1, shorts2;
+ vec_u8 perm1 = vec_lvsl(0, s1);
+ vec_u8 perm2 = vec_lvsl(0, s2);
+ const vec_u8 zero = (const vec_u8)vec_splat_u8(0);
+ vec_s16 shorts1, shorts2;
for (i = 0; i < 4; i++) {
/* Read potentially unaligned pixels.
* We're reading 16 pixels, and actually only want 8,
* but we simply ignore the extras. */
- vector unsigned char pixl = vec_ld(0, s1);
- vector unsigned char pixr = vec_ld(15, s1);
- vector unsigned char bytes = vec_perm(pixl, pixr, perm1);
+ vec_u8 pixl = vec_ld(0, s1);
+ vec_u8 pixr = vec_ld(15, s1);
+ vec_u8 bytes = vec_perm(pixl, pixr, perm1);
// Convert the bytes into shorts.
- shorts1 = (vector signed short) vec_mergeh(zero, bytes);
+ shorts1 = (vec_s16)vec_mergeh(zero, bytes);
// Do the same for the second block of pixels.
pixl = vec_ld(0, s2);
@@ -89,13 +86,13 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1,
bytes = vec_perm(pixl, pixr, perm2);
// Convert the bytes into shorts.
- shorts2 = (vector signed short) vec_mergeh(zero, bytes);
+ shorts2 = (vec_s16)vec_mergeh(zero, bytes);
// Do the subtraction.
shorts1 = vec_sub(shorts1, shorts2);
// Save the data to the block, we assume the block is 16-byte aligned.
- vec_st(shorts1, 0, (vector signed short *) block);
+ vec_st(shorts1, 0, (vec_s16 *)block);
s1 += stride;
s2 += stride;
@@ -112,7 +109,7 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1,
bytes = vec_perm(pixl, pixr, perm1);
// Convert the bytes into shorts.
- shorts1 = (vector signed short) vec_mergeh(zero, bytes);
+ shorts1 = (vec_s16)vec_mergeh(zero, bytes);
// Do the same for the second block of pixels.
pixl = vec_ld(0, s2);
@@ -120,13 +117,13 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1,
bytes = vec_perm(pixl, pixr, perm2);
// Convert the bytes into shorts.
- shorts2 = (vector signed short) vec_mergeh(zero, bytes);
+ shorts2 = (vec_s16)vec_mergeh(zero, bytes);
// Do the subtraction.
shorts1 = vec_sub(shorts1, shorts2);
// Save the data to the block, we assume the block is 16-byte aligned.
- vec_st(shorts1, 0, (vector signed short *) block);
+ vec_st(shorts1, 0, (vec_s16 *)block);
s1 += stride;
s2 += stride;
OpenPOWER on IntegriCloud