diff options
author | Luca Barbato <lu_zero@gentoo.org> | 2015-05-12 00:45:37 +0200 |
---|---|---|
committer | Luca Barbato <lu_zero@gentoo.org> | 2015-05-31 12:07:10 +0200 |
commit | 7014b65995b1fe6188fb53447bf61b08e3963355 (patch) | |
tree | 534bb2b02c1e5b0d204a4a5eb68de8a472c38509 /libavcodec/ppc | |
parent | 72cebae0d981dde144340cf51f3c323f01e215e5 (diff) | |
download | ffmpeg-7014b65995b1fe6188fb53447bf61b08e3963355.tar.gz |
ppc: pixblockdsp: Use the abriged vector types
Diffstat (limited to 'libavcodec/ppc')
-rw-r--r-- | libavcodec/ppc/pixblockdsp.c | 43 |
1 files changed, 20 insertions, 23 deletions
diff --git a/libavcodec/ppc/pixblockdsp.c b/libavcodec/ppc/pixblockdsp.c index 698d655fc6..a500a70885 100644 --- a/libavcodec/ppc/pixblockdsp.c +++ b/libavcodec/ppc/pixblockdsp.c @@ -39,24 +39,22 @@ static void get_pixels_altivec(int16_t *restrict block, const uint8_t *pixels, int line_size) { int i; - vector unsigned char perm = vec_lvsl(0, pixels); - const vector unsigned char zero = - (const vector unsigned char) vec_splat_u8(0); + vec_u8 perm = vec_lvsl(0, pixels); + const vec_u8 zero = (const vec_u8)vec_splat_u8(0); for (i = 0; i < 8; i++) { /* Read potentially unaligned pixels. * We're reading 16 pixels, and actually only want 8, * but we simply ignore the extras. */ - vector unsigned char pixl = vec_ld(0, pixels); - vector unsigned char pixr = vec_ld(7, pixels); - vector unsigned char bytes = vec_perm(pixl, pixr, perm); + vec_u8 pixl = vec_ld(0, pixels); + vec_u8 pixr = vec_ld(7, pixels); + vec_u8 bytes = vec_perm(pixl, pixr, perm); // Convert the bytes into shorts. - vector signed short shorts = (vector signed short) vec_mergeh(zero, - bytes); + vec_s16 shorts = (vec_s16)vec_mergeh(zero, bytes); // Save the data to the block, we assume the block is 16-byte aligned. - vec_st(shorts, i * 16, (vector signed short *) block); + vec_st(shorts, i * 16, (vec_s16 *)block); pixels += line_size; } @@ -66,22 +64,21 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1, const uint8_t *s2, int stride) { int i; - vector unsigned char perm1 = vec_lvsl(0, s1); - vector unsigned char perm2 = vec_lvsl(0, s2); - const vector unsigned char zero = - (const vector unsigned char) vec_splat_u8(0); - vector signed short shorts1, shorts2; + vec_u8 perm1 = vec_lvsl(0, s1); + vec_u8 perm2 = vec_lvsl(0, s2); + const vec_u8 zero = (const vec_u8)vec_splat_u8(0); + vec_s16 shorts1, shorts2; for (i = 0; i < 4; i++) { /* Read potentially unaligned pixels. * We're reading 16 pixels, and actually only want 8, * but we simply ignore the extras. */ - vector unsigned char pixl = vec_ld(0, s1); - vector unsigned char pixr = vec_ld(15, s1); - vector unsigned char bytes = vec_perm(pixl, pixr, perm1); + vec_u8 pixl = vec_ld(0, s1); + vec_u8 pixr = vec_ld(15, s1); + vec_u8 bytes = vec_perm(pixl, pixr, perm1); // Convert the bytes into shorts. - shorts1 = (vector signed short) vec_mergeh(zero, bytes); + shorts1 = (vec_s16)vec_mergeh(zero, bytes); // Do the same for the second block of pixels. pixl = vec_ld(0, s2); @@ -89,13 +86,13 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1, bytes = vec_perm(pixl, pixr, perm2); // Convert the bytes into shorts. - shorts2 = (vector signed short) vec_mergeh(zero, bytes); + shorts2 = (vec_s16)vec_mergeh(zero, bytes); // Do the subtraction. shorts1 = vec_sub(shorts1, shorts2); // Save the data to the block, we assume the block is 16-byte aligned. - vec_st(shorts1, 0, (vector signed short *) block); + vec_st(shorts1, 0, (vec_s16 *)block); s1 += stride; s2 += stride; @@ -112,7 +109,7 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1, bytes = vec_perm(pixl, pixr, perm1); // Convert the bytes into shorts. - shorts1 = (vector signed short) vec_mergeh(zero, bytes); + shorts1 = (vec_s16)vec_mergeh(zero, bytes); // Do the same for the second block of pixels. pixl = vec_ld(0, s2); @@ -120,13 +117,13 @@ static void diff_pixels_altivec(int16_t *restrict block, const uint8_t *s1, bytes = vec_perm(pixl, pixr, perm2); // Convert the bytes into shorts. - shorts2 = (vector signed short) vec_mergeh(zero, bytes); + shorts2 = (vec_s16)vec_mergeh(zero, bytes); // Do the subtraction. shorts1 = vec_sub(shorts1, shorts2); // Save the data to the block, we assume the block is 16-byte aligned. - vec_st(shorts1, 0, (vector signed short *) block); + vec_st(shorts1, 0, (vec_s16 *)block); s1 += stride; s2 += stride; |