Mercurial > libavcodec.hg
comparison ppc/h264_altivec.c @ 7333:a8a79f5385f6 libavcodec
cosmetics: Reformat PPC code in libavcodec according to style guidelines.
This includes indentation changes, comment reformatting, consistent brace
placement and some prettyprinting.
author | diego |
---|---|
date | Sun, 20 Jul 2008 18:58:30 +0000 |
parents | f7cbb7733146 |
children | 266d4949aa15 |
comparison
equal
deleted
inserted
replaced
7332:b1003e468c3d | 7333:a8a79f5385f6 |
---|---|
194 const vec_s16_t vB = vec_splat((vec_s16_t)vABCD, 3); | 194 const vec_s16_t vB = vec_splat((vec_s16_t)vABCD, 3); |
195 const vec_s16_t vC = vec_splat((vec_s16_t)vABCD, 5); | 195 const vec_s16_t vC = vec_splat((vec_s16_t)vABCD, 5); |
196 const vec_s16_t vD = vec_splat((vec_s16_t)vABCD, 7); | 196 const vec_s16_t vD = vec_splat((vec_s16_t)vABCD, 7); |
197 LOAD_ZERO; | 197 LOAD_ZERO; |
198 const vec_s16_t v28ss = vec_sub(vec_sl(vec_splat_s16(1),vec_splat_u16(5)),vec_splat_s16(4)); | 198 const vec_s16_t v28ss = vec_sub(vec_sl(vec_splat_s16(1),vec_splat_u16(5)),vec_splat_s16(4)); |
199 const vec_u16_t v6us = vec_splat_u16(6); | 199 const vec_u16_t v6us = vec_splat_u16(6); |
200 register int loadSecond = (((unsigned long)src) % 16) <= 7 ? 0 : 1; | 200 register int loadSecond = (((unsigned long)src) % 16) <= 7 ? 0 : 1; |
201 register int reallyBadAlign = (((unsigned long)src) % 16) == 15 ? 1 : 0; | 201 register int reallyBadAlign = (((unsigned long)src) % 16) == 15 ? 1 : 0; |
202 | 202 |
203 vec_u8_t vsrcAuc, vsrcBuc, vsrcperm0, vsrcperm1; | 203 vec_u8_t vsrcAuc, vsrcBuc, vsrcperm0, vsrcperm1; |
204 vec_u8_t vsrc0uc, vsrc1uc; | 204 vec_u8_t vsrc0uc, vsrc1uc; |
390 /* Implemented but could be faster | 390 /* Implemented but could be faster |
391 #define put_pixels16_l2_altivec(d,s1,s2,ds,s1s,h) put_pixels16_l2(d,s1,s2,ds,s1s,16,h) | 391 #define put_pixels16_l2_altivec(d,s1,s2,ds,s1s,h) put_pixels16_l2(d,s1,s2,ds,s1s,16,h) |
392 #define avg_pixels16_l2_altivec(d,s1,s2,ds,s1s,h) avg_pixels16_l2(d,s1,s2,ds,s1s,16,h) | 392 #define avg_pixels16_l2_altivec(d,s1,s2,ds,s1s,h) avg_pixels16_l2(d,s1,s2,ds,s1s,16,h) |
393 */ | 393 */ |
394 | 394 |
395 H264_MC(put_, 16, altivec) | 395 H264_MC(put_, 16, altivec) |
396 H264_MC(avg_, 16, altivec) | 396 H264_MC(avg_, 16, altivec) |
397 | 397 |
398 | 398 |
399 /**************************************************************************** | 399 /**************************************************************************** |
400 * IDCT transform: | 400 * IDCT transform: |
401 ****************************************************************************/ | 401 ****************************************************************************/ |
683 r13 = vec_mergel(r2, r6); /*2,6,10,14 set 1*/ \ | 683 r13 = vec_mergel(r2, r6); /*2,6,10,14 set 1*/ \ |
684 r14 = vec_mergeh(r3, r7); /*3,7,11,15 set 0*/ \ | 684 r14 = vec_mergeh(r3, r7); /*3,7,11,15 set 0*/ \ |
685 r15 = vec_mergel(r3, r7); /*3,7,11,15 set 1*/ \ | 685 r15 = vec_mergel(r3, r7); /*3,7,11,15 set 1*/ \ |
686 \ | 686 \ |
687 /*Third merge*/ \ | 687 /*Third merge*/ \ |
688 r0 = vec_mergeh(r8, r12); /*0,2,4,6,8,10,12,14 set 0*/ \ | 688 r0 = vec_mergeh(r8, r12); /*0,2,4,6,8,10,12,14 set 0*/ \ |
689 r1 = vec_mergel(r8, r12); /*0,2,4,6,8,10,12,14 set 1*/ \ | 689 r1 = vec_mergel(r8, r12); /*0,2,4,6,8,10,12,14 set 1*/ \ |
690 r2 = vec_mergeh(r9, r13); /*0,2,4,6,8,10,12,14 set 2*/ \ | 690 r2 = vec_mergeh(r9, r13); /*0,2,4,6,8,10,12,14 set 2*/ \ |
691 r4 = vec_mergeh(r10, r14); /*1,3,5,7,9,11,13,15 set 0*/ \ | 691 r4 = vec_mergeh(r10, r14); /*1,3,5,7,9,11,13,15 set 0*/ \ |
692 r5 = vec_mergel(r10, r14); /*1,3,5,7,9,11,13,15 set 1*/ \ | 692 r5 = vec_mergel(r10, r14); /*1,3,5,7,9,11,13,15 set 1*/ \ |
693 r6 = vec_mergeh(r11, r15); /*1,3,5,7,9,11,13,15 set 2*/ \ | 693 r6 = vec_mergeh(r11, r15); /*1,3,5,7,9,11,13,15 set 2*/ \ |
694 /* Don't need to compute 3 and 7*/ \ | 694 /* Don't need to compute 3 and 7*/ \ |
695 \ | 695 \ |