Mercurial > libavcodec.hg
comparison ppc/h264_template_altivec.c @ 6060:b1e3368c0d5e libavcodec
Reindent
author | lu_zero |
---|---|
date | Sat, 22 Dec 2007 02:53:36 +0000 |
parents | 8c1a381bddb6 |
children | 95ac40977cf7 |
comparison
equal
deleted
inserted
replaced
6059:8c1a381bddb6 | 6060:b1e3368c0d5e |
---|---|
81 vec_u8_t vdst, ppsum, vfdst, fsum; | 81 vec_u8_t vdst, ppsum, vfdst, fsum; |
82 | 82 |
83 POWERPC_PERF_START_COUNT(PREFIX_h264_chroma_mc8_num, 1); | 83 POWERPC_PERF_START_COUNT(PREFIX_h264_chroma_mc8_num, 1); |
84 | 84 |
85 if (((unsigned long)dst) % 16 == 0) { | 85 if (((unsigned long)dst) % 16 == 0) { |
86 fperm = (vec_u8_t)AVV(0x10, 0x11, 0x12, 0x13, | 86 fperm = (vec_u8_t)AVV(0x10, 0x11, 0x12, 0x13, |
87 0x14, 0x15, 0x16, 0x17, | 87 0x14, 0x15, 0x16, 0x17, |
88 0x08, 0x09, 0x0A, 0x0B, | 88 0x08, 0x09, 0x0A, 0x0B, |
89 0x0C, 0x0D, 0x0E, 0x0F); | 89 0x0C, 0x0D, 0x0E, 0x0F); |
90 } else { | 90 } else { |
91 fperm = (vec_u8_t)AVV(0x00, 0x01, 0x02, 0x03, | 91 fperm = (vec_u8_t)AVV(0x00, 0x01, 0x02, 0x03, |
92 0x04, 0x05, 0x06, 0x07, | 92 0x04, 0x05, 0x06, 0x07, |
93 0x18, 0x19, 0x1A, 0x1B, | 93 0x18, 0x19, 0x1A, 0x1B, |
94 0x1C, 0x1D, 0x1E, 0x1F); | 94 0x1C, 0x1D, 0x1E, 0x1F); |
95 } | 95 } |
96 | 96 |
97 vsrcAuc = vec_ld(0, src); | 97 vsrcAuc = vec_ld(0, src); |
98 | 98 |
99 if (loadSecond) | 99 if (loadSecond) |
100 vsrcBuc = vec_ld(16, src); | 100 vsrcBuc = vec_ld(16, src); |
101 vsrcperm0 = vec_lvsl(0, src); | 101 vsrcperm0 = vec_lvsl(0, src); |
102 vsrcperm1 = vec_lvsl(1, src); | 102 vsrcperm1 = vec_lvsl(1, src); |
103 | 103 |
104 vsrc0uc = vec_perm(vsrcAuc, vsrcBuc, vsrcperm0); | 104 vsrc0uc = vec_perm(vsrcAuc, vsrcBuc, vsrcperm0); |
105 if (reallyBadAlign) | 105 if (reallyBadAlign) |
106 vsrc1uc = vsrcBuc; | 106 vsrc1uc = vsrcBuc; |
107 else | 107 else |
108 vsrc1uc = vec_perm(vsrcAuc, vsrcBuc, vsrcperm1); | 108 vsrc1uc = vec_perm(vsrcAuc, vsrcBuc, vsrcperm1); |
109 | 109 |
110 vsrc0ssH = (vec_s16_t)vec_mergeh(zero_u8v,(vec_u8_t)vsrc0uc); | 110 vsrc0ssH = (vec_s16_t)vec_mergeh(zero_u8v,(vec_u8_t)vsrc0uc); |
111 vsrc1ssH = (vec_s16_t)vec_mergeh(zero_u8v,(vec_u8_t)vsrc1uc); | 111 vsrc1ssH = (vec_s16_t)vec_mergeh(zero_u8v,(vec_u8_t)vsrc1uc); |
112 | 112 |
113 if (!loadSecond) {// -> !reallyBadAlign | 113 if (!loadSecond) {// -> !reallyBadAlign |
114 for (i = 0 ; i < h ; i++) { | 114 for (i = 0 ; i < h ; i++) { |
115 | 115 vsrcCuc = vec_ld(stride + 0, src); |
116 | 116 vsrc2uc = vec_perm(vsrcCuc, vsrcCuc, vsrcperm0); |
117 vsrcCuc = vec_ld(stride + 0, src); | 117 vsrc3uc = vec_perm(vsrcCuc, vsrcCuc, vsrcperm1); |
118 | 118 |
119 vsrc2uc = vec_perm(vsrcCuc, vsrcCuc, vsrcperm0); | 119 CHROMA_MC8_ALTIVEC_CORE |
120 vsrc3uc = vec_perm(vsrcCuc, vsrcCuc, vsrcperm1); | 120 } |
121 | |
122 CHROMA_MC8_ALTIVEC_CORE | |
123 } | |
124 } else { | 121 } else { |
125 vec_u8_t vsrcDuc; | 122 vec_u8_t vsrcDuc; |
126 for (i = 0 ; i < h ; i++) { | 123 for (i = 0 ; i < h ; i++) { |
127 vsrcCuc = vec_ld(stride + 0, src); | 124 vsrcCuc = vec_ld(stride + 0, src); |
128 vsrcDuc = vec_ld(stride + 16, src); | 125 vsrcDuc = vec_ld(stride + 16, src); |
129 | 126 vsrc2uc = vec_perm(vsrcCuc, vsrcDuc, vsrcperm0); |
130 vsrc2uc = vec_perm(vsrcCuc, vsrcDuc, vsrcperm0); | 127 if (reallyBadAlign) |
131 if (reallyBadAlign) | 128 vsrc3uc = vsrcDuc; |
132 vsrc3uc = vsrcDuc; | 129 else |
133 else | 130 vsrc3uc = vec_perm(vsrcCuc, vsrcDuc, vsrcperm1); |
134 vsrc3uc = vec_perm(vsrcCuc, vsrcDuc, vsrcperm1); | 131 |
135 | 132 CHROMA_MC8_ALTIVEC_CORE |
136 CHROMA_MC8_ALTIVEC_CORE | 133 } |
137 } | |
138 } | 134 } |
139 POWERPC_PERF_STOP_COUNT(PREFIX_h264_chroma_mc8_num, 1); | 135 POWERPC_PERF_STOP_COUNT(PREFIX_h264_chroma_mc8_num, 1); |
140 } | 136 } |
141 | 137 |
142 #undef CHROMA_MC8_ALTIVEC_CORE | 138 #undef CHROMA_MC8_ALTIVEC_CORE |