Mercurial > libavcodec.hg
view arm/dsputil_arm_s.S @ 8823:70ea65c17597 libavcodec
Re-indent after last commit.
author | alexc |
---|---|
date | Fri, 13 Feb 2009 16:07:20 +0000 |
parents | 5698999894ec |
children | 47245bb9e85f |
line wrap: on
line source
@ @ ARMv4 optimized DSP utils @ Copyright (c) 2004 AGAWA Koji <i (AT) atty (DOT) jp> @ @ This file is part of FFmpeg. @ @ FFmpeg is free software; you can redistribute it and/or @ modify it under the terms of the GNU Lesser General Public @ License as published by the Free Software Foundation; either @ version 2.1 of the License, or (at your option) any later version. @ @ FFmpeg is distributed in the hope that it will be useful, @ but WITHOUT ANY WARRANTY; without even the implied warranty of @ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU @ Lesser General Public License for more details. @ @ You should have received a copy of the GNU Lesser General Public @ License along with FFmpeg; if not, write to the Free Software @ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA @ #include "config.h" #include "asm.S" preserve8 #if !HAVE_PLD .macro pld reg .endm #endif #if HAVE_ARMV5TE function ff_prefetch_arm, export=1 subs r2, r2, #1 pld [r0] add r0, r0, r1 bne ff_prefetch_arm bx lr .endfunc #endif .macro ADJ_ALIGN_QUADWORD_D shift, Rd0, Rd1, Rd2, Rd3, Rn0, Rn1, Rn2, Rn3, Rn4 mov \Rd0, \Rn0, lsr #(\shift * 8) mov \Rd1, \Rn1, lsr #(\shift * 8) mov \Rd2, \Rn2, lsr #(\shift * 8) mov \Rd3, \Rn3, lsr #(\shift * 8) orr \Rd0, \Rd0, \Rn1, lsl #(32 - \shift * 8) orr \Rd1, \Rd1, \Rn2, lsl #(32 - \shift * 8) orr \Rd2, \Rd2, \Rn3, lsl #(32 - \shift * 8) orr \Rd3, \Rd3, \Rn4, lsl #(32 - \shift * 8) .endm .macro ADJ_ALIGN_DOUBLEWORD shift, R0, R1, R2 mov \R0, \R0, lsr #(\shift * 8) orr \R0, \R0, \R1, lsl #(32 - \shift * 8) mov \R1, \R1, lsr #(\shift * 8) orr \R1, \R1, \R2, lsl #(32 - \shift * 8) .endm .macro ADJ_ALIGN_DOUBLEWORD_D shift, Rdst0, Rdst1, Rsrc0, Rsrc1, Rsrc2 mov \Rdst0, \Rsrc0, lsr #(\shift * 8) mov \Rdst1, \Rsrc1, lsr #(\shift * 8) orr \Rdst0, \Rdst0, \Rsrc1, lsl #(32 - (\shift * 8)) orr \Rdst1, \Rdst1, \Rsrc2, lsl #(32 - (\shift * 8)) .endm .macro RND_AVG32 Rd0, Rd1, Rn0, Rn1, Rm0, Rm1, Rmask @ Rd = (Rn | Rm) - (((Rn ^ Rm) & ~0x01010101) >> 1) @ Rmask = 0xFEFEFEFE @ Rn = destroy eor \Rd0, \Rn0, \Rm0 eor \Rd1, \Rn1, \Rm1 orr \Rn0, \Rn0, \Rm0 orr \Rn1, \Rn1, \Rm1 and \Rd0, \Rd0, \Rmask and \Rd1, \Rd1, \Rmask sub \Rd0, \Rn0, \Rd0, lsr #1 sub \Rd1, \Rn1, \Rd1, lsr #1 .endm .macro NO_RND_AVG32 Rd0, Rd1, Rn0, Rn1, Rm0, Rm1, Rmask @ Rd = (Rn & Rm) - (((Rn ^ Rm) & ~0x01010101) >> 1) @ Rmask = 0xFEFEFEFE @ Rn = destroy eor \Rd0, \Rn0, \Rm0 eor \Rd1, \Rn1, \Rm1 and \Rn0, \Rn0, \Rm0 and \Rn1, \Rn1, \Rm1 and \Rd0, \Rd0, \Rmask and \Rd1, \Rd1, \Rmask add \Rd0, \Rn0, \Rd0, lsr #1 add \Rd1, \Rn1, \Rd1, lsr #1 .endm .macro JMP_ALIGN tmp, reg ands \tmp, \reg, #3 bic \reg, \reg, #3 beq 1f subs \tmp, \tmp, #1 beq 2f subs \tmp, \tmp, #1 beq 3f b 4f .endm @ ---------------------------------------------------------------- .align 5 function put_pixels16_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r11, lr} @ R14 is also called LR JMP_ALIGN r5, r1 1: ldmia r1, {r4-r7} add r1, r1, r2 stmia r0, {r4-r7} pld [r1] subs r3, r3, #1 add r0, r0, r2 bne 1b ldmfd sp!, {r4-r11, pc} .align 5 2: ldmia r1, {r4-r8} add r1, r1, r2 ADJ_ALIGN_QUADWORD_D 1, r9, r10, r11, r12, r4, r5, r6, r7, r8 pld [r1] subs r3, r3, #1 stmia r0, {r9-r12} add r0, r0, r2 bne 2b ldmfd sp!, {r4-r11, pc} .align 5 3: ldmia r1, {r4-r8} add r1, r1, r2 ADJ_ALIGN_QUADWORD_D 2, r9, r10, r11, r12, r4, r5, r6, r7, r8 pld [r1] subs r3, r3, #1 stmia r0, {r9-r12} add r0, r0, r2 bne 3b ldmfd sp!, {r4-r11, pc} .align 5 4: ldmia r1, {r4-r8} add r1, r1, r2 ADJ_ALIGN_QUADWORD_D 3, r9, r10, r11, r12, r4, r5, r6, r7, r8 pld [r1] subs r3, r3, #1 stmia r0, {r9-r12} add r0, r0, r2 bne 4b ldmfd sp!, {r4-r11,pc} .endfunc @ ---------------------------------------------------------------- .align 5 function put_pixels8_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r5,lr} @ R14 is also called LR JMP_ALIGN r5, r1 1: ldmia r1, {r4-r5} add r1, r1, r2 subs r3, r3, #1 pld [r1] stmia r0, {r4-r5} add r0, r0, r2 bne 1b ldmfd sp!, {r4-r5,pc} .align 5 2: ldmia r1, {r4-r5, r12} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r12 pld [r1] subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 2b ldmfd sp!, {r4-r5,pc} .align 5 3: ldmia r1, {r4-r5, r12} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r12 pld [r1] subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 3b ldmfd sp!, {r4-r5,pc} .align 5 4: ldmia r1, {r4-r5, r12} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r12 pld [r1] subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 4b ldmfd sp!, {r4-r5,pc} .endfunc @ ---------------------------------------------------------------- .align 5 function put_pixels8_x2_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r10,lr} @ R14 is also called LR ldr r12, =0xfefefefe JMP_ALIGN r5, r1 1: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 pld [r1] RND_AVG32 r8, r9, r4, r5, r6, r7, r12 subs r3, r3, #1 stmia r0, {r8-r9} add r0, r0, r2 bne 1b ldmfd sp!, {r4-r10,pc} .align 5 2: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 ADJ_ALIGN_DOUBLEWORD_D 2, r8, r9, r4, r5, r10 pld [r1] RND_AVG32 r4, r5, r6, r7, r8, r9, r12 subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 2b ldmfd sp!, {r4-r10,pc} .align 5 3: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 2, r6, r7, r4, r5, r10 ADJ_ALIGN_DOUBLEWORD_D 3, r8, r9, r4, r5, r10 pld [r1] RND_AVG32 r4, r5, r6, r7, r8, r9, r12 subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 3b ldmfd sp!, {r4-r10,pc} .align 5 4: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 3, r6, r7, r4, r5, r10 pld [r1] RND_AVG32 r8, r9, r6, r7, r5, r10, r12 subs r3, r3, #1 stmia r0, {r8-r9} add r0, r0, r2 bne 4b ldmfd sp!, {r4-r10,pc} @@ update PC with LR content. .endfunc .align 5 function put_no_rnd_pixels8_x2_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r10,lr} @ R14 is also called LR ldr r12, =0xfefefefe JMP_ALIGN r5, r1 1: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 pld [r1] NO_RND_AVG32 r8, r9, r4, r5, r6, r7, r12 subs r3, r3, #1 stmia r0, {r8-r9} add r0, r0, r2 bne 1b ldmfd sp!, {r4-r10,pc} .align 5 2: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 ADJ_ALIGN_DOUBLEWORD_D 2, r8, r9, r4, r5, r10 pld [r1] NO_RND_AVG32 r4, r5, r6, r7, r8, r9, r12 subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 2b ldmfd sp!, {r4-r10,pc} .align 5 3: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 2, r6, r7, r4, r5, r10 ADJ_ALIGN_DOUBLEWORD_D 3, r8, r9, r4, r5, r10 pld [r1] NO_RND_AVG32 r4, r5, r6, r7, r8, r9, r12 subs r3, r3, #1 stmia r0, {r4-r5} add r0, r0, r2 bne 3b ldmfd sp!, {r4-r10,pc} .align 5 4: ldmia r1, {r4-r5, r10} add r1, r1, r2 ADJ_ALIGN_DOUBLEWORD_D 3, r6, r7, r4, r5, r10 pld [r1] NO_RND_AVG32 r8, r9, r6, r7, r5, r10, r12 subs r3, r3, #1 stmia r0, {r8-r9} add r0, r0, r2 bne 4b ldmfd sp!, {r4-r10,pc} @@ update PC with LR content. .endfunc @ ---------------------------------------------------------------- .align 5 function put_pixels8_y2_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r11,lr} @ R14 is also called LR mov r3, r3, lsr #1 ldr r12, =0xfefefefe JMP_ALIGN r5, r1 1: ldmia r1, {r4-r5} add r1, r1, r2 6: ldmia r1, {r6-r7} add r1, r1, r2 pld [r1] RND_AVG32 r8, r9, r4, r5, r6, r7, r12 ldmia r1, {r4-r5} add r1, r1, r2 stmia r0, {r8-r9} add r0, r0, r2 pld [r1] RND_AVG32 r8, r9, r6, r7, r4, r5, r12 subs r3, r3, #1 stmia r0, {r8-r9} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .align 5 2: ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 6: ldmia r1, {r7-r9} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 1, r7, r8, r9 RND_AVG32 r10, r11, r4, r5, r7, r8, r12 stmia r0, {r10-r11} add r0, r0, r2 ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 subs r3, r3, #1 RND_AVG32 r10, r11, r7, r8, r4, r5, r12 stmia r0, {r10-r11} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .align 5 3: ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 6: ldmia r1, {r7-r9} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 2, r7, r8, r9 RND_AVG32 r10, r11, r4, r5, r7, r8, r12 stmia r0, {r10-r11} add r0, r0, r2 ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 subs r3, r3, #1 RND_AVG32 r10, r11, r7, r8, r4, r5, r12 stmia r0, {r10-r11} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .align 5 4: ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 6: ldmia r1, {r7-r9} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 3, r7, r8, r9 RND_AVG32 r10, r11, r4, r5, r7, r8, r12 stmia r0, {r10-r11} add r0, r0, r2 ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 subs r3, r3, #1 RND_AVG32 r10, r11, r7, r8, r4, r5, r12 stmia r0, {r10-r11} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .endfunc .align 5 function put_no_rnd_pixels8_y2_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r11,lr} @ R14 is also called LR mov r3, r3, lsr #1 ldr r12, =0xfefefefe JMP_ALIGN r5, r1 1: ldmia r1, {r4-r5} add r1, r1, r2 6: ldmia r1, {r6-r7} add r1, r1, r2 pld [r1] NO_RND_AVG32 r8, r9, r4, r5, r6, r7, r12 ldmia r1, {r4-r5} add r1, r1, r2 stmia r0, {r8-r9} add r0, r0, r2 pld [r1] NO_RND_AVG32 r8, r9, r6, r7, r4, r5, r12 subs r3, r3, #1 stmia r0, {r8-r9} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .align 5 2: ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 6: ldmia r1, {r7-r9} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 1, r7, r8, r9 NO_RND_AVG32 r10, r11, r4, r5, r7, r8, r12 stmia r0, {r10-r11} add r0, r0, r2 ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 subs r3, r3, #1 NO_RND_AVG32 r10, r11, r7, r8, r4, r5, r12 stmia r0, {r10-r11} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .align 5 3: ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 6: ldmia r1, {r7-r9} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 2, r7, r8, r9 NO_RND_AVG32 r10, r11, r4, r5, r7, r8, r12 stmia r0, {r10-r11} add r0, r0, r2 ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 subs r3, r3, #1 NO_RND_AVG32 r10, r11, r7, r8, r4, r5, r12 stmia r0, {r10-r11} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .align 5 4: ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 6: ldmia r1, {r7-r9} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 3, r7, r8, r9 NO_RND_AVG32 r10, r11, r4, r5, r7, r8, r12 stmia r0, {r10-r11} add r0, r0, r2 ldmia r1, {r4-r6} add r1, r1, r2 pld [r1] ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 subs r3, r3, #1 NO_RND_AVG32 r10, r11, r7, r8, r4, r5, r12 stmia r0, {r10-r11} add r0, r0, r2 bne 6b ldmfd sp!, {r4-r11,pc} .endfunc .ltorg @ ---------------------------------------------------------------- .macro RND_XY2_IT align, rnd @ l1= (a & 0x03030303) + (b & 0x03030303) ?(+ 0x02020202) @ h1= ((a & 0xFCFCFCFCUL) >> 2) + ((b & 0xFCFCFCFCUL) >> 2) .if \align == 0 ldmia r1, {r6-r8} .elseif \align == 3 ldmia r1, {r5-r7} .else ldmia r1, {r8-r10} .endif add r1, r1, r2 pld [r1] .if \align == 0 ADJ_ALIGN_DOUBLEWORD_D 1, r4, r5, r6, r7, r8 .elseif \align == 1 ADJ_ALIGN_DOUBLEWORD_D 1, r4, r5, r8, r9, r10 ADJ_ALIGN_DOUBLEWORD_D 2, r6, r7, r8, r9, r10 .elseif \align == 2 ADJ_ALIGN_DOUBLEWORD_D 2, r4, r5, r8, r9, r10 ADJ_ALIGN_DOUBLEWORD_D 3, r6, r7, r8, r9, r10 .elseif \align == 3 ADJ_ALIGN_DOUBLEWORD_D 3, r4, r5, r5, r6, r7 .endif ldr r14, =0x03030303 tst r3, #1 and r8, r4, r14 and r9, r5, r14 and r10, r6, r14 and r11, r7, r14 andeq r14, r14, r14, \rnd #1 add r8, r8, r10 add r9, r9, r11 ldr r12, =0xfcfcfcfc >> 2 addeq r8, r8, r14 addeq r9, r9, r14 and r4, r12, r4, lsr #2 and r5, r12, r5, lsr #2 and r6, r12, r6, lsr #2 and r7, r12, r7, lsr #2 add r10, r4, r6 add r11, r5, r7 subs r3, r3, #1 .endm .macro RND_XY2_EXPAND align, rnd RND_XY2_IT \align, \rnd 6: stmfd sp!, {r8-r11} RND_XY2_IT \align, \rnd ldmfd sp!, {r4-r7} add r4, r4, r8 add r5, r5, r9 ldr r14, =0x0f0f0f0f add r6, r6, r10 add r7, r7, r11 and r4, r14, r4, lsr #2 and r5, r14, r5, lsr #2 add r4, r4, r6 add r5, r5, r7 stmia r0, {r4-r5} add r0, r0, r2 bge 6b ldmfd sp!, {r4-r11,pc} .endm .align 5 function put_pixels8_xy2_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r11,lr} @ R14 is also called LR JMP_ALIGN r5, r1 1: RND_XY2_EXPAND 0, lsl .align 5 2: RND_XY2_EXPAND 1, lsl .align 5 3: RND_XY2_EXPAND 2, lsl .align 5 4: RND_XY2_EXPAND 3, lsl .endfunc .align 5 function put_no_rnd_pixels8_xy2_arm, export=1 @ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) @ block = word aligned, pixles = unaligned pld [r1] stmfd sp!, {r4-r11,lr} @ R14 is also called LR JMP_ALIGN r5, r1 1: RND_XY2_EXPAND 0, lsr .align 5 2: RND_XY2_EXPAND 1, lsr .align 5 3: RND_XY2_EXPAND 2, lsr .align 5 4: RND_XY2_EXPAND 3, lsr .endfunc .align 5 @ void ff_add_pixels_clamped_ARM(int16_t *block, uint8_t *dest, int stride) function ff_add_pixels_clamped_ARM, export=1 push {r4-r10} mov r10, #8 1: ldr r4, [r1] /* load dest */ /* block[0] and block[1]*/ ldrsh r5, [r0] ldrsh r7, [r0, #2] and r6, r4, #0xFF and r8, r4, #0xFF00 add r6, r5, r6 add r8, r7, r8, lsr #8 mvn r5, r5 mvn r7, r7 tst r6, #0x100 movne r6, r5, lsr #24 tst r8, #0x100 movne r8, r7, lsr #24 mov r9, r6 ldrsh r5, [r0, #4] /* moved form [A] */ orr r9, r9, r8, lsl #8 /* block[2] and block[3] */ /* [A] */ ldrsh r7, [r0, #6] and r6, r4, #0xFF0000 and r8, r4, #0xFF000000 add r6, r5, r6, lsr #16 add r8, r7, r8, lsr #24 mvn r5, r5 mvn r7, r7 tst r6, #0x100 movne r6, r5, lsr #24 tst r8, #0x100 movne r8, r7, lsr #24 orr r9, r9, r6, lsl #16 ldr r4, [r1, #4] /* moved form [B] */ orr r9, r9, r8, lsl #24 /* store dest */ ldrsh r5, [r0, #8] /* moved form [C] */ str r9, [r1] /* load dest */ /* [B] */ /* block[4] and block[5] */ /* [C] */ ldrsh r7, [r0, #10] and r6, r4, #0xFF and r8, r4, #0xFF00 add r6, r5, r6 add r8, r7, r8, lsr #8 mvn r5, r5 mvn r7, r7 tst r6, #0x100 movne r6, r5, lsr #24 tst r8, #0x100 movne r8, r7, lsr #24 mov r9, r6 ldrsh r5, [r0, #12] /* moved from [D] */ orr r9, r9, r8, lsl #8 /* block[6] and block[7] */ /* [D] */ ldrsh r7, [r0, #14] and r6, r4, #0xFF0000 and r8, r4, #0xFF000000 add r6, r5, r6, lsr #16 add r8, r7, r8, lsr #24 mvn r5, r5 mvn r7, r7 tst r6, #0x100 movne r6, r5, lsr #24 tst r8, #0x100 movne r8, r7, lsr #24 orr r9, r9, r6, lsl #16 add r0, r0, #16 /* moved from [E] */ orr r9, r9, r8, lsl #24 subs r10, r10, #1 /* moved from [F] */ /* store dest */ str r9, [r1, #4] /* [E] */ /* [F] */ add r1, r1, r2 bne 1b pop {r4-r10} bx lr .endfunc