@ ScummVM Scumm Interpreter @ Copyright (C) 2007 The ScummVM project @ @ This program is free software@ you can redistribute it and/or @ modify it under the terms of the GNU General Public License @ as published by the Free Software Foundation@ either version 2 @ of the License, or (at your option) any later version. @ @ This program is distributed in the hope that it will be useful, @ but WITHOUT ANY WARRANTY; without even the implied warranty of @ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the @ GNU General Public License for more details. @ @ You should have received a copy of the GNU General Public License @ along with this program@ if not, write to the Free Software @ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. @ @ $URL: $ @ $Id: $ @ @ @author Robin Watts (robin@wss.co.uk) .global Rescale_320x256xPAL8_To_256x256x1555 .global Rescale_320x256x1555_To_256x256x1555 .section .itcm,"ax", %progbits .align 2 .code 32 @ ARM implementation of Rescale_320x256x1555_To_256x256x1555 @ @ C prototype would be: @ @ extern "C" void Rescale_320x256x1555_To_256x256x1555( @ u16 *dst, @ const u16 *src, @ int dstStride, @ int srcStride); Rescale_320x256x1555_To_256x256x1555: @ r0 = dst @ r1 = src @ r2 = dstStride @ r3 = srcStride STMFD r13!,{r4-r6,r8-r11,r14} SUB r2,r2,#64*4 @ srcStride -= line length SUB r3,r3,#64*5 @ dstStride -= line length MOV r8, #0x0000001F ORR r8, r8,#0x00007C00 ORR r8, r8,#0x03E00000 @ r8 = mask MOV r6, #0x8000 ORR r6, r6, r6, LSL #16 MOV r5, #200 @ r5 = y yLoop3: MOV r4, #64 @ r4 = x xLoop3: LDRH r9, [r1],#2 @ r9 = src0 LDRH r10,[r1],#2 @ r10= src1 LDRH r11,[r1],#2 @ r11= src2 LDRH r12,[r1],#2 @ r12= src3 LDRH r14,[r1],#2 @ r14= src4 ORR r9, r9, r9, LSL #16 @ r9 = src0 | src0 ORR r10,r10,r10,LSL #16 @ r10= src1 | src1 ORR r11,r11,r11,LSL #16 @ r11= src2 | src2 ORR r12,r12,r12,LSL #16 @ r12= src3 | src3 ORR r14,r14,r14,LSL #16 @ r13= src4 | src4 AND r9, r9, r8 @ r9 = 0 | G0 | 0 | B0 | 0 | R0 AND r10,r10,r8 @ r10= 0 | G1 | 0 | B1 | 0 | R1 AND r11,r11,r8 @ r11= 0 | G2 | 0 | B2 | 0 | R2 AND r12,r12,r8 @ r12= 0 | G3 | 0 | B3 | 0 | R3 AND r14,r14,r8 @ r14= 0 | G4 | 0 | B4 | 0 | R4 ADD r9, r9, r9, LSL #1 @ r9 = 3*src0 ADD r9, r9, r10 @ r9 = dst0<<2 ADD r10,r10,r11 @ r10= dst1 ADD r11,r11,r12 @ r11= dst2 ADD r12,r12,r14 @ r12= src3 + src4 ADD r12,r12,r14,LSL #1 @ r12= src3 + src4*3 = dst3<<2 AND r9, r8, r9, LSR #2 @ r9 = dst0 (split) AND r10,r8, r10,LSR #1 @ r10= dst1 (split) AND r11,r8, r11,LSR #1 @ r11= dst2 (split) AND r12,r8, r12,LSR #2 @ r12= dst3 (split) ORR r9, r9, r9, ROR #16 @ r9 = dst0 ORR r10,r10,r10,ROR #16 @ r10= dst1 ORR r11,r11,r11,ROR #16 @ r11= dst2 ORR r12,r12,r12,ROR #16 @ r12= dst3 ORR r10,r6, r10,LSL #16 ORR r9, r10,r9, LSR #16 ORR r12,r6, r12,LSL #16 ORR r11,r12,r11,LSR #16 STMIA r0!,{r9,r11} SUBS r4,r4,#1 BGT xLoop3 ADD r0,r0,r2,LSL #1 ADD r1,r1,r3,LSL #1 SUBS r5,r5,#1 BGT yLoop3 LDMFD r13!,{r4-r6,r8-r11,PC} @ ARM implementation of Rescale_320x256xPAL8_To_256x256x1555 @ @ C prototype would be: @ @ extern "C" void Rescale_320x256xPAL8_To_256x256x1555( @ u16 *dst, @ const u8 *src, @ int dstStride, @ int srcStride, @ const u16 *pal, @ u32 numLines); Rescale_320x256xPAL8_To_256x256x1555: @ r0 = dst @ r1 = src @ r2 = dstStride @ r3 = srcStride STMFD r13!,{r4-r11,r14} MOV r8, #0x0000001F ORR r8, r8,#0x0000FC00 ORR r8, r8,#0x03E00000 @ r8 = mask LDR r9, [r13,#9*4] @ r9 = palette LDR r7, [r13,#10*4] @ r7 = numLines SUB r13,r13,#256*4 @ r13 = 1K of space on the stack. MOV r5, r13 @ r5 points to this space MOV r14,#256 palLoop: LDRH r10,[r9],#2 @ r10 = palette entry SUBS r14,r14,#1 ORR r10,r10,r10,LSL #16 AND r10,r10,r8 @ r10 = separated palette entry ORR r10,r10,#0x00008000 STR r10,[r5], #4 BGT palLoop SUB r2,r2,#64*4 @ srcStride -= line length SUB r3,r3,#64*5 @ dstStride -= line length MOV r14,#0xFF @ r14= 255 MOV r5,r7 @ r5 = numLines yLoop4: MOV r4,#16 @ r4 = x xLoop4: LDMIA r1!,{r10,r11,r12} AND r6, r14,r10 @ r6 = src0 LDR r6, [r13,r6, LSL #2] @ r6 = pal[src0] AND r7, r14,r10,LSR #8 @ r7 = src1 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src1] ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src0] AND r9, r14,r10,LSR #16 @ r9 = src2 LDR r9, [r13,r9, LSL #2] @ r9 = pal[src2] MOV r10,r10,LSR #24 @ r10= src3 LDR r10,[r13,r10,LSL #2] @ r10= pal[src3] ADD r6, r6, r7 @ r6 = dst0<<2 AND r6, r8, r6, LSR #2 @ r6 = dst0 (split) ORR r6, r6, r6, ROR #16 @ r6 = dst0 (in both halves) ADD r7, r7, r9 @ r7 = dst1<<1 AND r7, r8, r7, LSR #1 @ r7 = dst1 (split) ORR r7, r7, r7, ROR #16 @ r7 = dst1 (in both halves) MOV r7, r7, LSL #16 @ r7 = dst1<<16 ORR r6, r7, r6, LSR #16 @ r6 = dst0 | dst1<<16 AND r7, r14,r11 @ r7 = src4 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src4] ADD r9, r9, r10 @ r9 = dst2<<1 AND r9, r8, r9, LSR #1 @ r9 = dst2 (split) ORR r9, r9, r9, ROR #16 @ r9 = dst2 (in both halves) ADD r10,r10,r7 @ r7 = pal[src3]+pal[src4] ADD r10,r10,r7, LSL #1 @ r10= dst3<<2 AND r10,r8, r10,LSR #2 @ r10= dst3 (split) ORR r10,r10,r10,ROR #16 @ r10= dst3 (in both halves) MOV r7, r9, LSR #16 ORR r7, r7, r10, LSL #16 @ r7 = dst2 | dst3<<16 STMIA r0!,{r6,r7} AND r6, r14,r11,LSR #8 @ r6 = src5 LDR r6, [r13,r6, LSL #2] @ r6 = pal[src5] AND r7, r14,r11,LSR #16 @ r7 = src6 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src6] ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src5] MOV r9, r11,LSR #24 @ r9 = src7 LDR r9, [r13,r9, LSL #2] @ r9 = pal[src7] AND r10,r14,r12 @ r10= src8 LDR r10,[r13,r10,LSL #2] @ r10= pal[src8] ADD r6, r6, r7 @ r6 = dst4<<2 AND r6, r8, r6, LSR #2 @ r6 = dst4 (split) ORR r6, r6, r6, ROR #16 @ r6 = dst4 (in both halves) ADD r7, r7, r9 @ r7 = dst5<<1 AND r7, r8, r7, LSR #1 @ r7 = dst5 (split) ORR r7, r7, r7, ROR #16 @ r7 = dst5 (in both halves) MOV r7, r7, LSL #16 @ r7 = dst5<<16 ORR r6, r7, r6, LSR #16 @ r6 = dst4 | dst5<<16 AND r7, r14,r12,LSR #8 @ r7 = src9 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src9] ADD r9, r9, r10 @ r9 = dst6<<1 AND r9, r8, r9, LSR #1 @ r9 = dst6 (split) ORR r9, r9, r9, ROR #16 @ r9 = dst6 (in both halves) ADD r10,r10,r7 @ r10= pal[src8]+pal[src9] ADD r10,r10,r7, LSL #1 @ r10= dst7<<2 AND r10,r8, r10,LSR #2 @ r10= dst7 (split) ORR r10,r10,r10,ROR #16 @ r10= dst7 (in both halves) MOV r7, r9, LSR #16 ORR r7, r7, r10, LSL #16 @ r7 = dst6 | dst7<<16 LDMIA r1!,{r10,r11} SUBS r4,r4,#1 STMIA r0!,{r6,r7} AND r6, r14,r12,LSR #16 @ r6 = src10 LDR r6, [r13,r6, LSL #2] @ r6 = pal[src10] MOV r7, r12,LSR #24 @ r7 = src11 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src11] ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src10] AND r9, r14,r10 @ r9 = src12 LDR r9, [r13,r9, LSL #2] @ r9 = pal[src12] AND r12,r14,r10,LSR #8 @ r11= src13 LDR r12,[r13,r12,LSL #2] @ r11= pal[src13] ADD r6, r6, r7 @ r6 = dst8<<2 AND r6, r8, r6, LSR #2 @ r6 = dst8 (split) ORR r6, r6, r6, ROR #16 @ r6 = dst8 (in both halves) ADD r7, r7, r9 @ r7 = dst9<<1 AND r7, r8, r7, LSR #1 @ r7 = dst9 (split) ORR r7, r7, r7, ROR #16 @ r7 = dst9 (in both halves) MOV r7, r7, LSL #16 @ r7 = dst9<<16 ORR r6, r7, r6, LSR #16 @ r6 = dst8 | dst9<<16 AND r7, r14,r10,LSR #16 @ r7 = src14 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src14] ADD r9, r9, r12 @ r9 = dst10<<1 AND r9, r8, r9, LSR #1 @ r9 = dst10 (split) ORR r9, r9, r9, ROR #16 @ r9 = dst10 (in both halves) ADD r12,r12,r7 @ r12= pal[src13]+pal[src14] ADD r12,r12,r7, LSL #1 @ r12= dst11<<2 AND r12,r8, r12,LSR #2 @ r12= dst11 (split) ORR r12,r12,r12,ROR #16 @ r12= dst11 (in both halves) MOV r7, r9, LSR #16 ORR r7, r7, r12, LSL #16 @ r7 = dst10 | dst11<<16 STMIA r0!,{r6,r7} MOV r6, r10,LSR #24 @ r6 = src15 LDR r6, [r13,r6, LSL #2] @ r6 = pal[src15] AND r7, r14,r11 @ r7 = src16 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src16] ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src15] AND r9, r14,r11,LSR #8 @ r9 = src17 LDR r9, [r13,r9, LSL #2] @ r9 = pal[src17] AND r12,r14,r11,LSR #16 @ r11= src18 LDR r12,[r13,r12,LSL #2] @ r11= pal[src18] ADD r6, r6, r7 @ r6 = dst12<<2 AND r6, r8, r6, LSR #2 @ r6 = dst12 (split) ORR r6, r6, r6, ROR #16 @ r6 = dst12 (in both halves) ADD r7, r7, r9 @ r7 = dst13<<1 AND r7, r8, r7, LSR #1 @ r7 = dst13 (split) ORR r7, r7, r7, ROR #16 @ r7 = dst13 (in both halves) MOV r7, r7, LSL #16 @ r7 = dst13<<16 ORR r6, r7, r6, LSR #16 @ r6 = dst12 | dst13<<16 MOV r7, r11,LSR #24 @ r7 = src19 LDR r7, [r13,r7, LSL #2] @ r7 = pal[src19] ADD r9, r9, r12 @ r9 = dst14<<1 AND r9, r8, r9, LSR #1 @ r9 = dst14 (split) ORR r9, r9, r9, ROR #16 @ r9 = dst14 (in both halves) ADD r12,r12,r7 @ r12= pal[src18]+pal[src19] ADD r12,r12,r7, LSL #1 @ r12= dst15<<2 AND r12,r8, r12,LSR #2 @ r12= dst15 (split) ORR r12,r12,r12,ROR #16 @ r12= dst15 (in both halves) MOV r7, r9, LSR #16 ORR r7, r7, r12, LSL #16 @ r7 = dst14 | dst15<<16 STMIA r0!,{r6,r7} BGT xLoop4 ADD r0,r0,r2,LSL #1 ADD r1,r1,r3 SUBS r5,r5,#1 BGT yLoop4 ADD r13,r13,#256*4 LDMFD r13!,{r4-r11,PC}