aboutsummaryrefslogtreecommitdiff
path: root/backends/platform/ds
diff options
context:
space:
mode:
authorRobin Watts2007-12-24 16:00:20 +0000
committerRobin Watts2007-12-24 16:00:20 +0000
commitaf754ff2c451eae3acd13d1b6d2ddf06a4d77348 (patch)
tree38c4ab2ef84201706c7eb7ddd88dff76f828bcbd /backends/platform/ds
parentb6ec364eae08f8dffef92808583729f6887a8914 (diff)
downloadscummvm-rg350-af754ff2c451eae3acd13d1b6d2ddf06a4d77348.tar.gz
scummvm-rg350-af754ff2c451eae3acd13d1b6d2ddf06a4d77348.tar.bz2
scummvm-rg350-af754ff2c451eae3acd13d1b6d2ddf06a4d77348.zip
New version of palettised screen scaler for DS; uses LDMs and STMs to improve
performance on slow VRAM. svn-id: r29978
Diffstat (limited to 'backends/platform/ds')
-rw-r--r--backends/platform/ds/arm9/source/blitters_arm.s216
1 files changed, 131 insertions, 85 deletions
diff --git a/backends/platform/ds/arm9/source/blitters_arm.s b/backends/platform/ds/arm9/source/blitters_arm.s
index be777993ab..93d5db7ea9 100644
--- a/backends/platform/ds/arm9/source/blitters_arm.s
+++ b/backends/platform/ds/arm9/source/blitters_arm.s
@@ -24,10 +24,10 @@
.global asmCopy8Col
.global Rescale_320x256xPAL8_To_256x256x1555
.global Rescale_320x256x1555_To_256x256x1555
- .section .itcm,"ax", %progbits
- .align 2
- .code 32
-
+ .section .itcm,"ax", %progbits
+ .align 2
+ .code 32
+
@ ARM implementation of asmDrawStripToScreen.
@
@ C prototype would be:
@@ -258,11 +258,11 @@ Rescale_320x256xPAL8_To_256x256x1555:
@ r1 = src
@ r2 = dstStride
@ r3 = srcStride
- STMFD r13!,{r4-r6,r8-r11,r14}
+ STMFD r13!,{r4-r11,r14}
MOV r8, #0x0000001F
ORR r8, r8,#0x0000FC00
ORR r8, r8,#0x03E00000 @ r8 = mask
- LDR r9, [r13,#8*4] @ r9 = palette
+ LDR r9, [r13,#9*4] @ r9 = palette
SUB r13,r13,#256*4 @ r13 = 1K of space on the stack.
MOV r5, r13 @ r5 points to this space
@@ -279,86 +279,134 @@ palLoop:
SUB r2,r2,#64*4 @ srcStride -= line length
SUB r3,r3,#64*5 @ dstStride -= line length
+ MOV r14,#0xFF @ r14= 255
MOV r5,#200 @ r5 = y
yLoop4:
- MOV r4,#32 @ r4 = x
+ MOV r4,#16 @ r4 = x
xLoop4:
- LDRH r9, [r1],#2
- LDRH r11,[r1],#2
- LDRH r6,[r1],#2
- MOV r10,r9, LSR #8
- AND r9, r9, #0xFF
- MOV r12,r11,LSR #8
- AND r11,r11,#0xFF
- AND r14,r6,#0xFF
-
- LDR r9, [r13,r9, LSL #2] @ r9 = pal[src0]
- LDR r10,[r13,r10,LSL #2] @ r10= pal[src1]
- LDR r11,[r13,r11,LSL #2] @ r11= pal[src2]
- LDR r12,[r13,r12,LSL #2] @ r12= pal[src3]
- LDR r14,[r13,r14,LSL #2] @ r13= pal[src4]
-
- ADD r9, r9, r9, LSL #1 @ r9 = 3*src0
- ADD r9, r9, r10 @ r9 = dst0<<2
- ADD r10,r10,r11 @ r10= dst1
- ADD r11,r11,r12 @ r11= dst2
- ADD r12,r12,r14 @ r12= src3 + src4
- ADD r12,r12,r14,LSL #1 @ r12= src3 + src4*3 = dst3<<2
-
- AND r9, r8, r9, LSR #2 @ r9 = dst0 (split)
- AND r10,r8, r10,LSR #1 @ r10= dst1 (split)
- AND r11,r8, r11,LSR #1 @ r11= dst2 (split)
- AND r12,r8, r12,LSR #2 @ r12= dst3 (split)
-
- ORR r9, r9, r9, ROR #16 @ r9 = dst0
- ORR r10,r10,r10,ROR #16 @ r10= dst1
- ORR r11,r11,r11,ROR #16 @ r11= dst2
- ORR r12,r12,r12,ROR #16 @ r12= dst3
-
- MOV r10,r10,LSL #16
- ORR r9, r10,r9, LSR #16
- MOV r12,r12,LSL #16
- ORR r11,r12,r11,LSR #16
- STMIA r0!,{r9,r11}
-
- LDRH r10,[r1],#2
- LDRH r12,[r1],#2
- MOV r9,r6,LSR #8
- MOV r11,r10,LSR #8
- AND r10,r10,#0xFF
- MOV r14,r12,LSR #8
- AND r12,r12,#0xFF
-
- LDR r9, [r13,r9, LSL #2] @ r9 = pal[src0]
- LDR r10,[r13,r10,LSL #2] @ r10= pal[src1]
- LDR r11,[r13,r11,LSL #2] @ r11= pal[src2]
- LDR r12,[r13,r12,LSL #2] @ r12= pal[src3]
- LDR r14,[r13,r14,LSL #2] @ r13= pal[src4]
-
- ADD r9, r9, r9, LSL #1 @ r9 = 3*src0
- ADD r9, r9, r10 @ r9 = dst0<<2
- ADD r10,r10,r11 @ r10= dst1
- ADD r11,r11,r12 @ r11= dst2
- ADD r12,r12,r14 @ r12= src3 + src4
- ADD r12,r12,r14,LSL #1 @ r12= src3 + src4*3 = dst3<<2
-
- AND r9, r8, r9, LSR #2 @ r9 = dst0 (split)
- AND r10,r8, r10,LSR #1 @ r10= dst1 (split)
- AND r11,r8, r11,LSR #1 @ r11= dst2 (split)
- AND r12,r8, r12,LSR #2 @ r12= dst3 (split)
-
- ORR r9, r9, r9, ROR #16 @ r9 = dst0
- ORR r10,r10,r10,ROR #16 @ r10= dst1
- ORR r11,r11,r11,ROR #16 @ r11= dst2
- ORR r12,r12,r12,ROR #16 @ r12= dst3
-
- MOV r10,r10,LSL #16
- ORR r9, r10,r9, LSR #16
- MOV r12,r12,LSL #16
- ORR r11,r12,r11,LSR #16
- STMIA r0!,{r9,r11}
-
+ LDMIA r1!,{r10,r11,r12}
+ AND r6, r14,r10 @ r6 = src0
+ LDR r6, [r13,r6, LSL #2] @ r6 = pal[src0]
+ AND r7, r14,r10,LSR #8 @ r7 = src1
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src1]
+ ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src0]
+ AND r9, r14,r10,LSR #16 @ r9 = src2
+ LDR r9, [r13,r9, LSL #2] @ r9 = pal[src2]
+ MOV r10,r10,LSR #24 @ r10= src3
+ LDR r10,[r13,r10,LSL #2] @ r10= pal[src3]
+ ADD r6, r6, r7 @ r6 = dst0<<2
+ AND r6, r8, r6, LSR #2 @ r6 = dst0 (split)
+ ORR r6, r6, r6, ROR #16 @ r6 = dst0 (in both halves)
+ ADD r7, r7, r9 @ r7 = dst1<<1
+ AND r7, r8, r7, LSR #1 @ r7 = dst1 (split)
+ ORR r7, r7, r7, ROR #16 @ r7 = dst1 (in both halves)
+ MOV r7, r7, LSL #16 @ r7 = dst1<<16
+ ORR r6, r7, r6, LSR #16 @ r6 = dst0 | dst1<<16
+ AND r7, r14,r11 @ r7 = src4
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src4]
+ ADD r9, r9, r10 @ r9 = dst2<<1
+ AND r9, r8, r9, LSR #1 @ r9 = dst2 (split)
+ ORR r9, r9, r9, ROR #16 @ r9 = dst2 (in both halves)
+ ADD r10,r10,r7 @ r7 = pal[src3]+pal[src4]
+ ADD r10,r10,r7, LSL #1 @ r10= dst3<<2
+ AND r10,r8, r10,LSR #2 @ r10= dst3 (split)
+ ORR r10,r10,r10,ROR #16 @ r10= dst3 (in both halves)
+ MOV r7, r9, LSR #16
+ ORR r7, r7, r10, LSL #16 @ r7 = dst2 | dst3<<16
+ STMIA r0!,{r6,r7}
+
+ AND r6, r14,r11,LSR #8 @ r6 = src5
+ LDR r6, [r13,r6, LSL #2] @ r6 = pal[src5]
+ AND r7, r14,r11,LSR #16 @ r7 = src6
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src6]
+ ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src5]
+ MOV r9, r11,LSR #24 @ r9 = src7
+ LDR r9, [r13,r9, LSL #2] @ r9 = pal[src7]
+ AND r10,r14,r12 @ r10= src8
+ LDR r10,[r13,r10,LSL #2] @ r10= pal[src8]
+ ADD r6, r6, r7 @ r6 = dst4<<2
+ AND r6, r8, r6, LSR #2 @ r6 = dst4 (split)
+ ORR r6, r6, r6, ROR #16 @ r6 = dst4 (in both halves)
+ ADD r7, r7, r9 @ r7 = dst5<<1
+ AND r7, r8, r7, LSR #1 @ r7 = dst5 (split)
+ ORR r7, r7, r7, ROR #16 @ r7 = dst5 (in both halves)
+ MOV r7, r7, LSL #16 @ r7 = dst5<<16
+ ORR r6, r7, r6, LSR #16 @ r6 = dst4 | dst5<<16
+ AND r7, r14,r12,LSR #8 @ r7 = src9
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src9]
+ ADD r9, r9, r10 @ r9 = dst6<<1
+ AND r9, r8, r9, LSR #1 @ r9 = dst6 (split)
+ ORR r9, r9, r9, ROR #16 @ r9 = dst6 (in both halves)
+ ADD r10,r10,r7 @ r10= pal[src8]+pal[src9]
+ ADD r10,r10,r7, LSL #1 @ r10= dst7<<2
+ AND r10,r8, r10,LSR #2 @ r10= dst7 (split)
+ ORR r10,r10,r10,ROR #16 @ r10= dst7 (in both halves)
+ MOV r7, r9, LSR #16
+ ORR r7, r7, r10, LSL #16 @ r7 = dst6 | dst7<<16
+ LDMIA r1!,{r10,r11}
SUBS r4,r4,#1
+ STMIA r0!,{r6,r7}
+
+ AND r6, r14,r12,LSR #16 @ r6 = src10
+ LDR r6, [r13,r6, LSL #2] @ r6 = pal[src10]
+ MOV r7, r12,LSR #24 @ r7 = src11
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src11]
+ ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src10]
+ AND r9, r14,r10 @ r9 = src12
+ LDR r9, [r13,r9, LSL #2] @ r9 = pal[src12]
+ AND r12,r14,r10,LSR #8 @ r11= src13
+ LDR r12,[r13,r12,LSL #2] @ r11= pal[src13]
+ ADD r6, r6, r7 @ r6 = dst8<<2
+ AND r6, r8, r6, LSR #2 @ r6 = dst8 (split)
+ ORR r6, r6, r6, ROR #16 @ r6 = dst8 (in both halves)
+ ADD r7, r7, r9 @ r7 = dst9<<1
+ AND r7, r8, r7, LSR #1 @ r7 = dst9 (split)
+ ORR r7, r7, r7, ROR #16 @ r7 = dst9 (in both halves)
+ MOV r7, r7, LSL #16 @ r7 = dst9<<16
+ ORR r6, r7, r6, LSR #16 @ r6 = dst8 | dst9<<16
+ AND r7, r14,r10,LSR #16 @ r7 = src14
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src14]
+ ADD r9, r9, r12 @ r9 = dst10<<1
+ AND r9, r8, r9, LSR #1 @ r9 = dst10 (split)
+ ORR r9, r9, r9, ROR #16 @ r9 = dst10 (in both halves)
+ ADD r12,r12,r7 @ r12= pal[src13]+pal[src14]
+ ADD r12,r12,r7, LSL #1 @ r12= dst11<<2
+ AND r12,r8, r12,LSR #2 @ r12= dst11 (split)
+ ORR r12,r12,r12,ROR #16 @ r12= dst11 (in both halves)
+ MOV r7, r9, LSR #16
+ ORR r7, r7, r12, LSL #16 @ r7 = dst10 | dst11<<16
+ STMIA r0!,{r6,r7}
+
+ MOV r6, r10,LSR #24 @ r6 = src15
+ LDR r6, [r13,r6, LSL #2] @ r6 = pal[src15]
+ AND r7, r14,r11 @ r7 = src16
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src16]
+ ADD r6, r6, r6, LSL #1 @ r6 = 3*pal[src15]
+ AND r9, r14,r11,LSR #8 @ r9 = src17
+ LDR r9, [r13,r9, LSL #2] @ r9 = pal[src17]
+ AND r12,r14,r11,LSR #16 @ r11= src18
+ LDR r12,[r13,r12,LSL #2] @ r11= pal[src18]
+ ADD r6, r6, r7 @ r6 = dst12<<2
+ AND r6, r8, r6, LSR #2 @ r6 = dst12 (split)
+ ORR r6, r6, r6, ROR #16 @ r6 = dst12 (in both halves)
+ ADD r7, r7, r9 @ r7 = dst13<<1
+ AND r7, r8, r7, LSR #1 @ r7 = dst13 (split)
+ ORR r7, r7, r7, ROR #16 @ r7 = dst13 (in both halves)
+ MOV r7, r7, LSL #16 @ r7 = dst13<<16
+ ORR r6, r7, r6, LSR #16 @ r6 = dst12 | dst13<<16
+ MOV r7, r11,LSR #24 @ r7 = src19
+ LDR r7, [r13,r7, LSL #2] @ r7 = pal[src19]
+ ADD r9, r9, r12 @ r9 = dst14<<1
+ AND r9, r8, r9, LSR #1 @ r9 = dst14 (split)
+ ORR r9, r9, r9, ROR #16 @ r9 = dst14 (in both halves)
+ ADD r12,r12,r7 @ r12= pal[src18]+pal[src19]
+ ADD r12,r12,r7, LSL #1 @ r12= dst15<<2
+ AND r12,r8, r12,LSR #2 @ r12= dst15 (split)
+ ORR r12,r12,r12,ROR #16 @ r12= dst15 (in both halves)
+ MOV r7, r9, LSR #16
+ ORR r7, r7, r12, LSL #16 @ r7 = dst14 | dst15<<16
+ STMIA r0!,{r6,r7}
+
BGT xLoop4
ADD r0,r0,r2,LSL #1
@@ -368,6 +416,4 @@ xLoop4:
ADD r13,r13,#256*4
- LDMFD r13!,{r4-r6,r8-r11,PC}
-
-
+ LDMFD r13!,{r4-r11,PC}