summaryrefslogtreecommitdiff
path: root/firmware/target/arm
diff options
context:
space:
mode:
authorMichael Sevakis <jethead71@rockbox.org>2007-09-25 02:59:42 +0000
committerMichael Sevakis <jethead71@rockbox.org>2007-09-25 02:59:42 +0000
commit287d6223d3e4f90043e5fd36cf49d7267b53023d (patch)
tree04771c28f2b826099805f5a5ad98b8c4e41091cc /firmware/target/arm
parent24327ddf7ba62c873bd392fa5abe5c341b4a47b9 (diff)
Dithering option for mpegplayer on gigabeat-f/x and e200. Assembly IDCT for ARm just to make it all work more nicely. Move UI simulator YUV code to its core to adapt it.
git-svn-id: svn://svn.rockbox.org/rockbox/trunk@14851 a1c6a512-1295-4272-9138-f99709370657
Diffstat (limited to 'firmware/target/arm')
-rw-r--r--firmware/target/arm/lcd-as-memframe.S554
-rw-r--r--firmware/target/arm/s3c2440/gigabeat-fx/lcd-as-meg-fx.S279
-rw-r--r--firmware/target/arm/s3c2440/gigabeat-fx/lcd-meg-fx.c45
-rw-r--r--firmware/target/arm/sandisk/sansa-e200/lcd-as-e200.S279
-rw-r--r--firmware/target/arm/sandisk/sansa-e200/lcd-e200.c45
5 files changed, 628 insertions, 574 deletions
diff --git a/firmware/target/arm/lcd-as-memframe.S b/firmware/target/arm/lcd-as-memframe.S
new file mode 100644
index 0000000000..7071bec7ac
--- /dev/null
+++ b/firmware/target/arm/lcd-as-memframe.S
@@ -0,0 +1,554 @@
+/***************************************************************************
+ * __________ __ ___.
+ * Open \______ \ ____ ____ | | _\_ |__ _______ ___
+ * Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
+ * Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
+ * Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
+ * \/ \/ \/ \/ \/
+ * $Id$
+ *
+ * Copyright (C) 2007 by Michael Sevakis
+ *
+ * ARM code for memory framebuffer LCDs
+ *
+ * All files in this archive are subject to the GNU General Public License.
+ * See the file COPYING in the source tree root for full license agreement.
+ *
+ * This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
+ * KIND, either express or implied.
+ *
+ ****************************************************************************/
+
+#include "config.h"
+#include "cpu.h"
+
+/****************************************************************************
+ * void lcd_copy_buffer_rect(fb_data *dst, fb_data *src, int width,
+ * int height);
+ */
+ .section .icode, "ax", %progbits
+ .align 2
+ .global lcd_copy_buffer_rect
+ .type lcd_copy_buffer_rect, %function
+ @ r0 = dst
+ @ r1 = src
+ @ r2 = width
+ @ r3 = height
+lcd_copy_buffer_rect: @
+ stmfd sp!, { r4-r12, lr } @ save non-scratch regs
+ mov r5, r2 @ r5 = cached width
+ rsb r4, r2, #LCD_WIDTH @ r4 = LCD_WIDTH - width
+10: @ copy line @
+ subs r2, r5, #1 @ r2 = width - 1
+ beq 40f @ finish line @ one halfword? skip to trailing copy
+ tst r0, #2 @ word aligned?
+ beq 20f @ rem copy @ yes? skip to word copy
+ ldrh r6, [r1], #2 @ copy leading halfword
+ subs r2, r2, #1 @
+ strh r6, [r0], #2 @
+ ble 40f @ finish line @ next line if lt or finish
+ @ trailing halfword if eq
+20: @ rem copy @
+ add r14, r2, #1 @ get remaining width mod 16 after word
+ @ align (rw)
+ and r14, r14, #0xe @ r14 = 0 (16), 2, 4, 6, 8, 10, 12, 14
+ add pc, pc, r14, lsl #3 @ branch to 32-byte align
+ nop @
+ b 30f @ rw % 16 = 0 or 1? use octword loop
+ nop @
+ nop @
+ nop @
+ ldr r6, [r1], #4 @ rw % 16 = 2 or 3
+ subs r2, r2, #2 @
+ str r6, [r0], #4 @
+ b 25f @ copy up done @
+ ldmia r1!, { r6-r7 } @ rw % 16 = 4 or 5
+ subs r2, r2, #4 @
+ stmia r0!, { r6-r7 } @
+ b 25f @ copy up done @
+ ldmia r1!, { r6-r8 } @ rw % 16 = 6 or 7
+ subs r2, r2, #6 @
+ stmia r0!, { r6-r8 } @
+ b 25f @ copy up done @
+ ldmia r1!, { r6-r9 } @ rw % 16 = 8 or 9
+ subs r2, r2, #8 @
+ stmia r0!, { r6-r9 } @
+ b 25f @ copy up done @
+ ldmia r1!, { r6-r10 } @ rw % 16 = 10 or 11
+ subs r2, r2, #10 @
+ stmia r0!, { r6-r10 } @
+ b 25f @ copy up done @
+ ldmia r1!, { r6-r11 } @ rw % 16 = 12 or 13
+ subs r2, r2, #12 @
+ stmia r0!, { r6-r11 } @
+ b 25f @ copy up done @
+ ldmia r1!, { r6-r12 } @ rw % 16 = 14 or 15
+ subs r2, r2, #14 @
+ stmia r0!, { r6-r12 } @
+25: @ copy up done @
+ ble 40f @ finish line @ no 32-byte segments remaining?
+30: @ octword loop @ copy 16 pixels per loop
+ ldmia r1!, { r6-r12, r14 } @
+ subs r2, r2, #16 @
+ stmia r0!, { r6-r12, r14 } @
+ bgt 30b @ octword loop @
+40: @ finish line @
+ ldreqh r6, [r1], #2 @ finish last halfword if eq ...
+ add r1, r1, r4, lsl #1 @
+ streqh r6, [r0], #2 @ ...
+ add r0, r0, r4, lsl #1 @
+ subs r3, r3, #1 @ next line
+ bgt 10b @ copy line @
+ ldmfd sp!, { r4-r12, pc } @ restore regs and return
+ .ltorg @ dump constant pool
+ .size lcd_copy_buffer_rect, .-lcd_copy_buffer_rect
+
+
+/****************************************************************************
+ * void lcd_write_yuv_420_lines(fb_data *dst,
+ * unsigned char const * const src[3],
+ * int width,
+ * int stride);
+ *
+ * |R| |1.000000 -0.000001 1.402000| |Y'|
+ * |G| = |1.000000 -0.334136 -0.714136| |Pb|
+ * |B| |1.000000 1.772000 0.000000| |Pr|
+ * Scaled, normalized, rounded and tweaked to yield RGB 565:
+ * |R| |74 0 101| |Y' - 16| >> 9
+ * |G| = |74 -24 -51| |Cb - 128| >> 8
+ * |B| |74 128 0| |Cr - 128| >> 9
+ *
+ * Write four RGB565 pixels in the following order on each loop:
+ * 1 3 + > down
+ * 2 4 \/ left
+ */
+ .section .icode, "ax", %progbits
+ .align 2
+ .global lcd_write_yuv420_lines
+ .type lcd_write_yuv420_lines, %function
+lcd_write_yuv420_lines:
+ @ r0 = dst
+ @ r1 = yuv_src
+ @ r2 = width
+ @ r3 = stride
+ stmfd sp!, { r4-r12 } @ save non-scratch
+ ldmia r1, { r4, r5, r6 } @ r4 = yuv_src[0] = Y'_p
+ @ r5 = yuv_src[1] = Cb_p
+ @ r6 = yuv_src[2] = Cr_p
+ @ r1 = scratch
+ sub r3, r3, #1 @
+10: @ loop line @
+ ldrb r7, [r4], #1 @ r7 = *Y'_p++;
+ ldrb r8, [r5], #1 @ r8 = *Cb_p++;
+ ldrb r9, [r6], #1 @ r9 = *Cr_p++;
+ @
+ sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
+ add r12, r7, r7, asl #2 @ actually (Y' - 16)*37 and shift right
+ add r7, r12, r7, asl #5 @ by one less when adding - same for all
+ @
+ sub r8, r8, #128 @ Cb -= 128
+ sub r9, r9, #128 @ Cr -= 128
+ @
+ add r10, r9, r9, asl #1 @ r10 = Cr*51 + Cb*24
+ add r10, r10, r10, asl #4 @
+ add r10, r10, r8, asl #3 @
+ add r10, r10, r8, asl #4 @
+ @
+ add r11, r9, r9, asl #2 @ r9 = Cr*101
+ add r11, r11, r9, asl #5 @
+ add r9, r11, r9, asl #6 @
+ @
+ add r8, r8, #2 @ r8 = bu = (Cb*128 + 128) >> 8
+ mov r8, r8, asr #2 @
+ add r9, r9, #256 @ r9 = rv = (r9 + 256) >> 9
+ mov r9, r9, asr #9 @
+ rsb r10, r10, #128 @ r10 = guv = (-r10 + 128) >> 8
+ mov r10, r10, asr #8 @
+ @ compute R, G, and B
+ add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
+ add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
+ add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
+ @
+ orr r12, r1, r11 @ check if clamping is needed...
+ orr r12, r12, r7, asr #1 @ ...at all
+ cmp r12, #31 @
+ bls 15f @ no clamp @
+ cmp r1, #31 @ clamp b
+ mvnhi r1, r1, asr #31 @
+ andhi r1, r1, #31 @
+ cmp r11, #31 @ clamp r
+ mvnhi r11, r11, asr #31 @
+ andhi r11, r11, #31 @
+ cmp r7, #63 @ clamp g
+ mvnhi r7, r7, asr #31 @
+ andhi r7, r7, #63 @
+15: @ no clamp @
+ @
+ ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride)
+ @
+ orr r1, r1, r7, lsl #5 @ r4 |= (g << 5)
+ orr r1, r1, r11, lsl #11 @ r4 = b | (r << 11)
+ strh r1, [r0], #LCD_WIDTH @ store pixel
+ @
+ sub r7, r12, #16 @ r7 = Y = (Y' - 16)*74
+ add r12, r7, r7, asl #2 @
+ add r7, r12, r7, asl #5 @
+ @ compute R, G, and B
+ add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
+ add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
+ add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
+ @
+ orr r12, r1, r11 @ check if clamping is needed...
+ orr r12, r12, r7, asr #1 @ ...at all
+ cmp r12, #31 @
+ bls 15f @ no clamp @
+ cmp r1, #31 @ clamp b
+ mvnhi r1, r1, asr #31 @
+ andhi r1, r1, #31 @
+ cmp r11, #31 @ clamp r
+ mvnhi r11, r11, asr #31 @
+ andhi r11, r11, #31 @
+ cmp r7, #63 @ clamp g
+ mvnhi r7, r7, asr #31 @
+ andhi r7, r7, #63 @
+15: @ no clamp @
+ @
+ ldrb r12, [r4], #1 @ r12 = Y' = *(Y'_p++)
+ @
+ orr r1, r1, r11, lsl #11 @ r1 = b | (r << 11)
+ orr r1, r1, r7, lsl #5 @ r1 |= (g << 5)
+ strh r1, [r0, #-LCD_WIDTH-2] @ store pixel
+ @
+ sub r7, r12, #16 @ r7 = Y = (Y' - 16)*74
+ add r12, r7, r7, asl #2 @
+ add r7, r12, r7, asl #5 @
+ @ compute R, G, and B
+ add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
+ add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
+ add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
+ @
+ orr r12, r1, r11 @ check if clamping is needed...
+ orr r12, r12, r7, asr #1 @ ...at all
+ cmp r12, #31 @
+ bls 15f @ no clamp @
+ cmp r1, #31 @ clamp b
+ mvnhi r1, r1, asr #31 @
+ andhi r1, r1, #31 @
+ cmp r11, #31 @ clamp r
+ mvnhi r11, r11, asr #31 @
+ andhi r11, r11, #31 @
+ cmp r7, #63 @ clamp g
+ mvnhi r7, r7, asr #31 @
+ andhi r7, r7, #63 @
+15: @ no clamp @
+ @
+ ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride)
+ @
+ orr r1, r1, r7, lsl #5 @ r1 = b | (g << 5)
+ orr r1, r1, r11, lsl #11 @ r1 |= (r << 11)
+ strh r1, [r0, #LCD_WIDTH]! @ store pixel
+ @
+ sub r7, r12, #16 @ r7 = Y = (Y' - 16)*74
+ add r12, r7, r7, asl #2 @
+ add r7, r12, r7, asl #5 @
+ @ compute R, G, and B
+ add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
+ add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
+ add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
+ @
+ orr r12, r1, r11 @ check if clamping is needed...
+ orr r12, r12, r7, asr #1 @ ...at all
+ cmp r12, #31 @
+ bls 15f @ no clamp @
+ cmp r1, #31 @ clamp b
+ mvnhi r1, r1, asr #31 @
+ andhi r1, r1, #31 @
+ cmp r11, #31 @ clamp r
+ mvnhi r11, r11, asr #31 @
+ andhi r11, r11, #31 @
+ cmp r7, #63 @ clamp g
+ mvnhi r7, r7, asr #31 @
+ andhi r7, r7, #63 @
+15: @ no clamp @
+ @
+ orr r12, r1, r11, lsl #11 @ r12 = b | (r << 11)
+ orr r12, r12, r7, lsl #5 @ r12 |= (g << 5)
+ strh r12, [r0, #-2] @ store pixel
+ add r0, r0, #2*LCD_WIDTH @
+ @
+ subs r2, r2, #2 @ subtract block from width
+ bgt 10b @ loop line @
+ @
+ ldmfd sp!, { r4-r12 } @ restore registers and return
+ bx lr @
+ .ltorg @ dump constant pool
+ .size lcd_write_yuv420_lines, .-lcd_write_yuv420_lines
+
+
+/****************************************************************************
+ * void lcd_write_yuv_420_lines_odither(fb_data *dst,
+ * unsigned char const * const src[3],
+ * int width,
+ * int stride,
+ * int x_screen,
+ * int y_screen);
+ *
+ * |R| |1.000000 -0.000001 1.402000| |Y'|
+ * |G| = |1.000000 -0.334136 -0.714136| |Pb|
+ * |B| |1.000000 1.772000 0.000000| |Pr|
+ * Red scaled at twice g & b but at same precision to place it in correct
+ * bit position after multiply and leave instruction count lower.
+ * |R| |258 0 408| |Y' - 16|
+ * |G| = |149 -49 -104| |Cb - 128|
+ * |B| |149 258 0| |Cr - 128|
+ *
+ * Write four RGB565 pixels in the following order on each loop:
+ * 1 3 + > down
+ * 2 4 \/ left
+ *
+ * Kernel pattern (raw|rotated|use order):
+ * 5 3 4 2 2 6 3 7 row0 row2 > down
+ * 1 7 0 6 | 4 0 5 1 | 2 4 6 0 3 5 7 1 col0 left
+ * 4 2 5 3 | 3 7 2 6 | 3 5 7 1 2 4 6 0 col2 \/
+ * 0 6 1 7 5 1 4 0
+ */
+ .section .icode, "ax", %progbits
+ .align 2
+ .global lcd_write_yuv420_lines_odither
+ .type lcd_write_yuv420_lines_odither, %function
+lcd_write_yuv420_lines_odither:
+ @ r0 = dst
+ @ r1 = yuv_src
+ @ r2 = width
+ @ r3 = stride
+ @ [sp] = x_screen
+ @ [sp+4] = y_screen
+ stmfd sp!, { r4-r12, lr } @ save non-scratch
+ ldmia r1, { r4, r5, r6 } @ r4 = yuv_src[0] = Y'_p
+ @ r5 = yuv_src[1] = Cb_p
+ @ r6 = yuv_src[2] = Cr_p
+ @
+ sub r3, r3, #1 @
+ add r1, sp, #40 @ Line up pattern and kernel quadrant
+ ldmia r1, { r12, r14 } @
+ eor r14, r14, r12 @
+ and r14, r14, #0x2 @
+ mov r14, r14, lsl #6 @ 0x00 or 0x80
+10: @ loop line @
+ @
+ ldrb r7, [r4], #1 @ r7 = *Y'_p++;
+ ldrb r8, [r5], #1 @ r8 = *Cb_p++;
+ ldrb r9, [r6], #1 @ r9 = *Cr_p++;
+ @
+ eor r14, r14, #0x80 @ flip pattern quadrant
+ @
+ sub r7, r7, #16 @ r7 = Y = (Y' - 16)*149
+ add r12, r7, r7, asl #2 @
+ add r12, r12, r12, asl #4 @
+ add r7, r12, r7, asl #6 @
+ @
+ sub r8, r8, #128 @ Cb -= 128
+ sub r9, r9, #128 @ Cr -= 128
+ @
+ add r10, r8, r8, asl #4 @ r10 = guv = Cr*104 + Cb*49
+ add r10, r10, r8, asl #5 @
+ add r10, r10, r9, asl #3 @
+ add r10, r10, r9, asl #5 @
+ add r10, r10, r9, asl #6 @
+ @
+ mov r8, r8, asl #1 @ r8 = bu = Cb*258
+ add r8, r8, r8, asl #7 @
+ @
+ add r9, r9, r9, asl #1 @ r9 = rv = Cr*408
+ add r9, r9, r9, asl #4 @
+ mov r9, r9, asl #3 @
+ @
+ @ compute R, G, and B
+ add r1, r8, r7 @ r1 = b' = Y + bu
+ add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv
+ rsb r7, r10, r7 @ r7 = g' = Y + guv
+ @
+ @ r8 = bu, r9 = rv, r10 = guv
+ @
+ sub r12, r1, r1, lsr #5 @ r1 = 31/32*b + b/256
+ add r1, r12, r1, lsr #8 @
+ @
+ sub r12, r11, r11, lsr #5 @ r11 = 31/32*r + r/256
+ add r11, r12, r11, lsr #8 @
+ @
+ sub r12, r7, r7, lsr #6 @ r7 = 63/64*g + g/256
+ add r7, r12, r7, lsr #8 @
+ @
+ add r12, r14, #0x100 @
+ @
+ add r1, r1, r12 @ b = r1 + delta
+ add r11, r11, r12, lsl #1 @ r = r11 + delta*2
+ add r7, r7, r12, lsr #1 @ g = r7 + delta/2
+ @
+ orr r12, r1, r11, asr #1 @ check if clamping is needed...
+ orr r12, r12, r7 @ ...at all
+ movs r12, r12, asr #15 @
+ beq 15f @ no clamp @
+ movs r12, r1, asr #15 @ clamp b
+ mvnne r1, r12, lsr #15 @
+ andne r1, r1, #0x7c00 @ mask b only if clamped
+ movs r12, r11, asr #16 @ clamp r
+ mvnne r11, r12, lsr #16 @
+ movs r12, r7, asr #15 @ clamp g
+ mvnne r7, r12, lsr #15 @
+15: @ no clamp @
+ @
+ ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride)
+ @
+ and r11, r11, #0xf800 @ pack pixel
+ and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) |
+ orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) |
+ orr r1, r11, r1, lsr #10 @ (b >> 10)
+ @
+ strh r1, [r0], #LCD_WIDTH @ store pixel
+ @
+ sub r7, r12, #16 @ r7 = Y = (Y' - 16)*149
+ add r12, r7, r7, asl #2 @
+ add r12, r12, r12, asl #4 @
+ add r7, r12, r7, asl #6 @
+ @ compute R, G, and B
+ add r1, r8, r7 @ r1 = b' = Y + bu
+ add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv
+ rsb r7, r10, r7 @ r7 = g' = Y + guv
+ @
+ sub r12, r1, r1, lsr #5 @ r1 = 31/32*b' + b'/256
+ add r1, r12, r1, lsr #8 @
+ @
+ sub r12, r11, r11, lsr #5 @ r11 = 31/32*r' + r'/256
+ add r11, r12, r11, lsr #8 @
+ @
+ sub r12, r7, r7, lsr #6 @ r7 = 63/64*g' + g'/256
+ add r7, r12, r7, lsr #8 @
+ @
+ add r12, r14, #0x200 @
+ @
+ add r1, r1, r12 @ b = r1 + delta
+ add r11, r11, r12, lsl #1 @ r = r11 + delta*2
+ add r7, r7, r12, lsr #1 @ g = r7 + delta/2
+ @
+ orr r12, r1, r11, asr #1 @ check if clamping is needed...
+ orr r12, r12, r7 @ ...at all
+ movs r12, r12, asr #15 @
+ beq 15f @ no clamp @
+ movs r12, r1, asr #15 @ clamp b
+ mvnne r1, r12, lsr #15 @
+ andne r1, r1, #0x7c00 @ mask b only if clamped
+ movs r12, r11, asr #16 @ clamp r
+ mvnne r11, r12, lsr #16 @
+ movs r12, r7, asr #15 @ clamp g
+ mvnne r7, r12, lsr #15 @
+15: @ no clamp @
+ @
+ ldrb r12, [r4], #1 @ r12 = Y' = *(Y'_p++)
+ @
+ and r11, r11, #0xf800 @ pack pixel
+ and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) |
+ orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) |
+ orr r1, r11, r1, lsr #10 @ (b >> 10)
+ @
+ strh r1, [r0, #-LCD_WIDTH-2] @ store pixel
+ @
+ sub r7, r12, #16 @ r7 = Y = (Y' - 16)*149
+ add r12, r7, r7, asl #2 @
+ add r12, r12, r12, asl #4 @
+ add r7, r12, r7, asl #6 @
+ @ compute R, G, and B
+ add r1, r8, r7 @ r1 = b' = Y + bu
+ add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv
+ rsb r7, r10, r7 @ r7 = g' = Y + guv
+ @
+ @ r8 = bu, r9 = rv, r10 = guv
+ @
+ sub r12, r1, r1, lsr #5 @ r1 = 31/32*b' + b'/256
+ add r1, r12, r1, lsr #8 @
+ @
+ sub r12, r11, r11, lsr #5 @ r11 = 31/32*r' + r'/256
+ add r11, r12, r11, lsr #8 @
+ @
+ sub r12, r7, r7, lsr #6 @ r7 = 63/64*g' + g'/256
+ add r7, r12, r7, lsr #8 @
+ @
+ add r12, r14, #0x300 @
+ @
+ add r1, r1, r12 @ b = r1 + delta
+ add r11, r11, r12, lsl #1 @ r = r11 + delta*2
+ add r7, r7, r12, lsr #1 @ g = r7 + delta/2
+ @
+ orr r12, r1, r11, asr #1 @ check if clamping is needed...
+ orr r12, r12, r7 @ ...at all
+ movs r12, r12, asr #15 @
+ beq 15f @ no clamp @
+ movs r12, r1, asr #15 @ clamp b
+ mvnne r1, r12, lsr #15 @
+ andne r1, r1, #0x7c00 @ mask b only if clamped
+ movs r12, r11, asr #16 @ clamp r
+ mvnne r11, r12, lsr #16 @
+ movs r12, r7, asr #15 @ clamp g
+ mvnne r7, r12, lsr #15 @
+15: @ no clamp @
+ @
+ ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride)
+ @
+ and r11, r11, #0xf800 @ pack pixel
+ and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) |
+ orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) |
+ orr r1, r11, r1, lsr #10 @ (b >> 10)
+ @
+ strh r1, [r0, #LCD_WIDTH]! @ store pixel
+ @
+ sub r7, r12, #16 @ r7 = Y = (Y' - 16)*149
+ add r12, r7, r7, asl #2 @
+ add r12, r12, r12, asl #4 @
+ add r7, r12, r7, asl #6 @
+ @ compute R, G, and B
+ add r1, r8, r7 @ r1 = b' = Y + bu
+ add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv
+ rsb r7, r10, r7 @ r7 = g' = Y + guv
+ @
+ sub r12, r1, r1, lsr #5 @ r1 = 31/32*b + b/256
+ add r1, r12, r1, lsr #8 @
+ @
+ sub r12, r11, r11, lsr #5 @ r11 = 31/32*r + r/256
+ add r11, r12, r11, lsr #8 @
+ @
+ sub r12, r7, r7, lsr #6 @ r7 = 63/64*g + g/256
+ add r7, r12, r7, lsr #8 @
+ @
+ @ This element is zero - use r14 @
+ @
+ add r1, r1, r14 @ b = r1 + delta
+ add r11, r11, r14, lsl #1 @ r = r11 + delta*2
+ add r7, r7, r14, lsr #1 @ g = r7 + delta/2
+ @
+ orr r12, r1, r11, asr #1 @ check if clamping is needed...
+ orr r12, r12, r7 @ ...at all
+ movs r12, r12, asr #15 @
+ beq 15f @ no clamp @
+ movs r12, r1, asr #15 @ clamp b
+ mvnne r1, r12, lsr #15 @
+ andne r1, r1, #0x7c00 @ mask b only if clamped
+ movs r12, r11, asr #16 @ clamp r
+ mvnne r11, r12, lsr #16 @
+ movs r12, r7, asr #15 @ clamp g
+ mvnne r7, r12, lsr #15 @
+15: @ no clamp @
+ @
+ and r11, r11, #0xf800 @ pack pixel
+ and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) |
+ orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) |
+ orr r1, r11, r1, lsr #10 @ (b >> 10)
+ @
+ strh r1, [r0, #-2] @ store pixel
+ add r0, r0, #2*LCD_WIDTH @
+ @
+ subs r2, r2, #2 @ subtract block from width
+ bgt 10b @ loop line @
+ @
+ ldmfd sp!, { r4-r12, pc } @ restore registers and return
+ .ltorg @ dump constant pool
+ .size lcd_write_yuv420_lines_odither, .-lcd_write_yuv420_lines_odither
diff --git a/firmware/target/arm/s3c2440/gigabeat-fx/lcd-as-meg-fx.S b/firmware/target/arm/s3c2440/gigabeat-fx/lcd-as-meg-fx.S
deleted file mode 100644
index 4926c7fa79..0000000000
--- a/firmware/target/arm/s3c2440/gigabeat-fx/lcd-as-meg-fx.S
+++ /dev/null
@@ -1,279 +0,0 @@
-/***************************************************************************
- * __________ __ ___.
- * Open \______ \ ____ ____ | | _\_ |__ _______ ___
- * Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
- * Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
- * Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
- * \/ \/ \/ \/ \/
- * $Id$
- *
- * Copyright (C) 2007 by Michael Sevakis
- *
- * All files in this archive are subject to the GNU General Public License.
- * See the file COPYING in the source tree root for full license agreement.
- *
- * This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
- * KIND, either express or implied.
- *
- ****************************************************************************/
-
-#include "config.h"
-#include "cpu.h"
-
-/****************************************************************************
- * void lcd_copy_buffer_rect(fb_data *dst, fb_data *src, int width,
- * int height);
- */
- .section .icode, "ax", %progbits
- .align 2
- .global lcd_copy_buffer_rect
- .type lcd_copy_buffer_rect, %function
- @ r0 = dst
- @ r1 = src
- @ r2 = width
- @ r3 = height
-lcd_copy_buffer_rect: @
- stmfd sp!, { r4-r12, lr } @ save non-scratch regs
- mov r5, r2 @ r5 = cached width
- rsb r4, r2, #LCD_WIDTH @ r4 = LCD_WIDTH - width
-10: @ copy line @
- subs r2, r5, #1 @ r2 = width - 1
- beq 40f @ finish line @ one halfword? skip to trailing copy
- tst r0, #2 @ word aligned?
- beq 20f @ rem copy @ yes? skip to word copy
- ldrh r6, [r1], #2 @ copy leading halfword
- subs r2, r2, #1 @
- strh r6, [r0], #2 @
- ble 40f @ finish line @ next line if lt or finish
- @ trailing halfword if eq
-20: @ rem copy @
- add r14, r2, #1 @ get remaining width mod 16 after word
- @ align (rw)
- and r14, r14, #0xe @ r14 = 0 (16), 2, 4, 6, 8, 10, 12, 14
- add pc, pc, r14, lsl #3 @ branch to 32-byte align
- nop @
- b 30f @ rw % 16 = 0 or 1? use octword loop
- nop @
- nop @
- nop @
- ldr r6, [r1], #4 @ rw % 16 = 2 or 3
- subs r2, r2, #2 @
- str r6, [r0], #4 @
- b 25f @ copy up done @
- ldmia r1!, { r6-r7 } @ rw % 16 = 4 or 5
- subs r2, r2, #4 @
- stmia r0!, { r6-r7 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r8 } @ rw % 16 = 6 or 7
- subs r2, r2, #6 @
- stmia r0!, { r6-r8 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r9 } @ rw % 16 = 8 or 9
- subs r2, r2, #8 @
- stmia r0!, { r6-r9 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r10 } @ rw % 16 = 10 or 11
- subs r2, r2, #10 @
- stmia r0!, { r6-r10 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r11 } @ rw % 16 = 12 or 13
- subs r2, r2, #12 @
- stmia r0!, { r6-r11 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r12 } @ rw % 16 = 14 or 15
- subs r2, r2, #14 @
- stmia r0!, { r6-r12 } @
-25: @ copy up done @
- ble 40f @ finish line @ no 32-byte segments remaining?
-30: @ octword loop @ copy 16 pixels per loop
- ldmia r1!, { r6-r12, r14 } @
- subs r2, r2, #16 @
- stmia r0!, { r6-r12, r14 } @
- bgt 30b @ octword loop @
-40: @ finish line @
- ldreqh r6, [r1], #2 @ finish last halfword if eq ...
- add r1, r1, r4, lsl #1 @
- streqh r6, [r0], #2 @ ...
- add r0, r0, r4, lsl #1 @
- subs r3, r3, #1 @ next line
- bgt 10b @ copy line @
- ldmfd sp!, { r4-r12, pc } @ restore regs and return
- .size lcd_copy_buffer_rect, .-lcd_copy_buffer_rect
-
-
-/****************************************************************************
- * void lcd_write_yuv_420_lines(fb_data *dst,
- * unsigned char const * const src[3],
- * int width,
- * int stride);
- *
- * |R| |1.000000 -0.000001 1.402000| |Y'|
- * |G| = |1.000000 -0.334136 -0.714136| |Pb|
- * |B| |1.000000 1.772000 0.000000| |Pr|
- * Scaled, normalized, rounded and tweaked to yield RGB 565:
- * |R| |74 0 101| |Y' - 16| >> 9
- * |G| = |74 -24 -51| |Cb - 128| >> 8
- * |B| |74 128 0| |Cr - 128| >> 9
- *
- * Write four RGB565 pixels in the following order on each loop:
- * 1 3 + > down
- * 2 4 \/ left
- */
- .section .icode, "ax", %progbits
- .align 2
- .global lcd_write_yuv420_lines
- .type lcd_write_yuv420_lines, %function
-lcd_write_yuv420_lines:
- @ r0 = dst
- @ r1 = yuv_src
- @ r2 = width
- @ r3 = stride
- stmfd sp!, { r4-r12 } @ save non-scratch
- ldmia r1, { r4, r5, r6 } @ r4 = yuv_src[0] = Y'_p
- @ r5 = yuv_src[1] = Cb_p
- @ r6 = yuv_src[2] = Cr_p
- @ r1 = scratch
-10: @ loop line @
- ldrb r7, [r4] @ r7 = *Y'_p;
- ldrb r8, [r5], #1 @ r8 = *Cb_p++;
- ldrb r9, [r6], #1 @ r9 = *Cr_p++;
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @ actually (Y' - 16)*37 and shift right
- add r7, r12, r7, asl #5 @ by one less when adding - same for all
- @
- sub r8, r8, #128 @ Cb -= 128
- sub r9, r9, #128 @ Cr -= 128
- @
- add r10, r9, r9, asl #1 @ r10 = Cr*51 + Cb*24
- add r10, r10, r10, asl #4 @
- add r10, r10, r8, asl #3 @
- add r10, r10, r8, asl #4 @
- @
- add r11, r9, r9, asl #2 @ r9 = Cr*101
- add r11, r11, r9, asl #5 @
- add r9, r11, r9, asl #6 @
- @
- add r8, r8, #2 @ r8 = bu = (Cb*128 + 128) >> 8
- mov r8, r8, asr #2 @
- add r9, r9, #256 @ r9 = rv = (r9 + 256) >> 9
- mov r9, r9, asr #9 @
- rsb r10, r10, #128 @ r10 = guv = (-r10 + 128) >> 8
- mov r10, r10, asr #8 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r7, lsl #5 @ r4 |= (g << 5)
- ldrb r7, [r4, r3] @ r7 = Y' = *(Y'_p + stride)
- orr r12, r12, r11, lsl #11 @ r4 = b | (r << 11)
- strh r12, [r0] @ store pixel
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @
- add r7, r12, r7, asl #5 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r11, lsl #11 @ r12 = b | (r << 11)
- orr r12, r12, r7, lsl #5 @ r12 |= (g << 5)
- ldrb r7, [r4, #1]! @ r7 = Y' = *(++Y'_p)
- strh r12, [r0, #-2] @ store pixel
- add r0, r0, #2*LCD_WIDTH @
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @
- add r7, r12, r7, asl #5 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r7, lsl #5 @ r12 = b | (g << 5)
- ldrb r7, [r4, r3] @ r7 = Y' = *(Y'_p + stride)
- orr r12, r12, r11, lsl #11 @ r12 |= (r << 11)
- strh r12, [r0] @ store pixel
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @
- add r7, r12, r7, asl #5 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r11, lsl #11 @ r12 = b | (r << 11)
- orr r12, r12, r7, lsl #5 @ r12 |= (g << 5)
- strh r12, [r0, #-2] @ store pixel
- add r0, r0, #2*LCD_WIDTH @
- add r4, r4, #1 @
- @
- subs r2, r2, #2 @ subtract block from width
- bgt 10b @ loop line @
- @
- ldmfd sp!, { r4-r12 } @ restore registers and return
- bx lr @
- .size lcd_write_yuv420_lines, .-lcd_write_yuv420_lines
diff --git a/firmware/target/arm/s3c2440/gigabeat-fx/lcd-meg-fx.c b/firmware/target/arm/s3c2440/gigabeat-fx/lcd-meg-fx.c
index ab7c91437c..91b2eae986 100644
--- a/firmware/target/arm/s3c2440/gigabeat-fx/lcd-meg-fx.c
+++ b/firmware/target/arm/s3c2440/gigabeat-fx/lcd-meg-fx.c
@@ -9,6 +9,7 @@
static volatile bool lcd_on = true;
volatile bool lcd_poweroff = false;
+static unsigned lcd_yuv_options = 0;
/*
** These are imported from lcd-16bit.c
*/
@@ -248,11 +249,22 @@ void lcd_bitmap_transparent_part(const fb_data *src, int src_x, int src_y,
);
}
+void lcd_yuv_set_options(unsigned options)
+{
+ lcd_yuv_options = options;
+}
+
/* Line write helper function for lcd_yuv_blit. Write two lines of yuv420. */
extern void lcd_write_yuv420_lines(fb_data *dst,
unsigned char const * const src[3],
int width,
int stride);
+extern void lcd_write_yuv420_lines_odither(fb_data *dst,
+ unsigned char const * const src[3],
+ int width,
+ int stride,
+ int x_screen, /* To align dither pattern */
+ int y_screen);
/* Performance function to blit a YUV bitmap directly to the LCD */
/* For the Gigabeat - show it rotated */
/* So the LCD_WIDTH is now the height */
@@ -272,22 +284,39 @@ void lcd_yuv_blit(unsigned char * const src[3],
width &= ~1;
height >>= 1;
- fb_data *dst = (fb_data*)FRAME + x * LCD_WIDTH + (LCD_WIDTH - y) - 1;
+ y = LCD_WIDTH - 1 - y;
+ fb_data *dst = (fb_data*)FRAME + x * LCD_WIDTH + y;
z = stride*src_y;
yuv_src[0] = src[0] + z + src_x;
yuv_src[1] = src[1] + (z >> 2) + (src_x >> 1);
yuv_src[2] = src[2] + (yuv_src[1] - src[1]);
- do
+ if (lcd_yuv_options & LCD_YUV_DITHER)
+ {
+ do
+ {
+ lcd_write_yuv420_lines_odither(dst, yuv_src, width, stride, y, x);
+ yuv_src[0] += stride << 1; /* Skip down two luma lines */
+ yuv_src[1] += stride >> 1; /* Skip down one chroma line */
+ yuv_src[2] += stride >> 1;
+ dst -= 2;
+ y -= 2;
+ }
+ while (--height > 0);
+ }
+ else
{
- lcd_write_yuv420_lines(dst, yuv_src, width, stride);
- yuv_src[0] += stride << 1; /* Skip down two luma lines */
- yuv_src[1] += stride >> 1; /* Skip down one chroma line */
- yuv_src[2] += stride >> 1;
- dst -= 2;
+ do
+ {
+ lcd_write_yuv420_lines(dst, yuv_src, width, stride);
+ yuv_src[0] += stride << 1; /* Skip down two luma lines */
+ yuv_src[1] += stride >> 1; /* Skip down one chroma line */
+ yuv_src[2] += stride >> 1;
+ dst -= 2;
+ }
+ while (--height > 0);
}
- while (--height > 0);
}
void lcd_set_contrast(int val) {
diff --git a/firmware/target/arm/sandisk/sansa-e200/lcd-as-e200.S b/firmware/target/arm/sandisk/sansa-e200/lcd-as-e200.S
deleted file mode 100644
index 4926c7fa79..0000000000
--- a/firmware/target/arm/sandisk/sansa-e200/lcd-as-e200.S
+++ /dev/null
@@ -1,279 +0,0 @@
-/***************************************************************************
- * __________ __ ___.
- * Open \______ \ ____ ____ | | _\_ |__ _______ ___
- * Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
- * Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
- * Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
- * \/ \/ \/ \/ \/
- * $Id$
- *
- * Copyright (C) 2007 by Michael Sevakis
- *
- * All files in this archive are subject to the GNU General Public License.
- * See the file COPYING in the source tree root for full license agreement.
- *
- * This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
- * KIND, either express or implied.
- *
- ****************************************************************************/
-
-#include "config.h"
-#include "cpu.h"
-
-/****************************************************************************
- * void lcd_copy_buffer_rect(fb_data *dst, fb_data *src, int width,
- * int height);
- */
- .section .icode, "ax", %progbits
- .align 2
- .global lcd_copy_buffer_rect
- .type lcd_copy_buffer_rect, %function
- @ r0 = dst
- @ r1 = src
- @ r2 = width
- @ r3 = height
-lcd_copy_buffer_rect: @
- stmfd sp!, { r4-r12, lr } @ save non-scratch regs
- mov r5, r2 @ r5 = cached width
- rsb r4, r2, #LCD_WIDTH @ r4 = LCD_WIDTH - width
-10: @ copy line @
- subs r2, r5, #1 @ r2 = width - 1
- beq 40f @ finish line @ one halfword? skip to trailing copy
- tst r0, #2 @ word aligned?
- beq 20f @ rem copy @ yes? skip to word copy
- ldrh r6, [r1], #2 @ copy leading halfword
- subs r2, r2, #1 @
- strh r6, [r0], #2 @
- ble 40f @ finish line @ next line if lt or finish
- @ trailing halfword if eq
-20: @ rem copy @
- add r14, r2, #1 @ get remaining width mod 16 after word
- @ align (rw)
- and r14, r14, #0xe @ r14 = 0 (16), 2, 4, 6, 8, 10, 12, 14
- add pc, pc, r14, lsl #3 @ branch to 32-byte align
- nop @
- b 30f @ rw % 16 = 0 or 1? use octword loop
- nop @
- nop @
- nop @
- ldr r6, [r1], #4 @ rw % 16 = 2 or 3
- subs r2, r2, #2 @
- str r6, [r0], #4 @
- b 25f @ copy up done @
- ldmia r1!, { r6-r7 } @ rw % 16 = 4 or 5
- subs r2, r2, #4 @
- stmia r0!, { r6-r7 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r8 } @ rw % 16 = 6 or 7
- subs r2, r2, #6 @
- stmia r0!, { r6-r8 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r9 } @ rw % 16 = 8 or 9
- subs r2, r2, #8 @
- stmia r0!, { r6-r9 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r10 } @ rw % 16 = 10 or 11
- subs r2, r2, #10 @
- stmia r0!, { r6-r10 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r11 } @ rw % 16 = 12 or 13
- subs r2, r2, #12 @
- stmia r0!, { r6-r11 } @
- b 25f @ copy up done @
- ldmia r1!, { r6-r12 } @ rw % 16 = 14 or 15
- subs r2, r2, #14 @
- stmia r0!, { r6-r12 } @
-25: @ copy up done @
- ble 40f @ finish line @ no 32-byte segments remaining?
-30: @ octword loop @ copy 16 pixels per loop
- ldmia r1!, { r6-r12, r14 } @
- subs r2, r2, #16 @
- stmia r0!, { r6-r12, r14 } @
- bgt 30b @ octword loop @
-40: @ finish line @
- ldreqh r6, [r1], #2 @ finish last halfword if eq ...
- add r1, r1, r4, lsl #1 @
- streqh r6, [r0], #2 @ ...
- add r0, r0, r4, lsl #1 @
- subs r3, r3, #1 @ next line
- bgt 10b @ copy line @
- ldmfd sp!, { r4-r12, pc } @ restore regs and return
- .size lcd_copy_buffer_rect, .-lcd_copy_buffer_rect
-
-
-/****************************************************************************
- * void lcd_write_yuv_420_lines(fb_data *dst,
- * unsigned char const * const src[3],
- * int width,
- * int stride);
- *
- * |R| |1.000000 -0.000001 1.402000| |Y'|
- * |G| = |1.000000 -0.334136 -0.714136| |Pb|
- * |B| |1.000000 1.772000 0.000000| |Pr|
- * Scaled, normalized, rounded and tweaked to yield RGB 565:
- * |R| |74 0 101| |Y' - 16| >> 9
- * |G| = |74 -24 -51| |Cb - 128| >> 8
- * |B| |74 128 0| |Cr - 128| >> 9
- *
- * Write four RGB565 pixels in the following order on each loop:
- * 1 3 + > down
- * 2 4 \/ left
- */
- .section .icode, "ax", %progbits
- .align 2
- .global lcd_write_yuv420_lines
- .type lcd_write_yuv420_lines, %function
-lcd_write_yuv420_lines:
- @ r0 = dst
- @ r1 = yuv_src
- @ r2 = width
- @ r3 = stride
- stmfd sp!, { r4-r12 } @ save non-scratch
- ldmia r1, { r4, r5, r6 } @ r4 = yuv_src[0] = Y'_p
- @ r5 = yuv_src[1] = Cb_p
- @ r6 = yuv_src[2] = Cr_p
- @ r1 = scratch
-10: @ loop line @
- ldrb r7, [r4] @ r7 = *Y'_p;
- ldrb r8, [r5], #1 @ r8 = *Cb_p++;
- ldrb r9, [r6], #1 @ r9 = *Cr_p++;
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @ actually (Y' - 16)*37 and shift right
- add r7, r12, r7, asl #5 @ by one less when adding - same for all
- @
- sub r8, r8, #128 @ Cb -= 128
- sub r9, r9, #128 @ Cr -= 128
- @
- add r10, r9, r9, asl #1 @ r10 = Cr*51 + Cb*24
- add r10, r10, r10, asl #4 @
- add r10, r10, r8, asl #3 @
- add r10, r10, r8, asl #4 @
- @
- add r11, r9, r9, asl #2 @ r9 = Cr*101
- add r11, r11, r9, asl #5 @
- add r9, r11, r9, asl #6 @
- @
- add r8, r8, #2 @ r8 = bu = (Cb*128 + 128) >> 8
- mov r8, r8, asr #2 @
- add r9, r9, #256 @ r9 = rv = (r9 + 256) >> 9
- mov r9, r9, asr #9 @
- rsb r10, r10, #128 @ r10 = guv = (-r10 + 128) >> 8
- mov r10, r10, asr #8 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r7, lsl #5 @ r4 |= (g << 5)
- ldrb r7, [r4, r3] @ r7 = Y' = *(Y'_p + stride)
- orr r12, r12, r11, lsl #11 @ r4 = b | (r << 11)
- strh r12, [r0] @ store pixel
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @
- add r7, r12, r7, asl #5 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r11, lsl #11 @ r12 = b | (r << 11)
- orr r12, r12, r7, lsl #5 @ r12 |= (g << 5)
- ldrb r7, [r4, #1]! @ r7 = Y' = *(++Y'_p)
- strh r12, [r0, #-2] @ store pixel
- add r0, r0, #2*LCD_WIDTH @
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @
- add r7, r12, r7, asl #5 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r7, lsl #5 @ r12 = b | (g << 5)
- ldrb r7, [r4, r3] @ r7 = Y' = *(Y'_p + stride)
- orr r12, r12, r11, lsl #11 @ r12 |= (r << 11)
- strh r12, [r0] @ store pixel
- @
- sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74
- add r12, r7, r7, asl #2 @
- add r7, r12, r7, asl #5 @
- @ compute R, G, and B
- add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu
- add r11, r9, r7, asr #8 @ r11 = r = (Y >> 9) + rv
- add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv
- @
- orr r12, r1, r11 @ check if clamping is needed...
- orr r12, r12, r7, asr #1 @ ...at all
- cmp r12, #31 @
- bls 15f @ no clamp @
- cmp r1, #31 @ clamp b
- mvnhi r1, r1, asr #31 @
- andhi r1, r1, #31 @
- cmp r11, #31 @ clamp r
- mvnhi r11, r11, asr #31 @
- andhi r11, r11, #31 @
- cmp r7, #63 @ clamp g
- mvnhi r7, r7, asr #31 @
- andhi r7, r7, #63 @
-15: @ no clamp @
- @
- orr r12, r1, r11, lsl #11 @ r12 = b | (r << 11)
- orr r12, r12, r7, lsl #5 @ r12 |= (g << 5)
- strh r12, [r0, #-2] @ store pixel
- add r0, r0, #2*LCD_WIDTH @
- add r4, r4, #1 @
- @
- subs r2, r2, #2 @ subtract block from width
- bgt 10b @ loop line @
- @
- ldmfd sp!, { r4-r12 } @ restore registers and return
- bx lr @
- .size lcd_write_yuv420_lines, .-lcd_write_yuv420_lines
diff --git a/firmware/target/arm/sandisk/sansa-e200/lcd-e200.c b/firmware/target/arm/sandisk/sansa-e200/lcd-e200.c
index 3e1f74d7a1..0d3a1a3049 100644
--- a/firmware/target/arm/sandisk/sansa-e200/lcd-e200.c
+++ b/firmware/target/arm/sandisk/sansa-e200/lcd-e200.c
@@ -29,6 +29,7 @@
/* Power and display status */
static bool power_on = false; /* Is the power turned on? */
static bool display_on NOCACHEBSS_ATTR = false; /* Is the display turned on? */
+static unsigned lcd_yuv_options NOCACHEBSS_ATTR = 0;
/* Reverse Flag */
#define R_DISP_CONTROL_NORMAL 0x0004
@@ -625,11 +626,22 @@ void lcd_blit(const fb_data* data, int x, int by, int width,
(void)stride;
}
+void lcd_yuv_set_options(unsigned options)
+{
+ lcd_yuv_options = options;
+}
+
/* Line write helper function for lcd_yuv_blit. Write two lines of yuv420. */
extern void lcd_write_yuv420_lines(fb_data *dst,
unsigned char const * const src[3],
int width,
int stride);
+extern void lcd_write_yuv420_lines_odither(fb_data *dst,
+ unsigned char const * const src[3],
+ int width,
+ int stride,
+ int x_screen, /* To align dither pattern */
+ int y_screen);
/* Performance function to blit a YUV bitmap directly to the LCD */
/* For the e200 - show it rotated */
/* So the LCD_WIDTH is now the height */
@@ -647,21 +659,38 @@ void lcd_yuv_blit(unsigned char * const src[3],
width &= ~1;
height >>= 1;
+ y = LCD_WIDTH - 1 - y;
fb_data *dst = (fb_data*)lcd_driver_framebuffer +
- x * LCD_WIDTH + (LCD_WIDTH - y) - 1;
+ x * LCD_WIDTH + y;
z = stride*src_y;
yuv_src[0] = src[0] + z + src_x;
yuv_src[1] = src[1] + (z >> 2) + (src_x >> 1);
yuv_src[2] = src[2] + (yuv_src[1] - src[1]);
- do
+ if (lcd_yuv_options & LCD_YUV_DITHER)
+ {
+ do
+ {
+ lcd_write_yuv420_lines_odither(dst, yuv_src, width, stride, y, x);
+ yuv_src[0] += stride << 1; /* Skip down two luma lines */
+ yuv_src[1] += stride >> 1; /* Skip down one chroma line */
+ yuv_src[2] += stride >> 1;
+ dst -= 2;
+ y -= 2;
+ }
+ while (--height > 0);
+ }
+ else
{
- lcd_write_yuv420_lines(dst, yuv_src, width, stride);
- yuv_src[0] += stride << 1; /* Skip down two luma lines */
- yuv_src[1] += stride >> 1; /* Skip down one chroma line */
- yuv_src[2] += stride >> 1;
- dst -= 2;
+ do
+ {
+ lcd_write_yuv420_lines(dst, yuv_src, width, stride);
+ yuv_src[0] += stride << 1; /* Skip down two luma lines */
+ yuv_src[1] += stride >> 1; /* Skip down one chroma line */
+ yuv_src[2] += stride >> 1;
+ dst -= 2;
+ }
+ while (--height > 0);
}
- while (--height > 0);
}