367 lines
10 KiB
NASM
367 lines
10 KiB
NASM
;
|
|
; jcsample.asm - downsampling (64-bit AVX2)
|
|
;
|
|
; Copyright 2009 Pierre Ossman <ossman@cendio.se> for Cendio AB
|
|
; Copyright (C) 2009, 2016, D. R. Commander.
|
|
; Copyright (C) 2015, Intel Corporation.
|
|
;
|
|
; Based on the x86 SIMD extension for IJG JPEG library
|
|
; Copyright (C) 1999-2006, MIYASAKA Masaru.
|
|
; For conditions of distribution and use, see copyright notice in jsimdext.inc
|
|
;
|
|
; This file should be assembled with NASM (Netwide Assembler),
|
|
; can *not* be assembled with Microsoft's MASM or any compatible
|
|
; assembler (including Borland's Turbo Assembler).
|
|
; NASM is available from http://nasm.sourceforge.net/ or
|
|
; http://sourceforge.net/project/showfiles.php?group_id=6208
|
|
|
|
%include "jsimdext.inc"
|
|
|
|
; --------------------------------------------------------------------------
|
|
SECTION SEG_TEXT
|
|
BITS 64
|
|
;
|
|
; Downsample pixel values of a single component.
|
|
; This version handles the common case of 2:1 horizontal and 1:1 vertical,
|
|
; without smoothing.
|
|
;
|
|
; GLOBAL(void)
|
|
; jsimd_h2v1_downsample_avx2(JDIMENSION image_width, int max_v_samp_factor,
|
|
; JDIMENSION v_samp_factor,
|
|
; JDIMENSION width_in_blocks, JSAMPARRAY input_data,
|
|
; JSAMPARRAY output_data);
|
|
;
|
|
|
|
; r10d = JDIMENSION image_width
|
|
; r11 = int max_v_samp_factor
|
|
; r12d = JDIMENSION v_samp_factor
|
|
; r13d = JDIMENSION width_in_blocks
|
|
; r14 = JSAMPARRAY input_data
|
|
; r15 = JSAMPARRAY output_data
|
|
|
|
align 32
|
|
GLOBAL_FUNCTION(jsimd_h2v1_downsample_avx2)
|
|
|
|
EXTN(jsimd_h2v1_downsample_avx2):
|
|
push rbp
|
|
mov rax, rsp
|
|
mov rbp, rsp
|
|
collect_args 6
|
|
|
|
mov ecx, r13d
|
|
shl rcx, 3 ; imul rcx,DCTSIZE (rcx = output_cols)
|
|
jz near .return
|
|
|
|
mov edx, r10d
|
|
|
|
; -- expand_right_edge
|
|
|
|
push rcx
|
|
shl rcx, 1 ; output_cols * 2
|
|
sub rcx, rdx
|
|
jle short .expand_end
|
|
|
|
mov rax, r11
|
|
test rax, rax
|
|
jle short .expand_end
|
|
|
|
cld
|
|
mov rsi, r14 ; input_data
|
|
.expandloop:
|
|
push rax
|
|
push rcx
|
|
|
|
mov rdi, JSAMPROW [rsi]
|
|
add rdi, rdx
|
|
mov al, JSAMPLE [rdi-1]
|
|
|
|
rep stosb
|
|
|
|
pop rcx
|
|
pop rax
|
|
|
|
add rsi, byte SIZEOF_JSAMPROW
|
|
dec rax
|
|
jg short .expandloop
|
|
|
|
.expand_end:
|
|
pop rcx ; output_cols
|
|
|
|
; -- h2v1_downsample
|
|
|
|
mov eax, r12d ; rowctr
|
|
test eax, eax
|
|
jle near .return
|
|
|
|
mov rdx, 0x00010000 ; bias pattern
|
|
vmovd xmm7, edx
|
|
vpshufd xmm7, xmm7, 0x00 ; xmm7={0, 1, 0, 1, 0, 1, 0, 1}
|
|
vperm2i128 ymm7, ymm7, ymm7, 0 ; ymm7={xmm7, xmm7}
|
|
vpcmpeqw ymm6, ymm6, ymm6
|
|
vpsrlw ymm6, ymm6, BYTE_BIT ; ymm6={0xFF 0x00 0xFF 0x00 ..}
|
|
|
|
mov rsi, r14 ; input_data
|
|
mov rdi, r15 ; output_data
|
|
.rowloop:
|
|
push rcx
|
|
push rdi
|
|
push rsi
|
|
|
|
mov rsi, JSAMPROW [rsi] ; inptr
|
|
mov rdi, JSAMPROW [rdi] ; outptr
|
|
|
|
cmp rcx, byte SIZEOF_YMMWORD
|
|
jae short .columnloop
|
|
|
|
.columnloop_r24:
|
|
; rcx can possibly be 8, 16, 24
|
|
cmp rcx, 24
|
|
jne .columnloop_r16
|
|
vmovdqu ymm0, YMMWORD [rsi+0*SIZEOF_YMMWORD]
|
|
vmovdqu xmm1, XMMWORD [rsi+1*SIZEOF_YMMWORD]
|
|
mov rcx, SIZEOF_YMMWORD
|
|
jmp short .downsample
|
|
|
|
.columnloop_r16:
|
|
cmp rcx, 16
|
|
jne .columnloop_r8
|
|
vmovdqu ymm0, YMMWORD [rsi+0*SIZEOF_YMMWORD]
|
|
vpxor ymm1, ymm1, ymm1
|
|
mov rcx, SIZEOF_YMMWORD
|
|
jmp short .downsample
|
|
|
|
.columnloop_r8:
|
|
vmovdqu xmm0, XMMWORD[rsi+0*SIZEOF_YMMWORD]
|
|
vpxor ymm1, ymm1, ymm1
|
|
mov rcx, SIZEOF_YMMWORD
|
|
jmp short .downsample
|
|
|
|
.columnloop:
|
|
vmovdqu ymm0, YMMWORD [rsi+0*SIZEOF_YMMWORD]
|
|
vmovdqu ymm1, YMMWORD [rsi+1*SIZEOF_YMMWORD]
|
|
|
|
.downsample:
|
|
vpsrlw ymm2, ymm0, BYTE_BIT
|
|
vpand ymm0, ymm0, ymm6
|
|
vpsrlw ymm3, ymm1, BYTE_BIT
|
|
vpand ymm1, ymm1, ymm6
|
|
|
|
vpaddw ymm0, ymm0, ymm2
|
|
vpaddw ymm1, ymm1, ymm3
|
|
vpaddw ymm0, ymm0, ymm7
|
|
vpaddw ymm1, ymm1, ymm7
|
|
vpsrlw ymm0, ymm0, 1
|
|
vpsrlw ymm1, ymm1, 1
|
|
|
|
vpackuswb ymm0, ymm0, ymm1
|
|
vpermq ymm0, ymm0, 0xd8
|
|
|
|
vmovdqu YMMWORD [rdi+0*SIZEOF_YMMWORD], ymm0
|
|
|
|
sub rcx, byte SIZEOF_YMMWORD ; outcol
|
|
add rsi, byte 2*SIZEOF_YMMWORD ; inptr
|
|
add rdi, byte 1*SIZEOF_YMMWORD ; outptr
|
|
cmp rcx, byte SIZEOF_YMMWORD
|
|
jae short .columnloop
|
|
test rcx, rcx
|
|
jnz near .columnloop_r24
|
|
|
|
pop rsi
|
|
pop rdi
|
|
pop rcx
|
|
|
|
add rsi, byte SIZEOF_JSAMPROW ; input_data
|
|
add rdi, byte SIZEOF_JSAMPROW ; output_data
|
|
dec rax ; rowctr
|
|
jg near .rowloop
|
|
|
|
.return:
|
|
vzeroupper
|
|
uncollect_args 6
|
|
pop rbp
|
|
ret
|
|
|
|
; --------------------------------------------------------------------------
|
|
;
|
|
; Downsample pixel values of a single component.
|
|
; This version handles the standard case of 2:1 horizontal and 2:1 vertical,
|
|
; without smoothing.
|
|
;
|
|
; GLOBAL(void)
|
|
; jsimd_h2v2_downsample_avx2(JDIMENSION image_width, int max_v_samp_factor,
|
|
; JDIMENSION v_samp_factor,
|
|
; JDIMENSION width_in_blocks, JSAMPARRAY input_data,
|
|
; JSAMPARRAY output_data);
|
|
;
|
|
|
|
; r10d = JDIMENSION image_width
|
|
; r11 = int max_v_samp_factor
|
|
; r12d = JDIMENSION v_samp_factor
|
|
; r13d = JDIMENSION width_in_blocks
|
|
; r14 = JSAMPARRAY input_data
|
|
; r15 = JSAMPARRAY output_data
|
|
|
|
align 32
|
|
GLOBAL_FUNCTION(jsimd_h2v2_downsample_avx2)
|
|
|
|
EXTN(jsimd_h2v2_downsample_avx2):
|
|
push rbp
|
|
mov rax, rsp
|
|
mov rbp, rsp
|
|
collect_args 6
|
|
|
|
mov ecx, r13d
|
|
shl rcx, 3 ; imul rcx,DCTSIZE (rcx = output_cols)
|
|
jz near .return
|
|
|
|
mov edx, r10d
|
|
|
|
; -- expand_right_edge
|
|
|
|
push rcx
|
|
shl rcx, 1 ; output_cols * 2
|
|
sub rcx, rdx
|
|
jle short .expand_end
|
|
|
|
mov rax, r11
|
|
test rax, rax
|
|
jle short .expand_end
|
|
|
|
cld
|
|
mov rsi, r14 ; input_data
|
|
.expandloop:
|
|
push rax
|
|
push rcx
|
|
|
|
mov rdi, JSAMPROW [rsi]
|
|
add rdi, rdx
|
|
mov al, JSAMPLE [rdi-1]
|
|
|
|
rep stosb
|
|
|
|
pop rcx
|
|
pop rax
|
|
|
|
add rsi, byte SIZEOF_JSAMPROW
|
|
dec rax
|
|
jg short .expandloop
|
|
|
|
.expand_end:
|
|
pop rcx ; output_cols
|
|
|
|
; -- h2v2_downsample
|
|
|
|
mov eax, r12d ; rowctr
|
|
test rax, rax
|
|
jle near .return
|
|
|
|
mov rdx, 0x00020001 ; bias pattern
|
|
vmovd xmm7, edx
|
|
vpcmpeqw ymm6, ymm6, ymm6
|
|
vpshufd xmm7, xmm7, 0x00 ; ymm7={1, 2, 1, 2, 1, 2, 1, 2}
|
|
vperm2i128 ymm7, ymm7, ymm7, 0
|
|
vpsrlw ymm6, ymm6, BYTE_BIT ; ymm6={0xFF 0x00 0xFF 0x00 ..}
|
|
|
|
mov rsi, r14 ; input_data
|
|
mov rdi, r15 ; output_data
|
|
.rowloop:
|
|
push rcx
|
|
push rdi
|
|
push rsi
|
|
|
|
mov rdx, JSAMPROW [rsi+0*SIZEOF_JSAMPROW] ; inptr0
|
|
mov rsi, JSAMPROW [rsi+1*SIZEOF_JSAMPROW] ; inptr1
|
|
mov rdi, JSAMPROW [rdi] ; outptr
|
|
|
|
cmp rcx, byte SIZEOF_YMMWORD
|
|
jae short .columnloop
|
|
|
|
.columnloop_r24:
|
|
cmp rcx, 24
|
|
jne .columnloop_r16
|
|
vmovdqu ymm0, YMMWORD [rdx+0*SIZEOF_YMMWORD]
|
|
vmovdqu ymm1, YMMWORD [rsi+0*SIZEOF_YMMWORD]
|
|
vmovdqu xmm2, XMMWORD [rdx+1*SIZEOF_YMMWORD]
|
|
vmovdqu xmm3, XMMWORD [rsi+1*SIZEOF_YMMWORD]
|
|
mov rcx, SIZEOF_YMMWORD
|
|
jmp short .downsample
|
|
|
|
.columnloop_r16:
|
|
cmp rcx, 16
|
|
jne .columnloop_r8
|
|
vmovdqu ymm0, YMMWORD [rdx+0*SIZEOF_YMMWORD]
|
|
vmovdqu ymm1, YMMWORD [rsi+0*SIZEOF_YMMWORD]
|
|
vpxor ymm2, ymm2, ymm2
|
|
vpxor ymm3, ymm3, ymm3
|
|
mov rcx, SIZEOF_YMMWORD
|
|
jmp short .downsample
|
|
|
|
.columnloop_r8:
|
|
vmovdqu xmm0, XMMWORD [rdx+0*SIZEOF_XMMWORD]
|
|
vmovdqu xmm1, XMMWORD [rsi+0*SIZEOF_XMMWORD]
|
|
vpxor ymm2, ymm2, ymm2
|
|
vpxor ymm3, ymm3, ymm3
|
|
mov rcx, SIZEOF_YMMWORD
|
|
jmp short .downsample
|
|
|
|
.columnloop:
|
|
vmovdqu ymm0, YMMWORD [rdx+0*SIZEOF_YMMWORD]
|
|
vmovdqu ymm1, YMMWORD [rsi+0*SIZEOF_YMMWORD]
|
|
vmovdqu ymm2, YMMWORD [rdx+1*SIZEOF_YMMWORD]
|
|
vmovdqu ymm3, YMMWORD [rsi+1*SIZEOF_YMMWORD]
|
|
|
|
.downsample:
|
|
vpand ymm4, ymm0, ymm6
|
|
vpsrlw ymm0, ymm0, BYTE_BIT
|
|
vpand ymm5, ymm1, ymm6
|
|
vpsrlw ymm1, ymm1, BYTE_BIT
|
|
vpaddw ymm0, ymm0, ymm4
|
|
vpaddw ymm1, ymm1, ymm5
|
|
|
|
vpand ymm4, ymm2, ymm6
|
|
vpsrlw ymm2, ymm2, BYTE_BIT
|
|
vpand ymm5, ymm3, ymm6
|
|
vpsrlw ymm3, ymm3, BYTE_BIT
|
|
vpaddw ymm2, ymm2, ymm4
|
|
vpaddw ymm3, ymm3, ymm5
|
|
|
|
vpaddw ymm0, ymm0, ymm1
|
|
vpaddw ymm2, ymm2, ymm3
|
|
vpaddw ymm0, ymm0, ymm7
|
|
vpaddw ymm2, ymm2, ymm7
|
|
vpsrlw ymm0, ymm0, 2
|
|
vpsrlw ymm2, ymm2, 2
|
|
|
|
vpackuswb ymm0, ymm0, ymm2
|
|
vpermq ymm0, ymm0, 0xd8
|
|
|
|
vmovdqu YMMWORD [rdi+0*SIZEOF_YMMWORD], ymm0
|
|
|
|
sub rcx, byte SIZEOF_YMMWORD ; outcol
|
|
add rdx, byte 2*SIZEOF_YMMWORD ; inptr0
|
|
add rsi, byte 2*SIZEOF_YMMWORD ; inptr1
|
|
add rdi, byte 1*SIZEOF_YMMWORD ; outptr
|
|
cmp rcx, byte SIZEOF_YMMWORD
|
|
jae near .columnloop
|
|
test rcx, rcx
|
|
jnz near .columnloop_r24
|
|
|
|
pop rsi
|
|
pop rdi
|
|
pop rcx
|
|
|
|
add rsi, byte 2*SIZEOF_JSAMPROW ; input_data
|
|
add rdi, byte 1*SIZEOF_JSAMPROW ; output_data
|
|
dec rax ; rowctr
|
|
jg near .rowloop
|
|
|
|
.return:
|
|
vzeroupper
|
|
uncollect_args 6
|
|
pop rbp
|
|
ret
|
|
|
|
; For some reason, the OS X linker does not honor the request to align the
|
|
; segment unless we do this.
|
|
align 32
|