| /* |
| * Loongson MMI optimizations for libjpeg-turbo |
| * |
| * Copyright 2009 Pierre Ossman <ossman@cendio.se> for Cendio AB |
| * Copyright (C) 2014-2015, D. R. Commander. All Rights Reserved. |
| * Copyright (C) 2016-2017, Loongson Technology Corporation Limited, BeiJing. |
| * All Rights Reserved. |
| * Authors: ZhuChen <zhuchen@loongson.cn> |
| * SunZhangzhi <sunzhangzhi-cq@loongson.cn> |
| * CaiWanwei <caiwanwei@loongson.cn> |
| * |
| * Based on the x86 SIMD extension for IJG JPEG library |
| * Copyright (C) 1999-2006, MIYASAKA Masaru. |
| * |
| * This software is provided 'as-is', without any express or implied |
| * warranty. In no event will the authors be held liable for any damages |
| * arising from the use of this software. |
| * |
| * Permission is granted to anyone to use this software for any purpose, |
| * including commercial applications, and to alter it and redistribute it |
| * freely, subject to the following restrictions: |
| * |
| * 1. The origin of this software must not be misrepresented; you must not |
| * claim that you wrote the original software. If you use this software |
| * in a product, an acknowledgment in the product documentation would be |
| * appreciated but is not required. |
| * 2. Altered source versions must be plainly marked as such, and must not be |
| * misrepresented as being the original software. |
| * 3. This notice may not be removed or altered from any source distribution. |
| */ |
| |
| /* This file is included by jccolor-mmi.c */ |
| |
| |
| #if RGB_RED == 0 |
| #define mmA mm0 |
| #define mmB mm1 |
| #elif RGB_GREEN == 0 |
| #define mmA mm2 |
| #define mmB mm3 |
| #elif RGB_BLUE == 0 |
| #define mmA mm4 |
| #define mmB mm5 |
| #else |
| #define mmA mm6 |
| #define mmB mm7 |
| #endif |
| |
| #if RGB_RED == 1 |
| #define mmC mm0 |
| #define mmD mm1 |
| #elif RGB_GREEN == 1 |
| #define mmC mm2 |
| #define mmD mm3 |
| #elif RGB_BLUE == 1 |
| #define mmC mm4 |
| #define mmD mm5 |
| #else |
| #define mmC mm6 |
| #define mmD mm7 |
| #endif |
| |
| #if RGB_RED == 2 |
| #define mmE mm0 |
| #define mmF mm1 |
| #elif RGB_GREEN == 2 |
| #define mmE mm2 |
| #define mmF mm3 |
| #elif RGB_BLUE == 2 |
| #define mmE mm4 |
| #define mmF mm5 |
| #else |
| #define mmE mm6 |
| #define mmF mm7 |
| #endif |
| |
| #if RGB_RED == 3 |
| #define mmG mm0 |
| #define mmH mm1 |
| #elif RGB_GREEN == 3 |
| #define mmG mm2 |
| #define mmH mm3 |
| #elif RGB_BLUE == 3 |
| #define mmG mm4 |
| #define mmH mm5 |
| #else |
| #define mmG mm6 |
| #define mmH mm7 |
| #endif |
| |
| |
| void jsimd_rgb_ycc_convert_mmi(JDIMENSION image_width, JSAMPARRAY input_buf, |
| JSAMPIMAGE output_buf, JDIMENSION output_row, |
| int num_rows) |
| { |
| JSAMPROW inptr, outptr0, outptr1, outptr2; |
| int num_cols, col; |
| __m64 mm0, mm1, mm2, mm3, mm4, mm5, mm6, mm7; |
| __m64 wk[7]; |
| __m64 Y_BG, Cb_RG, Cr_BG; |
| |
| while (--num_rows >= 0) { |
| inptr = *input_buf++; |
| outptr0 = output_buf[0][output_row]; |
| outptr1 = output_buf[1][output_row]; |
| outptr2 = output_buf[2][output_row]; |
| output_row++; |
| |
| for (num_cols = image_width; num_cols > 0; num_cols -= 8, |
| outptr0 += 8, outptr1 += 8, outptr2 += 8) { |
| |
| #if RGB_PIXELSIZE == 3 |
| |
| if (num_cols < 8) { |
| col = num_cols * 3; |
| asm(".set noreorder\r\n" |
| |
| "li $8, 1\r\n" |
| "move $9, %3\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 1f\r\n" |
| "nop \r\n" |
| "subu $9, $9, 1\r\n" |
| "xor $12, $12, $12\r\n" |
| "move $13, %5\r\n" |
| "dadd $13, $13, $9\r\n" |
| "lbu $12, 0($13)\r\n" |
| |
| "1: \r\n" |
| "li $8, 2\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 2f\r\n" |
| "nop \r\n" |
| "subu $9, $9, 2\r\n" |
| "xor $11, $11, $11\r\n" |
| "move $13, %5\r\n" |
| "dadd $13, $13, $9\r\n" |
| "lhu $11, 0($13)\r\n" |
| "sll $12, $12, 16\r\n" |
| "or $12, $12, $11\r\n" |
| |
| "2: \r\n" |
| "dmtc1 $12, %0\r\n" |
| "li $8, 4\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 3f\r\n" |
| "nop \r\n" |
| "subu $9, $9, 4\r\n" |
| "move $13, %5\r\n" |
| "dadd $13, $13, $9\r\n" |
| "lwu $14, 0($13)\r\n" |
| "dmtc1 $14, %1\r\n" |
| "dsll32 $12, $12, 0\r\n" |
| "or $12, $12, $14\r\n" |
| "dmtc1 $12, %0\r\n" |
| |
| "3: \r\n" |
| "li $8, 8\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 4f\r\n" |
| "nop \r\n" |
| "mov.s %1, %0\r\n" |
| "ldc1 %0, 0(%5)\r\n" |
| "li $9, 8\r\n" |
| "j 5f\r\n" |
| "nop \r\n" |
| |
| "4: \r\n" |
| "li $8, 16\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 5f\r\n" |
| "nop \r\n" |
| "mov.s %2, %0\r\n" |
| "ldc1 %0, 0(%5)\r\n" |
| "ldc1 %1, 8(%5)\r\n" |
| |
| "5: \r\n" |
| "nop \r\n" |
| ".set reorder\r\n" |
| |
| : "=f" (mmA), "=f" (mmG), "=f" (mmF) |
| : "r" (col), "r" (num_rows), "r" (inptr) |
| : "$f0", "$f2", "$f4", "$8", "$9", "$10", "$11", "$12", "$13", |
| "$14", "memory" |
| ); |
| } else { |
| mmA = _mm_load_si64((__m64 *)&inptr[0]); |
| mmG = _mm_load_si64((__m64 *)&inptr[8]); |
| mmF = _mm_load_si64((__m64 *)&inptr[16]); |
| inptr += RGB_PIXELSIZE * 8; |
| } |
| mmD = mmA; |
| mmA = _mm_slli_si64(mmA, 4 * BYTE_BIT); |
| mmD = _mm_srli_si64(mmD, 4 * BYTE_BIT); |
| |
| mmA = _mm_unpackhi_pi8(mmA, mmG); |
| mmG = _mm_slli_si64(mmG, 4 * BYTE_BIT); |
| |
| mmD = _mm_unpacklo_pi8(mmD, mmF); |
| mmG = _mm_unpackhi_pi8(mmG, mmF); |
| |
| mmE = mmA; |
| mmA = _mm_slli_si64(mmA, 4 * BYTE_BIT); |
| mmE = _mm_srli_si64(mmE, 4 * BYTE_BIT); |
| |
| mmA = _mm_unpackhi_pi8(mmA, mmD); |
| mmD = _mm_slli_si64(mmD, 4 * BYTE_BIT); |
| |
| mmE = _mm_unpacklo_pi8(mmE, mmG); |
| mmD = _mm_unpackhi_pi8(mmD, mmG); |
| mmC = mmA; |
| mmA = _mm_loadlo_pi8_f(mmA); |
| mmC = _mm_loadhi_pi8_f(mmC); |
| |
| mmB = mmE; |
| mmE = _mm_loadlo_pi8_f(mmE); |
| mmB = _mm_loadhi_pi8_f(mmB); |
| |
| mmF = mmD; |
| mmD = _mm_loadlo_pi8_f(mmD); |
| mmF = _mm_loadhi_pi8_f(mmF); |
| |
| #else /* RGB_PIXELSIZE == 4 */ |
| |
| if (num_cols < 8) { |
| col = num_cols; |
| asm(".set noreorder\r\n" |
| |
| "li $8, 1\r\n" |
| "move $9, %4\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 1f\r\n" |
| "nop \r\n" |
| "subu $9, $9, 1\r\n" |
| "dsll $11, $9, 2\r\n" |
| "move $13, %5\r\n" |
| "daddu $13, $13, $11\r\n" |
| "lwc1 %0, 0($13)\r\n" |
| |
| "1: \r\n" |
| "li $8, 2\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 2f\r\n" |
| "nop \r\n" |
| "subu $9, $9, 2\r\n" |
| "dsll $11, $9, 2\r\n" |
| "move $13, %5\r\n" |
| "daddu $13, $13, $11\r\n" |
| "mov.s %1, %0\r\n" |
| "ldc1 %0, 0($13)\r\n" |
| |
| "2: \r\n" |
| "li $8, 4\r\n" |
| "and $10, $9, $8\r\n" |
| "beqz $10, 3f\r\n" |
| "nop \r\n" |
| "mov.s %2, %0\r\n" |
| "mov.s %3, %1\r\n" |
| "ldc1 %0, 0(%5)\r\n" |
| "ldc1 %1, 8(%5)\r\n" |
| |
| "3: \r\n" |
| "nop \r\n" |
| ".set reorder\r\n" |
| |
| : "=f" (mmA), "=f" (mmF), "=f" (mmD), "=f" (mmC) |
| : "r" (col), "r" (inptr) |
| : "$f0", "$f2", "$8", "$9", "$10", "$11", "$13", "memory" |
| ); |
| } else { |
| mmA = _mm_load_si64((__m64 *)&inptr[0]); |
| mmF = _mm_load_si64((__m64 *)&inptr[8]); |
| mmD = _mm_load_si64((__m64 *)&inptr[16]); |
| mmC = _mm_load_si64((__m64 *)&inptr[24]); |
| inptr += RGB_PIXELSIZE * 8; |
| } |
| mmB = mmA; |
| mmA = _mm_unpacklo_pi8(mmA, mmF); |
| mmB = _mm_unpackhi_pi8(mmB, mmF); |
| |
| mmG = mmD; |
| mmD = _mm_unpacklo_pi8(mmD, mmC); |
| mmG = _mm_unpackhi_pi8(mmG, mmC); |
| |
| mmE = mmA; |
| mmA = _mm_unpacklo_pi16(mmA, mmD); |
| mmE = _mm_unpackhi_pi16(mmE, mmD); |
| |
| mmH = mmB; |
| mmB = _mm_unpacklo_pi16(mmB, mmG); |
| mmH = _mm_unpackhi_pi16(mmH, mmG); |
| |
| mmC = mmA; |
| mmA = _mm_loadlo_pi8_f(mmA); |
| mmC = _mm_loadhi_pi8_f(mmC); |
| |
| mmD = mmB; |
| mmB = _mm_loadlo_pi8_f(mmB); |
| mmD = _mm_loadhi_pi8_f(mmD); |
| |
| mmG = mmE; |
| mmE = _mm_loadlo_pi8_f(mmE); |
| mmG = _mm_loadhi_pi8_f(mmG); |
| |
| mmF = mmH; |
| mmF = _mm_unpacklo_pi8(mmF, mmH); |
| mmH = _mm_unpackhi_pi8(mmH, mmH); |
| mmF = _mm_srli_pi16(mmF, BYTE_BIT); |
| mmH = _mm_srli_pi16(mmH, BYTE_BIT); |
| |
| #endif |
| |
| wk[0] = mm0; |
| wk[1] = mm1; |
| wk[2] = mm4; |
| wk[3] = mm5; |
| |
| mm6 = mm1; |
| mm1 = _mm_unpacklo_pi16(mm1, mm3); |
| mm6 = _mm_unpackhi_pi16(mm6, mm3); |
| mm7 = mm1; |
| mm4 = mm6; |
| mm1 = _mm_madd_pi16(mm1, PW_F0299_F0337); |
| mm6 = _mm_madd_pi16(mm6, PW_F0299_F0337); |
| mm7 = _mm_madd_pi16(mm7, PW_MF016_MF033); |
| mm4 = _mm_madd_pi16(mm4, PW_MF016_MF033); |
| |
| wk[4] = mm1; |
| wk[5] = mm6; |
| |
| mm1 = _mm_loadlo_pi16_f(mm5); |
| mm6 = _mm_loadhi_pi16_f(mm5); |
| mm1 = _mm_srli_pi32(mm1, 1); |
| mm6 = _mm_srli_pi32(mm6, 1); |
| |
| mm5 = PD_ONEHALFM1_CJ; |
| mm7 = _mm_add_pi32(mm7, mm1); |
| mm4 = _mm_add_pi32(mm4, mm6); |
| mm7 = _mm_add_pi32(mm7, mm5); |
| mm4 = _mm_add_pi32(mm4, mm5); |
| mm7 = _mm_srli_pi32(mm7, SCALEBITS); |
| mm4 = _mm_srli_pi32(mm4, SCALEBITS); |
| mm7 = _mm_packs_pi32(mm7, mm4); |
| |
| mm1 = wk[2]; |
| mm6 = mm0; |
| mm0 = _mm_unpacklo_pi16(mm0, mm2); |
| mm6 = _mm_unpackhi_pi16(mm6, mm2); |
| mm5 = mm0; |
| mm4 = mm6; |
| mm0 = _mm_madd_pi16(mm0, PW_F0299_F0337); |
| mm6 = _mm_madd_pi16(mm6, PW_F0299_F0337); |
| mm5 = _mm_madd_pi16(mm5, PW_MF016_MF033); |
| mm4 = _mm_madd_pi16(mm4, PW_MF016_MF033); |
| |
| wk[6] = mm0; |
| wk[7] = mm6; |
| mm0 = _mm_loadlo_pi16_f(mm1); |
| mm6 = _mm_loadhi_pi16_f(mm1); |
| mm0 = _mm_srli_pi32(mm0, 1); |
| mm6 = _mm_srli_pi32(mm6, 1); |
| |
| mm1 = PD_ONEHALFM1_CJ; |
| mm5 = _mm_add_pi32(mm5, mm0); |
| mm4 = _mm_add_pi32(mm4, mm6); |
| mm5 = _mm_add_pi32(mm5, mm1); |
| mm4 = _mm_add_pi32(mm4, mm1); |
| mm5 = _mm_srli_pi32(mm5, SCALEBITS); |
| mm4 = _mm_srli_pi32(mm4, SCALEBITS); |
| mm5 = _mm_packs_pi32(mm5, mm4); |
| |
| mm7 = _mm_slli_pi16(mm7, BYTE_BIT); |
| mm5 = _mm_or_si64(mm5, mm7); |
| Cb_RG = mm5; |
| |
| mm0 = wk[3]; |
| mm6 = wk[2]; |
| mm1 = wk[1]; |
| |
| mm4 = mm0; |
| mm0 = _mm_unpacklo_pi16(mm0, mm3); |
| mm4 = _mm_unpackhi_pi16(mm4, mm3); |
| mm7 = mm0; |
| mm5 = mm4; |
| mm0 = _mm_madd_pi16(mm0, PW_F0114_F0250); |
| mm4 = _mm_madd_pi16(mm4, PW_F0114_F0250); |
| mm7 = _mm_madd_pi16(mm7, PW_MF008_MF041); |
| mm5 = _mm_madd_pi16(mm5, PW_MF008_MF041); |
| |
| mm3 = PD_ONEHALF; |
| mm0 = _mm_add_pi32(mm0, wk[4]); |
| mm4 = _mm_add_pi32(mm4, wk[5]); |
| mm0 = _mm_add_pi32(mm0, mm3); |
| mm4 = _mm_add_pi32(mm4, mm3); |
| mm0 = _mm_srli_pi32(mm0, SCALEBITS); |
| mm4 = _mm_srli_pi32(mm4, SCALEBITS); |
| mm0 = _mm_packs_pi32(mm0, mm4); |
| |
| mm3 = _mm_loadlo_pi16_f(mm1); |
| mm4 = _mm_loadhi_pi16_f(mm1); |
| mm3 = _mm_srli_pi32(mm3, 1); |
| mm4 = _mm_srli_pi32(mm4, 1); |
| |
| mm1 = PD_ONEHALFM1_CJ; |
| mm7 = _mm_add_pi32(mm7, mm3); |
| mm5 = _mm_add_pi32(mm5, mm4); |
| mm7 = _mm_add_pi32(mm7, mm1); |
| mm5 = _mm_add_pi32(mm5, mm1); |
| mm7 = _mm_srli_pi32(mm7, SCALEBITS); |
| mm5 = _mm_srli_pi32(mm5, SCALEBITS); |
| mm7 = _mm_packs_pi32(mm7, mm5); |
| |
| mm3 = wk[0]; |
| mm4 = mm6; |
| mm6 = _mm_unpacklo_pi16(mm6, mm2); |
| mm4 = _mm_unpackhi_pi16(mm4, mm2); |
| mm1 = mm6; |
| mm5 = mm4; |
| mm6 = _mm_madd_pi16(mm6, PW_F0114_F0250); |
| mm4 = _mm_madd_pi16(mm4, PW_F0114_F0250); |
| mm1 = _mm_madd_pi16(mm1, PW_MF008_MF041); |
| mm5 = _mm_madd_pi16(mm5, PW_MF008_MF041); |
| |
| mm2 = PD_ONEHALF; |
| mm6 = _mm_add_pi32(mm6, wk[6]); |
| mm4 = _mm_add_pi32(mm4, wk[7]); |
| mm6 = _mm_add_pi32(mm6, mm2); |
| mm4 = _mm_add_pi32(mm4, mm2); |
| mm6 = _mm_srli_pi32(mm6, SCALEBITS); |
| mm4 = _mm_srli_pi32(mm4, SCALEBITS); |
| mm6 = _mm_packs_pi32(mm6, mm4); |
| |
| mm0 = _mm_slli_pi16(mm0, BYTE_BIT); |
| mm6 = _mm_or_si64(mm6, mm0); |
| Y_BG = mm6; |
| |
| mm2 = _mm_loadlo_pi16_f(mm3); |
| mm4 = _mm_loadhi_pi16_f(mm3); |
| mm2 = _mm_srli_pi32(mm2, 1); |
| mm4 = _mm_srli_pi32(mm4, 1); |
| |
| mm0 = PD_ONEHALFM1_CJ; |
| mm1 = _mm_add_pi32(mm1, mm2); |
| mm5 = _mm_add_pi32(mm5, mm4); |
| mm1 = _mm_add_pi32(mm1, mm0); |
| mm5 = _mm_add_pi32(mm5, mm0); |
| mm1 = _mm_srli_pi32(mm1, SCALEBITS); |
| mm5 = _mm_srli_pi32(mm5, SCALEBITS); |
| mm1 = _mm_packs_pi32(mm1, mm5); |
| |
| mm7 = _mm_slli_pi16(mm7, BYTE_BIT); |
| mm1 = _mm_or_si64(mm1, mm7); |
| Cr_BG = mm1; |
| |
| _mm_store_si64((__m64 *)&outptr0[0], Y_BG); |
| _mm_store_si64((__m64 *)&outptr1[0], Cb_RG); |
| _mm_store_si64((__m64 *)&outptr2[0], Cr_BG); |
| } |
| } |
| } |
| |
| #undef mmA |
| #undef mmB |
| #undef mmC |
| #undef mmD |
| #undef mmE |
| #undef mmF |
| #undef mmG |
| #undef mmH |