// Copyright 2015 Google Inc. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // kernel_SSE.h: a collection of Intel SSE optimized kernels. // Check in kernel_default.h which one(s) are actually used by default. // Others are mere experiments; they are still covered by tests // in case they might be useful some day. // #ifndef GEMMLOWP_INTERNAL_KERNEL_SSE_H_ #define GEMMLOWP_INTERNAL_KERNEL_SSE_H_ #include "kernel.h" #include <string.h> #include <cassert> namespace gemmlowp { #ifdef GEMMLOWP_SSE4_32 struct SSE4_32_Kernel4x4Depth2 : KernelBase { typedef KernelFormat< KernelSideFormat<CellFormat<4, 2, CellOrder::WidthMajor>, 1>, KernelSideFormat<CellFormat<4, 2, CellOrder::WidthMajor>, 1> > Format; const char* Name() const override { return "SSE, 4x4, depth 2"; } void Run(std::int32_t* dst_ptr, std::size_t dst_row_stride, std::size_t dst_col_stride, const std::uint8_t* lhs_ptr, const std::uint8_t* rhs_ptr, std::size_t start_depth, std::size_t run_depth) const override { ScopedProfilingLabel label("optimized kernel"); assert(dst_row_stride == 1); std::int32_t run_depth_cells = run_depth / Format::kDepth; /* Main loop */ // A 2x4 cell of Rhs is stored in 16bit in xmm1 . // A 4x2 block Lhs is stored in 16bit in xmm0. // A 4x4 block of accumulators is stored in 32bit in xmm4--xmm7. // // +-------+-------+-------+-------+ // |xmm1[0]|xmm1[2]|xmm1[4]|xmm1[6]| // Rhs +-------+---------------+-------+ // |xmm1[1]|xmm1[3]|xmm1[5]|xmm1[7]| // +-------+-------+-------+-------+ // // | | | | | // // Lhs | | | | | // // +--+--+ - - - - +-------+-------+-------+-------+ // |xmm0 | | xmm4 | xmm5 | xmm6 | xmm7 | // |xmm0 | (Iter1) | xmm4 | xmm5 | xmm6 | xmm7 | // |xmm0 | | xmm4 | xmm5 | xmm6 | xmm7 | // |xmm0 | | xmm4 | xmm5 | xmm6 | xmm7 | // +--+--+ - - - - +-------+-------+-------+-------+ // // Accumulator asm volatile( // set accumulators to zero. "pxor %%xmm4 , %%xmm4 \n\t" "pxor %%xmm5 , %%xmm5 \n\t" "pxor %%xmm6 , %%xmm6 \n\t" "pxor %%xmm7 , %%xmm7 \n\t" "movl %[run_depth_cells], %%eax\n\t" "subl $2, %%eax\n\t" "js outerLoop1%=\n\t" // Loop for K unrolled by 4 "outerLoop2%=:\n\t" // K = 1,2 // RHS cell to xmm1 "pmovzxbw (%[rhs_ptr]), %%xmm1\n\t" // LHS cell "pmovzxbw 0x00(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm4 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm5 \n\t" "prefetcht0 0x80(%[lhs_ptr]) \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm6 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm7 \n\t" "prefetcht0 0x80(%[rhs_ptr]) \n\t" // K = 3,4 // RHS cell to xmm1 "pmovzxbw 0x08(%[rhs_ptr]), %%xmm1\n\t" // LHS cell "pmovzxbw 0x08(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm4 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm5 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm6 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm7 \n\t" "addl $0x10, %[lhs_ptr]\n\t" "addl $0x10, %[rhs_ptr]\n\t" "subl $2, %[run_depth_cells]\n\t" "jnz outerLoop2%=\n\t" "movl %[run_depth_cells], %%eax\n\t" "decl %%eax\n\t" "js finish%=\n\t" // Loop for K unrolled by 2 "outerLoop1%=:\n\t" // RHS cell to xmm1 "pmovzxbw (%[rhs_ptr]), %%xmm1\n\t" // LHS cell "pmovzxbw 0x00(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm4 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm5 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm6 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm7 \n\t" "addl $0x08, %[lhs_ptr]\n\t" "addl $0x08, %[rhs_ptr]\n\t" "decl %[run_depth_cells]\n\t" "jnz outerLoop1%=\n\t" "finish%=:\n\t" "movl %[dst_col_stride], %%eax\n\t" "shll $2, %%eax\n\t" "movl %[start_depth], %%ecx\n\t" "test %%ecx, %%ecx\n\t" "jz storeDst%=\n\t" "leal (%%eax,%%eax,0x2), %%ecx\n\t" "paddd 0x00(%[dst_ptr]) , %%xmm4 \n\t" "paddd 0x00(%[dst_ptr], %%eax, 1) , %%xmm5 \n\t" "paddd 0x00(%[dst_ptr], %%eax, 2) , %%xmm6 \n\t" "paddd 0x00(%[dst_ptr], %%ecx, 1) , %%xmm7 \n\t" "storeDst%=:\n\t" "leal (%%eax,%%eax,0x2), %%ecx\n\t" "movdqu %%xmm4 , 0x00(%[dst_ptr]) \n\t" "movdqu %%xmm5 , 0x00(%[dst_ptr], %%eax, 1)\n\t" "movdqu %%xmm6 , 0x00(%[dst_ptr], %%eax, 2)\n\t" "movdqu %%xmm7 , 0x00(%[dst_ptr], %%ecx, 1)\n\t" : // outputs [lhs_ptr] "+r"(lhs_ptr), [rhs_ptr] "+r"(rhs_ptr), [dst_ptr] "+r"(dst_ptr) : // inputs [start_depth] "g"(start_depth), [dst_col_stride] "g"(dst_col_stride), [run_depth_cells] "g"(run_depth_cells) : // clobbers "cc", "memory", "%xmm0", "%xmm1", "%xmm3", "%xmm2", "%xmm4", "%xmm5", "%xmm6", "%xmm7", "%eax", "%ecx"); } }; #endif #ifdef GEMMLOWP_SSE4_64 struct SSE4_64_Kernel12x4Depth2 : KernelBase { typedef KernelFormat< KernelSideFormat<CellFormat<4, 2, CellOrder::WidthMajor>, 3>, KernelSideFormat<CellFormat<4, 2, CellOrder::WidthMajor>, 1> > Format; const char* Name() const override { return "SSE, 12x4, depth 2"; } void Run(std::int32_t* dst_ptr, std::size_t dst_row_stride, std::size_t dst_col_stride, const std::uint8_t* lhs_ptr, const std::uint8_t* rhs_ptr, std::size_t start_depth, std::size_t run_depth) const override { ScopedProfilingLabel label("optimized kernel"); assert(dst_row_stride == 1); const std::int64_t run_depth_cells = run_depth / Format::kDepth; const std::int64_t dst_col_stride_q = dst_col_stride; /* Main loop */ // A 2x4 cell of Rhs is stored in 16bit in xmm1 . // A 12x2 block of 3 4x2 cells Lhs is stored in 16bit in xmm0, replaced // every Iteration. // A 12x4 block of accumulators is stored in 32bit in xmm4--xmm15. // // +-------+-------+-------+-------+ // |xmm1[0]|xmm1[2]|xmm1[4]|xmm1[6]| // Rhs +-------+---------------+-------+ // |xmm1[1]|xmm1[3]|xmm1[5]|xmm1[7]| // +-------+-------+-------+-------+ // // | | | | | // // Lhs | | | | | // // +--+--+ - - - - +-------+-------+-------+-------+ // |xmm0 | | xmm4 | xmm5 | xmm6 | xmm7 | // |xmm0 | (Iter1) | xmm4 | xmm5 | xmm6 | xmm7 | // |xmm0 | | xmm4 | xmm5 | xmm6 | xmm7 | // |xmm0 | | xmm4 | xmm5 | xmm6 | xmm7 | // +--+--+ - - - - +-------+-------+-------+-------+ // |xmm0 | | xmm8 | xmm9 | xmm10 | xmm11 | // |xmm0 | (Iter2) | xmm8 | xmm9 | xmm10 | xmm11 | // |xmm0 | | xmm8 | xmm9 | xmm10 | xmm11 | // |xmm0 | | xmm8 | xmm9 | xmm10 | xmm11 | // +--+--+ - - - - +-------+-------+-------+-------+ // |xmm0 | | xmm12 | xmm13 | xmm14 | xmm15 | // |xmm0 | (Iter3) | xmm12 | xmm13 | xmm14 | xmm15 | // |xmm0 | | xmm12 | xmm13 | xmm14 | xmm15 | // |xmm0 | | xmm12 | xmm13 | xmm14 | xmm15 | // +--+--+ - - - - +-------+-------+-------+-------+ // // Accumulator asm volatile( // Set registers for destination "movq %[dst_col_stride_q], %%r12\n\t" "shlq $2, %%r12\n\t" "leaq (%%r12,%%r12,0x2), %%r13\n\t" // Set accumulators to zero. "pxor %%xmm4 , %%xmm4 \n\t" "pxor %%xmm5 , %%xmm5 \n\t" "pxor %%xmm6 , %%xmm6 \n\t" "pxor %%xmm7 , %%xmm7 \n\t" "pxor %%xmm8 , %%xmm8 \n\t" "pxor %%xmm9 , %%xmm9 \n\t" "pxor %%xmm10 , %%xmm10\n\t" "pxor %%xmm11 , %%xmm11\n\t" "pxor %%xmm12 , %%xmm12\n\t" "pxor %%xmm13 , %%xmm13\n\t" "pxor %%xmm14 , %%xmm14\n\t" "pxor %%xmm15 , %%xmm15\n\t" "movq %[run_depth_cells], %%r14\n\t" "subq $2, %%r14\n\t" "js outerLoop1%=\n\t" // Loop for K unrolled by 4 "outerLoop2%=:\n\t" // K = 1,2 // RHS cell to xmm1 "pmovzxbw (%[rhs_ptr]), %%xmm1\n\t" // LHS cell "pmovzxbw 0x00(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm4 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm5 \n\t" "prefetcht0 0x80(%[lhs_ptr]) \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm6 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm7 \n\t" // next LHS cell "pmovzxbw 0x08(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm8 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm9 \n\t" "prefetcht0 0x80(%[rhs_ptr]) \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm10 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm11 \n\t" // next LHS cell "pmovzxbw 0x10(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm12 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm13 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm14 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm15 \n\t" // K = 3,4 // RHS cell to xmm1 "pmovzxbw 0x08(%[rhs_ptr]), %%xmm1\n\t" // LHS cell "pmovzxbw 0x18(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm4 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm5 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm6 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm7 \n\t" // next LHS cell "pmovzxbw 0x20(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm8 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm9 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm10 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm11 \n\t" // next LHS cell "pmovzxbw 0x28(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm12 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm13 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm14 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm15 \n\t" "addq $0x30, %[lhs_ptr]\n\t" "addq $0x10, %[rhs_ptr]\n\t" "subq $2, %[run_depth_cells]\n\t" "jnz outerLoop2%=\n\t" "movq %[run_depth_cells], %%r14\n\t" "decq %%r14\n\t" "js finish%=\n\t" // Loop for K unrolled by 2 "outerLoop1%=:\n\t" // RHS cell to xmm1 "pmovzxbw (%[rhs_ptr]), %%xmm1\n\t" // LHS cell "pmovzxbw 0x00(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm4 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm5 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm6 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm7 \n\t" // next LHS cell "pmovzxbw 0x08(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm8 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm9 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm10 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm11 \n\t" // next LHS cell "pmovzxbw 0x10(%[lhs_ptr]), %%xmm0\n\t" "pshufd $0x00,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm12 \n\t" "pshufd $0x55,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm13 \n\t" "pshufd $0xaa,%%xmm1,%%xmm2 \n\t" "pmaddwd %%xmm0, %%xmm2 \n\t" "paddd %%xmm2, %%xmm14 \n\t" "pshufd $0xff,%%xmm1,%%xmm3 \n\t" "pmaddwd %%xmm0, %%xmm3 \n\t" "paddd %%xmm3, %%xmm15 \n\t" "addq $0x18, %[lhs_ptr]\n\t" "addq $0x08, %[rhs_ptr]\n\t" "decq %[run_depth_cells]\n\t" "jnz outerLoop1%=\n\t" "finish%=:\n\t" "test %[start_depth], %[start_depth]\n\t" "jz storeDst%=\n\t" "paddd 0x00(%[dst_ptr]) , %%xmm4 \n\t" "paddd 0x10(%[dst_ptr]) , %%xmm8 \n\t" "paddd 0x20(%[dst_ptr]) , %%xmm12\n\t" "paddd 0x00(%[dst_ptr], %%r12, 1) , %%xmm5 \n\t" "paddd 0x10(%[dst_ptr], %%r12, 1) , %%xmm9 \n\t" "paddd 0x20(%[dst_ptr], %%r12, 1) , %%xmm13\n\t" "paddd 0x00(%[dst_ptr], %%r12, 2) , %%xmm6 \n\t" "paddd 0x10(%[dst_ptr], %%r12, 2) , %%xmm10\n\t" "paddd 0x20(%[dst_ptr], %%r12, 2) , %%xmm14\n\t" "paddd 0x00(%[dst_ptr], %%r13, 1) , %%xmm7 \n\t" "paddd 0x10(%[dst_ptr], %%r13, 1) , %%xmm11\n\t" "paddd 0x20(%[dst_ptr], %%r13, 1) , %%xmm15\n\t" "storeDst%=:\n\t" "movdqu %%xmm4 , 0x00(%[dst_ptr]) \n\t" "movdqu %%xmm8 , 0x10(%[dst_ptr]) \n\t" "movdqu %%xmm12 , 0x20(%[dst_ptr]) \n\t" "movdqu %%xmm5 , 0x00(%[dst_ptr], %%r12, 1)\n\t" "movdqu %%xmm9 , 0x10(%[dst_ptr], %%r12, 1)\n\t" "movdqu %%xmm13 , 0x20(%[dst_ptr], %%r12, 1)\n\t" "movdqu %%xmm6 , 0x00(%[dst_ptr], %%r12, 2)\n\t" "movdqu %%xmm10 , 0x10(%[dst_ptr], %%r12, 2)\n\t" "movdqu %%xmm14 , 0x20(%[dst_ptr], %%r12, 2)\n\t" "movdqu %%xmm7 , 0x00(%[dst_ptr], %%r13, 1)\n\t" "movdqu %%xmm11 , 0x10(%[dst_ptr], %%r13, 1)\n\t" "movdqu %%xmm15 , 0x20(%[dst_ptr], %%r13, 1)\n\t" : // outputs [lhs_ptr] "+r"(lhs_ptr), [rhs_ptr] "+r"(rhs_ptr), [dst_ptr] "+r"(dst_ptr) : // inputs [start_depth] "r"(start_depth), [dst_col_stride_q] "r"(dst_col_stride_q), [run_depth_cells] "r"(run_depth_cells) : // clobbers "cc", "memory", "%xmm0", "%xmm1", "%xmm3", "%xmm2", "%xmm4", "%xmm5", "%xmm6", "%xmm7", "%xmm8", "%xmm9", "%xmm10", "%r12", "%r13", "%r14", "%xmm11", "%xmm12", "%xmm13", "%xmm14", "%xmm15"); } }; #endif } // namespace gemmlowp #endif // GEMMLOWP_INTERNAL_KERNEL_SSE_H_