// Copyright (c) 2013 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include <algorithm>
#include "skia/ext/convolver.h"
#include "skia/ext/convolver_mips_dspr2.h"
#include "third_party/skia/include/core/SkTypes.h"
namespace skia {
// Convolves horizontally along a single row. The row data is given in
// |src_data| and continues for the num_values() of the filter.
void ConvolveHorizontally_mips_dspr2(const unsigned char* src_data,
const ConvolutionFilter1D& filter,
unsigned char* out_row,
bool has_alpha) {
#if SIMD_MIPS_DSPR2
int row_to_filter = 0;
int num_values = filter.num_values();
if (has_alpha) {
for (int out_x = 0; out_x < num_values; out_x++) {
// Get the filter that determines the current output pixel.
int filter_offset, filter_length;
const ConvolutionFilter1D::Fixed* filter_values =
filter.FilterForValue(out_x, &filter_offset, &filter_length);
int filter_x = 0;
__asm__ __volatile__ (
".set push \n"
".set noreorder \n"
"beqz %[filter_len], 3f \n"
" sll $t0, %[filter_offset], 2 \n"
"addu %[rtf], %[src_data], $t0 \n"
"mtlo $0, $ac0 \n"
"mtlo $0, $ac1 \n"
"mtlo $0, $ac2 \n"
"mtlo $0, $ac3 \n"
"srl $t7, %[filter_len], 2 \n"
"beqz $t7, 2f \n"
" li %[fx], 0 \n"
"11: \n"
"addu $t4, %[filter_val], %[fx] \n"
"sll $t5, %[fx], 1 \n"
"ulw $t6, 0($t4) \n" // t6 = |cur[1]|cur[0]|
"ulw $t8, 4($t4) \n" // t8 = |cur[3]|cur[2]|
"addu $t0, %[rtf], $t5 \n"
"lw $t1, 0($t0) \n" // t1 = |a0|b0|g0|r0|
"lw $t2, 4($t0) \n" // t2 = |a1|b1|g1|r1|
"lw $t3, 8($t0) \n" // t3 = |a2|b2|g2|r2|
"lw $t4, 12($t0) \n" // t4 = |a3|b3|g3|r3|
"precrq.qb.ph $t0, $t2, $t1 \n" // t0 = |a1|g1|a0|g0|
"precr.qb.ph $t5, $t2, $t1 \n" // t5 = |b1|r1|b0|r0|
"preceu.ph.qbla $t1, $t0 \n" // t1 = |0|a1|0|a0|
"preceu.ph.qbra $t2, $t0 \n" // t2 = |0|g1|0|g0|
"preceu.ph.qbla $t0, $t5 \n" // t0 = |0|b1|0|b0|
"preceu.ph.qbra $t5, $t5 \n" // t5 = |0|r1|0|r0|
"dpa.w.ph $ac0, $t1, $t6 \n" // ac0+(cur*a1)+(cur*a0)
"dpa.w.ph $ac1, $t0, $t6 \n" // ac1+(cur*b1)+(cur*b0)
"dpa.w.ph $ac2, $t2, $t6 \n" // ac2+(cur*g1)+(cur*g0)
"dpa.w.ph $ac3, $t5, $t6 \n" // ac3+(cur*r1)+(cur*r0)
"precrq.qb.ph $t0, $t4, $t3 \n" // t0 = |a3|g3|a2|g2|
"precr.qb.ph $t5, $t4, $t3 \n" // t5 = |b3|r3|b2|r2|
"preceu.ph.qbla $t1, $t0 \n" // t1 = |0|a3|0|a2|
"preceu.ph.qbra $t2, $t0 \n" // t2 = |0|g3|0|g2|
"preceu.ph.qbla $t0, $t5 \n" // t0 = |0|b3|0|b2|
"preceu.ph.qbra $t5, $t5 \n" // t5 = |0|r3|0|r2|
"dpa.w.ph $ac0, $t1, $t8 \n" // ac0+(cur*a3)+(cur*a2)
"dpa.w.ph $ac1, $t0, $t8 \n" // ac1+(cur*b3)+(cur*b2)
"dpa.w.ph $ac2, $t2, $t8 \n" // ac2+(cur*g3)+(cur*g2)
"dpa.w.ph $ac3, $t5, $t8 \n" // ac3+(cur*r3)+(cur*r2)
"addiu $t7, $t7, -1 \n"
"bgtz $t7, 11b \n"
" addiu %[fx], %[fx], 8 \n"
"2: \n"
"andi $t7, %[filter_len], 0x3 \n" // residual
"beqz $t7, 3f \n"
" nop \n"
"21: \n"
"sll $t1, %[fx], 1 \n"
"addu $t2, %[filter_val], %[fx] \n"
"addu $t0, %[rtf], $t1 \n"
"lh $t6, 0($t2) \n" // t6 = filter_val[fx]
"lbu $t1, 0($t0) \n" // t1 = row[fx * 4 + 0]
"lbu $t2, 1($t0) \n" // t2 = row[fx * 4 + 1]
"lbu $t3, 2($t0) \n" // t3 = row[fx * 4 + 2]
"lbu $t4, 3($t0) \n" // t4 = row[fx * 4 + 2]
"maddu $ac3, $t6, $t1 \n"
"maddu $ac2, $t6, $t2 \n"
"maddu $ac1, $t6, $t3 \n"
"maddu $ac0, $t6, $t4 \n"
"addiu $t7, $t7, -1 \n"
"bgtz $t7, 21b \n"
" addiu %[fx], %[fx], 2 \n"
"3: \n"
"extrv.w $t0, $ac0, %[kShiftBits] \n" // a >> kShiftBits
"extrv.w $t1, $ac1, %[kShiftBits] \n" // b >> kShiftBits
"extrv.w $t2, $ac2, %[kShiftBits] \n" // g >> kShiftBits
"extrv.w $t3, $ac3, %[kShiftBits] \n" // r >> kShiftBits
"sll $t5, %[out_x], 2 \n"
"repl.ph $t6, 128 \n" // t6 = | 128 | 128 |
"addu $t5, %[out_row], $t5 \n"
"append $t2, $t3, 16 \n"
"append $t0, $t1, 16 \n"
"subu.ph $t1, $t0, $t6 \n"
"shll_s.ph $t1, $t1, 8 \n"
"shra.ph $t1, $t1, 8 \n"
"addu.ph $t1, $t1, $t6 \n"
"subu.ph $t3, $t2, $t6 \n"
"shll_s.ph $t3, $t3, 8 \n"
"shra.ph $t3, $t3, 8 \n"
"addu.ph $t3, $t3, $t6 \n"
"precr.qb.ph $t0, $t1, $t3 \n"
"usw $t0, 0($t5) \n"
".set pop \n"
: [fx] "+r" (filter_x), [out_x] "+r" (out_x), [out_row] "+r" (out_row),
[rtf] "+r" (row_to_filter)
: [filter_val] "r" (filter_values), [filter_len] "r" (filter_length),
[kShiftBits] "r" (ConvolutionFilter1D::kShiftBits),
[filter_offset] "r" (filter_offset), [src_data] "r" (src_data)
: "lo", "hi", "$ac1lo", "$ac1hi", "$ac2lo", "$ac2hi", "$ac3lo", "$ac3hi",
"t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8"
);
}
} else {
for (int out_x = 0; out_x < num_values; out_x++) {
// Get the filter that determines the current output pixel.
int filter_offset, filter_length;
const ConvolutionFilter1D::Fixed* filter_values =
filter.FilterForValue(out_x, &filter_offset, &filter_length);
int filter_x = 0;
__asm__ __volatile__ (
".set push \n"
".set noreorder \n"
"beqz %[filter_len], 3f \n"
" sll $t0, %[filter_offset], 2 \n"
"addu %[rtf], %[src_data], $t0 \n"
"mtlo $0, $ac1 \n"
"mtlo $0, $ac2 \n"
"mtlo $0, $ac3 \n"
"srl $t7, %[filter_len], 2 \n"
"beqz $t7, 2f \n"
" li %[fx], 0 \n"
"11: \n"
"addu $t4, %[filter_val], %[fx] \n"
"sll $t5, %[fx], 1 \n"
"ulw $t6, 0($t4) \n" // t6 = |cur[1]|cur[0]|
"ulw $t8, 4($t4) \n" // t8 = |cur[3]|cur[2]|
"addu $t0, %[rtf], $t5 \n"
"lw $t1, 0($t0) \n" // t1 = |a0|b0|g0|r0|
"lw $t2, 4($t0) \n" // t2 = |a1|b1|g1|r1|
"lw $t3, 8($t0) \n" // t3 = |a2|b2|g2|r2|
"lw $t4, 12($t0) \n" // t4 = |a3|b3|g3|r3|
"precrq.qb.ph $t0, $t2, $t1 \n" // t0 = |a1|g1|a0|g0|
"precr.qb.ph $t5, $t2, $t1 \n" // t5 = |b1|r1|b0|r0|
"preceu.ph.qbra $t2, $t0 \n" // t2 = |0|g1|0|g0|
"preceu.ph.qbla $t0, $t5 \n" // t0 = |0|b1|0|b0|
"preceu.ph.qbra $t5, $t5 \n" // t5 = |0|r1|0|r0|
"dpa.w.ph $ac1, $t0, $t6 \n" // ac1+(cur*b1)+(cur*b0)
"dpa.w.ph $ac2, $t2, $t6 \n" // ac2+(cur*g1)+(cur*g0)
"dpa.w.ph $ac3, $t5, $t6 \n" // ac3+(cur*r1)+(cur*r0)
"precrq.qb.ph $t0, $t4, $t3 \n" // t0 = |a3|g3|a2|g2|
"precr.qb.ph $t5, $t4, $t3 \n" // t5 = |b3|r3|b2|r2|
"preceu.ph.qbra $t2, $t0 \n" // t2 = |0|g3|0|g2|
"preceu.ph.qbla $t0, $t5 \n" // t0 = |0|b3|0|b2|
"preceu.ph.qbra $t5, $t5 \n" // t5 = |0|r3|0|r2|
"dpa.w.ph $ac1, $t0, $t8 \n" // ac1+(cur*b3)+(cur*b2)
"dpa.w.ph $ac2, $t2, $t8 \n" // ac2+(cur*g3)+(cur*g2)
"dpa.w.ph $ac3, $t5, $t8 \n" // ac3+(cur*r3)+(cur*r2)
"addiu $t7, $t7, -1 \n"
"bgtz $t7, 11b \n"
" addiu %[fx], %[fx], 8 \n"
"2: \n"
"andi $t7, %[filter_len], 0x3 \n" // residual
"beqz $t7, 3f \n"
" nop \n"
"21: \n"
"sll $t1, %[fx], 1 \n"
"addu $t2, %[filter_val], %[fx] \n"
"addu $t0, %[rtf], $t1 \n"
"lh $t6, 0($t2) \n" // t6 = filter_val[fx]
"lbu $t1, 0($t0) \n" // t1 = row[fx * 4 + 0]
"lbu $t2, 1($t0) \n" // t2 = row[fx * 4 + 1]
"lbu $t3, 2($t0) \n" // t3 = row[fx * 4 + 2]
"maddu $ac3, $t6, $t1 \n"
"maddu $ac2, $t6, $t2 \n"
"maddu $ac1, $t6, $t3 \n"
"addiu $t7, $t7, -1 \n"
"bgtz $t7, 21b \n"
" addiu %[fx], %[fx], 2 \n"
"3: \n"
"extrv.w $t1, $ac1, %[kShiftBits] \n" // b >> kShiftBits
"extrv.w $t2, $ac2, %[kShiftBits] \n" // g >> kShiftBits
"extrv.w $t3, $ac3, %[kShiftBits] \n" // r >> kShiftBits
"repl.ph $t6, 128 \n" // t6 = | 128 | 128 |
"sll $t8, %[out_x], 2 \n"
"addu $t8, %[out_row], $t8 \n"
"append $t2, $t3, 16 \n"
"andi $t1, 0xFFFF \n"
"subu.ph $t5, $t1, $t6 \n"
"shll_s.ph $t5, $t5, 8 \n"
"shra.ph $t5, $t5, 8 \n"
"addu.ph $t5, $t5, $t6 \n"
"subu.ph $t4, $t2, $t6 \n"
"shll_s.ph $t4, $t4, 8 \n"
"shra.ph $t4, $t4, 8 \n"
"addu.ph $t4, $t4, $t6 \n"
"precr.qb.ph $t0, $t5, $t4 \n"
"usw $t0, 0($t8) \n"
".set pop \n"
: [fx] "+r" (filter_x), [out_x] "+r" (out_x), [out_row] "+r" (out_row),
[rtf] "+r" (row_to_filter)
: [filter_val] "r" (filter_values), [filter_len] "r" (filter_length),
[kShiftBits] "r" (ConvolutionFilter1D::kShiftBits),
[filter_offset] "r" (filter_offset), [src_data] "r" (src_data)
: "lo", "hi", "$ac1lo", "$ac1hi", "$ac2lo", "$ac2hi", "$ac3lo", "$ac3hi",
"t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8"
);
}
}
#endif
}
void ConvolveVertically_mips_dspr2(const ConvolutionFilter1D::Fixed* filter_val,
int filter_length,
unsigned char* const* source_data_rows,
int pixel_width,
unsigned char* out_row,
bool has_alpha) {
#if SIMD_MIPS_DSPR2
// We go through each column in the output and do a vertical convolution,
// generating one output pixel each time.
int byte_offset;
int cnt;
int filter_y;
if (has_alpha) {
for (int out_x = 0; out_x < pixel_width; out_x++) {
__asm__ __volatile__ (
".set push \n"
".set noreorder \n"
"beqz %[filter_len], 3f \n"
" sll %[offset], %[out_x], 2 \n"
"mtlo $0, $ac0 \n"
"mtlo $0, $ac1 \n"
"mtlo $0, $ac2 \n"
"mtlo $0, $ac3 \n"
"srl %[cnt], %[filter_len], 2 \n"
"beqz %[cnt], 2f \n"
" li %[fy], 0 \n"
"11: \n"
"sll $t1, %[fy], 1 \n"
"addu $t0, %[src_data_rows], $t1 \n"
"lw $t1, 0($t0) \n"
"lw $t2, 4($t0) \n"
"lw $t3, 8($t0) \n"
"lw $t4, 12($t0) \n"
"addu $t1, $t1, %[offset] \n"
"addu $t2, $t2, %[offset] \n"
"addu $t3, $t3, %[offset] \n"
"addu $t4, $t4, %[offset] \n"
"lw $t1, 0($t1) \n" // t1 = |a0|b0|g0|r0|
"lw $t2, 0($t2) \n" // t2 = |a1|b1|g1|r1|
"lw $t3, 0($t3) \n" // t3 = |a0|b0|g0|r0|
"lw $t4, 0($t4) \n" // t4 = |a1|b1|g1|r1|
"precrq.qb.ph $t5, $t2, $t1 \n" // t5 = |a1|g1|a0|g0|
"precr.qb.ph $t6, $t2, $t1 \n" // t6 = |b1|r1|b0|r0|
"preceu.ph.qbla $t0, $t5 \n" // t0 = |0|a1|0|a0|
"preceu.ph.qbra $t1, $t5 \n" // t1 = |0|g1|0|g0|
"preceu.ph.qbla $t2, $t6 \n" // t2 = |0|b1|0|b0|
"preceu.ph.qbra $t5, $t6 \n" // t5 = |0|r1|0|r0|
"addu $t6, %[filter_val], %[fy] \n"
"ulw $t7, 0($t6) \n" // t7 = |cur_1|cur_0|
"ulw $t6, 4($t6) \n" // t6 = |cur_3|cur_2|
"dpa.w.ph $ac0, $t5, $t7 \n" // (cur*r1)+(cur*r0)
"dpa.w.ph $ac1, $t1, $t7 \n" // (cur*g1)+(cur*g0)
"dpa.w.ph $ac2, $t2, $t7 \n" // (cur*b1)+(cur*b0)
"dpa.w.ph $ac3, $t0, $t7 \n" // (cur*a1)+(cur*a0)
"precrq.qb.ph $t5, $t4, $t3 \n" // t5 = |a3|g3|a2|g2|
"precr.qb.ph $t7, $t4, $t3 \n" // t7 = |b3|r3|b2|r2|
"preceu.ph.qbla $t0, $t5 \n" // t0 = |0|a3|0|a2|
"preceu.ph.qbra $t1, $t5 \n" // t1 = |0|g3|0|g2|
"preceu.ph.qbla $t2, $t7 \n" // t2 = |0|b3|0|b2|
"preceu.ph.qbra $t5, $t7 \n" // t5 = |0|r3|0|r2|
"dpa.w.ph $ac0, $t5, $t6 \n" // (cur*r3)+(cur*r2)
"dpa.w.ph $ac1, $t1, $t6 \n" // (cur*g3)+(cur*g2)
"dpa.w.ph $ac2, $t2, $t6 \n" // (cur*b3)+(cur*b2)
"dpa.w.ph $ac3, $t0, $t6 \n" // (cur*a3)+(cur*a2)
"addiu %[cnt], %[cnt], -1 \n"
"bgtz %[cnt], 11b \n"
" addiu %[fy], %[fy], 8 \n"
"2: \n"
"andi %[cnt], %[filter_len], 0x3 \n" // residual
"beqz %[cnt], 3f \n"
" nop \n"
"21: \n"
"addu $t0, %[filter_val], %[fy] \n"
"lh $t4, 0($t0) \n" // t4=filter_val[fx]
"sll $t1, %[fy], 1 \n"
"addu $t0, %[src_data_rows], $t1 \n"
"lw $t1, 0($t0) \n"
"addu $t0, $t1, %[offset] \n"
"lbu $t1, 0($t0) \n" // t1 = row[fx*4 + 0]
"lbu $t2, 1($t0) \n" // t2 = row[fx*4 + 1]
"lbu $t3, 2($t0) \n" // t3 = row[fx*4 + 2]
"lbu $t0, 3($t0) \n" // t4 = row[fx*4 + 2]
"maddu $ac0, $t4, $t1 \n"
"maddu $ac1, $t4, $t2 \n"
"maddu $ac2, $t4, $t3 \n"
"maddu $ac3, $t4, $t0 \n"
"addiu %[cnt], %[cnt], -1 \n"
"bgtz %[cnt], 21b \n"
" addiu %[fy], %[fy], 2 \n"
"3: \n"
"extrv.w $t3, $ac0, %[kShiftBits] \n" // a >> kShiftBits
"extrv.w $t2, $ac1, %[kShiftBits] \n" // b >> kShiftBits
"extrv.w $t1, $ac2, %[kShiftBits] \n" // g >> kShiftBits
"extrv.w $t0, $ac3, %[kShiftBits] \n" // r >> kShiftBits
"repl.ph $t4, 128 \n" // t4 = | 128 | 128 |
"addu $t5, %[out_row], %[offset] \n"
"append $t2, $t3, 16 \n" // t2 = |0|g|0|r|
"append $t0, $t1, 16 \n" // t0 = |0|a|0|b|
"subu.ph $t1, $t0, $t4 \n"
"shll_s.ph $t1, $t1, 8 \n"
"shra.ph $t1, $t1, 8 \n"
"addu.ph $t1, $t1, $t4 \n" // Clamp(a)|Clamp(b)
"subu.ph $t2, $t2, $t4 \n"
"shll_s.ph $t2, $t2, 8 \n"
"shra.ph $t2, $t2, 8 \n"
"addu.ph $t2, $t2, $t4 \n" // Clamp(g)|Clamp(r)
"andi $t3, $t1, 0xFF \n" // t3 = ClampTo8(b)
"cmp.lt.ph $t3, $t2 \n" // cmp b, g, r
"pick.ph $t0, $t2, $t3 \n"
"andi $t3, $t0, 0xFF \n"
"srl $t4, $t0, 16 \n"
"cmp.lt.ph $t3, $t4 \n"
"pick.ph $t0, $t4, $t3 \n" // t0 = max_color_ch
"srl $t3, $t1, 16 \n" // t1 = ClampTo8(a)
"cmp.lt.ph $t3, $t0 \n"
"pick.ph $t0, $t0, $t3 \n"
"ins $t1, $t0, 16, 8 \n"
"precr.qb.ph $t0, $t1, $t2 \n" // t0 = |a|b|g|r|
"usw $t0, 0($t5) \n"
".set pop \n"
: [filter_val] "+r" (filter_val), [filter_len] "+r" (filter_length),
[offset] "+r" (byte_offset), [fy] "+r" (filter_y), [cnt] "+r" (cnt),
[out_x] "+r" (out_x), [pixel_width] "+r" (pixel_width)
: [src_data_rows] "r" (source_data_rows), [out_row] "r" (out_row),
[kShiftBits] "r" (ConvolutionFilter1D::kShiftBits)
: "lo", "hi", "$ac1lo", "$ac1hi", "$ac2lo", "$ac2hi", "$ac3lo", "$ac3hi",
"t0", "t1", "t2", "t3", "t4", "t5", "t6","t7", "memory"
);
}
} else {
for (int out_x = 0; out_x < pixel_width; out_x++) {
__asm__ __volatile__ (
".set push \n"
".set noreorder \n"
"beqz %[filter_len], 3f \n"
" sll %[offset], %[out_x], 2 \n"
"mtlo $0, $ac0 \n"
"mtlo $0, $ac1 \n"
"mtlo $0, $ac2 \n"
"srl %[cnt], %[filter_len], 2 \n"
"beqz %[cnt], 2f \n"
" li %[fy], 0 \n"
"11: \n"
"sll $t1, %[fy], 1 \n"
"addu $t0, %[src_data_rows], $t1 \n"
"lw $t1, 0($t0) \n"
"lw $t2, 4($t0) \n"
"lw $t3, 8($t0) \n"
"lw $t4, 12($t0) \n"
"addu $t1, $t1, %[offset] \n"
"addu $t2, $t2, %[offset] \n"
"addu $t3, $t3, %[offset] \n"
"addu $t4, $t4, %[offset] \n"
"lw $t1, 0($t1) \n" // t1 = |a0|b0|g0|r0|
"lw $t2, 0($t2) \n" // t2 = |a1|b1|g1|r1|
"lw $t3, 0($t3) \n" // t3 = |a0|b0|g0|r0|
"lw $t4, 0($t4) \n" // t4 = |a1|b1|g1|r1|
"precrq.qb.ph $t5, $t2, $t1 \n" // t5 = |a1|g1|a0|g0|
"precr.qb.ph $t6, $t2, $t1 \n" // t6 = |b1|r1|b0|r0|
"preceu.ph.qbra $t1, $t5 \n" // t1 = |0|g1|0|g0|
"preceu.ph.qbla $t2, $t6 \n" // t2 = |0|b1|0|b0|
"preceu.ph.qbra $t5, $t6 \n" // t5 = |0|r1|0|r0|
"addu $t6, %[filter_val], %[fy] \n"
"ulw $t0, 0($t6) \n" // t0 = |cur_1|cur_0|
"ulw $t6, 4($t6) \n" // t6 = |cur_1|cur_0|
"dpa.w.ph $ac0, $t5, $t0 \n" // (cur*r1)+(cur*r0)
"dpa.w.ph $ac1, $t1, $t0 \n" // (cur*g1)+(cur*g0)
"dpa.w.ph $ac2, $t2, $t0 \n" // (cur*b1)+(cur*b0)
"precrq.qb.ph $t5, $t4, $t3 \n" // t5 = |a3|g3|a2|g2|
"precr.qb.ph $t0, $t4, $t3 \n" // t0 = |b3|r3|b2|r2|
"preceu.ph.qbra $t1, $t5 \n" // t1 = |0|g3|0|g2|
"preceu.ph.qbla $t2, $t0 \n" // t2 = |0|b3|0|b2|
"preceu.ph.qbra $t5, $t0 \n" // t5 = |0|r3|0|r2|
"dpa.w.ph $ac0, $t5, $t6 \n" // (cur*r1)+(cur*r0)
"dpa.w.ph $ac1, $t1, $t6 \n" // (cur*g1)+(cur*g0)
"dpa.w.ph $ac2, $t2, $t6 \n" // (cur*b1)+(cur*b0)
"addiu %[cnt], %[cnt], -1 \n"
"bgtz %[cnt], 11b \n"
" addiu %[fy], %[fy], 8 \n"
"2: \n"
"andi %[cnt], %[filter_len], 0x3 \n" // residual
"beqz %[cnt], 3f \n"
" nop \n"
"21: \n"
"addu $t0, %[filter_val], %[fy] \n"
"lh $t4, 0($t0) \n" // filter_val[fx]
"sll $t1, %[fy], 1 \n"
"addu $t0, %[src_data_rows], $t1 \n"
"lw $t1, 0($t0) \n"
"addu $t0, $t1, %[offset] \n"
"lbu $t1, 0($t0) \n" // t1 = row[fx*4 + 0]
"lbu $t2, 1($t0) \n" // t2 = row[fx*4 + 1]
"lbu $t3, 2($t0) \n" // t3 = row[fx*4 + 2]
"maddu $ac0, $t4, $t1 \n"
"maddu $ac1, $t4, $t2 \n"
"maddu $ac2, $t4, $t3 \n"
"addiu %[cnt], %[cnt], -1 \n"
"bgtz %[cnt], 21b \n"
" addiu %[fy], %[fy], 2 \n"
"3: \n"
"extrv.w $t3, $ac0, %[kShiftBits] \n" // r >> kShiftBits
"extrv.w $t2, $ac1, %[kShiftBits] \n" // g >> kShiftBits
"extrv.w $t1, $ac2, %[kShiftBits] \n" // b >> kShiftBits
"repl.ph $t6, 128 \n" // t6 = | 128 | 128 |
"addu $t5, %[out_row], %[offset] \n"
"append $t2, $t3, 16 \n" // t2 = |0|g|0|r|
"andi $t1, $t1, 0xFFFF \n"
"subu.ph $t1, $t1, $t6 \n"
"shll_s.ph $t1, $t1, 8 \n"
"shra.ph $t1, $t1, 8 \n"
"addu.ph $t1, $t1, $t6 \n" // Clamp(a)|Clamp(b)
"subu.ph $t2, $t2, $t6 \n"
"shll_s.ph $t2, $t2, 8 \n"
"shra.ph $t2, $t2, 8 \n"
"addu.ph $t2, $t2, $t6 \n" // Clamp(g)|Clamp(r)
"li $t0, 0xFF \n"
"ins $t1, $t0, 16, 8 \n"
"precr.qb.ph $t0, $t1, $t2 \n" // t0 = |a|b|g|r|
"usw $t0, 0($t5) \n"
".set pop \n"
: [filter_val] "+r" (filter_val), [filter_len] "+r" (filter_length),
[offset] "+r" (byte_offset), [fy] "+r" (filter_y), [cnt] "+r" (cnt),
[out_x] "+r" (out_x), [pixel_width] "+r" (pixel_width)
: [src_data_rows] "r" (source_data_rows), [out_row] "r" (out_row),
[kShiftBits] "r" (ConvolutionFilter1D::kShiftBits)
: "lo", "hi", "$ac1lo", "$ac1hi", "$ac2lo", "$ac2hi", "$ac3lo", "$ac3hi",
"t0", "t1", "t2", "t3", "t4", "t5", "t6", "memory"
);
}
}
#endif
}
} // namespace skia