668 lines
36 KiB
C++
668 lines
36 KiB
C++
/*
|
|
* Copyright 2012 The LibYuv Project Authors. All rights reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "libyuv/basic_types.h"
|
|
#include "libyuv/row.h"
|
|
|
|
#ifdef __cplusplus
|
|
namespace libyuv {
|
|
extern "C" {
|
|
#endif
|
|
|
|
// This module is for GCC MIPS DSPR2
|
|
#if !defined(LIBYUV_DISABLE_DSPR2) && defined(__mips_dsp) && \
|
|
(__mips_dsp_rev >= 2) && (_MIPS_SIM == _MIPS_SIM_ABI32)
|
|
|
|
void ScaleRowDown2_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst,
|
|
int dst_width) {
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"srl $t9, %[dst_width], 4 \n" // iterations -> by 16
|
|
"beqz $t9, 2f \n"
|
|
" nop \n"
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t1, 4(%[src_ptr]) \n" // |7|6|5|4|
|
|
"lw $t2, 8(%[src_ptr]) \n" // |11|10|9|8|
|
|
"lw $t3, 12(%[src_ptr]) \n" // |15|14|13|12|
|
|
"lw $t4, 16(%[src_ptr]) \n" // |19|18|17|16|
|
|
"lw $t5, 20(%[src_ptr]) \n" // |23|22|21|20|
|
|
"lw $t6, 24(%[src_ptr]) \n" // |27|26|25|24|
|
|
"lw $t7, 28(%[src_ptr]) \n" // |31|30|29|28|
|
|
// TODO(fbarchard): Use odd pixels instead of even.
|
|
"precrq.qb.ph $t8, $t1, $t0 \n" // |7|5|3|1|
|
|
"precrq.qb.ph $t0, $t3, $t2 \n" // |15|13|11|9|
|
|
"precrq.qb.ph $t1, $t5, $t4 \n" // |23|21|19|17|
|
|
"precrq.qb.ph $t2, $t7, $t6 \n" // |31|29|27|25|
|
|
"addiu %[src_ptr], %[src_ptr], 32 \n"
|
|
"addiu $t9, $t9, -1 \n"
|
|
"sw $t8, 0(%[dst]) \n"
|
|
"sw $t0, 4(%[dst]) \n"
|
|
"sw $t1, 8(%[dst]) \n"
|
|
"sw $t2, 12(%[dst]) \n"
|
|
"bgtz $t9, 1b \n"
|
|
" addiu %[dst], %[dst], 16 \n"
|
|
|
|
"2: \n"
|
|
"andi $t9, %[dst_width], 0xf \n" // residue
|
|
"beqz $t9, 3f \n"
|
|
" nop \n"
|
|
|
|
"21: \n"
|
|
"lbu $t0, 1(%[src_ptr]) \n"
|
|
"addiu %[src_ptr], %[src_ptr], 2 \n"
|
|
"addiu $t9, $t9, -1 \n"
|
|
"sb $t0, 0(%[dst]) \n"
|
|
"bgtz $t9, 21b \n"
|
|
" addiu %[dst], %[dst], 1 \n"
|
|
|
|
"3: \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [dst] "+r"(dst)
|
|
: [dst_width] "r"(dst_width)
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8", "t9");
|
|
}
|
|
|
|
void ScaleRowDown2Box_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst,
|
|
int dst_width) {
|
|
const uint8* t = src_ptr + src_stride;
|
|
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"srl $t9, %[dst_width], 3 \n" // iterations -> step 8
|
|
"bltz $t9, 2f \n"
|
|
" nop \n"
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t1, 4(%[src_ptr]) \n" // |7|6|5|4|
|
|
"lw $t2, 8(%[src_ptr]) \n" // |11|10|9|8|
|
|
"lw $t3, 12(%[src_ptr]) \n" // |15|14|13|12|
|
|
"lw $t4, 0(%[t]) \n" // |19|18|17|16|
|
|
"lw $t5, 4(%[t]) \n" // |23|22|21|20|
|
|
"lw $t6, 8(%[t]) \n" // |27|26|25|24|
|
|
"lw $t7, 12(%[t]) \n" // |31|30|29|28|
|
|
"addiu $t9, $t9, -1 \n"
|
|
"srl $t8, $t0, 16 \n" // |X|X|3|2|
|
|
"ins $t0, $t4, 16, 16 \n" // |17|16|1|0|
|
|
"ins $t4, $t8, 0, 16 \n" // |19|18|3|2|
|
|
"raddu.w.qb $t0, $t0 \n" // |17+16+1+0|
|
|
"raddu.w.qb $t4, $t4 \n" // |19+18+3+2|
|
|
"shra_r.w $t0, $t0, 2 \n" // |t0+2|>>2
|
|
"shra_r.w $t4, $t4, 2 \n" // |t4+2|>>2
|
|
"srl $t8, $t1, 16 \n" // |X|X|7|6|
|
|
"ins $t1, $t5, 16, 16 \n" // |21|20|5|4|
|
|
"ins $t5, $t8, 0, 16 \n" // |22|23|7|6|
|
|
"raddu.w.qb $t1, $t1 \n" // |21+20+5+4|
|
|
"raddu.w.qb $t5, $t5 \n" // |23+22+7+6|
|
|
"shra_r.w $t1, $t1, 2 \n" // |t1+2|>>2
|
|
"shra_r.w $t5, $t5, 2 \n" // |t5+2|>>2
|
|
"srl $t8, $t2, 16 \n" // |X|X|11|10|
|
|
"ins $t2, $t6, 16, 16 \n" // |25|24|9|8|
|
|
"ins $t6, $t8, 0, 16 \n" // |27|26|11|10|
|
|
"raddu.w.qb $t2, $t2 \n" // |25+24+9+8|
|
|
"raddu.w.qb $t6, $t6 \n" // |27+26+11+10|
|
|
"shra_r.w $t2, $t2, 2 \n" // |t2+2|>>2
|
|
"shra_r.w $t6, $t6, 2 \n" // |t5+2|>>2
|
|
"srl $t8, $t3, 16 \n" // |X|X|15|14|
|
|
"ins $t3, $t7, 16, 16 \n" // |29|28|13|12|
|
|
"ins $t7, $t8, 0, 16 \n" // |31|30|15|14|
|
|
"raddu.w.qb $t3, $t3 \n" // |29+28+13+12|
|
|
"raddu.w.qb $t7, $t7 \n" // |31+30+15+14|
|
|
"shra_r.w $t3, $t3, 2 \n" // |t3+2|>>2
|
|
"shra_r.w $t7, $t7, 2 \n" // |t7+2|>>2
|
|
"addiu %[src_ptr], %[src_ptr], 16 \n"
|
|
"addiu %[t], %[t], 16 \n"
|
|
"sb $t0, 0(%[dst]) \n"
|
|
"sb $t4, 1(%[dst]) \n"
|
|
"sb $t1, 2(%[dst]) \n"
|
|
"sb $t5, 3(%[dst]) \n"
|
|
"sb $t2, 4(%[dst]) \n"
|
|
"sb $t6, 5(%[dst]) \n"
|
|
"sb $t3, 6(%[dst]) \n"
|
|
"sb $t7, 7(%[dst]) \n"
|
|
"bgtz $t9, 1b \n"
|
|
" addiu %[dst], %[dst], 8 \n"
|
|
|
|
"2: \n"
|
|
"andi $t9, %[dst_width], 0x7 \n" // x = residue
|
|
"beqz $t9, 3f \n"
|
|
" nop \n"
|
|
|
|
"21: \n"
|
|
"lwr $t1, 0(%[src_ptr]) \n"
|
|
"lwl $t1, 3(%[src_ptr]) \n"
|
|
"lwr $t2, 0(%[t]) \n"
|
|
"lwl $t2, 3(%[t]) \n"
|
|
"srl $t8, $t1, 16 \n"
|
|
"ins $t1, $t2, 16, 16 \n"
|
|
"ins $t2, $t8, 0, 16 \n"
|
|
"raddu.w.qb $t1, $t1 \n"
|
|
"raddu.w.qb $t2, $t2 \n"
|
|
"shra_r.w $t1, $t1, 2 \n"
|
|
"shra_r.w $t2, $t2, 2 \n"
|
|
"sb $t1, 0(%[dst]) \n"
|
|
"sb $t2, 1(%[dst]) \n"
|
|
"addiu %[src_ptr], %[src_ptr], 4 \n"
|
|
"addiu $t9, $t9, -2 \n"
|
|
"addiu %[t], %[t], 4 \n"
|
|
"bgtz $t9, 21b \n"
|
|
" addiu %[dst], %[dst], 2 \n"
|
|
|
|
"3: \n"
|
|
".set pop \n"
|
|
|
|
: [src_ptr] "+r"(src_ptr), [dst] "+r"(dst), [t] "+r"(t)
|
|
: [dst_width] "r"(dst_width)
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8", "t9");
|
|
}
|
|
|
|
void ScaleRowDown4_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst,
|
|
int dst_width) {
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"srl $t9, %[dst_width], 3 \n"
|
|
"beqz $t9, 2f \n"
|
|
" nop \n"
|
|
|
|
"1: \n"
|
|
"lw $t1, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t2, 4(%[src_ptr]) \n" // |7|6|5|4|
|
|
"lw $t3, 8(%[src_ptr]) \n" // |11|10|9|8|
|
|
"lw $t4, 12(%[src_ptr]) \n" // |15|14|13|12|
|
|
"lw $t5, 16(%[src_ptr]) \n" // |19|18|17|16|
|
|
"lw $t6, 20(%[src_ptr]) \n" // |23|22|21|20|
|
|
"lw $t7, 24(%[src_ptr]) \n" // |27|26|25|24|
|
|
"lw $t8, 28(%[src_ptr]) \n" // |31|30|29|28|
|
|
"precr.qb.ph $t1, $t2, $t1 \n" // |6|4|2|0|
|
|
"precr.qb.ph $t2, $t4, $t3 \n" // |14|12|10|8|
|
|
"precr.qb.ph $t5, $t6, $t5 \n" // |22|20|18|16|
|
|
"precr.qb.ph $t6, $t8, $t7 \n" // |30|28|26|24|
|
|
"precrq.qb.ph $t1, $t2, $t1 \n" // |14|10|6|2|
|
|
"precrq.qb.ph $t5, $t6, $t5 \n" // |30|26|22|18|
|
|
"addiu %[src_ptr], %[src_ptr], 32 \n"
|
|
"addiu $t9, $t9, -1 \n"
|
|
"sw $t1, 0(%[dst]) \n"
|
|
"sw $t5, 4(%[dst]) \n"
|
|
"bgtz $t9, 1b \n"
|
|
" addiu %[dst], %[dst], 8 \n"
|
|
|
|
"2: \n"
|
|
"andi $t9, %[dst_width], 7 \n" // residue
|
|
"beqz $t9, 3f \n"
|
|
" nop \n"
|
|
|
|
"21: \n"
|
|
"lbu $t1, 2(%[src_ptr]) \n"
|
|
"addiu %[src_ptr], %[src_ptr], 4 \n"
|
|
"addiu $t9, $t9, -1 \n"
|
|
"sb $t1, 0(%[dst]) \n"
|
|
"bgtz $t9, 21b \n"
|
|
" addiu %[dst], %[dst], 1 \n"
|
|
|
|
"3: \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [dst] "+r"(dst)
|
|
: [dst_width] "r"(dst_width)
|
|
: "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8", "t9");
|
|
}
|
|
|
|
void ScaleRowDown4Box_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst,
|
|
int dst_width) {
|
|
intptr_t stride = src_stride;
|
|
const uint8* s1 = src_ptr + stride;
|
|
const uint8* s2 = s1 + stride;
|
|
const uint8* s3 = s2 + stride;
|
|
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"srl $t9, %[dst_width], 1 \n"
|
|
"andi $t8, %[dst_width], 1 \n"
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t1, 0(%[s1]) \n" // |7|6|5|4|
|
|
"lw $t2, 0(%[s2]) \n" // |11|10|9|8|
|
|
"lw $t3, 0(%[s3]) \n" // |15|14|13|12|
|
|
"lw $t4, 4(%[src_ptr]) \n" // |19|18|17|16|
|
|
"lw $t5, 4(%[s1]) \n" // |23|22|21|20|
|
|
"lw $t6, 4(%[s2]) \n" // |27|26|25|24|
|
|
"lw $t7, 4(%[s3]) \n" // |31|30|29|28|
|
|
"raddu.w.qb $t0, $t0 \n" // |3 + 2 + 1 + 0|
|
|
"raddu.w.qb $t1, $t1 \n" // |7 + 6 + 5 + 4|
|
|
"raddu.w.qb $t2, $t2 \n" // |11 + 10 + 9 + 8|
|
|
"raddu.w.qb $t3, $t3 \n" // |15 + 14 + 13 + 12|
|
|
"raddu.w.qb $t4, $t4 \n" // |19 + 18 + 17 + 16|
|
|
"raddu.w.qb $t5, $t5 \n" // |23 + 22 + 21 + 20|
|
|
"raddu.w.qb $t6, $t6 \n" // |27 + 26 + 25 + 24|
|
|
"raddu.w.qb $t7, $t7 \n" // |31 + 30 + 29 + 28|
|
|
"add $t0, $t0, $t1 \n"
|
|
"add $t1, $t2, $t3 \n"
|
|
"add $t0, $t0, $t1 \n"
|
|
"add $t4, $t4, $t5 \n"
|
|
"add $t6, $t6, $t7 \n"
|
|
"add $t4, $t4, $t6 \n"
|
|
"shra_r.w $t0, $t0, 4 \n"
|
|
"shra_r.w $t4, $t4, 4 \n"
|
|
"sb $t0, 0(%[dst]) \n"
|
|
"sb $t4, 1(%[dst]) \n"
|
|
"addiu %[src_ptr], %[src_ptr], 8 \n"
|
|
"addiu %[s1], %[s1], 8 \n"
|
|
"addiu %[s2], %[s2], 8 \n"
|
|
"addiu %[s3], %[s3], 8 \n"
|
|
"addiu $t9, $t9, -1 \n"
|
|
"bgtz $t9, 1b \n"
|
|
" addiu %[dst], %[dst], 2 \n"
|
|
"beqz $t8, 2f \n"
|
|
" nop \n"
|
|
|
|
"lw $t0, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t1, 0(%[s1]) \n" // |7|6|5|4|
|
|
"lw $t2, 0(%[s2]) \n" // |11|10|9|8|
|
|
"lw $t3, 0(%[s3]) \n" // |15|14|13|12|
|
|
"raddu.w.qb $t0, $t0 \n" // |3 + 2 + 1 + 0|
|
|
"raddu.w.qb $t1, $t1 \n" // |7 + 6 + 5 + 4|
|
|
"raddu.w.qb $t2, $t2 \n" // |11 + 10 + 9 + 8|
|
|
"raddu.w.qb $t3, $t3 \n" // |15 + 14 + 13 + 12|
|
|
"add $t0, $t0, $t1 \n"
|
|
"add $t1, $t2, $t3 \n"
|
|
"add $t0, $t0, $t1 \n"
|
|
"shra_r.w $t0, $t0, 4 \n"
|
|
"sb $t0, 0(%[dst]) \n"
|
|
|
|
"2: \n"
|
|
".set pop \n"
|
|
|
|
: [src_ptr] "+r"(src_ptr), [dst] "+r"(dst), [s1] "+r"(s1), [s2] "+r"(s2),
|
|
[s3] "+r"(s3)
|
|
: [dst_width] "r"(dst_width)
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8", "t9");
|
|
}
|
|
|
|
void ScaleRowDown34_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst,
|
|
int dst_width) {
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
"1: \n"
|
|
"lw $t1, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t2, 4(%[src_ptr]) \n" // |7|6|5|4|
|
|
"lw $t3, 8(%[src_ptr]) \n" // |11|10|9|8|
|
|
"lw $t4, 12(%[src_ptr]) \n" // |15|14|13|12|
|
|
"lw $t5, 16(%[src_ptr]) \n" // |19|18|17|16|
|
|
"lw $t6, 20(%[src_ptr]) \n" // |23|22|21|20|
|
|
"lw $t7, 24(%[src_ptr]) \n" // |27|26|25|24|
|
|
"lw $t8, 28(%[src_ptr]) \n" // |31|30|29|28|
|
|
"precrq.qb.ph $t0, $t2, $t4 \n" // |7|5|15|13|
|
|
"precrq.qb.ph $t9, $t6, $t8 \n" // |23|21|31|30|
|
|
"addiu %[dst_width], %[dst_width], -24 \n"
|
|
"ins $t1, $t1, 8, 16 \n" // |3|1|0|X|
|
|
"ins $t4, $t0, 8, 16 \n" // |X|15|13|12|
|
|
"ins $t5, $t5, 8, 16 \n" // |19|17|16|X|
|
|
"ins $t8, $t9, 8, 16 \n" // |X|31|29|28|
|
|
"addiu %[src_ptr], %[src_ptr], 32 \n"
|
|
"packrl.ph $t0, $t3, $t0 \n" // |9|8|7|5|
|
|
"packrl.ph $t9, $t7, $t9 \n" // |25|24|23|21|
|
|
"prepend $t1, $t2, 8 \n" // |4|3|1|0|
|
|
"prepend $t3, $t4, 24 \n" // |15|13|12|11|
|
|
"prepend $t5, $t6, 8 \n" // |20|19|17|16|
|
|
"prepend $t7, $t8, 24 \n" // |31|29|28|27|
|
|
"sw $t1, 0(%[dst]) \n"
|
|
"sw $t0, 4(%[dst]) \n"
|
|
"sw $t3, 8(%[dst]) \n"
|
|
"sw $t5, 12(%[dst]) \n"
|
|
"sw $t9, 16(%[dst]) \n"
|
|
"sw $t7, 20(%[dst]) \n"
|
|
"bnez %[dst_width], 1b \n"
|
|
" addiu %[dst], %[dst], 24 \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [dst] "+r"(dst), [dst_width] "+r"(dst_width)
|
|
:
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8", "t9");
|
|
}
|
|
|
|
void ScaleRowDown34_0_Box_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* d,
|
|
int dst_width) {
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
"repl.ph $t3, 3 \n" // 0x00030003
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |S3|S2|S1|S0|
|
|
"lwx $t1, %[src_stride](%[src_ptr]) \n" // |T3|T2|T1|T0|
|
|
"rotr $t2, $t0, 8 \n" // |S0|S3|S2|S1|
|
|
"rotr $t6, $t1, 8 \n" // |T0|T3|T2|T1|
|
|
"muleu_s.ph.qbl $t4, $t2, $t3 \n" // |S0*3|S3*3|
|
|
"muleu_s.ph.qbl $t5, $t6, $t3 \n" // |T0*3|T3*3|
|
|
"andi $t0, $t2, 0xFFFF \n" // |0|0|S2|S1|
|
|
"andi $t1, $t6, 0xFFFF \n" // |0|0|T2|T1|
|
|
"raddu.w.qb $t0, $t0 \n"
|
|
"raddu.w.qb $t1, $t1 \n"
|
|
"shra_r.w $t0, $t0, 1 \n"
|
|
"shra_r.w $t1, $t1, 1 \n"
|
|
"preceu.ph.qbr $t2, $t2 \n" // |0|S2|0|S1|
|
|
"preceu.ph.qbr $t6, $t6 \n" // |0|T2|0|T1|
|
|
"rotr $t2, $t2, 16 \n" // |0|S1|0|S2|
|
|
"rotr $t6, $t6, 16 \n" // |0|T1|0|T2|
|
|
"addu.ph $t2, $t2, $t4 \n"
|
|
"addu.ph $t6, $t6, $t5 \n"
|
|
"sll $t5, $t0, 1 \n"
|
|
"add $t0, $t5, $t0 \n"
|
|
"shra_r.ph $t2, $t2, 2 \n"
|
|
"shra_r.ph $t6, $t6, 2 \n"
|
|
"shll.ph $t4, $t2, 1 \n"
|
|
"addq.ph $t4, $t4, $t2 \n"
|
|
"addu $t0, $t0, $t1 \n"
|
|
"addiu %[src_ptr], %[src_ptr], 4 \n"
|
|
"shra_r.w $t0, $t0, 2 \n"
|
|
"addu.ph $t6, $t6, $t4 \n"
|
|
"shra_r.ph $t6, $t6, 2 \n"
|
|
"srl $t1, $t6, 16 \n"
|
|
"addiu %[dst_width], %[dst_width], -3 \n"
|
|
"sb $t1, 0(%[d]) \n"
|
|
"sb $t0, 1(%[d]) \n"
|
|
"sb $t6, 2(%[d]) \n"
|
|
"bgtz %[dst_width], 1b \n"
|
|
" addiu %[d], %[d], 3 \n"
|
|
"3: \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [src_stride] "+r"(src_stride), [d] "+r"(d),
|
|
[dst_width] "+r"(dst_width)
|
|
:
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6");
|
|
}
|
|
|
|
void ScaleRowDown34_1_Box_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* d,
|
|
int dst_width) {
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
"repl.ph $t2, 3 \n" // 0x00030003
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |S3|S2|S1|S0|
|
|
"lwx $t1, %[src_stride](%[src_ptr]) \n" // |T3|T2|T1|T0|
|
|
"rotr $t4, $t0, 8 \n" // |S0|S3|S2|S1|
|
|
"rotr $t6, $t1, 8 \n" // |T0|T3|T2|T1|
|
|
"muleu_s.ph.qbl $t3, $t4, $t2 \n" // |S0*3|S3*3|
|
|
"muleu_s.ph.qbl $t5, $t6, $t2 \n" // |T0*3|T3*3|
|
|
"andi $t0, $t4, 0xFFFF \n" // |0|0|S2|S1|
|
|
"andi $t1, $t6, 0xFFFF \n" // |0|0|T2|T1|
|
|
"raddu.w.qb $t0, $t0 \n"
|
|
"raddu.w.qb $t1, $t1 \n"
|
|
"shra_r.w $t0, $t0, 1 \n"
|
|
"shra_r.w $t1, $t1, 1 \n"
|
|
"preceu.ph.qbr $t4, $t4 \n" // |0|S2|0|S1|
|
|
"preceu.ph.qbr $t6, $t6 \n" // |0|T2|0|T1|
|
|
"rotr $t4, $t4, 16 \n" // |0|S1|0|S2|
|
|
"rotr $t6, $t6, 16 \n" // |0|T1|0|T2|
|
|
"addu.ph $t4, $t4, $t3 \n"
|
|
"addu.ph $t6, $t6, $t5 \n"
|
|
"shra_r.ph $t6, $t6, 2 \n"
|
|
"shra_r.ph $t4, $t4, 2 \n"
|
|
"addu.ph $t6, $t6, $t4 \n"
|
|
"addiu %[src_ptr], %[src_ptr], 4 \n"
|
|
"shra_r.ph $t6, $t6, 1 \n"
|
|
"addu $t0, $t0, $t1 \n"
|
|
"addiu %[dst_width], %[dst_width], -3 \n"
|
|
"shra_r.w $t0, $t0, 1 \n"
|
|
"srl $t1, $t6, 16 \n"
|
|
"sb $t1, 0(%[d]) \n"
|
|
"sb $t0, 1(%[d]) \n"
|
|
"sb $t6, 2(%[d]) \n"
|
|
"bgtz %[dst_width], 1b \n"
|
|
" addiu %[d], %[d], 3 \n"
|
|
"3: \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [src_stride] "+r"(src_stride), [d] "+r"(d),
|
|
[dst_width] "+r"(dst_width)
|
|
:
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6");
|
|
}
|
|
|
|
void ScaleRowDown38_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst,
|
|
int dst_width) {
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |3|2|1|0|
|
|
"lw $t1, 4(%[src_ptr]) \n" // |7|6|5|4|
|
|
"lw $t2, 8(%[src_ptr]) \n" // |11|10|9|8|
|
|
"lw $t3, 12(%[src_ptr]) \n" // |15|14|13|12|
|
|
"lw $t4, 16(%[src_ptr]) \n" // |19|18|17|16|
|
|
"lw $t5, 20(%[src_ptr]) \n" // |23|22|21|20|
|
|
"lw $t6, 24(%[src_ptr]) \n" // |27|26|25|24|
|
|
"lw $t7, 28(%[src_ptr]) \n" // |31|30|29|28|
|
|
"wsbh $t0, $t0 \n" // |2|3|0|1|
|
|
"wsbh $t6, $t6 \n" // |26|27|24|25|
|
|
"srl $t0, $t0, 8 \n" // |X|2|3|0|
|
|
"srl $t3, $t3, 16 \n" // |X|X|15|14|
|
|
"srl $t5, $t5, 16 \n" // |X|X|23|22|
|
|
"srl $t7, $t7, 16 \n" // |X|X|31|30|
|
|
"ins $t1, $t2, 24, 8 \n" // |8|6|5|4|
|
|
"ins $t6, $t5, 0, 8 \n" // |26|27|24|22|
|
|
"ins $t1, $t0, 0, 16 \n" // |8|6|3|0|
|
|
"ins $t6, $t7, 24, 8 \n" // |30|27|24|22|
|
|
"prepend $t2, $t3, 24 \n" // |X|15|14|11|
|
|
"ins $t4, $t4, 16, 8 \n" // |19|16|17|X|
|
|
"ins $t4, $t2, 0, 16 \n" // |19|16|14|11|
|
|
"addiu %[src_ptr], %[src_ptr], 32 \n"
|
|
"addiu %[dst_width], %[dst_width], -12 \n"
|
|
"addiu $t8,%[dst_width], -12 \n"
|
|
"sw $t1, 0(%[dst]) \n"
|
|
"sw $t4, 4(%[dst]) \n"
|
|
"sw $t6, 8(%[dst]) \n"
|
|
"bgez $t8, 1b \n"
|
|
" addiu %[dst], %[dst], 12 \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [dst] "+r"(dst), [dst_width] "+r"(dst_width)
|
|
:
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8");
|
|
}
|
|
|
|
void ScaleRowDown38_2_Box_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst_ptr,
|
|
int dst_width) {
|
|
intptr_t stride = src_stride;
|
|
const uint8* t = src_ptr + stride;
|
|
const int c = 0x2AAA;
|
|
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |S3|S2|S1|S0|
|
|
"lw $t1, 4(%[src_ptr]) \n" // |S7|S6|S5|S4|
|
|
"lw $t2, 0(%[t]) \n" // |T3|T2|T1|T0|
|
|
"lw $t3, 4(%[t]) \n" // |T7|T6|T5|T4|
|
|
"rotr $t1, $t1, 16 \n" // |S5|S4|S7|S6|
|
|
"packrl.ph $t4, $t1, $t3 \n" // |S7|S6|T7|T6|
|
|
"packrl.ph $t5, $t3, $t1 \n" // |T5|T4|S5|S4|
|
|
"raddu.w.qb $t4, $t4 \n" // S7+S6+T7+T6
|
|
"raddu.w.qb $t5, $t5 \n" // T5+T4+S5+S4
|
|
"precrq.qb.ph $t6, $t0, $t2 \n" // |S3|S1|T3|T1|
|
|
"precrq.qb.ph $t6, $t6, $t6 \n" // |S3|T3|S3|T3|
|
|
"srl $t4, $t4, 2 \n" // t4 / 4
|
|
"srl $t6, $t6, 16 \n" // |0|0|S3|T3|
|
|
"raddu.w.qb $t6, $t6 \n" // 0+0+S3+T3
|
|
"addu $t6, $t5, $t6 \n"
|
|
"mul $t6, $t6, %[c] \n" // t6 * 0x2AAA
|
|
"sll $t0, $t0, 8 \n" // |S2|S1|S0|0|
|
|
"sll $t2, $t2, 8 \n" // |T2|T1|T0|0|
|
|
"raddu.w.qb $t0, $t0 \n" // S2+S1+S0+0
|
|
"raddu.w.qb $t2, $t2 \n" // T2+T1+T0+0
|
|
"addu $t0, $t0, $t2 \n"
|
|
"mul $t0, $t0, %[c] \n" // t0 * 0x2AAA
|
|
"addiu %[src_ptr], %[src_ptr], 8 \n"
|
|
"addiu %[t], %[t], 8 \n"
|
|
"addiu %[dst_width], %[dst_width], -3 \n"
|
|
"addiu %[dst_ptr], %[dst_ptr], 3 \n"
|
|
"srl $t6, $t6, 16 \n"
|
|
"srl $t0, $t0, 16 \n"
|
|
"sb $t4, -1(%[dst_ptr]) \n"
|
|
"sb $t6, -2(%[dst_ptr]) \n"
|
|
"bgtz %[dst_width], 1b \n"
|
|
" sb $t0, -3(%[dst_ptr]) \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [dst_ptr] "+r"(dst_ptr), [t] "+r"(t),
|
|
[dst_width] "+r"(dst_width)
|
|
: [c] "r"(c)
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6");
|
|
}
|
|
|
|
void ScaleRowDown38_3_Box_DSPR2(const uint8* src_ptr,
|
|
ptrdiff_t src_stride,
|
|
uint8* dst_ptr,
|
|
int dst_width) {
|
|
intptr_t stride = src_stride;
|
|
const uint8* s1 = src_ptr + stride;
|
|
stride += stride;
|
|
const uint8* s2 = src_ptr + stride;
|
|
const int c1 = 0x1C71;
|
|
const int c2 = 0x2AAA;
|
|
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
|
|
"1: \n"
|
|
"lw $t0, 0(%[src_ptr]) \n" // |S3|S2|S1|S0|
|
|
"lw $t1, 4(%[src_ptr]) \n" // |S7|S6|S5|S4|
|
|
"lw $t2, 0(%[s1]) \n" // |T3|T2|T1|T0|
|
|
"lw $t3, 4(%[s1]) \n" // |T7|T6|T5|T4|
|
|
"lw $t4, 0(%[s2]) \n" // |R3|R2|R1|R0|
|
|
"lw $t5, 4(%[s2]) \n" // |R7|R6|R5|R4|
|
|
"rotr $t1, $t1, 16 \n" // |S5|S4|S7|S6|
|
|
"packrl.ph $t6, $t1, $t3 \n" // |S7|S6|T7|T6|
|
|
"raddu.w.qb $t6, $t6 \n" // S7+S6+T7+T6
|
|
"packrl.ph $t7, $t3, $t1 \n" // |T5|T4|S5|S4|
|
|
"raddu.w.qb $t7, $t7 \n" // T5+T4+S5+S4
|
|
"sll $t8, $t5, 16 \n" // |R5|R4|0|0|
|
|
"raddu.w.qb $t8, $t8 \n" // R5+R4
|
|
"addu $t7, $t7, $t8 \n"
|
|
"srl $t8, $t5, 16 \n" // |0|0|R7|R6|
|
|
"raddu.w.qb $t8, $t8 \n" // R7 + R6
|
|
"addu $t6, $t6, $t8 \n"
|
|
"mul $t6, $t6, %[c2] \n" // t6 * 0x2AAA
|
|
"precrq.qb.ph $t8, $t0, $t2 \n" // |S3|S1|T3|T1|
|
|
"precrq.qb.ph $t8, $t8, $t4 \n" // |S3|T3|R3|R1|
|
|
"srl $t8, $t8, 8 \n" // |0|S3|T3|R3|
|
|
"raddu.w.qb $t8, $t8 \n" // S3 + T3 + R3
|
|
"addu $t7, $t7, $t8 \n"
|
|
"mul $t7, $t7, %[c1] \n" // t7 * 0x1C71
|
|
"sll $t0, $t0, 8 \n" // |S2|S1|S0|0|
|
|
"sll $t2, $t2, 8 \n" // |T2|T1|T0|0|
|
|
"sll $t4, $t4, 8 \n" // |R2|R1|R0|0|
|
|
"raddu.w.qb $t0, $t0 \n"
|
|
"raddu.w.qb $t2, $t2 \n"
|
|
"raddu.w.qb $t4, $t4 \n"
|
|
"addu $t0, $t0, $t2 \n"
|
|
"addu $t0, $t0, $t4 \n"
|
|
"mul $t0, $t0, %[c1] \n" // t0 * 0x1C71
|
|
"addiu %[src_ptr], %[src_ptr], 8 \n"
|
|
"addiu %[s1], %[s1], 8 \n"
|
|
"addiu %[s2], %[s2], 8 \n"
|
|
"addiu %[dst_width], %[dst_width], -3 \n"
|
|
"addiu %[dst_ptr], %[dst_ptr], 3 \n"
|
|
"srl $t6, $t6, 16 \n"
|
|
"srl $t7, $t7, 16 \n"
|
|
"srl $t0, $t0, 16 \n"
|
|
"sb $t6, -1(%[dst_ptr]) \n"
|
|
"sb $t7, -2(%[dst_ptr]) \n"
|
|
"bgtz %[dst_width], 1b \n"
|
|
" sb $t0, -3(%[dst_ptr]) \n"
|
|
".set pop \n"
|
|
: [src_ptr] "+r"(src_ptr), [dst_ptr] "+r"(dst_ptr), [s1] "+r"(s1),
|
|
[s2] "+r"(s2), [dst_width] "+r"(dst_width)
|
|
: [c1] "r"(c1), [c2] "r"(c2)
|
|
: "t0", "t1", "t2", "t3", "t4", "t5", "t6", "t7", "t8");
|
|
}
|
|
|
|
void ScaleAddRow_DSPR2(const uint8* src_ptr, uint16* dst_ptr, int src_width) {
|
|
int x;
|
|
for (x = 0; x < ((src_width - 1)); x += 8) {
|
|
uint32 tmp_t1, tmp_t2, tmp_t3, tmp_t4;
|
|
uint32 tmp_t5, tmp_t6, tmp_t7, tmp_t8;
|
|
__asm__ __volatile__(
|
|
".set push \n"
|
|
".set noreorder \n"
|
|
"lw %[tmp_t5], 0(%[src_ptr]) \n"
|
|
"lw %[tmp_t6], 4(%[src_ptr]) \n"
|
|
"lw %[tmp_t1], 0(%[dst_ptr]) \n"
|
|
"lw %[tmp_t2], 4(%[dst_ptr]) \n"
|
|
"lw %[tmp_t3], 8(%[dst_ptr]) \n"
|
|
"lw %[tmp_t4], 12(%[dst_ptr]) \n"
|
|
"preceu.ph.qbr %[tmp_t7], %[tmp_t5] \n"
|
|
"preceu.ph.qbl %[tmp_t8], %[tmp_t5] \n"
|
|
"addu.ph %[tmp_t1], %[tmp_t1], %[tmp_t7] \n"
|
|
"addu.ph %[tmp_t2], %[tmp_t2], %[tmp_t8] \n"
|
|
"preceu.ph.qbr %[tmp_t7], %[tmp_t6] \n"
|
|
"preceu.ph.qbl %[tmp_t8], %[tmp_t6] \n"
|
|
"addu.ph %[tmp_t3], %[tmp_t3], %[tmp_t7] \n"
|
|
"addu.ph %[tmp_t4], %[tmp_t4], %[tmp_t8] \n"
|
|
"sw %[tmp_t1], 0(%[dst_ptr]) \n"
|
|
"sw %[tmp_t2], 4(%[dst_ptr]) \n"
|
|
"sw %[tmp_t3], 8(%[dst_ptr]) \n"
|
|
"sw %[tmp_t4], 12(%[dst_ptr]) \n"
|
|
".set pop \n"
|
|
:
|
|
[tmp_t1] "=&r"(tmp_t1), [tmp_t2] "=&r"(tmp_t2), [tmp_t3] "=&r"(tmp_t3),
|
|
[tmp_t4] "=&r"(tmp_t4), [tmp_t5] "=&r"(tmp_t5), [tmp_t6] "=&r"(tmp_t6),
|
|
[tmp_t7] "=&r"(tmp_t7), [tmp_t8] "=&r"(tmp_t8), [src_ptr] "+r"(src_ptr)
|
|
: [dst_ptr] "r"(dst_ptr));
|
|
src_ptr += 8;
|
|
dst_ptr += 8;
|
|
}
|
|
|
|
if ((src_width)&7) {
|
|
for (x = 0; x < ((src_width - 1) & 7); x += 1) {
|
|
dst_ptr[0] += src_ptr[0];
|
|
src_ptr += 1;
|
|
dst_ptr += 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
#endif // defined(__mips_dsp) && (__mips_dsp_rev >= 2)
|
|
|
|
#ifdef __cplusplus
|
|
} // extern "C"
|
|
} // namespace libyuv
|
|
#endif
|