/* libs/opengles/matrix.h
**
** Copyright 2006, The Android Open Source Project
**
** Licensed under the Apache License, Version 2.0 (the "License");
** you may not use this file except in compliance with the License.
** You may obtain a copy of the License at
**
** http://www.apache.org/licenses/LICENSE-2.0
**
** Unless required by applicable law or agreed to in writing, software
** distributed under the License is distributed on an "AS IS" BASIS,
** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
** See the License for the specific language governing permissions and
** limitations under the License.
*/
#ifndef ANDROID_OPENGLES_MATRIX_H
#define ANDROID_OPENGLES_MATRIX_H
#include <stdint.h>
#include <stddef.h>
#include <sys/types.h>
#include <utils/Log.h>
#include <private/pixelflinger/ggl_context.h>
#include <GLES/gl.h>
namespace android {
const int OGLES_MODELVIEW_STACK_DEPTH = 16;
const int OGLES_PROJECTION_STACK_DEPTH = 2;
const int OGLES_TEXTURE_STACK_DEPTH = 2;
void ogles_init_matrix(ogles_context_t*);
void ogles_uninit_matrix(ogles_context_t*);
void ogles_invalidate_perspective(ogles_context_t* c);
void ogles_validate_transform_impl(ogles_context_t* c, uint32_t want);
int ogles_surfaceport(ogles_context_t* c, GLint x, GLint y);
void ogles_scissor(ogles_context_t* c,
GLint x, GLint y, GLsizei w, GLsizei h);
void ogles_viewport(ogles_context_t* c,
GLint x, GLint y, GLsizei w, GLsizei h);
inline void ogles_validate_transform(
ogles_context_t* c, uint32_t want)
{
if (c->transforms.dirty & want)
ogles_validate_transform_impl(c, want);
}
// ----------------------------------------------------------------------------
inline
GLfixed vsquare3(GLfixed a, GLfixed b, GLfixed c)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
int32_t t;
asm(
"smull %0, %1, %2, %2 \n"
"smlal %0, %1, %3, %3 \n"
"smlal %0, %1, %4, %4 \n"
"movs %0, %0, lsr #16 \n"
"adc %0, %0, %1, lsl #16 \n"
: "=&r"(r), "=&r"(t)
: "%r"(a), "r"(b), "r"(c)
: "cc"
);
return r;
#elif defined(__mips__)
GLfixed res;
int32_t t1,t2,t3;
asm(
"mult %[a], %[a] \r\n"
"li %[res],0x8000 \r\n"
"madd %[b],%[b] \r\n"
"move %[t3],$zero \r\n"
"madd %[c],%[c] \r\n"
"mflo %[t1]\r\n"
"mfhi %[t2]\r\n"
"addu %[t1],%[res],%[t1]\r\n" /*add 0x8000*/
"sltu %[t3],%[t1],%[res]\r\n"
"addu %[t2],%[t2],%[t3]\r\n"
"srl %[res],%[t1],16\r\n"
"sll %[t2],%[t2],16\r\n"
"or %[res],%[res],%[t2]\r\n"
: [res]"=&r"(res),[t1]"=&r"(t1),[t2]"=&r"(t2),[t3]"=&r"(t3)
: [a] "r" (a),[b] "r" (b),[c] "r" (c)
: "%hi","%lo"
);
return res;
#else
return (( int64_t(a)*a +
int64_t(b)*b +
int64_t(c)*c + 0x8000)>>16);
#endif
}
static inline GLfixed mla2a( GLfixed a0, GLfixed b0,
GLfixed a1, GLfixed b1,
GLfixed c)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
int32_t t;
asm(
"smull %0, %1, %2, %3 \n"
"smlal %0, %1, %4, %5 \n"
"add %0, %6, %0, lsr #16 \n"
"add %0, %0, %1, lsl #16 \n"
: "=&r"(r), "=&r"(t)
: "%r"(a0), "r"(b0),
"%r"(a1), "r"(b1),
"r"(c)
:
);
return r;
#else
return (( int64_t(a0)*b0 +
int64_t(a1)*b1)>>16) + c;
#endif
}
static inline GLfixed mla3a( GLfixed a0, GLfixed b0,
GLfixed a1, GLfixed b1,
GLfixed a2, GLfixed b2,
GLfixed c)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
int32_t t;
asm(
"smull %0, %1, %2, %3 \n"
"smlal %0, %1, %4, %5 \n"
"smlal %0, %1, %6, %7 \n"
"add %0, %8, %0, lsr #16 \n"
"add %0, %0, %1, lsl #16 \n"
: "=&r"(r), "=&r"(t)
: "%r"(a0), "r"(b0),
"%r"(a1), "r"(b1),
"%r"(a2), "r"(b2),
"r"(c)
:
);
return r;
#elif defined(__mips__)
GLfixed res;
int32_t t1,t2;
asm(
"mult %[a0],%[b0] \r\n"
"madd %[a1],%[b1] \r\n"
"madd %[a2],%[b2] \r\n"
"mflo %[t2]\r\n"
"mfhi %[t1]\r\n"
"srl %[t2],%[t2],16\r\n"
"sll %[t1],%[t1],16\r\n"
"or %[t2],%[t2],%[t1]\r\n"
"addu %[res],%[t2],%[c]"
: [res]"=&r"(res),[t1]"=&r"(t1),[t2]"=&r"(t2)
: [a0] "r" (a0),[b0] "r" (b0),[a1] "r" (a1),[b1] "r" (b1),[a2] "r" (a2),[b2] "r" (b2),[c] "r" (c)
: "%hi","%lo"
);
return res;
#else
return (( int64_t(a0)*b0 +
int64_t(a1)*b1 +
int64_t(a2)*b2)>>16) + c;
#endif
}
// b0, b1, b2 are signed 16-bit quanities
// that have been shifted right by 'shift' bits relative to normal
// S16.16 fixed point
static inline GLfixed mla3a16( GLfixed a0, int32_t b1b0,
GLfixed a1,
GLfixed a2, int32_t b2,
GLint shift,
GLfixed c)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
asm(
"smulwb %0, %1, %2 \n"
"smlawt %0, %3, %2, %0 \n"
"smlawb %0, %4, %5, %0 \n"
"add %0, %7, %0, lsl %6 \n"
: "=&r"(r)
: "r"(a0), "r"(b1b0),
"r"(a1),
"r"(a2), "r"(b2),
"r"(shift),
"r"(c)
:
);
return r;
#else
int32_t accum;
int16_t b0 = b1b0 & 0xffff;
int16_t b1 = (b1b0 >> 16) & 0xffff;
accum = int64_t(a0)*int16_t(b0) >> 16;
accum += int64_t(a1)*int16_t(b1) >> 16;
accum += int64_t(a2)*int16_t(b2) >> 16;
accum = (accum << shift) + c;
return accum;
#endif
}
static inline GLfixed mla3a16_btb( GLfixed a0,
GLfixed a1,
GLfixed a2,
int32_t b1b0, int32_t xxb2,
GLint shift,
GLfixed c)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
asm(
"smulwb %0, %1, %4 \n"
"smlawt %0, %2, %4, %0 \n"
"smlawb %0, %3, %5, %0 \n"
"add %0, %7, %0, lsl %6 \n"
: "=&r"(r)
: "r"(a0),
"r"(a1),
"r"(a2),
"r"(b1b0), "r"(xxb2),
"r"(shift),
"r"(c)
:
);
return r;
#else
int32_t accum;
int16_t b0 = b1b0 & 0xffff;
int16_t b1 = (b1b0 >> 16) & 0xffff;
int16_t b2 = xxb2 & 0xffff;
accum = int64_t(a0)*int16_t(b0) >> 16;
accum += int64_t(a1)*int16_t(b1) >> 16;
accum += int64_t(a2)*int16_t(b2) >> 16;
accum = (accum << shift) + c;
return accum;
#endif
}
static inline GLfixed mla3a16_btt( GLfixed a0,
GLfixed a1,
GLfixed a2,
int32_t b1b0, int32_t b2xx,
GLint shift,
GLfixed c)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
asm(
"smulwb %0, %1, %4 \n"
"smlawt %0, %2, %4, %0 \n"
"smlawt %0, %3, %5, %0 \n"
"add %0, %7, %0, lsl %6 \n"
: "=&r"(r)
: "r"(a0),
"r"(a1),
"r"(a2),
"r"(b1b0), "r"(b2xx),
"r"(shift),
"r"(c)
:
);
return r;
#else
int32_t accum;
int16_t b0 = b1b0 & 0xffff;
int16_t b1 = (b1b0 >> 16) & 0xffff;
int16_t b2 = (b2xx >> 16) & 0xffff;
accum = int64_t(a0)*int16_t(b0) >> 16;
accum += int64_t(a1)*int16_t(b1) >> 16;
accum += int64_t(a2)*int16_t(b2) >> 16;
accum = (accum << shift) + c;
return accum;
#endif
}
static inline GLfixed mla3( GLfixed a0, GLfixed b0,
GLfixed a1, GLfixed b1,
GLfixed a2, GLfixed b2)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
int32_t t;
asm(
"smull %0, %1, %2, %3 \n"
"smlal %0, %1, %4, %5 \n"
"smlal %0, %1, %6, %7 \n"
"movs %0, %0, lsr #16 \n"
"adc %0, %0, %1, lsl #16 \n"
: "=&r"(r), "=&r"(t)
: "%r"(a0), "r"(b0),
"%r"(a1), "r"(b1),
"%r"(a2), "r"(b2)
: "cc"
);
return r;
#else
return (( int64_t(a0)*b0 +
int64_t(a1)*b1 +
int64_t(a2)*b2 + 0x8000)>>16);
#endif
}
static inline GLfixed mla4( GLfixed a0, GLfixed b0,
GLfixed a1, GLfixed b1,
GLfixed a2, GLfixed b2,
GLfixed a3, GLfixed b3)
{
#if defined(__arm__) && !defined(__thumb__)
GLfixed r;
int32_t t;
asm(
"smull %0, %1, %2, %3 \n"
"smlal %0, %1, %4, %5 \n"
"smlal %0, %1, %6, %7 \n"
"smlal %0, %1, %8, %9 \n"
"movs %0, %0, lsr #16 \n"
"adc %0, %0, %1, lsl #16 \n"
: "=&r"(r), "=&r"(t)
: "%r"(a0), "r"(b0),
"%r"(a1), "r"(b1),
"%r"(a2), "r"(b2),
"%r"(a3), "r"(b3)
: "cc"
);
return r;
#else
return (( int64_t(a0)*b0 +
int64_t(a1)*b1 +
int64_t(a2)*b2 +
int64_t(a3)*b3 + 0x8000)>>16);
#endif
}
inline
GLfixed dot4(const GLfixed* a, const GLfixed* b)
{
return mla4(a[0], b[0], a[1], b[1], a[2], b[2], a[3], b[3]);
}
inline
GLfixed dot3(const GLfixed* a, const GLfixed* b)
{
return mla3(a[0], b[0], a[1], b[1], a[2], b[2]);
}
}; // namespace android
#endif // ANDROID_OPENGLES_MATRIX_H