aboutsummaryrefslogtreecommitdiff
path: root/pixman/pixman/pixman-x64-mmx-emulation.h
diff options
context:
space:
mode:
Diffstat (limited to 'pixman/pixman/pixman-x64-mmx-emulation.h')
-rw-r--r--pixman/pixman/pixman-x64-mmx-emulation.h263
1 files changed, 263 insertions, 0 deletions
diff --git a/pixman/pixman/pixman-x64-mmx-emulation.h b/pixman/pixman/pixman-x64-mmx-emulation.h
new file mode 100644
index 000000000..231030fba
--- /dev/null
+++ b/pixman/pixman/pixman-x64-mmx-emulation.h
@@ -0,0 +1,263 @@
+#ifndef MMX_X64_H_INCLUDED
+#define MMX_X64_H_INCLUDED
+
+/* Implementation of x64 MMX substitition functions, before
+ * pixman is reimplemented not to use __m64 type on Visual C++
+ *
+ * Copyright (C)2009 by George Yohng
+ * Released in public domain.
+ */
+
+#include <intrin.h>
+
+#define M64C(a) (*(const __m64 *)(&a))
+#define M64U(a) (*(const unsigned long long *)(&a))
+
+__inline __m64
+_m_from_int (int a)
+{
+ long long i64 = a;
+
+ return M64C (i64);
+}
+
+__inline __m64
+_mm_setzero_si64 ()
+{
+ long long i64 = 0;
+
+ return M64C (i64);
+}
+
+__inline __m64
+_mm_set_pi32 (int i1, int i0)
+{
+ unsigned long long i64 = ((unsigned)i0) + (((unsigned long long)(unsigned)i1) << 32);
+
+ return M64C (i64);
+}
+
+__inline void
+_m_empty ()
+{
+}
+
+__inline __m64
+_mm_set1_pi16 (short w)
+{
+ unsigned long long i64 = ((unsigned long long)(unsigned short)(w)) * 0x0001000100010001ULL;
+
+ return M64C (i64);
+}
+
+__inline int
+_m_to_int (__m64 m)
+{
+ return m.m64_i32[0];
+}
+
+__inline __m64
+_mm_movepi64_pi64 (__m128i a)
+{
+ return M64C (a.m128i_i64[0]);
+}
+
+__inline __m64
+_m_pand (__m64 a, __m64 b)
+{
+ unsigned long long i64 = M64U (a) & M64U (b);
+
+ return M64C (i64);
+}
+
+__inline __m64
+_m_por (__m64 a, __m64 b)
+{
+ unsigned long long i64 = M64U (a) | M64U (b);
+
+ return M64C (i64);
+}
+
+__inline __m64
+_m_pxor (__m64 a, __m64 b)
+{
+ unsigned long long i64 = M64U (a) ^ M64U (b);
+
+ return M64C (i64);
+}
+
+__inline __m64
+_m_pmulhuw (__m64 a, __m64 b) /* unoptimized */
+{
+ unsigned short d[4] =
+ {
+ (unsigned short)((((unsigned)a.m64_u16[0]) * b.m64_u16[0]) >> 16),
+ (unsigned short)((((unsigned)a.m64_u16[1]) * b.m64_u16[1]) >> 16),
+ (unsigned short)((((unsigned)a.m64_u16[2]) * b.m64_u16[2]) >> 16),
+ (unsigned short)((((unsigned)a.m64_u16[3]) * b.m64_u16[3]) >> 16)
+ };
+
+ return M64C (d[0]);
+}
+
+__inline __m64
+_m_pmullw2 (__m64 a, __m64 b) /* unoptimized */
+{
+ unsigned short d[4] =
+ {
+ (unsigned short)((((unsigned)a.m64_u16[0]) * b.m64_u16[0])),
+ (unsigned short)((((unsigned)a.m64_u16[1]) * b.m64_u16[1])),
+ (unsigned short)((((unsigned)a.m64_u16[2]) * b.m64_u16[2])),
+ (unsigned short)((((unsigned)a.m64_u16[3]) * b.m64_u16[3]))
+ };
+
+ return M64C (d[0]);
+}
+
+__inline __m64
+_m_pmullw (__m64 a, __m64 b) /* unoptimized */
+{
+ unsigned long long x =
+ ((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[0]) * b.m64_u16[0]))) +
+ (((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[1]) * b.m64_u16[1]))) << 16) +
+ (((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[2]) * b.m64_u16[2]))) << 32) +
+ (((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[3]) * b.m64_u16[3]))) << 48);
+
+ return M64C (x);
+}
+
+__inline __m64
+_m_paddusb (__m64 a, __m64 b) /* unoptimized */
+{
+ unsigned long long x = (M64U (a) & 0x00FF00FF00FF00FFULL) +
+ (M64U (b) & 0x00FF00FF00FF00FFULL);
+
+ unsigned long long y = ((M64U (a) >> 8) & 0x00FF00FF00FF00FFULL) +
+ ((M64U (b) >> 8) & 0x00FF00FF00FF00FFULL);
+
+ x | = ((x & 0xFF00FF00FF00FF00ULL) >> 8) * 0xFF;
+ y | = ((y & 0xFF00FF00FF00FF00ULL) >> 8) * 0xFF;
+
+ x = (x & 0x00FF00FF00FF00FFULL) | ((y & 0x00FF00FF00FF00FFULL) << 8);
+
+ return M64C (x);
+}
+
+__inline __m64
+_m_paddusw (__m64 a, __m64 b) /* unoptimized */
+{
+ unsigned long long x = (M64U (a) & 0x0000FFFF0000FFFFULL) +
+ (M64U (b) & 0x0000FFFF0000FFFFULL);
+
+ unsigned long long y = ((M64U (a) >> 16) & 0x0000FFFF0000FFFFULL) +
+ ((M64U (b) >> 16) & 0x0000FFFF0000FFFFULL);
+
+ x | = ((x & 0xFFFF0000FFFF0000) >> 16) * 0xFFFF;
+ y | = ((y & 0xFFFF0000FFFF0000) >> 16) * 0xFFFF;
+
+ x = (x & 0x0000FFFF0000FFFFULL) | ((y & 0x0000FFFF0000FFFFULL) << 16);
+
+ return M64C (x);
+}
+
+__inline __m64
+_m_pshufw (__m64 a, int n) /* unoptimized */
+{
+ unsigned short d[4] =
+ {
+ a.m64_u16[n & 3],
+ a.m64_u16[(n >> 2) & 3],
+ a.m64_u16[(n >> 4) & 3],
+ a.m64_u16[(n >> 6) & 3]
+ };
+
+ return M64C (d[0]);
+}
+
+__inline unsigned char
+sat16 (unsigned short d)
+{
+ if (d > 0xFF) return 0xFF;
+ else return d & 0xFF;
+}
+
+__inline __m64
+_m_packuswb (__m64 m1, __m64 m2) /* unoptimized */
+{
+ unsigned char d[8] =
+ {
+ sat16 (m1.m64_u16[0]),
+ sat16 (m1.m64_u16[1]),
+ sat16 (m1.m64_u16[2]),
+ sat16 (m1.m64_u16[3]),
+ sat16 (m2.m64_u16[0]),
+ sat16 (m2.m64_u16[1]),
+ sat16 (m2.m64_u16[2]),
+ sat16 (m2.m64_u16[3])
+ };
+
+ return M64C (d[0]);
+}
+
+__inline __m64 _m_punpcklbw (__m64 m1, __m64 m2) /* unoptimized */
+{
+ unsigned char d[8] =
+ {
+ m1.m64_u8[0],
+ m2.m64_u8[0],
+ m1.m64_u8[1],
+ m2.m64_u8[1],
+ m1.m64_u8[2],
+ m2.m64_u8[2],
+ m1.m64_u8[3],
+ m2.m64_u8[3],
+ };
+
+ return M64C (d[0]);
+}
+
+__inline __m64 _m_punpckhbw (__m64 m1, __m64 m2) /* unoptimized */
+{
+ unsigned char d[8] =
+ {
+ m1.m64_u8[4],
+ m2.m64_u8[4],
+ m1.m64_u8[5],
+ m2.m64_u8[5],
+ m1.m64_u8[6],
+ m2.m64_u8[6],
+ m1.m64_u8[7],
+ m2.m64_u8[7],
+ };
+
+ return M64C (d[0]);
+}
+
+__inline __m64 _m_psrlwi (__m64 a, int n) /* unoptimized */
+{
+ unsigned short d[4] =
+ {
+ a.m64_u16[0] >> n,
+ a.m64_u16[1] >> n,
+ a.m64_u16[2] >> n,
+ a.m64_u16[3] >> n
+ };
+
+ return M64C (d[0]);
+}
+
+__inline __m64 _m_psrlqi (__m64 m, int n)
+{
+ unsigned long long x = M64U (m) >> n;
+
+ return M64C (x);
+}
+
+__inline __m64 _m_psllqi (__m64 m, int n)
+{
+ unsigned long long x = M64U (m) << n;
+
+ return M64C (x);
+}
+
+#endif /* MMX_X64_H_INCLUDED */