pixman-x64-mmx-emulation.h   [plain text]


#ifndef MMX_X64_H_INCLUDED
#define MMX_X64_H_INCLUDED

/* Implementation of x64 MMX substitition functions, before
 * pixman is reimplemented not to use __m64 type on Visual C++
 *
 * Copyright (C)2009 by George Yohng
 * Released in public domain.
 */

#include <intrin.h>

#define M64C(a) (*(const __m64 *)(&a))
#define M64U(a) (*(const unsigned long long *)(&a))

__inline __m64
_m_from_int (int a)
{
    long long i64 = a;

    return M64C (i64);
}

__inline __m64
_mm_setzero_si64 ()
{
    long long i64 = 0;

    return M64C (i64);
}

__inline __m64
_mm_set_pi32 (int i1,   int i0)
{
    unsigned long long i64 = ((unsigned)i0) + (((unsigned long long)(unsigned)i1) << 32);

    return M64C (i64);
}

__inline void
_m_empty ()
{
}

__inline __m64
_mm_set1_pi16 (short w)
{
    unsigned long long i64 = ((unsigned long long)(unsigned short)(w)) * 0x0001000100010001ULL;

    return M64C (i64);
}

__inline int
_m_to_int (__m64 m)
{
    return m.m64_i32[0];
}

__inline __m64
_mm_movepi64_pi64 (__m128i a)
{
    return M64C (a.m128i_i64[0]);
}

__inline __m64
_m_pand (__m64 a, __m64 b)
{
    unsigned long long i64 = M64U (a) & M64U (b);

    return M64C (i64);
}

__inline __m64
_m_por (__m64 a, __m64 b)
{
    unsigned long long i64 = M64U (a) | M64U (b);

    return M64C (i64);
}

__inline __m64
_m_pxor (__m64 a, __m64 b)
{
    unsigned long long i64 = M64U (a) ^ M64U (b);

    return M64C (i64);
}

__inline __m64
_m_pmulhuw (__m64 a, __m64 b)        /* unoptimized */
{
    unsigned short d[4] =
    {
	(unsigned short)((((unsigned)a.m64_u16[0]) * b.m64_u16[0]) >> 16),
	(unsigned short)((((unsigned)a.m64_u16[1]) * b.m64_u16[1]) >> 16),
	(unsigned short)((((unsigned)a.m64_u16[2]) * b.m64_u16[2]) >> 16),
	(unsigned short)((((unsigned)a.m64_u16[3]) * b.m64_u16[3]) >> 16)
    };

    return M64C (d[0]);
}

__inline __m64
_m_pmullw2 (__m64 a, __m64 b)        /* unoptimized */
{
    unsigned short d[4] =
    {
	(unsigned short)((((unsigned)a.m64_u16[0]) * b.m64_u16[0])),
	(unsigned short)((((unsigned)a.m64_u16[1]) * b.m64_u16[1])),
	(unsigned short)((((unsigned)a.m64_u16[2]) * b.m64_u16[2])),
	(unsigned short)((((unsigned)a.m64_u16[3]) * b.m64_u16[3]))
    };

    return M64C (d[0]);
}

__inline __m64
_m_pmullw (__m64 a, __m64 b)        /* unoptimized */
{
    unsigned long long x =
	((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[0]) * b.m64_u16[0])))  +
	(((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[1]) * b.m64_u16[1]))) << 16)  +
	(((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[2]) * b.m64_u16[2]))) << 32)  +
	(((unsigned long long)(unsigned short)((((unsigned)a.m64_u16[3]) * b.m64_u16[3]))) << 48);

    return M64C (x);
}

__inline __m64
_m_paddusb (__m64 a, __m64 b)        /* unoptimized */
{
    unsigned long long x = (M64U (a) & 0x00FF00FF00FF00FFULL) +
                           (M64U (b) & 0x00FF00FF00FF00FFULL);

    unsigned long long y = ((M64U (a) >> 8) & 0x00FF00FF00FF00FFULL) +
                           ((M64U (b) >> 8) & 0x00FF00FF00FF00FFULL);

    x |= ((x & 0xFF00FF00FF00FF00ULL) >> 8) * 0xFF;
    y |= ((y & 0xFF00FF00FF00FF00ULL) >> 8) * 0xFF;

    x = (x & 0x00FF00FF00FF00FFULL) | ((y & 0x00FF00FF00FF00FFULL) << 8);

    return M64C (x);
}

__inline __m64
_m_paddusw (__m64 a, __m64 b)        /* unoptimized */
{
    unsigned long long x = (M64U (a) & 0x0000FFFF0000FFFFULL) +
                           (M64U (b) & 0x0000FFFF0000FFFFULL);

    unsigned long long y = ((M64U (a) >> 16) & 0x0000FFFF0000FFFFULL) +
                           ((M64U (b) >> 16) & 0x0000FFFF0000FFFFULL);

    x |= ((x & 0xFFFF0000FFFF0000) >> 16) * 0xFFFF;
    y |= ((y & 0xFFFF0000FFFF0000) >> 16) * 0xFFFF;

    x = (x & 0x0000FFFF0000FFFFULL) | ((y & 0x0000FFFF0000FFFFULL) << 16);

    return M64C (x);
}

__inline __m64
_m_pshufw (__m64 a, int n)         /* unoptimized */
{
    unsigned short d[4] =
    {
	a.m64_u16[n & 3],
	a.m64_u16[(n >> 2) & 3],
	a.m64_u16[(n >> 4) & 3],
	a.m64_u16[(n >> 6) & 3]
    };

    return M64C (d[0]);
}

__inline unsigned char
sat16 (unsigned short d)
{
    if (d > 0xFF) return 0xFF;
    else return d & 0xFF;
}

__inline __m64
_m_packuswb (__m64 m1, __m64 m2)          /* unoptimized */
{
    unsigned char d[8] =
    {
	sat16 (m1.m64_u16[0]),
	sat16 (m1.m64_u16[1]),
	sat16 (m1.m64_u16[2]),
	sat16 (m1.m64_u16[3]),
	sat16 (m2.m64_u16[0]),
	sat16 (m2.m64_u16[1]),
	sat16 (m2.m64_u16[2]),
	sat16 (m2.m64_u16[3])
    };

    return M64C (d[0]);
}

__inline __m64 _m_punpcklbw (__m64 m1, __m64 m2)          /* unoptimized */
{
    unsigned char d[8] =
    {
	m1.m64_u8[0],
	m2.m64_u8[0],
	m1.m64_u8[1],
	m2.m64_u8[1],
	m1.m64_u8[2],
	m2.m64_u8[2],
	m1.m64_u8[3],
	m2.m64_u8[3],
    };

    return M64C (d[0]);
}

__inline __m64 _m_punpckhbw (__m64 m1, __m64 m2)          /* unoptimized */
{
    unsigned char d[8] =
    {
	m1.m64_u8[4],
	m2.m64_u8[4],
	m1.m64_u8[5],
	m2.m64_u8[5],
	m1.m64_u8[6],
	m2.m64_u8[6],
	m1.m64_u8[7],
	m2.m64_u8[7],
    };

    return M64C (d[0]);
}

__inline __m64 _m_psrlwi (__m64 a, int n)       /* unoptimized */
{
    unsigned short d[4] =
    {
	a.m64_u16[0] >> n,
	a.m64_u16[1] >> n,
	a.m64_u16[2] >> n,
	a.m64_u16[3] >> n
    };

    return M64C (d[0]);
}

__inline __m64 _m_psrlqi (__m64 m, int n)
{
    unsigned long long x = M64U (m) >> n;

    return M64C (x);
}

__inline __m64 _m_psllqi (__m64 m, int n)
{
    unsigned long long x = M64U (m) << n;

    return M64C (x);
}

#endif /* MMX_X64_H_INCLUDED */