Bug 476042 - Integrate sparc nanojit intro tracemonkey. r=gal, r=jim.

This commit is contained in:
Leon Sha<leon.sha@sun.com> 2009-02-16 10:10:03 +08:00
parent ddb759753a
commit 128a82a274
9 changed files with 2115 additions and 1 deletions

View File

@ -2502,6 +2502,10 @@ arm*-*)
ENABLE_JIT=1
NANOJIT_ARCH=ARM
;;
sparc*-*)
ENABLE_JIT=1
NANOJIT_ARCH=Sparc
;;
esac
MOZ_ARG_DISABLE_BOOL(jit,
@ -2521,6 +2525,9 @@ i?86-*)
arm*-*)
AC_DEFINE(AVMPLUS_ARM)
;;
sparc-*)
AC_DEFINE(AVMPLUS_SPARC)
;;
esac
case "$target_os" in

View File

@ -2122,7 +2122,11 @@ class RegExpNativeCompiler {
LIns* compileFlatDoubleChar(jschar ch1, jschar ch2, LIns* pos,
LInsList& fails)
{
#ifdef IS_BIG_ENDIAN
uint32 word = (ch1 << 16) | ch2;
#else
uint32 word = (ch2 << 16) | ch1;
#endif
/*
* Fast case-insensitive test for ASCII letters: convert text
* char to lower case by bit-or-ing in 32 and compare.

View File

@ -49,6 +49,10 @@
extern "C" void __clear_cache(char *BEG, char *END);
#endif
#ifdef AVMPLUS_SPARC
extern "C" void sync_instruction_memory(caddr_t v, u_int len);
#endif
namespace nanojit
{
@ -932,6 +936,20 @@ namespace nanojit
// to execute junk
# if defined(UNDER_CE)
FlushInstructionCache(GetCurrentProcess(), NULL, NULL);
#elif defined AVMPLUS_SPARC
// Clear Instruction Cache
for (int i = 0; i < 2; i++) {
Page *p = (i == 0) ? _nativePages : _nativeExitPages;
Page *first = p;
while (p) {
if (!p->next || p->next != p+1) {
sync_instruction_memory((char *)first, NJ_PAGE_SIZE);
first = p->next;
}
p = p->next;
}
}
# elif defined(AVMPLUS_UNIX)
for (int i = 0; i < 2; i++) {
Page *p = (i == 0) ? _nativePages : _nativeExitPages;

View File

@ -689,8 +689,13 @@ namespace nanojit
return *(const uint64_t*)ptr;
#else
union { uint64_t tmp; int32_t dst[2]; } u;
#ifdef AVMPLUS_BIG_ENDIAN
u.dst[0] = l->v[1];
u.dst[1] = l->v[0];
#else
u.dst[0] = l->v[0];
u.dst[1] = l->v[1];
#endif
return u.tmp;
#endif
}
@ -704,8 +709,13 @@ namespace nanojit
return *(const double*)ptr;
#else
union { uint32_t dst[2]; double tmpf; } u;
#ifdef AVMPLUS_BIG_ENDIAN
u.dst[0] = l->v[1];
u.dst[1] = l->v[0];
#else
u.dst[0] = l->v[0];
u.dst[1] = l->v[1];
#endif
return u.tmpf;
#endif
}

View File

@ -51,6 +51,8 @@
#endif
#elif defined(NANOJIT_PPC)
#include "NativePpc.h"
#elif defined(NANOJIT_SPARC)
#include "NativeSparc.h"
#elif defined(NANOJIT_AMD64)
#include "NativeAMD64.h"
#else
@ -115,7 +117,7 @@ namespace nanojit {
#define gpn(r) regNames[(r)]
#define fpn(r) regNames[(r)]
#else
#define asm_output(f, ...)
#define asm_output(...)
#define gpn(r)
#define fpn(r)
#endif /* NJ_VERBOSE */

File diff suppressed because it is too large Load Diff

View File

@ -0,0 +1,936 @@
/* -*- Mode: C++; c-basic-offset: 4; indent-tabs-mode: t; tab-width: 4 -*- */
/* ***** BEGIN LICENSE BLOCK *****
* Version: MPL 1.1/GPL 2.0/LGPL 2.1
*
* The contents of this file are subject to the Mozilla Public License Version
* 1.1 (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
* http://www.mozilla.org/MPL/
*
* Software distributed under the License is distributed on an "AS IS" basis,
* WITHOUT WARRANTY OF ANY KIND, either express or implied. See the License
* for the specific language governing rights and limitations under the
* License.
*
* The Original Code is [Open Source Virtual Machine].
*
* The Initial Developer of the Original Code is
* Adobe System Incorporated.
* Portions created by the Initial Developer are Copyright (C) 2004-2007
* the Initial Developer. All Rights Reserved.
*
* Contributor(s):
* Adobe AS3 Team
* leon.sha@sun.com
*
* Alternatively, the contents of this file may be used under the terms of
* either the GNU General Public License Version 2 or later (the "GPL"), or
* the GNU Lesser General Public License Version 2.1 or later (the "LGPL"),
* in which case the provisions of the GPL or the LGPL are applicable instead
* of those above. If you wish to allow use of your version of this file only
* under the terms of either the GPL or the LGPL, and not to allow others to
* use your version of this file under the terms of the MPL, indicate your
* decision by deleting the provisions above and replace them with the notice
* and other provisions required by the GPL or the LGPL. If you do not delete
* the provisions above, a recipient may use your version of this file under
* the terms of any one of the MPL, the GPL or the LGPL.
*
* ***** END LICENSE BLOCK ***** */
#ifndef __nanojit_NativeSparc__
#define __nanojit_NativeSparc__
#define count_instr()
#define count_ret()
#define count_push()
#define count_pop()
#define count_st()
#define count_stq()
#define count_ld()
#define count_ldq()
#define count_call()
#define count_calli()
#define count_prolog()
#define count_alu()
#define count_mov()
#define count_fpu()
#define count_jmp()
#define count_jcc()
#define count_fpuld()
#define count_aluld()
#define count_alust()
#define count_pushld()
#define count_imt()
namespace nanojit
{
#ifdef MMGC_SPARC
const int NJ_LOG2_PAGE_SIZE = 12; // 4K
#else
const int NJ_LOG2_PAGE_SIZE = 13; // 8K
#endif
const int NJ_MAX_REGISTERS = 30; // L0 - L7, I0 - I5, F2 - F14
const int NJ_STACK_OFFSET = 0;
const int LARGEST_UNDERRUN_PROT = 32; // largest value passed to underrunProtect
#define NJ_MAX_STACK_ENTRY 256
#define NJ_MAX_PARAMETERS 1
const int NJ_ALIGN_STACK = 16;
typedef uint32_t NIns;
// These are used as register numbers in various parts of the code
typedef enum
{
G0 = 0,
G1 = 1,
G2 = 2,
G3 = 3,
G4 = 4,
G5 = 5, // Reserved for system
G6 = 6, // Reserved for system
G7 = 7, // Reserved for system
O0 = 8,
O1 = 9,
O2 = 10,
O3 = 11,
O4 = 12,
O5 = 13,
O6 = 14, // SP
O7 = 15, // Not used.
L0 = 16,
L1 = 17,
L2 = 18,
L3 = 19,
L4 = 20,
L5 = 21,
L6 = 22,
L7 = 23,
I0 = 24,
I1 = 25,
I2 = 26,
I3 = 27,
I4 = 28,
I5 = 29,
I6 = 30, // FP
I7 = 31, // Not used
SP = O6,
FP = I6,
F0 = 0,
F1 = 1,
F2 = 2,
F3 = 3,
F4 = 4,
F5 = 5,
F6 = 6,
F7 = 7,
F8 = 8,
F9 = 9,
F10 = 10,
F11 = 11,
F12 = 12,
F13 = 13,
F14 = 14,
F15 = 15,
F16 = 16,
F17 = 17,
F18 = 18,
F19 = 19,
F20 = 20,
F21 = 21,
F22 = 22,
F23 = 23,
F24 = 24,
F25 = 25,
F26 = 26,
F27 = 27,
F28 = 28,
F29 = 29,
F30 = 30,
F31 = 31,
// helpers
FRAME_PTR = G4,
FirstReg = 0,
LastReg = 29,
UnknownReg = 30
}
Register;
// We only use 32 registers to be managed.
// So we choose some of them.
// And other unmanaged registers we can use them directly.
// The registers that can be used directly are G0-G4, L0, L2, L4, L6
// SP, FP, F8-F12, F24-F28
typedef int RegisterMask;
#define _rmask_(r) (1<<(r))
static const int NumSavedRegs = 6;
static const RegisterMask SavedRegs = 1<<L1 | 1<<L3 | 1<<L5 | 1<<L7 |
1<<I0 | 1<<I1 | 1<<I2 | 1<<I3 |
1<<I4 | 1<<I5;
static const RegisterMask GpRegs = SavedRegs | 1<<O0 | 1<<O1 | 1<<O2 |
1<<O3 | 1<<O4 | 1<<O5;
static const RegisterMask FpRegs = 1<<F0 | 1<<F2 | 1<<F4 | 1<<F6 |
1<<F14 | 1<<F16 | 1<<F18 | 1<<F20 |
1<<F22;
static const RegisterMask AllowableFlagRegs = GpRegs;
#define nextreg(r) Register(r+1)
#define prevreg(r) Register(r-1)
verbose_only( extern const char* regNames[]; )
#define DECLARE_PLATFORM_STATS()
#define DECLARE_PLATFORM_REGALLOC()
#define DECLARE_PLATFORM_ASSEMBLER() \
const static Register argRegs[6], retRegs[1]; \
bool has_cmov; \
void nativePageReset(); \
void nativePageSetup(); \
void underrunProtect(int bytes); \
void asm_align_code();
#define swapptrs() { NIns* _tins = _nIns; _nIns=_nExitIns; _nExitIns=_tins; }
#define IMM32(i) \
--_nIns; \
*((int32_t*)_nIns) = (int32_t)(i)
#define CALL(c) do { \
int offset = (c->_address) - ((int)_nIns) + 4; \
int i = 0x40000000 | ((offset >> 2) & 0x3FFFFFFF); \
IMM32(i); \
verbose_only(asm_output("call %s",(c->_name));) \
} while (0)
#define Format_2_1(rd, op2, imm22) do { \
int i = rd << 25 | op2 << 22 | (imm22 & 0x3FFFFF); \
IMM32(i); \
} while (0)
#define Format_2_2(a, cond, op2, disp22) \
Format_2_1((a & 0x1) << 4 | (cond & 0xF), op2, disp22)
#define Format_2_3(a, cond, op2, cc1, cc0, p, disp19) \
Format_2_2(a, cond, op2, (cc1 & 0x1) << 21 | (cc0 & 0x1) << 20 | (p & 0x1) << 19 | (disp19 & 0x7FFFF))
#define Format_2_4(a, rcond, op2, d16hi, p, rs1, d16lo) \
Format_2_2(a, (rcond & 0x7), op2, (d16hi & 0x3) << 20 | (p & 0x1) << 19 | rs1 << 14 | (d16lo & 0x3FFF))
#define Format_3(op1, rd, op3, bits19) do { \
int i = op1 << 30 | rd << 25 | op3 << 19 | (bits19 & 0x7FFFF); \
IMM32(i); \
} while (0)
#define Format_3_1(op1, rd, op3, rs1, bit8, rs2) \
Format_3(op1, rd, op3, rs1 << 14 | (bit8 & 0xFF) << 5 | rs2)
#define Format_3_1I(op1, rd, op3, rs1, simm13) \
Format_3(op1, rd, op3, rs1 << 14 | 1 << 13 | (simm13 & 0x1FFF))
#define Format_3_2(op1, rd, op3, rs1, rcond, rs2) \
Format_3(op1, rd, op3, rs1 << 14 | (rcond & 0x3) << 10 | rs2)
#define Format_3_2I(op1, rd, op3, rs1, rcond, simm10) \
Format_3(op1, rd, op3, rs1 << 14 | 1 << 13 | (rcond & 0x3) << 10 | (simm10 & 0x1FFF))
#define Format_3_3(op1, rd, op3, rs1, cmask, mmask) \
Format_3(op1, rd, op3, rs1 << 14 | 1 << 13 | (cmask & 0x7) << 5 | (mmask & 0xF))
#define Format_3_4(op1, rd, op3, bits19) \
Format_3(op1, rd, op3, bits19)
#define Format_3_5(op1, rd, op3, rs1, x, rs2) \
Format_3(op1, rd, op3, rs1 << 14 | (x & 0x1) << 12 | rs2)
#define Format_3_6(op1, rd, op3, rs1, shcnt32) \
Format_3(op1, rd, op3, rs1 << 14 | 1 << 13 | (shcnt32 & 0x3F))
#define Format_3_7(op1, rd, op3, rs1, shcnt64) \
Format_3(op1, rd, op3, rs1 << 14 | 1 << 13 | 1 << 12 | (shcnt64 & 0x7F))
#define Format_3_8(op1, rd, op3, rs1, bits9, rs2) \
Format_3(op1, rd, op3, rs1 << 14 | (bits9 & 0x1FF) << 5 | rs2)
#define Format_3_9(op1, cc1, cc0, op3, rs1, bits9, rs2) \
Format_3(op1, (cc1 & 0x1) << 1 | (cc0 & 0x1), op3, rs1 << 14 | (bits9 & 0x1FF) << 5 | rs2)
#define Format_4_1(rd, op3, rs1, cc1, cc0, rs2) \
Format_3(2, rd, op3, rs1 << 14 | (cc1 & 0x1) << 12 | (cc0 & 0x1) << 11 | rs2)
#define Format_4_1I(rd, op3, rs1, cc1, cc0, simm11) \
Format_3(2, rd, op3, rs1 << 14 | (cc1 & 0x1) << 12 | 1 << 13 |(cc0 & 0x1) << 11 | (simm11 & 0x7FF))
#define Format_4_2(rd, op3, cc2, cond, cc1, cc0, rs2) \
Format_3(2, rd, op3, (cc2 & 0x1) << 18 | (cond & 0xF) << 14 | (cc1 & 0x1) << 12 | (cc0 & 0x1) << 11 | rs2)
#define Format_4_2I(rd, op3, cc2, cond, cc1, cc0, simm11) \
Format_3(2, rd, op3, (cc2 & 0x1) << 18 | (cond & 0xF) << 14 | 1 << 13 | (cc1 & 0x1) << 12 | (cc0 & 0x1) << 11 | (simm11 & 0x7FF))
#define Format_4_3(rd, op3, rs1, cc1, cc0, swap_trap) \
Format_3(2, rd, op3, rs1 << 14 | 1 << 13 | (cc1 & 0x1) << 12 | (cc0 & 0x1) << 11 | (swap_trap & 0x7F))
#define Format_4_4(rd, op3, rs1, rcond, opf_low, rs2) \
Format_3(2, rd, op3, rs1 << 14 | (rcond & 0x7) << 10 | (opf_low & 0x1F) << 5 | rs2)
#define Format_4_5(rd, op3, cond, opf_cc, opf_low, rs2) \
Format_3(2, rd, op3, (cond & 0xF) << 14 | (opf_cc & 0x7) << 11 | (opf_low & 0x3F) << 5 | rs2)
#define ADD(rs1, rs2, rd) \
do { \
asm_output("add %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0, rs1, 0, rs2); \
} while (0)
#define AND(rs1, rs2, rd) \
do { \
asm_output("and %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x1, rs1, 0, rs2); \
} while (0)
#define BA(a, dsp22) \
do { \
asm_output("ba %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x8, 0x2, dsp22); \
} while (0)
#define BE(a, dsp22) \
do { \
asm_output("be %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x1, 0x2, dsp22); \
} while (0)
#define BG(a, dsp22) \
do { \
asm_output("bg %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xA, 0x2, dsp22); \
} while (0)
#define BGU(a, dsp22) \
do { \
asm_output("bgu %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xC, 0x2, dsp22); \
} while (0)
#define BGE(a, dsp22) \
do { \
asm_output("bge %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xB, 0x2, dsp22); \
} while (0)
#define BL(a, dsp22) \
do { \
asm_output("bl %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x3, 0x2, dsp22); \
} while (0)
#define BLE(a, dsp22) \
do { \
asm_output("ble %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x2, 0x2, dsp22); \
} while (0)
#define BLEU(a, dsp22) \
do { \
asm_output("bleu %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x4, 0x2, dsp22); \
} while (0)
#define BCC(a, dsp22) \
do { \
asm_output("bcc %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xd, 0x2, dsp22); \
} while (0)
#define BCS(a, dsp22) \
do { \
asm_output("bcs %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x5, 0x2, dsp22); \
} while (0)
#define BVC(a, dsp22) \
do { \
asm_output("bvc %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xf, 0x2, dsp22); \
} while (0)
#define BVS(a, dsp22) \
do { \
asm_output("bvc %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x7, 0x2, dsp22); \
} while (0)
#define BNE(a, dsp22) \
do { \
asm_output("bne %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x9, 0x2, dsp22); \
} while (0)
#define FABSS(rs2, rd) \
do { \
asm_output("fabs %s, %s", gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, 0, 0x9, rs2); \
} while (0)
#define FADDD(rs1, rs2, rd) \
do { \
asm_output("faddd %s, %s, %s", gpn(rs1+32), gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, rs1, 0x42, rs2); \
} while (0)
#define FBE(a, dsp22) \
do { \
asm_output("fbe %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x9, 0x6, dsp22); \
} while(0)
#define FBNE(a, dsp22) \
do { \
asm_output("fbne %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x1, 0x6, dsp22); \
} while(0)
#define FBUE(a, dsp22) \
do { \
asm_output("fbue %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xA, 0x6, dsp22); \
} while(0)
#define FBG(a, dsp22) \
do { \
asm_output("fng %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x6, 0x6, dsp22); \
} while(0)
#define FBUG(a, dsp22) \
do { \
asm_output("fbug %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x5, 0x6, dsp22); \
} while(0)
#define FBGE(a, dsp22) \
do { \
asm_output("fbge %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xB, 0x6, dsp22); \
} while(0)
#define FBUGE(a, dsp22) \
do { \
asm_output("fbuge %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xC, 0x6, dsp22); \
} while(0)
#define FBL(a, dsp22) \
do { \
asm_output("fbl %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0x4, 0x6, dsp22); \
} while(0)
#define FBLE(a, dsp22) \
do { \
asm_output("fble %p", _nIns + dsp22 - 1); \
Format_2_2(a, 0xD, 0x6, dsp22); \
} while(0)
#define FCMPD(rs1, rs2) \
do { \
asm_output("fcmpd %s, %s", gpn(rs1+32), gpn(rs2+32)); \
Format_3_9(2, 0, 0, 0x35, rs1, 0x52, rs2); \
} while (0)
#define FSUBD(rs1, rs2, rd) \
do { \
asm_output("fsubd %s, %s, %s", gpn(rs1+32), gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, rs1, 0x46, rs2); \
} while (0)
#define FMULD(rs1, rs2, rd) \
do { \
asm_output("fmuld %s, %s, %s", gpn(rs1+32), gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, rs1, 0x4a, rs2); \
} while (0)
#define FDIVD(rs1, rs2, rd) \
do { \
asm_output("fdivd %s, %s, %s", gpn(rs1+32), gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, rs1, 0x4e, rs2); \
} while (0)
#define FMOVD(rs2, rd) \
do { \
asm_output("fmovd %s, %s", gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, 0, 0x2, rs2); \
} while (0)
#define FNEGD(rs2, rd) \
do { \
asm_output("fnegd %s, %s", gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, 0, 0x6, rs2); \
} while (0)
#define FITOD(rs2, rd) \
do { \
asm_output("fitod %s, %s", gpn(rs2+32), gpn(rd+32)); \
Format_3_8(2, rd, 0x34, 0, 0xc8, rs2); \
} while (0)
#define JMPL(rs1, rs2, rd) \
do { \
asm_output("jmpl [%s + %s]", gpn(rs1), gpn(rs2)); \
Format_3_1(2, rd, 0x38, rs1, 0, rs2); \
} while (0)
#define JMPLI(rs1, simm13, rd) \
do { \
asm_output("jmpl [%s + %d]", gpn(rs1), simm13); \
Format_3_1I(2, rd, 0x38, rs1, simm13); \
} while (0)
#define LDF(rs1, rs2, rd) \
do { \
asm_output("ld [%s + %s], %s", gpn(rs1), gpn(rs2), gpn(rd+32)); \
Format_3_1(3, rd, 0x20, rs1, 0, rs2); \
} while (0)
#define LDFI(rs1, simm13, rd) \
do { \
asm_output("ld [%s + %d], %s", gpn(rs1), simm13, gpn(rd+32)); \
Format_3_1I(3, rd, 0x20, rs1, simm13); \
} while (0)
#define LDSB(rs1, rs2, rd) \
do { \
asm_output("ld [%s + %s], %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(3, rd, 0x9, rs1, 0, rs2); \
} while (0)
#define LDSBI(rs1, simm13, rd) \
do { \
asm_output("ld [%s + %d], %s", gpn(rs1), simm13, gpn(rd)); \
Format_3_1I(3, rd, 0x9, rs1, simm13); \
} while (0)
#define LDSH(rs1, rs2, rd) \
do { \
asm_output("ld [%s + %s], %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(3, rd, 0xa, rs1, 0, rs2); \
} while (0)
#define LDSHI(rs1, simm13, rd) \
do { \
asm_output("ld [%s + %d], %s", gpn(rs1), simm13, gpn(rd)); \
Format_3_1I(3, rd, 0xa, rs1, simm13); \
} while (0)
#define LDSW(rs1, rs2, rd) \
do { \
asm_output("ld [%s + %s], %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(3, rd, 0x8, rs1, 0, rs2); \
} while (0)
#define LDSWI(rs1, simm13, rd) \
do { \
asm_output("ld [%s + %d], %s", gpn(rs1), simm13, gpn(rd)); \
Format_3_1I(3, rd, 0x8, rs1, simm13); \
} while (0)
#define MOVE(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("move %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 1, cc1, cc0, rs); \
} while (0)
#define MOVNE(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movne %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 9, cc1, cc0, rs); \
} while (0)
#define MOVL(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movl %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 3, cc1, cc0, rs); \
} while (0)
#define MOVLE(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movle %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 2, cc1, cc0, rs); \
} while (0)
#define MOVG(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movg %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 0xa, cc1, cc0, rs); \
} while (0)
#define MOVGE(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movge %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 0xb, cc1, cc0, rs); \
} while (0)
#define MOVCS(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movcs %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 5, cc1, cc0, rs); \
} while (0)
#define MOVLEU(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movleu %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 4, cc1, cc0, rs); \
} while (0)
#define MOVGU(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movgu %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 0xc, cc1, cc0, rs); \
} while (0)
#define MOVCC(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movcc %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 0xd, cc1, cc0, rs); \
} while (0)
#define MOVVC(rs, cc2, cc1, cc0, rd) \
do { \
asm_output("movvc %s, %s", gpn(rs), gpn(rd)); \
Format_4_2(rd, 0x2c, cc2, 0xf, cc1, cc0, rs); \
} while (0)
#define MOVEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("move %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 1, cc1, cc0, simm11); \
} while (0)
#define MOVFEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("move %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 9, cc1, cc0, simm11); \
} while (0)
#define MOVNEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("move %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 9, cc1, cc0, simm11); \
} while (0)
#define MOVLI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("move %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 3, cc1, cc0, simm11); \
} while (0)
#define MOVFLI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("move %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 4, cc1, cc0, simm11); \
} while (0)
#define MOVLEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movle %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 2, cc1, cc0, simm11); \
} while (0)
#define MOVFLEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movle %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 0xd, cc1, cc0, simm11); \
} while (0)
#define MOVGI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movg %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 0xa, cc1, cc0, simm11); \
} while (0)
#define MOVFGI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movg %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 6, cc1, cc0, simm11); \
} while (0)
#define MOVGEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movge %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 0xb, cc1, cc0, simm11); \
} while (0)
#define MOVFGEI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movge %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 0xb, cc1, cc0, simm11); \
} while (0)
#define MOVCSI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movcs %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 5, cc1, cc0, simm11); \
} while (0)
#define MOVLEUI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movleu %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 4, cc1, cc0, simm11); \
} while (0)
#define MOVGUI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movgu %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 0xc, cc1, cc0, simm11); \
} while (0)
#define MOVCCI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movcc %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 0xd, cc1, cc0, simm11); \
} while (0)
#define MOVVSI(simm11, cc2, cc1, cc0, rd) \
do { \
asm_output("movvs %d, %s", simm11, gpn(rd)); \
Format_4_2I(rd, 0x2c, cc2, 7, cc1, cc0, simm11); \
} while (0)
#define MULX(rs1, rs2, rd) \
do { \
asm_output("mul %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x9, rs1, 0, rs2); \
} while (0)
#define NOP() \
do { \
asm_output("nop"); \
Format_2_1(0, 0x4, 0); \
} while (0)
#define ORI(rs1, simm13, rd) \
do { \
asm_output("or %s, %d, %s", gpn(rs1), simm13, gpn(rd)); \
Format_3_1I(2, rd, 0x2, rs1, simm13); \
} while (0)
#define OR(rs1, rs2, rd) \
do { \
asm_output("or %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x2, rs1, 0, rs2); \
} while (0)
#define ORN(rs1, rs2, rd) \
do { \
asm_output("orn %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x6, rs1, 0, rs2); \
} while (0)
#define ANDCC(rs1, rs2, rd) \
do { \
asm_output("andcc %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x11, rs1, 0, rs2); \
} while (0)
#define RESTORE(rs1, rs2, rd) \
do { \
asm_output("restore"); \
Format_3_1(2, rd, 0x3D, rs1, 0, rs2); \
} while (0)
#define SAVEI(rs1, simm13, rd) \
do { \
asm_output("save %s, %d, %s", gpn(rs1), simm13, gpn(rd)); \
Format_3_1I(2, rd, 0x3C, rs1, simm13); \
} while (0)
#define SAVE(rs1, rs2, rd) \
do { \
asm_output("save %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x3C, rs1, 0, rs2); \
} while (0)
#define SETHI(imm22, rd) \
do { \
asm_output("sethi %p, %s", imm22, gpn(rd)); \
Format_2_1(rd, 0x4, (imm22 >> 10)); \
} while (0)
#define SLL(rs1, rs2, rd) \
do { \
asm_output("sll %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_5(2, rd, 0x25, rs1, 0, rs2); \
} while (0)
#define SRA(rs1, rs2, rd) \
do { \
asm_output("sra %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_5(2, rd, 0x27, rs1, 0, rs2); \
} while (0)
#define SRL(rs1, rs2, rd) \
do { \
asm_output("srl %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_5(2, rd, 0x26, rs1, 0, rs2); \
} while (0)
#define STF(rd, rs1, rs2) \
do { \
asm_output("st %s, [%s + %s]", gpn(rd+32), gpn(rs1), gpn(rs2)); \
Format_3_1(3, rd, 0x24, rs1, 0, rs2); \
} while (0)
#define STFI(rd, simm13, rs1) \
do { \
asm_output("st %s, [%s + %d]", gpn(rd+32), gpn(rs1), simm13); \
Format_3_1I(3, rd, 0x24, rs1, simm13); \
} while (0)
#define STW(rd, rs2, rs1) \
do { \
asm_output("st %s, [%s + %s]", gpn(rd), gpn(rs1), gpn(rs2)); \
Format_3_1(3, rd, 0x4, rs1, 0, rs2); \
} while (0)
#define STWI(rd, simm13, rs1) \
do { \
asm_output("st %s, [%s + %d]", gpn(rd), gpn(rs1), simm13); \
Format_3_1I(3, rd, 0x4, rs1, simm13); \
} while (0)
#define SUBCC(rs1, rs2, rd) \
do { \
asm_output("subcc %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x14, rs1, 0, rs2); \
} while (0)
#define SUB(rs1, rs2, rd) \
do { \
asm_output("sub %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x4, rs1, 0, rs2); \
} while (0)
#define XOR(rs1, rs2, rd) \
do { \
asm_output("xor %s, %s, %s", gpn(rs1), gpn(rs2), gpn(rd)); \
Format_3_1(2, rd, 0x3, rs1, 0, rs2); \
} while (0)
// Returns true if imm below 13-bit unsigned immediate)
#define isIMM13(imm) \
(imm) <= 4095 && (imm) >= -4096
#define isIMM19(imm) \
(imm) <= 0x3ffff && (imm) >= -0x40000
#define isIMM22(imm) \
(imm) <= 0x1fffff && (imm) >= -0x200000
#define SET32(imm32, rd) \
if(isIMM13(imm32)) { \
ORI(G0, imm32, rd); \
} else { \
ORI(rd, imm32 & 0x3FF, rd); \
SETHI(imm32, rd); \
}
#define STDF32(rd, imm32, rs1) \
if(isIMM13(imm32+4)) { \
STFI(rd+1, imm32+4, rs1); \
STFI(rd, imm32, rs1); \
} else { \
STF(rd+1, L0, rs1); \
SET32(imm32+4, L0); \
STF(rd, L0, rs1); \
SET32(imm32, L0); \
}
#define LDDF32(rs1, imm32, rd) \
if(isIMM13(imm32+4)) { \
LDFI(rs1, imm32+4, rd+1); \
LDFI(rs1, imm32, rd); \
} else { \
LDF(rs1, L0, rd+1); \
SET32(imm32+4, L0); \
LDF(rs1, L0, rd); \
SET32(imm32, L0); \
}
#define STW32(rd, imm32, rs1) \
if(isIMM13(imm32)) { \
STWI(rd, imm32, rs1); \
} else { \
STW(rd, L0, rs1); \
SET32(imm32, L0); \
}
#define LDSB32(rs1, imm32, rd) \
if(isIMM13(imm32)) { \
LDSBI(rs1, imm32, rd); \
} else { \
LDSB(rs1, L0, rd); \
SET32(imm32, L0); \
}
#define LDSH32(rs1, imm32, rd) \
if(isIMM13(imm32)) { \
LDSHI(rs1, imm32, rd); \
} else { \
LDSH(rs1, L0, rd); \
SET32(imm32, L0); \
}
#define LDSW32(rs1, imm32, rd) \
if(isIMM13(imm32)) { \
LDSWI(rs1, imm32, rd); \
} else { \
LDSW(rs1, L0, rd); \
SET32(imm32, L0); \
}
// general Assemble
#define JMP_long_nocheck(t) do { NOP(); JMPL(G0, G2, G0); ORI(G2, t & 0x3FF, G2); SETHI(t, G2); } while(0)
#define JMP_long(t) do { underrunProtect(16); JMP_long_nocheck(t); } while(0)
#define JMP_long_placeholder() JMP_long(0)
#define JCC(t, tt) \
underrunProtect(32); \
tt = ((intptr_t)t - (intptr_t)_nIns + 8) >> 2; \
if( !(isIMM22((int32_t)tt)) ) { \
NOP(); \
JMPL(G0, G2, G0); \
SET32((intptr_t)t, G2); \
NOP(); \
BA(0, 5); \
tt = 4; \
} \
NOP()
#define JMP(t) do { \
if (!t) { \
JMP_long_placeholder(); \
} else { \
intptr_t tt; \
JCC(t, tt); \
BA(0, tt); \
} \
} while(0)
#define MR(d, s) do { underrunProtect(4); ORI(s, 0, d); } while(0)
}
#endif // __nanojit_NativeSparc__

View File

@ -46,6 +46,14 @@
#include "jstypes.h"
#if !defined(AVMPLUS_LITTLE_ENDIAN) && !defined(AVMPLUS_BIG_ENDIAN)
#ifdef IS_BIG_ENDIAN
#define AVMPLUS_BIG_ENDIAN
#else
#define AVMPLUS_LITTLE_ENDIAN
#endif
#endif
#define FASTCALL JS_FASTCALL
#if defined(JS_NO_FASTCALL)

View File

@ -50,6 +50,8 @@
#define NANOJIT_ARM
#elif AVMPLUS_PPC
#define NANOJIT_PPC
#elif AVMPLUS_SPARC
#define NANOJIT_SPARC
#elif AVMPLUS_AMD64
#define NANOJIT_AMD64
#define NANOJIT_64BIT