diff options
author | jle <jle@138bc75d-0d04-0410-961f-82ee72b054a4> | 2003-08-23 02:02:18 +0000 |
---|---|---|
committer | jle <jle@138bc75d-0d04-0410-961f-82ee72b054a4> | 2003-08-23 02:02:18 +0000 |
commit | 8dd25e426afa10b3e79b3bc1e9c2fd8bb2b73e47 (patch) | |
tree | d3af9a9bf50559d84be1c7c057f153b1e5c86d66 /gcc/config/i860/i860.c | |
parent | 54e4edf280efceddf550ff961754db2ce6239b0c (diff) | |
download | gcc-8dd25e426afa10b3e79b3bc1e9c2fd8bb2b73e47.tar.gz |
2003-08-22 Jason Eckhardt <jle@rice.edu>
ChangeLog
* MAINTAINERS: Resurrect the i860 maintainer.
gcc/ChangeLog:
* gcc/config.gcc (i860-*-sysv4*): Add target.
* config/i860/i860-protos.h: New.
* config/i860/i860.c: New.
* config/i860/i860.h: New.
* config/i860/i860.md: New.
* config/i860/sysv4.h: New.
* config/i860/varargs.asm: New.
* config/i860/x-sysv4: New.
git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@70719 138bc75d-0d04-0410-961f-82ee72b054a4
Diffstat (limited to 'gcc/config/i860/i860.c')
-rw-r--r-- | gcc/config/i860/i860.c | 2361 |
1 files changed, 2361 insertions, 0 deletions
diff --git a/gcc/config/i860/i860.c b/gcc/config/i860/i860.c new file mode 100644 index 00000000000..18493976615 --- /dev/null +++ b/gcc/config/i860/i860.c @@ -0,0 +1,2361 @@ +/* Subroutines for insn-output.c for Intel 860 + Copyright (C) 1989, 1991, 1997, 1998, 1999, 2000, 2001, 2002 + Free Software Foundation, Inc. + Derived from sparc.c. + + Written by Richard Stallman (rms@ai.mit.edu). + + Hacked substantially by Ron Guilmette (rfg@netcom.com) to cater + to the whims of the System V Release 4 assembler. + +This file is part of GNU CC. + +GNU CC is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; either version 2, or (at your option) +any later version. + +GNU CC is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with GNU CC; see the file COPYING. If not, write to +the Free Software Foundation, 59 Temple Place - Suite 330, +Boston, MA 02111-1307, USA. */ + + +#include "config.h" +#include "system.h" +#include "flags.h" +#include "rtl.h" +#include "tree.h" +#include "regs.h" +#include "hard-reg-set.h" +#include "real.h" +#include "insn-config.h" +#include "conditions.h" +#include "output.h" +#include "recog.h" +#include "insn-attr.h" +#include "function.h" +#include "expr.h" +#include "tm_p.h" +#include "target.h" +#include "target-def.h" + +static rtx find_addr_reg PARAMS ((rtx)); +static int reg_clobbered_p PARAMS ((rtx, rtx)); +static const char *singlemove_string PARAMS ((rtx *)); +static const char *load_opcode PARAMS ((enum machine_mode, const char *, rtx)); +static const char *store_opcode PARAMS ((enum machine_mode, const char *, rtx)); +static void output_size_for_block_move PARAMS ((rtx, rtx, rtx)); +static void i860_output_function_prologue PARAMS ((FILE *, HOST_WIDE_INT)); +static void i860_output_function_epilogue PARAMS ((FILE *, HOST_WIDE_INT)); + +#ifndef I860_REG_PREFIX +#define I860_REG_PREFIX "" +#endif + +const char *i860_reg_prefix = I860_REG_PREFIX; + +/* Save information from a "cmpxx" operation until the branch is emitted. */ + +rtx i860_compare_op0, i860_compare_op1; + +/* Initialize the GCC target structure. */ +#undef TARGET_ASM_FUNCTION_PROLOGUE +#define TARGET_ASM_FUNCTION_PROLOGUE i860_output_function_prologue +#undef TARGET_ASM_FUNCTION_EPILOGUE +#define TARGET_ASM_FUNCTION_EPILOGUE i860_output_function_epilogue + +struct gcc_target targetm = TARGET_INITIALIZER; + +/* Return non-zero if this pattern, can be evaluated safely, even if it + was not asked for. */ +int +safe_insn_src_p (op, mode) + rtx op; + enum machine_mode mode; +{ + /* Just experimenting. */ + + /* No floating point src is safe if it contains an arithmetic + operation, since that operation may trap. */ + switch (GET_CODE (op)) + { + case CONST_INT: + case LABEL_REF: + case SYMBOL_REF: + case CONST: + return 1; + + case REG: + return 1; + + case MEM: + return CONSTANT_ADDRESS_P (XEXP (op, 0)); + + /* We never need to negate or complement constants. */ + case NEG: + return (mode != SFmode && mode != DFmode); + case NOT: + case ZERO_EXTEND: + return 1; + + case EQ: + case NE: + case LT: + case GT: + case LE: + case GE: + case LTU: + case GTU: + case LEU: + case GEU: + case MINUS: + case PLUS: + return (mode != SFmode && mode != DFmode); + case AND: + case IOR: + case XOR: + case ASHIFT: + case ASHIFTRT: + case LSHIFTRT: + if ((GET_CODE (XEXP (op, 0)) == CONST_INT && ! SMALL_INT (XEXP (op, 0))) + || (GET_CODE (XEXP (op, 1)) == CONST_INT && ! SMALL_INT (XEXP (op, 1)))) + return 0; + return 1; + + default: + return 0; + } +} + +/* Return 1 if REG is clobbered in IN. + Return 2 if REG is used in IN. + Return 3 if REG is both used and clobbered in IN. + Return 0 if neither. */ + +static int +reg_clobbered_p (reg, in) + rtx reg; + rtx in; +{ + register enum rtx_code code; + + if (in == 0) + return 0; + + code = GET_CODE (in); + + if (code == SET || code == CLOBBER) + { + rtx dest = SET_DEST (in); + int set = 0; + int used = 0; + + while (GET_CODE (dest) == STRICT_LOW_PART + || GET_CODE (dest) == SUBREG + || GET_CODE (dest) == SIGN_EXTRACT + || GET_CODE (dest) == ZERO_EXTRACT) + dest = XEXP (dest, 0); + + if (dest == reg) + set = 1; + else if (GET_CODE (dest) == REG + && refers_to_regno_p (REGNO (reg), + REGNO (reg) + HARD_REGNO_NREGS (reg, GET_MODE (reg)), + SET_DEST (in), 0)) + { + set = 1; + /* Anything that sets just part of the register + is considered using as well as setting it. + But note that a straight SUBREG of a single-word value + clobbers the entire value. */ + if (dest != SET_DEST (in) + && ! (GET_CODE (SET_DEST (in)) == SUBREG + || UNITS_PER_WORD >= GET_MODE_SIZE (GET_MODE (dest)))) + used = 1; + } + + if (code == SET) + { + if (set) + used = refers_to_regno_p (REGNO (reg), + REGNO (reg) + HARD_REGNO_NREGS (reg, GET_MODE (reg)), + SET_SRC (in), 0); + else + used = refers_to_regno_p (REGNO (reg), + REGNO (reg) + HARD_REGNO_NREGS (reg, GET_MODE (reg)), + in, 0); + } + + return set + used * 2; + } + + if (refers_to_regno_p (REGNO (reg), + REGNO (reg) + HARD_REGNO_NREGS (reg, GET_MODE (reg)), + in, 0)) + return 2; + return 0; +} + +/* Return non-zero if OP can be written to without screwing up + GCC's model of what's going on. It is assumed that this operand + appears in the dest position of a SET insn in a conditional + branch's delay slot. AFTER is the label to start looking from. */ +int +operand_clobbered_before_used_after (op, after) + rtx op; + rtx after; +{ + /* Just experimenting. */ + if (GET_CODE (op) == CC0) + return 1; + if (GET_CODE (op) == REG) + { + rtx insn; + + if (op == stack_pointer_rtx) + return 0; + + /* Scan forward from the label, to see if the value of OP + is clobbered before the first use. */ + + for (insn = NEXT_INSN (after); insn; insn = NEXT_INSN (insn)) + { + if (GET_CODE (insn) == NOTE) + continue; + if (GET_CODE (insn) == INSN + || GET_CODE (insn) == JUMP_INSN + || GET_CODE (insn) == CALL_INSN) + { + switch (reg_clobbered_p (op, PATTERN (insn))) + { + default: + return 0; + case 1: + return 1; + case 0: + break; + } + } + /* If we reach another label without clobbering OP, + then we cannot safely write it here. */ + else if (GET_CODE (insn) == CODE_LABEL) + return 0; + if (GET_CODE (insn) == JUMP_INSN) + { + if (condjump_p (insn)) + return 0; + /* This is a jump insn which has already + been mangled. We can't tell what it does. */ + if (GET_CODE (PATTERN (insn)) == PARALLEL) + return 0; + if (! JUMP_LABEL (insn)) + return 0; + /* Keep following jumps. */ + insn = JUMP_LABEL (insn); + } + } + return 1; + } + + /* In both of these cases, the first insn executed + for this op will be a orh whatever%h,%?r0,%?r31, + which is tolerable. */ + if (GET_CODE (op) == MEM) + return (CONSTANT_ADDRESS_P (XEXP (op, 0))); + + return 0; +} + +/* Return non-zero if this pattern, as a source to a "SET", + is known to yield an instruction of unit size. */ +int +single_insn_src_p (op, mode) + rtx op; + enum machine_mode mode; +{ + switch (GET_CODE (op)) + { + case CONST_INT: + /* This is not always a single insn src, technically, + but output_delayed_branch knows how to deal with it. */ + return 1; + + case SYMBOL_REF: + case CONST: + /* This is not a single insn src, technically, + but output_delayed_branch knows how to deal with it. */ + return 1; + + case REG: + return 1; + + case MEM: + return 1; + + /* We never need to negate or complement constants. */ + case NEG: + return (mode != DFmode); + case NOT: + case ZERO_EXTEND: + return 1; + + case PLUS: + case MINUS: + /* Detect cases that require multiple instructions. */ + if (CONSTANT_P (XEXP (op, 1)) + && !(GET_CODE (XEXP (op, 1)) == CONST_INT + && SMALL_INT (XEXP (op, 1)))) + return 0; + case EQ: + case NE: + case LT: + case GT: + case LE: + case GE: + case LTU: + case GTU: + case LEU: + case GEU: + /* Not doing floating point, since they probably + take longer than the branch slot they might fill. */ + return (mode != SFmode && mode != DFmode); + + case AND: + if (GET_CODE (XEXP (op, 1)) == NOT) + { + rtx arg = XEXP (XEXP (op, 1), 0); + if (CONSTANT_P (arg) + && !(GET_CODE (arg) == CONST_INT + && (SMALL_INT (arg) + || (INTVAL (arg) & 0xffff) == 0))) + return 0; + } + case IOR: + case XOR: + /* Both small and round numbers take one instruction; + others take two. */ + if (CONSTANT_P (XEXP (op, 1)) + && !(GET_CODE (XEXP (op, 1)) == CONST_INT + && (SMALL_INT (XEXP (op, 1)) + || (INTVAL (XEXP (op, 1)) & 0xffff) == 0))) + return 0; + + case ASHIFT: + case ASHIFTRT: + case LSHIFTRT: + return 1; + + case SUBREG: + if (SUBREG_BYTE (op) != 0) + return 0; + return single_insn_src_p (SUBREG_REG (op), mode); + + /* Not doing floating point, since they probably + take longer than the branch slot they might fill. */ + case FLOAT_EXTEND: + case FLOAT_TRUNCATE: + case FLOAT: + case FIX: + case UNSIGNED_FLOAT: + case UNSIGNED_FIX: + return 0; + + default: + return 0; + } +} + +/* Return non-zero only if OP is a register of mode MODE, + or const0_rtx. */ +int +reg_or_0_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (op == const0_rtx || register_operand (op, mode) + || op == CONST0_RTX (mode)); +} + +/* Return truth value of whether OP can be used as an operands in a three + address add/subtract insn (such as add %o1,7,%l2) of mode MODE. */ + +int +arith_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (register_operand (op, mode) + || (GET_CODE (op) == CONST_INT && SMALL_INT (op))); +} + +/* Return 1 if OP is a valid first operand for a logical insn of mode MODE. */ + +int +logic_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (register_operand (op, mode) + || (GET_CODE (op) == CONST_INT && LOGIC_INT (op))); +} + +/* Return 1 if OP is a valid first operand for a shift insn of mode MODE. */ + +int +shift_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (register_operand (op, mode) + || (GET_CODE (op) == CONST_INT)); +} + +/* Return 1 if OP is a valid first operand for either a logical insn + or an add insn of mode MODE. */ + +int +compare_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (register_operand (op, mode) + || (GET_CODE (op) == CONST_INT && SMALL_INT (op) && LOGIC_INT (op))); +} + +/* Return truth value of whether OP can be used as the 5-bit immediate + operand of a bte or btne insn. */ + +int +bte_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (register_operand (op, mode) + || (GET_CODE (op) == CONST_INT + && (unsigned) INTVAL (op) < 0x20)); +} + +/* Return 1 if OP is an indexed memory reference of mode MODE. */ + +int +indexed_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (GET_CODE (op) == MEM && GET_MODE (op) == mode + && GET_CODE (XEXP (op, 0)) == PLUS + && GET_MODE (XEXP (op, 0)) == SImode + && register_operand (XEXP (XEXP (op, 0), 0), SImode) + && register_operand (XEXP (XEXP (op, 0), 1), SImode)); +} + +/* Return 1 if OP is a suitable source operand for a load insn + with mode MODE. */ + +int +load_operand (op, mode) + rtx op; + enum machine_mode mode; +{ + return (memory_operand (op, mode) || indexed_operand (op, mode)); +} + +/* Return truth value of whether OP is an integer which fits the + range constraining immediate operands in add/subtract insns. */ + +int +small_int (op, mode) + rtx op; + enum machine_mode mode ATTRIBUTE_UNUSED; +{ + return (GET_CODE (op) == CONST_INT && SMALL_INT (op)); +} + +/* Return truth value of whether OP is an integer which fits the + range constraining immediate operands in logic insns. */ + +int +logic_int (op, mode) + rtx op; + enum machine_mode mode ATTRIBUTE_UNUSED; +{ + return (GET_CODE (op) == CONST_INT && LOGIC_INT (op)); +} + +/* Test for a valid operand for a call instruction. + Don't allow the arg pointer register or virtual regs + since they may change into reg + const, which the patterns + can't handle yet. */ + +int +call_insn_operand (op, mode) + rtx op; + enum machine_mode mode ATTRIBUTE_UNUSED; +{ + if (GET_CODE (op) == MEM + && (CONSTANT_ADDRESS_P (XEXP (op, 0)) + || (GET_CODE (XEXP (op, 0)) == REG + && XEXP (op, 0) != arg_pointer_rtx + && !(REGNO (XEXP (op, 0)) >= FIRST_PSEUDO_REGISTER + && REGNO (XEXP (op, 0)) <= LAST_VIRTUAL_REGISTER)))) + return 1; + return 0; +} + +/* Return the best assembler insn template + for moving operands[1] into operands[0] as a fullword. */ + +static const char * +singlemove_string (operands) + rtx *operands; +{ + if (GET_CODE (operands[0]) == MEM) + { + if (GET_CODE (operands[1]) != MEM) + if (CONSTANT_ADDRESS_P (XEXP (operands[0], 0))) + { + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && cc_prev_status.mdep == XEXP (operands[0], 0))) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h0,%?r0,%?r31", operands); + } + cc_status.flags |= CC_KNOW_HI_R31 | CC_HI_R31_ADJ; + cc_status.mdep = XEXP (operands[0], 0); + return "st.l %r1,%L0(%?r31)"; + } + else + return "st.l %r1,%0"; + else + abort (); +#if 0 + { + rtx xoperands[2]; + + cc_status.flags &= ~CC_F0_IS_0; + xoperands[0] = gen_rtx_REG (SFmode, 32); + xoperands[1] = operands[1]; + output_asm_insn (singlemove_string (xoperands), xoperands); + xoperands[1] = xoperands[0]; + xoperands[0] = operands[0]; + output_asm_insn (singlemove_string (xoperands), xoperands); + return ""; + } +#endif + } + if (GET_CODE (operands[1]) == MEM) + { + if (CONSTANT_ADDRESS_P (XEXP (operands[1], 0))) + { + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && cc_prev_status.mdep == XEXP (operands[1], 0))) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h1,%?r0,%?r31", operands); + } + cc_status.flags |= CC_KNOW_HI_R31 | CC_HI_R31_ADJ; + cc_status.mdep = XEXP (operands[1], 0); + return "ld.l %L1(%?r31),%0"; + } + return "ld.l %m1,%0"; + } + if (GET_CODE (operands[1]) == CONST_INT) + { + if (operands[1] == const0_rtx) + return "mov %?r0,%0"; + if((INTVAL (operands[1]) & 0xffff0000) == 0) + return "or %L1,%?r0,%0"; + if((INTVAL (operands[1]) & 0xffff8000) == 0xffff8000) + return "adds %1,%?r0,%0"; + if((INTVAL (operands[1]) & 0x0000ffff) == 0) + return "orh %H1,%?r0,%0"; + + return "orh %H1,%?r0,%0\n\tor %L1,%0,%0"; + } + return "mov %1,%0"; +} + +/* Output assembler code to perform a doubleword move insn + with operands OPERANDS. */ + +const char * +output_move_double (operands) + rtx *operands; +{ + enum { REGOP, OFFSOP, MEMOP, PUSHOP, POPOP, CNSTOP, RNDOP } optype0, optype1; + rtx latehalf[2]; + rtx addreg0 = 0, addreg1 = 0; + int highest_first = 0; + int no_addreg1_decrement = 0; + + /* First classify both operands. */ + + if (REG_P (operands[0])) + optype0 = REGOP; + else if (offsettable_memref_p (operands[0])) + optype0 = OFFSOP; + else if (GET_CODE (operands[0]) == MEM) + optype0 = MEMOP; + else + optype0 = RNDOP; + + if (REG_P (operands[1])) + optype1 = REGOP; + else if (CONSTANT_P (operands[1])) + optype1 = CNSTOP; + else if (offsettable_memref_p (operands[1])) + optype1 = OFFSOP; + else if (GET_CODE (operands[1]) == MEM) + optype1 = MEMOP; + else + optype1 = RNDOP; + + /* Check for the cases that the operand constraints are not + supposed to allow to happen. Abort if we get one, + because generating code for these cases is painful. */ + + if (optype0 == RNDOP || optype1 == RNDOP) + abort (); + + /* If an operand is an unoffsettable memory ref, find a register + we can increment temporarily to make it refer to the second word. */ + + if (optype0 == MEMOP) + addreg0 = find_addr_reg (XEXP (operands[0], 0)); + + if (optype1 == MEMOP) + addreg1 = find_addr_reg (XEXP (operands[1], 0)); + +/* ??? Perhaps in some cases move double words + if there is a spare pair of floating regs. */ + + /* Ok, we can do one word at a time. + Normally we do the low-numbered word first, + but if either operand is autodecrementing then we + do the high-numbered word first. + + In either case, set up in LATEHALF the operands to use + for the high-numbered word and in some cases alter the + operands in OPERANDS to be suitable for the low-numbered word. */ + + if (optype0 == REGOP) + latehalf[0] = gen_rtx_REG (SImode, REGNO (operands[0]) + 1); + else if (optype0 == OFFSOP) + latehalf[0] = adjust_address (operands[0], SImode, 4); + else + latehalf[0] = operands[0]; + + if (optype1 == REGOP) + latehalf[1] = gen_rtx_REG (SImode, REGNO (operands[1]) + 1); + else if (optype1 == OFFSOP) + latehalf[1] = adjust_address (operands[1], SImode, 4); + else if (optype1 == CNSTOP) + { + if (GET_CODE (operands[1]) == CONST_DOUBLE) + split_double (operands[1], &operands[1], &latehalf[1]); + else if (CONSTANT_P (operands[1])) + latehalf[1] = const0_rtx; + } + else + latehalf[1] = operands[1]; + + /* If the first move would clobber the source of the second one, + do them in the other order. + + RMS says "This happens only for registers; + such overlap can't happen in memory unless the user explicitly + sets it up, and that is an undefined circumstance." + + but it happens on the sparc when loading parameter registers, + so I am going to define that circumstance, and make it work + as expected. */ + + if (optype0 == REGOP && optype1 == REGOP + && REGNO (operands[0]) == REGNO (latehalf[1])) + { + CC_STATUS_PARTIAL_INIT; + /* Make any unoffsettable addresses point at high-numbered word. */ + if (addreg0) + output_asm_insn ("adds 0x4,%0,%0", &addreg0); + if (addreg1) + output_asm_insn ("adds 0x4,%0,%0", &addreg1); + + /* Do that word. */ + output_asm_insn (singlemove_string (latehalf), latehalf); + + /* Undo the adds we just did. */ + if (addreg0) + output_asm_insn ("adds -0x4,%0,%0", &addreg0); + if (addreg1) + output_asm_insn ("adds -0x4,%0,%0", &addreg1); + + /* Do low-numbered word. */ + return singlemove_string (operands); + } + else if (optype0 == REGOP && optype1 != REGOP + && reg_overlap_mentioned_p (operands[0], operands[1])) + { + /* If both halves of dest are used in the src memory address, + add the two regs and put them in the low reg (operands[0]). + Then it works to load latehalf first. */ + if (reg_mentioned_p (operands[0], XEXP (operands[1], 0)) + && reg_mentioned_p (latehalf[0], XEXP (operands[1], 0))) + { + rtx xops[2]; + xops[0] = latehalf[0]; + xops[1] = operands[0]; + output_asm_insn ("adds %1,%0,%1", xops); + operands[1] = gen_rtx_MEM (DImode, operands[0]); + latehalf[1] = adjust_address (operands[1], SImode, 4); + addreg1 = 0; + highest_first = 1; + } + /* Only one register in the dest is used in the src memory address, + and this is the first register of the dest, so we want to do + the late half first here also. */ + else if (! reg_mentioned_p (latehalf[0], XEXP (operands[1], 0))) + highest_first = 1; + /* Only one register in the dest is used in the src memory address, + and this is the second register of the dest, so we want to do + the late half last. If addreg1 is set, and addreg1 is the same + register as latehalf, then we must suppress the trailing decrement, + because it would clobber the value just loaded. */ + else if (addreg1 && reg_mentioned_p (addreg1, latehalf[0])) + no_addreg1_decrement = 1; + } + + /* Normal case: do the two words, low-numbered first. + Overlap case (highest_first set): do high-numbered word first. */ + + if (! highest_first) + output_asm_insn (singlemove_string (operands), operands); + + CC_STATUS_PARTIAL_INIT; + /* Make any unoffsettable addresses point at high-numbered word. */ + if (addreg0) + output_asm_insn ("adds 0x4,%0,%0", &addreg0); + if (addreg1) + output_asm_insn ("adds 0x4,%0,%0", &addreg1); + + /* Do that word. */ + output_asm_insn (singlemove_string (latehalf), latehalf); + + /* Undo the adds we just did. */ + if (addreg0) + output_asm_insn ("adds -0x4,%0,%0", &addreg0); + if (addreg1 && !no_addreg1_decrement) + output_asm_insn ("adds -0x4,%0,%0", &addreg1); + + if (highest_first) + output_asm_insn (singlemove_string (operands), operands); + + return ""; +} + +const char * +output_fp_move_double (operands) + rtx *operands; +{ + /* If the source operand is any sort of zero, use f0 instead. */ + + if (operands[1] == CONST0_RTX (GET_MODE (operands[1]))) + operands[1] = gen_rtx_REG (DFmode, F0_REGNUM); + + if (FP_REG_P (operands[0])) + { + if (FP_REG_P (operands[1])) + return "fmov.dd %1,%0"; + if (GET_CODE (operands[1]) == REG) + { + output_asm_insn ("ixfr %1,%0", operands); + operands[0] = gen_rtx_REG (VOIDmode, REGNO (operands[0]) + 1); + operands[1] = gen_rtx_REG (VOIDmode, REGNO (operands[1]) + 1); + return "ixfr %1,%0"; + } + if (operands[1] == CONST0_RTX (DFmode)) + return "fmov.dd f0,%0"; + if (CONSTANT_ADDRESS_P (XEXP (operands[1], 0))) + { + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && cc_prev_status.mdep == XEXP (operands[1], 0))) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h1,%?r0,%?r31", operands); + } + cc_status.flags |= CC_KNOW_HI_R31 | CC_HI_R31_ADJ; + cc_status.mdep = XEXP (operands[1], 0); + return "fld.d %L1(%?r31),%0"; + } + return "fld.d %1,%0"; + } + else if (FP_REG_P (operands[1])) + { + if (GET_CODE (operands[0]) == REG) + { + output_asm_insn ("fxfr %1,%0", operands); + operands[0] = gen_rtx_REG (VOIDmode, REGNO (operands[0]) + 1); + operands[1] = gen_rtx_REG (VOIDmode, REGNO (operands[1]) + 1); + return "fxfr %1,%0"; + } + if (CONSTANT_ADDRESS_P (XEXP (operands[0], 0))) + { + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && cc_prev_status.mdep == XEXP (operands[0], 0))) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h0,%?r0,%?r31", operands); + } + cc_status.flags |= CC_KNOW_HI_R31 | CC_HI_R31_ADJ; + cc_status.mdep = XEXP (operands[0], 0); + return "fst.d %1,%L0(%?r31)"; + } + return "fst.d %1,%0"; + } + else + abort (); + /* NOTREACHED */ + return NULL; +} + +/* Return a REG that occurs in ADDR with coefficient 1. + ADDR can be effectively incremented by incrementing REG. */ + +static rtx +find_addr_reg (addr) + rtx addr; +{ + while (GET_CODE (addr) == PLUS) + { + if (GET_CODE (XEXP (addr, 0)) == REG) + addr = XEXP (addr, 0); + else if (GET_CODE (XEXP (addr, 1)) == REG) + addr = XEXP (addr, 1); + else if (CONSTANT_P (XEXP (addr, 0))) + addr = XEXP (addr, 1); + else if (CONSTANT_P (XEXP (addr, 1))) + addr = XEXP (addr, 0); + else + abort (); + } + if (GET_CODE (addr) == REG) + return addr; + abort (); + /* NOTREACHED */ + return NULL; +} + +/* Return a template for a load instruction with mode MODE and + arguments from the string ARGS. + + This string is in static storage. */ + +static const char * +load_opcode (mode, args, reg) + enum machine_mode mode; + const char *args; + rtx reg; +{ + static char buf[30]; + const char *opcode; + + switch (mode) + { + case QImode: + opcode = "ld.b"; + break; + + case HImode: + opcode = "ld.s"; + break; + + case SImode: + case SFmode: + if (FP_REG_P (reg)) + opcode = "fld.l"; + else + opcode = "ld.l"; + break; + + case DImode: + if (!FP_REG_P (reg)) + abort (); + case DFmode: + opcode = "fld.d"; + break; + + default: + abort (); + } + + sprintf (buf, "%s %s", opcode, args); + return buf; +} + +/* Return a template for a store instruction with mode MODE and + arguments from the string ARGS. + + This string is in static storage. */ + +static const char * +store_opcode (mode, args, reg) + enum machine_mode mode; + const char *args; + rtx reg; +{ + static char buf[30]; + const char *opcode; + + switch (mode) + { + case QImode: + opcode = "st.b"; + break; + + case HImode: + opcode = "st.s"; + break; + + case SImode: + case SFmode: + if (FP_REG_P (reg)) + opcode = "fst.l"; + else + opcode = "st.l"; + break; + + case DImode: + if (!FP_REG_P (reg)) + abort (); + case DFmode: + opcode = "fst.d"; + break; + + default: + abort (); + } + + sprintf (buf, "%s %s", opcode, args); + return buf; +} + +/* Output a store-in-memory whose operands are OPERANDS[0,1]. + OPERANDS[0] is a MEM, and OPERANDS[1] is a reg or zero. + + This function returns a template for an insn. + This is in static storage. + + It may also output some insns directly. + It may alter the values of operands[0] and operands[1]. */ + +const char * +output_store (operands) + rtx *operands; +{ + enum machine_mode mode = GET_MODE (operands[0]); + rtx address = XEXP (operands[0], 0); + + cc_status.flags |= CC_KNOW_HI_R31 | CC_HI_R31_ADJ; + cc_status.mdep = address; + + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && address == cc_prev_status.mdep)) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h0,%?r0,%?r31", operands); + cc_prev_status.mdep = address; + } + + /* Store zero in two parts when appropriate. */ + if (mode == DFmode && operands[1] == CONST0_RTX (DFmode)) + return store_opcode (DFmode, "%r1,%L0(%?r31)", operands[1]); + + /* Code below isn't smart enough to move a doubleword in two parts, + so use output_move_double to do that in the cases that require it. */ + if ((mode == DImode || mode == DFmode) + && ! FP_REG_P (operands[1])) + return output_move_double (operands); + + return store_opcode (mode, "%r1,%L0(%?r31)", operands[1]); +} + +/* Output a load-from-memory whose operands are OPERANDS[0,1]. + OPERANDS[0] is a reg, and OPERANDS[1] is a mem. + + This function returns a template for an insn. + This is in static storage. + + It may also output some insns directly. + It may alter the values of operands[0] and operands[1]. */ + +const char * +output_load (operands) + rtx *operands; +{ + enum machine_mode mode = GET_MODE (operands[0]); + rtx address = XEXP (operands[1], 0); + + /* We don't bother trying to see if we know %hi(address). + This is because we are doing a load, and if we know the + %hi value, we probably also know that value in memory. */ + cc_status.flags |= CC_KNOW_HI_R31 | CC_HI_R31_ADJ; + cc_status.mdep = address; + + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && address == cc_prev_status.mdep + && cc_prev_status.mdep == cc_status.mdep)) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h1,%?r0,%?r31", operands); + cc_prev_status.mdep = address; + } + + /* Code below isn't smart enough to move a doubleword in two parts, + so use output_move_double to do that in the cases that require it. */ + if ((mode == DImode || mode == DFmode) + && ! FP_REG_P (operands[0])) + return output_move_double (operands); + + return load_opcode (mode, "%L1(%?r31),%0", operands[0]); +} + +#if 0 +/* Load the address specified by OPERANDS[3] into the register + specified by OPERANDS[0]. + + OPERANDS[3] may be the result of a sum, hence it could either be: + + (1) CONST + (2) REG + (2) REG + CONST_INT + (3) REG + REG + CONST_INT + (4) REG + REG (special case of 3). + + Note that (3) is not a legitimate address. + All cases are handled here. */ + +void +output_load_address (operands) + rtx *operands; +{ + rtx base, offset; + + if (CONSTANT_P (operands[3])) + { + output_asm_insn ("mov %3,%0", operands); + return; + } + + if (REG_P (operands[3])) + { + if (REGNO (operands[0]) != REGNO (operands[3])) + output_asm_insn ("shl %?r0,%3,%0", operands); + return; + } + + if (GET_CODE (operands[3]) != PLUS) + abort (); + + base = XEXP (operands[3], 0); + offset = XEXP (operands[3], 1); + + if (GET_CODE (base) == CONST_INT) + { + rtx tmp = base; + base = offset; + offset = tmp; + } + + if (GET_CODE (offset) != CONST_INT) + { + /* Operand is (PLUS (REG) (REG)). */ + base = operands[3]; + offset = const0_rtx; + } + + if (REG_P (base)) + { + operands[6] = base; + operands[7] = offset; + CC_STATUS_PARTIAL_INIT; + if (SMALL_INT (offset)) + output_asm_insn ("adds %7,%6,%0", operands); + else + output_asm_insn ("mov %7,%0\n\tadds %0,%6,%0", operands); + } + else if (GET_CODE (base) == PLUS) + { + operands[6] = XEXP (base, 0); + operands[7] = XEXP (base, 1); + operands[8] = offset; + + CC_STATUS_PARTIAL_INIT; + if (SMALL_INT (offset)) + output_asm_insn ("adds %6,%7,%0\n\tadds %8,%0,%0", operands); + else + output_asm_insn ("mov %8,%0\n\tadds %0,%6,%0\n\tadds %0,%7,%0", operands); + } + else + abort (); +} +#endif + +/* Output code to place a size count SIZE in register REG. + Because block moves are pipelined, we don't include the + first element in the transfer of SIZE to REG. + For this, we subtract ALIGN. (Actually, I think it is not + right to subtract on this machine, so right now we don't.) */ + +static void +output_size_for_block_move (size, reg, align) + rtx size, reg, align; +{ + rtx xoperands[3]; + + xoperands[0] = reg; + xoperands[1] = size; + xoperands[2] = align; + +#if 1 + cc_status.flags &= ~ CC_KNOW_HI_R31; + output_asm_insn (singlemove_string (xoperands), xoperands); +#else + if (GET_CODE (size) == REG) + output_asm_insn ("sub %2,%1,%0", xoperands); + else + { + xoperands[1] = GEN_INT (INTVAL (size) - INTVAL (align)); + cc_status.flags &= ~ CC_KNOW_HI_R31; + output_asm_insn ("mov %1,%0", xoperands); + } +#endif +} + +/* Emit code to perform a block move. + + OPERANDS[0] is the destination. + OPERANDS[1] is the source. + OPERANDS[2] is the size. + OPERANDS[3] is the known safe alignment. + OPERANDS[4..6] are pseudos we can safely clobber as temps. */ + +const char * +output_block_move (operands) + rtx *operands; +{ + /* A vector for our computed operands. Note that load_output_address + makes use of (and can clobber) up to the 8th element of this vector. */ + rtx xoperands[10]; +#if 0 + rtx zoperands[10]; +#endif + static int movstrsi_label = 0; + int i; + rtx temp1 = operands[4]; + rtx alignrtx = operands[3]; + int align = INTVAL (alignrtx); + int chunk_size; + + xoperands[0] = operands[0]; + xoperands[1] = operands[1]; + xoperands[2] = temp1; + + /* We can't move more than four bytes at a time + because we have only one register to move them through. */ + if (align > 4) + { + align = 4; + alignrtx = GEN_INT (4); + } + + /* Recognize special cases of block moves. These occur + when GNU C++ is forced to treat something as BLKmode + to keep it in memory, when its mode could be represented + with something smaller. + + We cannot do this for global variables, since we don't know + what pages they don't cross. Sigh. */ + if (GET_CODE (operands[2]) == CONST_INT + && ! CONSTANT_ADDRESS_P (operands[0]) + && ! CONSTANT_ADDRESS_P (operands[1])) + { + int size = INTVAL (operands[2]); + rtx op0 = xoperands[0]; + rtx op1 = xoperands[1]; + + if ((align & 3) == 0 && (size & 3) == 0 && (size >> 2) <= 16) + { + if (memory_address_p (SImode, plus_constant (op0, size)) + && memory_address_p (SImode, plus_constant (op1, size))) + { + cc_status.flags &= ~CC_KNOW_HI_R31; + for (i = (size>>2)-1; i >= 0; i--) + { + xoperands[0] = plus_constant (op0, i * 4); + xoperands[1] = plus_constant (op1, i * 4); + output_asm_insn ("ld.l %a1,%?r31\n\tst.l %?r31,%a0", + xoperands); + } + return ""; + } + } + else if ((align & 1) == 0 && (size & 1) == 0 && (size >> 1) <= 16) + { + if (memory_address_p (HImode, plus_constant (op0, size)) + && memory_address_p (HImode, plus_constant (op1, size))) + { + cc_status.flags &= ~CC_KNOW_HI_R31; + for (i = (size>>1)-1; i >= 0; i--) + { + xoperands[0] = plus_constant (op0, i * 2); + xoperands[1] = plus_constant (op1, i * 2); + output_asm_insn ("ld.s %a1,%?r31\n\tst.s %?r31,%a0", + xoperands); + } + return ""; + } + } + else if (size <= 16) + { + if (memory_address_p (QImode, plus_constant (op0, size)) + && memory_address_p (QImode, plus_constant (op1, size))) + { + cc_status.flags &= ~CC_KNOW_HI_R31; + for (i = size-1; i >= 0; i--) + { + xoperands[0] = plus_constant (op0, i); + xoperands[1] = plus_constant (op1, i); + output_asm_insn ("ld.b %a1,%?r31\n\tst.b %?r31,%a0", + xoperands); + } + return ""; + } + } + } + + /* Since we clobber untold things, nix the condition codes. */ + CC_STATUS_INIT; + + /* This is the size of the transfer. + Either use the register which already contains the size, + or use a free register (used by no operands). */ + output_size_for_block_move (operands[2], operands[4], alignrtx); + +#if 0 + /* Also emit code to decrement the size value by ALIGN. */ + zoperands[0] = operands[0]; + zoperands[3] = plus_constant (operands[0], align); + output_load_address (zoperands); +#endif + + /* Generate number for unique label. */ + + xoperands[3] = GEN_INT (movstrsi_label++); + + /* Calculate the size of the chunks we will be trying to move first. */ + +#if 0 + if ((align & 3) == 0) + chunk_size = 4; + else if ((align & 1) == 0) + chunk_size = 2; + else +#endif + chunk_size = 1; + + /* Copy the increment (negative) to a register for bla insn. */ + + xoperands[4] = GEN_INT (- chunk_size); + xoperands[5] = operands[5]; + output_asm_insn ("adds %4,%?r0,%5", xoperands); + + /* Predecrement the loop counter. This happens again also in the `bla' + instruction which precedes the loop, but we need to have it done + two times before we enter the loop because of the bizarre semantics + of the bla instruction. */ + + output_asm_insn ("adds %5,%2,%2", xoperands); + + /* Check for the case where the original count was less than or equal to + zero. Avoid going through the loop at all if the original count was + indeed less than or equal to zero. Note that we treat the count as + if it were a signed 32-bit quantity here, rather than an unsigned one, + even though we really shouldn't. We have to do this because of the + semantics of the `ble' instruction, which assume that the count is + a signed 32-bit value. Anyway, in practice it won't matter because + nobody is going to try to do a memcpy() of more than half of the + entire address space (i.e. 2 gigabytes) anyway. */ + + output_asm_insn ("bc .Le%3", xoperands); + + /* Make available a register which is a temporary. */ + + xoperands[6] = operands[6]; + + /* Now the actual loop. + In xoperands, elements 1 and 0 are the input and output vectors. + Element 2 is the loop index. Element 5 is the increment. */ + + output_asm_insn ("subs %1,%5,%1", xoperands); + output_asm_insn ("bla %5,%2,.Lm%3", xoperands); + output_asm_insn ("adds %0,%2,%6", xoperands); + output_asm_insn ("\n.Lm%3:", xoperands); /* Label for bla above. */ + output_asm_insn ("\n.Ls%3:", xoperands); /* Loop start label. */ + output_asm_insn ("adds %5,%6,%6", xoperands); + + /* NOTE: The code here which is supposed to handle the cases where the + sources and destinations are known to start on a 4 or 2 byte boundary + are currently broken. They fail to do anything about the overflow + bytes which might still need to be copied even after we have copied + some number of words or halfwords. Thus, for now we use the lowest + common denominator, i.e. the code which just copies some number of + totally unaligned individual bytes. (See the calculation of + chunk_size above. */ + + if (chunk_size == 4) + { + output_asm_insn ("ld.l %2(%1),%?r31", xoperands); + output_asm_insn ("bla %5,%2,.Ls%3", xoperands); + output_asm_insn ("st.l %?r31,8(%6)", xoperands); + } + else if (chunk_size == 2) + { + output_asm_insn ("ld.s %2(%1),%?r31", xoperands); + output_asm_insn ("bla %5,%2,.Ls%3", xoperands); + output_asm_insn ("st.s %?r31,4(%6)", xoperands); + } + else /* chunk_size == 1 */ + { + output_asm_insn ("ld.b %2(%1),%?r31", xoperands); + output_asm_insn ("bla %5,%2,.Ls%3", xoperands); + output_asm_insn ("st.b %?r31,2(%6)", xoperands); + } + output_asm_insn ("\n.Le%3:", xoperands); /* Here if count <= 0. */ + + return ""; +} + +#if 0 +/* Output a delayed branch insn with the delay insn in its + branch slot. The delayed branch insn template is in TEMPLATE, + with operands OPERANDS. The insn in its delay slot is INSN. + + As a special case, since we know that all memory transfers are via + ld/st insns, if we see a (MEM (SYMBOL_REF ...)) we divide the memory + reference around the branch as + + orh ha%x,%?r0,%?r31 + b ... + ld/st l%x(%?r31),... + + As another special case, we handle loading (SYMBOL_REF ...) and + other large constants around branches as well: + + orh h%x,%?r0,%0 + b ... + or l%x,%0,%1 + + */ +/* ??? Disabled because this re-recognition is incomplete and causes + constrain_operands to segfault. Anyone who cares should fix up + the code to use the DBR pass. */ + +const char * +output_delayed_branch (template, operands, insn) + const char *template; + rtx *operands; + rtx insn; +{ + rtx src = XVECEXP (PATTERN (insn), 0, 1); + rtx dest = XVECEXP (PATTERN (insn), 0, 0); + + /* See if we are doing some branch together with setting some register + to some 32-bit value which does (or may) have some of the high-order + 16 bits set. If so, we need to set the register in two stages. One + stage must be done before the branch, and the other one can be done + in the delay slot. */ + + if ( (GET_CODE (src) == CONST_INT + && ((unsigned) INTVAL (src) & (unsigned) 0xffff0000) != (unsigned) 0) + || (GET_CODE (src) == SYMBOL_REF) + || (GET_CODE (src) == LABEL_REF) + || (GET_CODE (src) == CONST)) + { + rtx xoperands[2]; + xoperands[0] = dest; + xoperands[1] = src; + + CC_STATUS_PARTIAL_INIT; + /* Output the `orh' insn. */ + output_asm_insn ("orh %H1,%?r0,%0", xoperands); + + /* Output the branch instruction next. */ + output_asm_insn (template, operands); + + /* Now output the `or' insn. */ + output_asm_insn ("or %L1,%0,%0", xoperands); + } + else if ((GET_CODE (src) == MEM + && CONSTANT_ADDRESS_P (XEXP (src, 0))) + || (GET_CODE (dest) == MEM + && CONSTANT_ADDRESS_P (XEXP (dest, 0)))) + { + rtx xoperands[2]; + const char *split_template; + xoperands[0] = dest; + xoperands[1] = src; + + /* Output the `orh' insn. */ + if (GET_CODE (src) == MEM) + { + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && cc_prev_status.mdep == XEXP (operands[1], 0))) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h1,%?r0,%?r31", xoperands); + } + split_template = load_opcode (GET_MODE (dest), + "%L1(%?r31),%0", dest); + } + else + { + if (! ((cc_prev_status.flags & CC_KNOW_HI_R31) + && (cc_prev_status.flags & CC_HI_R31_ADJ) + && cc_prev_status.mdep == XEXP (operands[0], 0))) + { + CC_STATUS_INIT; + output_asm_insn ("orh %h0,%?r0,%?r31", xoperands); + } + split_template = store_opcode (GET_MODE (dest), + "%r1,%L0(%?r31)", src); + } + + /* Output the branch instruction next. */ + output_asm_insn (template, operands); + + /* Now output the load or store. + No need to do a CC_STATUS_INIT, because we are branching anyway. */ + output_asm_insn (split_template, xoperands); + } + else + { + int insn_code_number; + rtx pat = gen_rtx_SET (VOIDmode, dest, src); + rtx delay_insn = gen_rtx_INSN (VOIDmode, 0, 0, 0, pat, -1, 0, 0); + int i; + + /* Output the branch instruction first. */ + output_asm_insn (template, operands); + + /* Now recognize the insn which we put in its delay slot. + We must do this after outputting the branch insn, + since operands may just be a pointer to `recog_data.operand'. */ + INSN_CODE (delay_insn) = insn_code_number + = recog (pat, delay_insn, NULL); + if (insn_code_number == -1) + abort (); + + for (i = 0; i < insn_data[insn_code_number].n_operands; i++) + { + if (GET_CODE (recog_data.operand[i]) == SUBREG) + alter_subreg (&recog_data.operand[i]); + } + + insn_extract (delay_insn); + if (! constrain_operands (1)) + fatal_insn_not_found (delay_insn); + + template = get_insn_template (insn_code_number, delay_insn); + output_asm_insn (template, recog_data.operand); + } + CC_STATUS_INIT; + return ""; +} + +/* Output a newly constructed insn DELAY_INSN. */ +const char * +output_delay_insn (delay_insn) + rtx delay_insn; +{ + const char *template; + int insn_code_number; + int i; + + /* Now recognize the insn which we put in its delay slot. + We must do this after outputting the branch insn, + since operands may just be a pointer to `recog_data.operand'. */ + insn_code_number = recog_memoized (delay_insn); + if (insn_code_number == -1) + abort (); + + /* Extract the operands of this delay insn. */ + INSN_CODE (delay_insn) = insn_code_number; + insn_extract (delay_insn); + + /* It is possible that this insn has not been properly scanned by final + yet. If this insn's operands don't appear in the peephole's + actual operands, then they won't be fixed up by final, so we + make sure they get fixed up here. -- This is a kludge. */ + for (i = 0; i < insn_data[insn_code_number].n_operands; i++) + { + if (GET_CODE (recog_data.operand[i]) == SUBREG) + alter_subreg (&recog_data.operand[i]); + } + + if (! constrain_operands (1)) + abort (); + + cc_prev_status = cc_status; + + /* Update `cc_status' for this instruction. + The instruction's output routine may change it further. + If the output routine for a jump insn needs to depend + on the cc status, it should look at cc_prev_status. */ + + NOTICE_UPDATE_CC (PATTERN (delay_insn), delay_insn); + + /* Now get the template for what this insn would + have been, without the branch. */ + + template = get_insn_template (insn_code_number, delay_insn); + output_asm_insn (template, recog_data.operand); + return ""; +} +#endif + +/* Special routine to convert an SFmode value represented as a + CONST_DOUBLE into its equivalent unsigned long bit pattern. + We convert the value from a double precision floating-point + value to single precision first, and thence to a bit-wise + equivalent unsigned long value. This routine is used when + generating an immediate move of an SFmode value directly + into a general register because the svr4 assembler doesn't + grok floating literals in instruction operand contexts. */ + +unsigned long +sfmode_constant_to_ulong (x) + rtx x; +{ + REAL_VALUE_TYPE d; + unsigned long l; + + if (GET_CODE (x) != CONST_DOUBLE || GET_MODE (x) != SFmode) + abort (); + + REAL_VALUE_FROM_CONST_DOUBLE (d, x); + REAL_VALUE_TO_TARGET_SINGLE (d, l); + return l; +} + +/* This function generates the assembly code for function entry. + + ASM_FILE is a stdio stream to output the code to. + SIZE is an int: how many units of temporary storage to allocate. + + Refer to the array `regs_ever_live' to determine which registers + to save; `regs_ever_live[I]' is nonzero if register number I + is ever used in the function. This macro is responsible for + knowing which registers should not be saved even if used. + + NOTE: `frame_lower_bytes' is the count of bytes which will lie + between the new `fp' value and the new `sp' value after the + prologue is done. `frame_upper_bytes' is the count of bytes + that will lie between the new `fp' and the *old* `sp' value + after the new `fp' is setup (in the prologue). The upper + part of each frame always includes at least 2 words (8 bytes) + to hold the saved frame pointer and the saved return address. + + The svr4 ABI for the i860 now requires that the values of the + stack pointer and frame pointer registers be kept aligned to + 16-byte boundaries at all times. We obey that restriction here. + + The svr4 ABI for the i860 is entirely vague when it comes to specifying + exactly where the "preserved" registers should be saved. The native + svr4 C compiler I now have doesn't help to clarify the requirements + very much because it is plainly out-of-date and non-ABI-compliant + (in at least one important way, i.e. how it generates function + epilogues). + + The native svr4 C compiler saves the "preserved" registers (i.e. + r4-r15 and f2-f7) in the lower part of a frame (i.e. at negative + offsets from the frame pointer). + + Previous versions of GCC also saved the "preserved" registers in the + "negative" part of the frame, but they saved them using positive + offsets from the (adjusted) stack pointer (after it had been adjusted + to allocate space for the new frame). That's just plain wrong + because if the current function calls alloca(), the stack pointer + will get moved, and it will be impossible to restore the registers + properly again after that. + + Both compilers handled parameter registers (i.e. r16-r27 and f8-f15) + by copying their values either into various "preserved" registers or + into stack slots in the lower part of the current frame (as seemed + appropriate, depending upon subsequent usage of these values). + + Here we want to save the preserved registers at some offset from the + frame pointer register so as to avoid any possible problems arising + from calls to alloca(). We can either save them at small positive + offsets from the frame pointer, or at small negative offsets from + the frame pointer. If we save them at small negative offsets from + the frame pointer (i.e. in the lower part of the frame) then we + must tell the rest of GCC (via STARTING_FRAME_OFFSET) exactly how + many bytes of space we plan to use in the lower part of the frame + for this purpose. Since other parts of the compiler reference the + value of STARTING_FRAME_OFFSET long before final() calls this function, + we would have to go ahead and assume the worst-case storage requirements + for saving all of the "preserved" registers (and use that number, i.e. + `80', to define STARTING_FRAME_OFFSET) if we wanted to save them in + the lower part of the frame. That could potentially be very wasteful, + and that wastefulness could really hamper people compiling for embedded + i860 targets with very tight limits on stack space. Thus, we choose + here to save the preserved registers in the upper part of the + frame, so that we can decide at the very last minute how much (or how + little) space we must allocate for this purpose. + + To satisfy the needs of the svr4 ABI "tdesc" scheme, preserved + registers must always be saved so that the saved values of registers + with higher numbers are at higher addresses. We obey that restriction + here. + + There are two somewhat different ways that you can generate prologues + here... i.e. pedantically ABI-compliant, and the "other" way. The + "other" way is more consistent with what is currently generated by the + "native" svr4 C compiler for the i860. That's important if you want + to use the current (as of 8/91) incarnation of svr4 SDB for the i860. + The SVR4 SDB for the i860 insists on having function prologues be + non-ABI-compliant! + + To get fully ABI-compliant prologues, define I860_STRICT_ABI_PROLOGUES + in the i860svr4.h file. (By default this is *not* defined). + + The differences between the ABI-compliant and non-ABI-compliant prologues + are that (a) the ABI version seems to require the use of *signed* + (rather than unsigned) adds and subtracts, and (b) the ordering of + the various steps (e.g. saving preserved registers, saving the + return address, setting up the new frame pointer value) is different. + + For strict ABI compliance, it seems to be the case that the very last + thing that is supposed to happen in the prologue is getting the frame + pointer set to its new value (but only after everything else has + already been properly setup). We do that here, but only if the symbol + I860_STRICT_ABI_PROLOGUES is defined. +*/ + +#ifndef STACK_ALIGNMENT +#define STACK_ALIGNMENT 16 +#endif + +const char *current_function_original_name; + +static int must_preserve_r1; +static unsigned must_preserve_bytes; + +static void +i860_output_function_prologue (asm_file, local_bytes) + register FILE *asm_file; + register HOST_WIDE_INT local_bytes; +{ + register HOST_WIDE_INT frame_lower_bytes; + register HOST_WIDE_INT frame_upper_bytes; + register HOST_WIDE_INT total_fsize; + register unsigned preserved_reg_bytes = 0; + register unsigned i; + register unsigned preserved_so_far = 0; + + must_preserve_r1 = (optimize < 2 || ! leaf_function_p ()); + must_preserve_bytes = 4 + (must_preserve_r1 ? 4 : 0); + + /* Count registers that need preserving. Ignore r0. It never needs + preserving. */ + + for (i = 1; i < FIRST_PSEUDO_REGISTER; i++) + { + if (regs_ever_live[i] && ! call_used_regs[i]) + preserved_reg_bytes += 4; + } + + /* Round-up the frame_lower_bytes so that it's a multiple of 16. */ + + frame_lower_bytes = (local_bytes + STACK_ALIGNMENT - 1) & -STACK_ALIGNMENT; + + /* The upper part of each frame will contain the saved fp, + the saved r1, and stack slots for all of the other "preserved" + registers that we find we will need to save & restore. */ + + frame_upper_bytes = must_preserve_bytes + preserved_reg_bytes; + + /* Round-up the frame_upper_bytes so that it's a multiple of 16. */ + + frame_upper_bytes + = (frame_upper_bytes + STACK_ALIGNMENT - 1) & -STACK_ALIGNMENT; + + total_fsize = frame_upper_bytes + frame_lower_bytes; + +#ifndef I860_STRICT_ABI_PROLOGUES + + /* There are two kinds of function prologues. + You use the "small" version if the total frame size is + small enough so that it can fit into an immediate 16-bit + value in one instruction. Otherwise, you use the "large" + version of the function prologue. */ + + if (total_fsize > 0x7fff) + { + /* Adjust the stack pointer. The ABI sez to do this using `adds', + but the native C compiler on svr4 uses `addu'. */ + + fprintf (asm_file, "\taddu -%d,%ssp,%ssp\n", + frame_upper_bytes, i860_reg_prefix, i860_reg_prefix); + + /* Save the old frame pointer. */ + + fprintf (asm_file, "\tst.l %sfp,0(%ssp)\n", + i860_reg_prefix, i860_reg_prefix); + + /* Setup the new frame pointer. The ABI sez to do this after + preserving registers (using adds), but that's not what the + native C compiler on svr4 does. */ + + fprintf (asm_file, "\taddu 0,%ssp,%sfp\n", + i860_reg_prefix, i860_reg_prefix); + + /* Get the value of frame_lower_bytes into r31. */ + + fprintf (asm_file, "\torh %d,%sr0,%sr31\n", + frame_lower_bytes >> 16, i860_reg_prefix, i860_reg_prefix); + fprintf (asm_file, "\tor %d,%sr31,%sr31\n", + frame_lower_bytes & 0xffff, i860_reg_prefix, i860_reg_prefix); + + /* Now re-adjust the stack pointer using the value in r31. + The ABI sez to do this with `subs' but SDB may prefer `subu'. */ + + fprintf (asm_file, "\tsubu %ssp,%sr31,%ssp\n", + i860_reg_prefix, i860_reg_prefix, i860_reg_prefix); + + /* Preserve registers. The ABI sez to do this before setting + up the new frame pointer, but that's not what the native + C compiler on svr4 does. */ + + for (i = 1; i < 32; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tst.l %s%s,%d(%sfp)\n", + i860_reg_prefix, reg_names[i], + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + for (i = 32; i < 64; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tfst.l %s%s,%d(%sfp)\n", + i860_reg_prefix, reg_names[i], + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + /* Save the return address. */ + + if (must_preserve_r1) + fprintf (asm_file, "\tst.l %sr1,4(%sfp)\n", + i860_reg_prefix, i860_reg_prefix); + } + else + { + /* Adjust the stack pointer. The ABI sez to do this using `adds', + but the native C compiler on svr4 uses `addu'. */ + + fprintf (asm_file, "\taddu -%d,%ssp,%ssp\n", + total_fsize, i860_reg_prefix, i860_reg_prefix); + + /* Save the old frame pointer. */ + + fprintf (asm_file, "\tst.l %sfp,%d(%ssp)\n", + i860_reg_prefix, frame_lower_bytes, i860_reg_prefix); + + /* Setup the new frame pointer. The ABI sez to do this after + preserving registers and after saving the return address, + (and its saz to do this using adds), but that's not what the + native C compiler on svr4 does. */ + + fprintf (asm_file, "\taddu %d,%ssp,%sfp\n", + frame_lower_bytes, i860_reg_prefix, i860_reg_prefix); + + /* Preserve registers. The ABI sez to do this before setting + up the new frame pointer, but that's not what the native + compiler on svr4 does. */ + + for (i = 1; i < 32; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tst.l %s%s,%d(%sfp)\n", + i860_reg_prefix, reg_names[i], + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + for (i = 32; i < 64; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tfst.l %s%s,%d(%sfp)\n", + i860_reg_prefix, reg_names[i], + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + /* Save the return address. The ABI sez to do this earlier, + and also via an offset from %sp, but the native C compiler + on svr4 does it later (i.e. now) and uses an offset from + %fp. */ + + if (must_preserve_r1) + fprintf (asm_file, "\tst.l %sr1,4(%sfp)\n", + i860_reg_prefix, i860_reg_prefix); + } + +#else /* defined(I860_STRICT_ABI_PROLOGUES) */ + + /* There are two kinds of function prologues. + You use the "small" version if the total frame size is + small enough so that it can fit into an immediate 16-bit + value in one instruction. Otherwise, you use the "large" + version of the function prologue. */ + + if (total_fsize > 0x7fff) + { + /* Adjust the stack pointer (thereby allocating a new frame). */ + + fprintf (asm_file, "\tadds -%d,%ssp,%ssp\n", + frame_upper_bytes, i860_reg_prefix, i860_reg_prefix); + + /* Save the caller's frame pointer. */ + + fprintf (asm_file, "\tst.l %sfp,0(%ssp)\n", + i860_reg_prefix, i860_reg_prefix); + + /* Save return address. */ + + if (must_preserve_r1) + fprintf (asm_file, "\tst.l %sr1,4(%ssp)\n", + i860_reg_prefix, i860_reg_prefix); + + /* Get the value of frame_lower_bytes into r31 for later use. */ + + fprintf (asm_file, "\torh %d,%sr0,%sr31\n", + frame_lower_bytes >> 16, i860_reg_prefix, i860_reg_prefix); + fprintf (asm_file, "\tor %d,%sr31,%sr31\n", + frame_lower_bytes & 0xffff, i860_reg_prefix, i860_reg_prefix); + + /* Now re-adjust the stack pointer using the value in r31. */ + + fprintf (asm_file, "\tsubs %ssp,%sr31,%ssp\n", + i860_reg_prefix, i860_reg_prefix, i860_reg_prefix); + + /* Pre-compute value to be used as the new frame pointer. */ + + fprintf (asm_file, "\tadds %ssp,%sr31,%sr31\n", + i860_reg_prefix, i860_reg_prefix, i860_reg_prefix); + + /* Preserve registers. */ + + for (i = 1; i < 32; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tst.l %s%s,%d(%sr31)\n", + i860_reg_prefix, reg_names[i], + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + for (i = 32; i < 64; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tfst.l %s%s,%d(%sr31)\n", + i860_reg_prefix, reg_names[i], + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + /* Actually set the new value of the frame pointer. */ + + fprintf (asm_file, "\tmov %sr31,%sfp\n", + i860_reg_prefix, i860_reg_prefix); + } + else + { + /* Adjust the stack pointer. */ + + fprintf (asm_file, "\tadds -%d,%ssp,%ssp\n", + total_fsize, i860_reg_prefix, i860_reg_prefix); + + /* Save the caller's frame pointer. */ + + fprintf (asm_file, "\tst.l %sfp,%d(%ssp)\n", + i860_reg_prefix, frame_lower_bytes, i860_reg_prefix); + + /* Save the return address. */ + + if (must_preserve_r1) + fprintf (asm_file, "\tst.l %sr1,%d(%ssp)\n", + i860_reg_prefix, frame_lower_bytes + 4, i860_reg_prefix); + + /* Preserve registers. */ + + for (i = 1; i < 32; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tst.l %s%s,%d(%ssp)\n", + i860_reg_prefix, reg_names[i], + frame_lower_bytes + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + for (i = 32; i < 64; i++) + if (regs_ever_live[i] && ! call_used_regs[i]) + fprintf (asm_file, "\tfst.l %s%s,%d(%ssp)\n", + i860_reg_prefix, reg_names[i], + frame_lower_bytes + must_preserve_bytes + (4 * preserved_so_far++), + i860_reg_prefix); + + /* Setup the new frame pointer. */ + + fprintf (asm_file, "\tadds %d,%ssp,%sfp\n", + frame_lower_bytes, i860_reg_prefix, i860_reg_prefix); + } +#endif /* defined(I860_STRICT_ABI_PROLOGUES) */ + +#ifdef ASM_OUTPUT_PROLOGUE_SUFFIX + ASM_OUTPUT_PROLOGUE_SUFFIX (asm_file); +#endif /* defined(ASM_OUTPUT_PROLOGUE_SUFFIX) */ +} + +/* This function generates the assembly code for function exit. + + ASM_FILE is a stdio stream to output the code to. + SIZE is an int: how many units of temporary storage to allocate. + + The function epilogue should not depend on the current stack pointer! + It should use the frame pointer only. This is mandatory because + of alloca; we also take advantage of it to omit stack adjustments + before returning. + + Note that when we go to restore the preserved register values we must + not try to address their slots by using offsets from the stack pointer. + That's because the stack pointer may have been moved during the function + execution due to a call to alloca(). Rather, we must restore all + preserved registers via offsets from the frame pointer value. + + Note also that when the current frame is being "popped" (by adjusting + the value of the stack pointer) on function exit, we must (for the + sake of alloca) set the new value of the stack pointer based upon + the current value of the frame pointer. We can't just add what we + believe to be the (static) frame size to the stack pointer because + if we did that, and alloca() had been called during this function, + we would end up returning *without* having fully deallocated all of + the space grabbed by alloca. If that happened, and a function + containing one or more alloca() calls was called over and over again, + then the stack would grow without limit! + + Finally note that the epilogues generated here are completely ABI + compliant. They go out of their way to insure that the value in + the frame pointer register is never less than the value in the stack + pointer register. It's not clear why this relationship needs to be + maintained at all times, but maintaining it only costs one extra + instruction, so what the hell. +*/ + +/* This corresponds to a version 4 TDESC structure. Lower numbered + versions successively omit the last word of the structure. We + don't try to handle version 5 here. */ + +typedef struct TDESC_flags { + int version:4; + int reg_packing:1; + int callable_block:1; + int reserved:4; + int fregs:6; /* fp regs 2-7 */ + int iregs:16; /* regs 0-15 */ +} TDESC_flags; + +typedef struct TDESC { + TDESC_flags flags; + int integer_reg_offset; /* same as must_preserve_bytes */ + int floating_point_reg_offset; + unsigned int positive_frame_size; /* same as frame_upper_bytes */ + unsigned int negative_frame_size; /* same as frame_lower_bytes */ +} TDESC; + +static void +i860_output_function_epilogue (asm_file, local_bytes) + register FILE *asm_file; + register HOST_WIDE_INT local_bytes; +{ + register HOST_WIDE_INT frame_upper_bytes; + register HOST_WIDE_INT frame_lower_bytes; + register HOST_WIDE_INT preserved_reg_bytes = 0; + register unsigned i; + register unsigned restored_so_far = 0; + register unsigned int_restored; + register unsigned mask; + unsigned intflags=0; + register TDESC_flags *flags = (TDESC_flags *) &intflags; +#ifdef OUTPUT_TDESC /* Output an ABI-compliant TDESC entry */ + const char *long_op = integer_asm_op (4, TRUE); +#endif + + flags->version = 4; + flags->reg_packing = 1; + flags->iregs = 8; /* old fp always gets saved */ + + /* Round-up the frame_lower_bytes so that it's a multiple of 16. */ + + frame_lower_bytes = (local_bytes + STACK_ALIGNMENT - 1) & -STACK_ALIGNMENT; + + /* Count the number of registers that were preserved in the prologue. + Ignore r0. It is never preserved. */ + + for (i = 1; i < FIRST_PSEUDO_REGISTER; i++) + { + if (regs_ever_live[i] && ! call_used_regs[i]) + preserved_reg_bytes += 4; + } + + /* The upper part of each frame will contain only saved fp, + the saved r1, and stack slots for all of the other "preserved" + registers that we find we will need to save & restore. */ + + frame_upper_bytes = must_preserve_bytes + preserved_reg_bytes; + + /* Round-up frame_upper_bytes so that t is a multiple of 16. */ + + frame_upper_bytes + = (frame_upper_bytes + STACK_ALIGNMENT - 1) & -STACK_ALIGNMENT; + + /* Restore all of the "preserved" registers that need restoring. */ + + mask = 2; + + for (i = 1; i < 32; i++, mask<<=1) + if (regs_ever_live[i] && ! call_used_regs[i]) { + fprintf (asm_file, "\tld.l %d(%sfp),%s%s\n", + must_preserve_bytes + (4 * restored_so_far++), + i860_reg_prefix, i860_reg_prefix, reg_names[i]); + if (i > 3 && i < 16) + flags->iregs |= mask; + } + + int_restored = restored_so_far; + mask = 1; + + for (i = 32; i < 64; i++) { + if (regs_ever_live[i] && ! call_used_regs[i]) { + fprintf (asm_file, "\tfld.l %d(%sfp),%s%s\n", + must_preserve_bytes + (4 * restored_so_far++), + i860_reg_prefix, i860_reg_prefix, reg_names[i]); + if (i > 33 && i < 40) + flags->fregs |= mask; + } + if (i > 33 && i < 40) + mask<<=1; + } + + /* Get the value we plan to use to restore the stack pointer into r31. */ + + fprintf (asm_file, "\tadds %d,%sfp,%sr31\n", + frame_upper_bytes, i860_reg_prefix, i860_reg_prefix); + + /* Restore the return address and the old frame pointer. */ + + if (must_preserve_r1) { + fprintf (asm_file, "\tld.l 4(%sfp),%sr1\n", + i860_reg_prefix, i860_reg_prefix); + flags->iregs |= 2; + } + + fprintf (asm_file, "\tld.l 0(%sfp),%sfp\n", + i860_reg_prefix, i860_reg_prefix); + + /* Return and restore the old stack pointer value. */ + + fprintf (asm_file, "\tbri %sr1\n\tmov %sr31,%ssp\n", + i860_reg_prefix, i860_reg_prefix, i860_reg_prefix); + +#ifdef OUTPUT_TDESC /* Output an ABI-compliant TDESC entry */ + if (! frame_lower_bytes) { + flags->version--; + if (! frame_upper_bytes) { + flags->version--; + if (restored_so_far == int_restored) /* No FP saves */ + flags->version--; + } + } + assemble_name(asm_file,current_function_original_name); + fputs(".TDESC:\n", asm_file); + fprintf(asm_file, "%s 0x%0x\n", long_op, intflags); + fprintf(asm_file, "%s %d\n", long_op, + int_restored ? must_preserve_bytes : 0); + if (flags->version > 1) { + fprintf(asm_file, "%s %d\n", long_op, + (restored_so_far == int_restored) ? 0 : must_preserve_bytes + + (4 * int_restored)); + if (flags->version > 2) { + fprintf(asm_file, "%s %d\n", long_op, frame_upper_bytes); + if (flags->version > 3) + fprintf(asm_file, "%s %d\n", long_op, frame_lower_bytes); + } + } + tdesc_section(); + fprintf(asm_file, "%s ", long_op); + assemble_name(asm_file, current_function_original_name); + fprintf(asm_file, "\n%s ", long_op); + assemble_name(asm_file, current_function_original_name); + fputs(".TDESC\n", asm_file); + text_section(); +#endif +} + + +/* Expand a library call to __builtin_saveregs. */ +rtx +i860_saveregs () +{ + rtx fn = gen_rtx_SYMBOL_REF (Pmode, "__builtin_saveregs"); + rtx save = gen_reg_rtx (Pmode); + rtx valreg = LIBCALL_VALUE (Pmode); + rtx ret; + + /* The return value register overlaps the first argument register. + Save and restore it around the call. */ + emit_move_insn (save, valreg); + ret = emit_library_call_value (fn, NULL_RTX, 1, Pmode, 0); + if (GET_CODE (ret) != REG || REGNO (ret) < FIRST_PSEUDO_REGISTER) + ret = copy_to_reg (ret); + emit_move_insn (valreg, save); + + return ret; +} + +tree +i860_build_va_list () +{ + tree field_ireg_used, field_freg_used, field_reg_base, field_mem_ptr; + tree record; + + record = make_node (RECORD_TYPE); + + field_ireg_used = build_decl (FIELD_DECL, get_identifier ("__ireg_used"), + unsigned_type_node); + field_freg_used = build_decl (FIELD_DECL, get_identifier ("__freg_used"), + unsigned_type_node); + field_reg_base = build_decl (FIELD_DECL, get_identifier ("__reg_base"), + ptr_type_node); + field_mem_ptr = build_decl (FIELD_DECL, get_identifier ("__mem_ptr"), + ptr_type_node); + + DECL_FIELD_CONTEXT (field_ireg_used) = record; + DECL_FIELD_CONTEXT (field_freg_used) = record; + DECL_FIELD_CONTEXT (field_reg_base) = record; + DECL_FIELD_CONTEXT (field_mem_ptr) = record; + +#ifdef I860_SVR4_VA_LIST + TYPE_FIELDS (record) = field_ireg_used; + TREE_CHAIN (field_ireg_used) = field_freg_used; + TREE_CHAIN (field_freg_used) = field_reg_base; + TREE_CHAIN (field_reg_base) = field_mem_ptr; +#else + TYPE_FIELDS (record) = field_reg_base; + TREE_CHAIN (field_reg_base) = field_mem_ptr; + TREE_CHAIN (field_mem_ptr) = field_ireg_used; + TREE_CHAIN (field_ireg_used) = field_freg_used; +#endif + + layout_type (record); + return record; +} + +void +i860_va_start (stdarg_p, valist, nextarg) + int stdarg_p; + tree valist; + rtx nextarg; +{ + tree saveregs, t; + + saveregs = make_tree (build_pointer_type (va_list_type_node), + expand_builtin_saveregs ()); + saveregs = build1 (INDIRECT_REF, va_list_type_node, saveregs); + + if (stdarg_p) + { + tree field_ireg_used, field_freg_used, field_reg_base, field_mem_ptr; + tree ireg_used, freg_used, reg_base, mem_ptr; + +#ifdef I860_SVR4_VA_LIST + field_ireg_used = TYPE_FIELDS (va_list_type_node); + field_freg_used = TREE_CHAIN (field_ireg_used); + field_reg_base = TREE_CHAIN (field_freg_used); + field_mem_ptr = TREE_CHAIN (field_reg_base); +#else + field_reg_base = TYPE_FIELDS (va_list_type_node); + field_mem_ptr = TREE_CHAIN (field_reg_base); + field_ireg_used = TREE_CHAIN (field_mem_ptr); + field_freg_used = TREE_CHAIN (field_ireg_used); +#endif + + ireg_used = build (COMPONENT_REF, TREE_TYPE (field_ireg_used), + valist, field_ireg_used); + freg_used = build (COMPONENT_REF, TREE_TYPE (field_freg_used), + valist, field_freg_used); + reg_base = build (COMPONENT_REF, TREE_TYPE (field_reg_base), + valist, field_reg_base); + mem_ptr = build (COMPONENT_REF, TREE_TYPE (field_mem_ptr), + valist, field_mem_ptr); + + t = build_int_2 (current_function_args_info.ints, 0); + t = build (MODIFY_EXPR, TREE_TYPE (ireg_used), ireg_used, t); + TREE_SIDE_EFFECTS (t) = 1; + expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); + + t = build_int_2 (ROUNDUP (current_function_args_info.floats, 8), 0); + t = build (MODIFY_EXPR, TREE_TYPE (freg_used), freg_used, t); + TREE_SIDE_EFFECTS (t) = 1; + expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); + + t = build (COMPONENT_REF, TREE_TYPE (field_reg_base), + saveregs, field_reg_base); + t = build (MODIFY_EXPR, TREE_TYPE (reg_base), reg_base, t); + TREE_SIDE_EFFECTS (t) = 1; + expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); + + t = make_tree (ptr_type_node, nextarg); + t = build (MODIFY_EXPR, TREE_TYPE (mem_ptr), mem_ptr, t); + TREE_SIDE_EFFECTS (t) = 1; + expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); + } + else + { + t = build (MODIFY_EXPR, va_list_type_node, valist, saveregs); + TREE_SIDE_EFFECTS (t) = 1; + expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); + } +} + +#define NUM_PARM_FREGS 8 +#define NUM_PARM_IREGS 12 +#ifdef I860_SVR4_VARARGS +#define FREG_OFFSET 0 +#define IREG_OFFSET (NUM_PARM_FREGS * UNITS_PER_WORD) +#else +#define FREG_OFFSET (NUM_PARM_IREGS * UNITS_PER_WORD) +#define IREG_OFFSET 0 +#endif + +rtx +i860_va_arg (valist, type) + tree valist, type; +{ + tree field_ireg_used, field_freg_used, field_reg_base, field_mem_ptr; + tree type_ptr_node, t; + rtx lab_over = NULL_RTX; + rtx ret, val; + HOST_WIDE_INT align; + +#ifdef I860_SVR4_VA_LIST + field_ireg_used = TYPE_FIELDS (va_list_type_node); + field_freg_used = TREE_CHAIN (field_ireg_used); + field_reg_base = TREE_CHAIN (field_freg_used); + field_mem_ptr = TREE_CHAIN (field_reg_base); +#else + field_reg_base = TYPE_FIELDS (va_list_type_node); + field_mem_ptr = TREE_CHAIN (field_reg_base); + field_ireg_used = TREE_CHAIN (field_mem_ptr); + field_freg_used = TREE_CHAIN (field_ireg_used); +#endif + + field_ireg_used = build (COMPONENT_REF, TREE_TYPE (field_ireg_used), + valist, field_ireg_used); + field_freg_used = build (COMPONENT_REF, TREE_TYPE (field_freg_used), + valist, field_freg_used); + field_reg_base = build (COMPONENT_REF, TREE_TYPE (field_reg_base), + valist, field_reg_base); + field_mem_ptr = build (COMPONENT_REF, TREE_TYPE (field_mem_ptr), + valist, field_mem_ptr); + + ret = gen_reg_rtx (Pmode); + type_ptr_node = build_pointer_type (type); + + if (! AGGREGATE_TYPE_P (type)) + { + int nparm, incr, ofs; + tree field; + rtx lab_false; + + if (FLOAT_TYPE_P (type)) + { + field = field_freg_used; + nparm = NUM_PARM_FREGS; + incr = 2; + ofs = FREG_OFFSET; + } + else + { + field = field_ireg_used; + nparm = NUM_PARM_IREGS; + incr = int_size_in_bytes (type) / UNITS_PER_WORD; + ofs = IREG_OFFSET; + } + + lab_false = gen_label_rtx (); + lab_over = gen_label_rtx (); + + emit_cmp_and_jump_insns (expand_expr (field, NULL_RTX, 0, 0), + GEN_INT (nparm - incr), GT, const0_rtx, + TYPE_MODE (TREE_TYPE (field)), + TREE_UNSIGNED (field), lab_false); + + t = fold (build (POSTINCREMENT_EXPR, TREE_TYPE (field), field, + build_int_2 (incr, 0))); + TREE_SIDE_EFFECTS (t) = 1; + + t = fold (build (MULT_EXPR, TREE_TYPE (field), field, + build_int_2 (UNITS_PER_WORD, 0))); + TREE_SIDE_EFFECTS (t) = 1; + + t = fold (build (PLUS_EXPR, ptr_type_node, field_reg_base, + fold (build (PLUS_EXPR, TREE_TYPE (field), t, + build_int_2 (ofs, 0))))); + TREE_SIDE_EFFECTS (t) = 1; + + val = expand_expr (t, ret, VOIDmode, EXPAND_NORMAL); + if (val != ret) + emit_move_insn (ret, val); + + emit_jump_insn (gen_jump (lab_over)); + emit_barrier (); + emit_label (lab_false); + } + + align = TYPE_ALIGN (type); + if (align < BITS_PER_WORD) + align = BITS_PER_WORD; + align /= BITS_PER_UNIT; + + t = build (PLUS_EXPR, ptr_type_node, field_mem_ptr, + build_int_2 (align - 1, 0)); + t = build (BIT_AND_EXPR, ptr_type_node, t, build_int_2 (-align, -1)); + + val = expand_expr (t, ret, VOIDmode, EXPAND_NORMAL); + if (val != ret) + emit_move_insn (ret, val); + + t = fold (build (PLUS_EXPR, ptr_type_node, + make_tree (ptr_type_node, ret), + build_int_2 (int_size_in_bytes (type), 0))); + t = build (MODIFY_EXPR, ptr_type_node, field_mem_ptr, t); + TREE_SIDE_EFFECTS (t) = 1; + expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); + + if (lab_over) + emit_label (lab_over); + + return ret; +} |