2002-02-01 18:16:02 +00:00
|
|
|
|
/* Expand builtin functions.
|
|
|
|
|
Copyright (C) 1988, 1992, 1993, 1994, 1995, 1996, 1997, 1998,
|
|
|
|
|
1999, 2000, 2001, 2002 Free Software Foundation, Inc.
|
|
|
|
|
|
|
|
|
|
This file is part of GCC.
|
|
|
|
|
|
|
|
|
|
GCC is free software; you can redistribute it and/or modify it under
|
|
|
|
|
the terms of the GNU General Public License as published by the Free
|
|
|
|
|
Software Foundation; either version 2, or (at your option) any later
|
|
|
|
|
version.
|
|
|
|
|
|
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
|
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
|
|
|
for more details.
|
|
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
|
along with GCC; see the file COPYING. If not, write to the Free
|
|
|
|
|
Software Foundation, 59 Temple Place - Suite 330, Boston, MA
|
|
|
|
|
02111-1307, USA. */
|
|
|
|
|
|
Gcc 3.1 (-O) now generates broken inline code for memset in some cases.
This broke newfs (newfs left some garbage in a bitmap).
The ASM for:
#include <string.h>
int x, foo[100];
main()
{
memset(&foo[0], 0, x);
}
is (at least if you have fixed function alignment):
.file "z.c"
.text
.p2align 2,,3
.globl main
.type main,@function
main:
pushl %ebp
movl %esp, %ebp
pushl %edi
pushl %eax
movl x, %ecx
xorl %eax, %eax
shrl $2, %ecx
movl $foo, %edi
cld
rep
stosl
andl $-16, %esp
<-- the lower bits of `len' should be loaded
near here
testl $2, %edi <-- this seems to be meant to test the 2^1
bit in `len' (not alignment of the pointer
like it actually does). %edi is the wrong
register for holding the bits, since it is
still needed for the pointer.
je .L2
stosw
.L2:
testl $1, %edi <-- similarly for the 2^0 bit.
je .L3
stosb
.L3:
movl -4(%ebp), %edi
leave
ret
.Lfe1:
.size main,.Lfe1-main
.comm foo,400,32
.comm x,4,4
.ident "GCC: (GNU) 3.1 [FreeBSD] 20020509 (prerelease)"
This seems to only result in (len % 3) bytes not being cleared, since gcc
doesn't seem to use the builtin memset unless it knows that the pointer is
aligned. If %edi could be misaligned, then too many bytes would be set.
Submitted by: BDE
2002-06-04 18:04:27 +00:00
|
|
|
|
/* $FreeBSD$ */
|
|
|
|
|
|
2002-02-01 18:16:02 +00:00
|
|
|
|
#include "config.h"
|
|
|
|
|
#include "system.h"
|
|
|
|
|
#include "machmode.h"
|
|
|
|
|
#include "rtl.h"
|
|
|
|
|
#include "tree.h"
|
|
|
|
|
#include "obstack.h"
|
|
|
|
|
#include "flags.h"
|
|
|
|
|
#include "regs.h"
|
|
|
|
|
#include "hard-reg-set.h"
|
|
|
|
|
#include "except.h"
|
|
|
|
|
#include "function.h"
|
|
|
|
|
#include "insn-config.h"
|
|
|
|
|
#include "expr.h"
|
|
|
|
|
#include "optabs.h"
|
|
|
|
|
#include "libfuncs.h"
|
|
|
|
|
#include "recog.h"
|
|
|
|
|
#include "output.h"
|
|
|
|
|
#include "typeclass.h"
|
|
|
|
|
#include "toplev.h"
|
|
|
|
|
#include "predict.h"
|
|
|
|
|
#include "tm_p.h"
|
|
|
|
|
#include "target.h"
|
|
|
|
|
|
|
|
|
|
#define CALLED_AS_BUILT_IN(NODE) \
|
|
|
|
|
(!strncmp (IDENTIFIER_POINTER (DECL_NAME (NODE)), "__builtin_", 10))
|
|
|
|
|
|
|
|
|
|
/* Register mappings for target machines without register windows. */
|
|
|
|
|
#ifndef INCOMING_REGNO
|
|
|
|
|
#define INCOMING_REGNO(OUT) (OUT)
|
|
|
|
|
#endif
|
|
|
|
|
#ifndef OUTGOING_REGNO
|
|
|
|
|
#define OUTGOING_REGNO(IN) (IN)
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#ifndef PAD_VARARGS_DOWN
|
|
|
|
|
#define PAD_VARARGS_DOWN BYTES_BIG_ENDIAN
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Define the names of the builtin function types and codes. */
|
|
|
|
|
const char *const built_in_class_names[4]
|
|
|
|
|
= {"NOT_BUILT_IN", "BUILT_IN_FRONTEND", "BUILT_IN_MD", "BUILT_IN_NORMAL"};
|
|
|
|
|
|
|
|
|
|
#define DEF_BUILTIN(X, N, C, T, LT, B, F, NA) STRINGX(X),
|
|
|
|
|
const char *const built_in_names[(int) END_BUILTINS] =
|
|
|
|
|
{
|
|
|
|
|
#include "builtins.def"
|
|
|
|
|
};
|
|
|
|
|
#undef DEF_BUILTIN
|
|
|
|
|
|
|
|
|
|
/* Setup an array of _DECL trees, make sure each element is
|
|
|
|
|
initialized to NULL_TREE. */
|
|
|
|
|
tree built_in_decls[(int) END_BUILTINS];
|
|
|
|
|
|
|
|
|
|
tree (*lang_type_promotes_to) PARAMS ((tree));
|
|
|
|
|
|
|
|
|
|
static int get_pointer_alignment PARAMS ((tree, unsigned int));
|
|
|
|
|
static tree c_strlen PARAMS ((tree));
|
|
|
|
|
static const char *c_getstr PARAMS ((tree));
|
|
|
|
|
static rtx c_readstr PARAMS ((const char *,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static int target_char_cast PARAMS ((tree, char *));
|
|
|
|
|
static rtx get_memory_rtx PARAMS ((tree));
|
|
|
|
|
static int apply_args_size PARAMS ((void));
|
|
|
|
|
static int apply_result_size PARAMS ((void));
|
|
|
|
|
#if defined (HAVE_untyped_call) || defined (HAVE_untyped_return)
|
|
|
|
|
static rtx result_vector PARAMS ((int, rtx));
|
|
|
|
|
#endif
|
|
|
|
|
static rtx expand_builtin_setjmp PARAMS ((tree, rtx));
|
|
|
|
|
static void expand_builtin_prefetch PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_apply_args PARAMS ((void));
|
|
|
|
|
static rtx expand_builtin_apply_args_1 PARAMS ((void));
|
|
|
|
|
static rtx expand_builtin_apply PARAMS ((rtx, rtx, rtx));
|
|
|
|
|
static void expand_builtin_return PARAMS ((rtx));
|
|
|
|
|
static enum type_class type_to_class PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_classify_type PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_mathfn PARAMS ((tree, rtx, rtx));
|
|
|
|
|
static rtx expand_builtin_constant_p PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_args_info PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_next_arg PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_va_start PARAMS ((int, tree));
|
|
|
|
|
static rtx expand_builtin_va_end PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_va_copy PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_memcmp PARAMS ((tree, tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strcmp PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strncmp PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx builtin_memcpy_read_str PARAMS ((PTR, HOST_WIDE_INT,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strcat PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strncat PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strspn PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strcspn PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_memcpy PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strcpy PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx builtin_strncpy_read_str PARAMS ((PTR, HOST_WIDE_INT,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strncpy PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx builtin_memset_read_str PARAMS ((PTR, HOST_WIDE_INT,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_memset PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_bzero PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_strlen PARAMS ((tree, rtx));
|
|
|
|
|
static rtx expand_builtin_strstr PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strpbrk PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strchr PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_strrchr PARAMS ((tree, rtx,
|
|
|
|
|
enum machine_mode));
|
|
|
|
|
static rtx expand_builtin_alloca PARAMS ((tree, rtx));
|
|
|
|
|
static rtx expand_builtin_ffs PARAMS ((tree, rtx, rtx));
|
|
|
|
|
static rtx expand_builtin_frame_address PARAMS ((tree));
|
|
|
|
|
static rtx expand_builtin_fputs PARAMS ((tree, int, int));
|
|
|
|
|
static tree stabilize_va_list PARAMS ((tree, int));
|
|
|
|
|
static rtx expand_builtin_expect PARAMS ((tree, rtx));
|
|
|
|
|
static tree fold_builtin_constant_p PARAMS ((tree));
|
|
|
|
|
static tree fold_builtin_classify_type PARAMS ((tree));
|
|
|
|
|
static tree build_function_call_expr PARAMS ((tree, tree));
|
|
|
|
|
static int validate_arglist PARAMS ((tree, ...));
|
|
|
|
|
|
|
|
|
|
/* Return the alignment in bits of EXP, a pointer valued expression.
|
|
|
|
|
But don't return more than MAX_ALIGN no matter what.
|
|
|
|
|
The alignment returned is, by default, the alignment of the thing that
|
|
|
|
|
EXP points to. If it is not a POINTER_TYPE, 0 is returned.
|
|
|
|
|
|
|
|
|
|
Otherwise, look at the expression to see if we can do better, i.e., if the
|
|
|
|
|
expression is actually pointing at an object whose alignment is tighter. */
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
|
get_pointer_alignment (exp, max_align)
|
|
|
|
|
tree exp;
|
|
|
|
|
unsigned int max_align;
|
|
|
|
|
{
|
|
|
|
|
unsigned int align, inner;
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (TREE_TYPE (exp)) != POINTER_TYPE)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
align = TYPE_ALIGN (TREE_TYPE (TREE_TYPE (exp)));
|
|
|
|
|
align = MIN (align, max_align);
|
|
|
|
|
|
|
|
|
|
while (1)
|
|
|
|
|
{
|
|
|
|
|
switch (TREE_CODE (exp))
|
|
|
|
|
{
|
|
|
|
|
case NOP_EXPR:
|
|
|
|
|
case CONVERT_EXPR:
|
|
|
|
|
case NON_LVALUE_EXPR:
|
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
|
|
|
if (TREE_CODE (TREE_TYPE (exp)) != POINTER_TYPE)
|
|
|
|
|
return align;
|
|
|
|
|
|
|
|
|
|
inner = TYPE_ALIGN (TREE_TYPE (TREE_TYPE (exp)));
|
|
|
|
|
align = MIN (inner, max_align);
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case PLUS_EXPR:
|
|
|
|
|
/* If sum of pointer + int, restrict our maximum alignment to that
|
|
|
|
|
imposed by the integer. If not, we can't do any better than
|
|
|
|
|
ALIGN. */
|
|
|
|
|
if (! host_integerp (TREE_OPERAND (exp, 1), 1))
|
|
|
|
|
return align;
|
|
|
|
|
|
|
|
|
|
while (((tree_low_cst (TREE_OPERAND (exp, 1), 1))
|
|
|
|
|
& (max_align / BITS_PER_UNIT - 1))
|
|
|
|
|
!= 0)
|
|
|
|
|
max_align >>= 1;
|
|
|
|
|
|
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case ADDR_EXPR:
|
|
|
|
|
/* See what we are pointing at and look at its alignment. */
|
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
|
|
|
if (TREE_CODE (exp) == FUNCTION_DECL)
|
|
|
|
|
align = FUNCTION_BOUNDARY;
|
|
|
|
|
else if (DECL_P (exp))
|
|
|
|
|
align = DECL_ALIGN (exp);
|
|
|
|
|
#ifdef CONSTANT_ALIGNMENT
|
|
|
|
|
else if (TREE_CODE_CLASS (TREE_CODE (exp)) == 'c')
|
|
|
|
|
align = CONSTANT_ALIGNMENT (exp, align);
|
|
|
|
|
#endif
|
|
|
|
|
return MIN (align, max_align);
|
|
|
|
|
|
|
|
|
|
default:
|
|
|
|
|
return align;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Compute the length of a C string. TREE_STRING_LENGTH is not the right
|
|
|
|
|
way, because it could contain a zero byte in the middle.
|
|
|
|
|
TREE_STRING_LENGTH is the size of the character array, not the string.
|
|
|
|
|
|
|
|
|
|
The value returned is of type `ssizetype'.
|
|
|
|
|
|
|
|
|
|
Unfortunately, string_constant can't access the values of const char
|
|
|
|
|
arrays with initializers, so neither can we do so here. */
|
|
|
|
|
|
|
|
|
|
static tree
|
|
|
|
|
c_strlen (src)
|
|
|
|
|
tree src;
|
|
|
|
|
{
|
|
|
|
|
tree offset_node;
|
|
|
|
|
HOST_WIDE_INT offset;
|
|
|
|
|
int max;
|
|
|
|
|
const char *ptr;
|
|
|
|
|
|
|
|
|
|
src = string_constant (src, &offset_node);
|
|
|
|
|
if (src == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
max = TREE_STRING_LENGTH (src) - 1;
|
|
|
|
|
ptr = TREE_STRING_POINTER (src);
|
|
|
|
|
|
|
|
|
|
if (offset_node && TREE_CODE (offset_node) != INTEGER_CST)
|
|
|
|
|
{
|
|
|
|
|
/* If the string has an internal zero byte (e.g., "foo\0bar"), we can't
|
|
|
|
|
compute the offset to the following null if we don't know where to
|
|
|
|
|
start searching for it. */
|
|
|
|
|
int i;
|
|
|
|
|
|
|
|
|
|
for (i = 0; i < max; i++)
|
|
|
|
|
if (ptr[i] == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* We don't know the starting offset, but we do know that the string
|
|
|
|
|
has no internal zero bytes. We can assume that the offset falls
|
|
|
|
|
within the bounds of the string; otherwise, the programmer deserves
|
|
|
|
|
what he gets. Subtract the offset from the length of the string,
|
|
|
|
|
and return that. This would perhaps not be valid if we were dealing
|
|
|
|
|
with named arrays in addition to literal string constants. */
|
|
|
|
|
|
|
|
|
|
return size_diffop (size_int (max), offset_node);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* We have a known offset into the string. Start searching there for
|
|
|
|
|
a null character if we can represent it as a single HOST_WIDE_INT. */
|
|
|
|
|
if (offset_node == 0)
|
|
|
|
|
offset = 0;
|
|
|
|
|
else if (! host_integerp (offset_node, 0))
|
|
|
|
|
offset = -1;
|
|
|
|
|
else
|
|
|
|
|
offset = tree_low_cst (offset_node, 0);
|
|
|
|
|
|
|
|
|
|
/* If the offset is known to be out of bounds, warn, and call strlen at
|
|
|
|
|
runtime. */
|
|
|
|
|
if (offset < 0 || offset > max)
|
|
|
|
|
{
|
|
|
|
|
warning ("offset outside bounds of constant string");
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Use strlen to search for the first zero byte. Since any strings
|
|
|
|
|
constructed with build_string will have nulls appended, we win even
|
|
|
|
|
if we get handed something like (char[4])"abcd".
|
|
|
|
|
|
|
|
|
|
Since OFFSET is our starting index into the string, no further
|
|
|
|
|
calculation is needed. */
|
|
|
|
|
return ssize_int (strlen (ptr + offset));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Return a char pointer for a C string if it is a string constant
|
|
|
|
|
or sum of string constant and integer constant. */
|
|
|
|
|
|
|
|
|
|
static const char *
|
|
|
|
|
c_getstr (src)
|
|
|
|
|
tree src;
|
|
|
|
|
{
|
|
|
|
|
tree offset_node;
|
|
|
|
|
|
|
|
|
|
src = string_constant (src, &offset_node);
|
|
|
|
|
if (src == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
if (offset_node == 0)
|
|
|
|
|
return TREE_STRING_POINTER (src);
|
|
|
|
|
else if (!host_integerp (offset_node, 1)
|
|
|
|
|
|| compare_tree_int (offset_node, TREE_STRING_LENGTH (src) - 1) > 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
return TREE_STRING_POINTER (src) + tree_low_cst (offset_node, 1);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Return a CONST_INT or CONST_DOUBLE corresponding to target reading
|
|
|
|
|
GET_MODE_BITSIZE (MODE) bits from string constant STR. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
c_readstr (str, mode)
|
|
|
|
|
const char *str;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
HOST_WIDE_INT c[2];
|
|
|
|
|
HOST_WIDE_INT ch;
|
|
|
|
|
unsigned int i, j;
|
|
|
|
|
|
|
|
|
|
if (GET_MODE_CLASS (mode) != MODE_INT)
|
|
|
|
|
abort ();
|
|
|
|
|
c[0] = 0;
|
|
|
|
|
c[1] = 0;
|
|
|
|
|
ch = 1;
|
|
|
|
|
for (i = 0; i < GET_MODE_SIZE (mode); i++)
|
|
|
|
|
{
|
|
|
|
|
j = i;
|
|
|
|
|
if (WORDS_BIG_ENDIAN)
|
|
|
|
|
j = GET_MODE_SIZE (mode) - i - 1;
|
|
|
|
|
if (BYTES_BIG_ENDIAN != WORDS_BIG_ENDIAN
|
|
|
|
|
&& GET_MODE_SIZE (mode) > UNITS_PER_WORD)
|
|
|
|
|
j = j + UNITS_PER_WORD - 2 * (j % UNITS_PER_WORD) - 1;
|
|
|
|
|
j *= BITS_PER_UNIT;
|
|
|
|
|
if (j > 2 * HOST_BITS_PER_WIDE_INT)
|
|
|
|
|
abort ();
|
|
|
|
|
if (ch)
|
|
|
|
|
ch = (unsigned char) str[i];
|
|
|
|
|
c[j / HOST_BITS_PER_WIDE_INT] |= ch << (j % HOST_BITS_PER_WIDE_INT);
|
|
|
|
|
}
|
|
|
|
|
return immed_double_const (c[0], c[1], mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Cast a target constant CST to target CHAR and if that value fits into
|
|
|
|
|
host char type, return zero and put that value into variable pointed by
|
|
|
|
|
P. */
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
|
target_char_cast (cst, p)
|
|
|
|
|
tree cst;
|
|
|
|
|
char *p;
|
|
|
|
|
{
|
|
|
|
|
unsigned HOST_WIDE_INT val, hostval;
|
|
|
|
|
|
|
|
|
|
if (!host_integerp (cst, 1)
|
|
|
|
|
|| CHAR_TYPE_SIZE > HOST_BITS_PER_WIDE_INT)
|
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
|
|
val = tree_low_cst (cst, 1);
|
|
|
|
|
if (CHAR_TYPE_SIZE < HOST_BITS_PER_WIDE_INT)
|
|
|
|
|
val &= (((unsigned HOST_WIDE_INT) 1) << CHAR_TYPE_SIZE) - 1;
|
|
|
|
|
|
|
|
|
|
hostval = val;
|
|
|
|
|
if (HOST_BITS_PER_CHAR < HOST_BITS_PER_WIDE_INT)
|
|
|
|
|
hostval &= (((unsigned HOST_WIDE_INT) 1) << HOST_BITS_PER_CHAR) - 1;
|
|
|
|
|
|
|
|
|
|
if (val != hostval)
|
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
|
|
*p = hostval;
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Given TEM, a pointer to a stack frame, follow the dynamic chain COUNT
|
|
|
|
|
times to get the address of either a higher stack frame, or a return
|
|
|
|
|
address located within it (depending on FNDECL_CODE). */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
expand_builtin_return_addr (fndecl_code, count, tem)
|
|
|
|
|
enum built_in_function fndecl_code;
|
|
|
|
|
int count;
|
|
|
|
|
rtx tem;
|
|
|
|
|
{
|
|
|
|
|
int i;
|
|
|
|
|
|
|
|
|
|
/* Some machines need special handling before we can access
|
|
|
|
|
arbitrary frames. For example, on the sparc, we must first flush
|
|
|
|
|
all register windows to the stack. */
|
|
|
|
|
#ifdef SETUP_FRAME_ADDRESSES
|
|
|
|
|
if (count > 0)
|
|
|
|
|
SETUP_FRAME_ADDRESSES ();
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* On the sparc, the return address is not in the frame, it is in a
|
|
|
|
|
register. There is no way to access it off of the current frame
|
|
|
|
|
pointer, but it can be accessed off the previous frame pointer by
|
|
|
|
|
reading the value from the register window save area. */
|
|
|
|
|
#ifdef RETURN_ADDR_IN_PREVIOUS_FRAME
|
|
|
|
|
if (fndecl_code == BUILT_IN_RETURN_ADDRESS)
|
|
|
|
|
count--;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Scan back COUNT frames to the specified frame. */
|
|
|
|
|
for (i = 0; i < count; i++)
|
|
|
|
|
{
|
|
|
|
|
/* Assume the dynamic chain pointer is in the word that the
|
|
|
|
|
frame address points to, unless otherwise specified. */
|
|
|
|
|
#ifdef DYNAMIC_CHAIN_ADDRESS
|
|
|
|
|
tem = DYNAMIC_CHAIN_ADDRESS (tem);
|
|
|
|
|
#endif
|
|
|
|
|
tem = memory_address (Pmode, tem);
|
|
|
|
|
tem = gen_rtx_MEM (Pmode, tem);
|
|
|
|
|
set_mem_alias_set (tem, get_frame_alias_set ());
|
|
|
|
|
tem = copy_to_reg (tem);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* For __builtin_frame_address, return what we've got. */
|
|
|
|
|
if (fndecl_code == BUILT_IN_FRAME_ADDRESS)
|
|
|
|
|
return tem;
|
|
|
|
|
|
|
|
|
|
/* For __builtin_return_address, Get the return address from that
|
|
|
|
|
frame. */
|
|
|
|
|
#ifdef RETURN_ADDR_RTX
|
|
|
|
|
tem = RETURN_ADDR_RTX (count, tem);
|
|
|
|
|
#else
|
|
|
|
|
tem = memory_address (Pmode,
|
|
|
|
|
plus_constant (tem, GET_MODE_SIZE (Pmode)));
|
|
|
|
|
tem = gen_rtx_MEM (Pmode, tem);
|
|
|
|
|
set_mem_alias_set (tem, get_frame_alias_set ());
|
|
|
|
|
#endif
|
|
|
|
|
return tem;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Alias set used for setjmp buffer. */
|
|
|
|
|
static HOST_WIDE_INT setjmp_alias_set = -1;
|
|
|
|
|
|
|
|
|
|
/* Construct the leading half of a __builtin_setjmp call. Control will
|
|
|
|
|
return to RECEIVER_LABEL. This is used directly by sjlj exception
|
|
|
|
|
handling code. */
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
expand_builtin_setjmp_setup (buf_addr, receiver_label)
|
|
|
|
|
rtx buf_addr;
|
|
|
|
|
rtx receiver_label;
|
|
|
|
|
{
|
|
|
|
|
enum machine_mode sa_mode = STACK_SAVEAREA_MODE (SAVE_NONLOCAL);
|
|
|
|
|
rtx stack_save;
|
|
|
|
|
rtx mem;
|
|
|
|
|
|
|
|
|
|
if (setjmp_alias_set == -1)
|
|
|
|
|
setjmp_alias_set = new_alias_set ();
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (buf_addr) != Pmode)
|
|
|
|
|
buf_addr = convert_memory_address (Pmode, buf_addr);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
buf_addr = force_reg (Pmode, force_operand (buf_addr, NULL_RTX));
|
|
|
|
|
|
|
|
|
|
emit_queue ();
|
|
|
|
|
|
|
|
|
|
/* We store the frame pointer and the address of receiver_label in
|
|
|
|
|
the buffer and use the rest of it for the stack save area, which
|
|
|
|
|
is machine-dependent. */
|
|
|
|
|
|
|
|
|
|
#ifndef BUILTIN_SETJMP_FRAME_VALUE
|
|
|
|
|
#define BUILTIN_SETJMP_FRAME_VALUE virtual_stack_vars_rtx
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
mem = gen_rtx_MEM (Pmode, buf_addr);
|
|
|
|
|
set_mem_alias_set (mem, setjmp_alias_set);
|
|
|
|
|
emit_move_insn (mem, BUILTIN_SETJMP_FRAME_VALUE);
|
|
|
|
|
|
|
|
|
|
mem = gen_rtx_MEM (Pmode, plus_constant (buf_addr, GET_MODE_SIZE (Pmode))),
|
|
|
|
|
set_mem_alias_set (mem, setjmp_alias_set);
|
|
|
|
|
|
|
|
|
|
emit_move_insn (validize_mem (mem),
|
|
|
|
|
force_reg (Pmode, gen_rtx_LABEL_REF (Pmode, receiver_label)));
|
|
|
|
|
|
|
|
|
|
stack_save = gen_rtx_MEM (sa_mode,
|
|
|
|
|
plus_constant (buf_addr,
|
|
|
|
|
2 * GET_MODE_SIZE (Pmode)));
|
|
|
|
|
set_mem_alias_set (stack_save, setjmp_alias_set);
|
|
|
|
|
emit_stack_save (SAVE_NONLOCAL, &stack_save, NULL_RTX);
|
|
|
|
|
|
|
|
|
|
/* If there is further processing to do, do it. */
|
|
|
|
|
#ifdef HAVE_builtin_setjmp_setup
|
|
|
|
|
if (HAVE_builtin_setjmp_setup)
|
|
|
|
|
emit_insn (gen_builtin_setjmp_setup (buf_addr));
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Tell optimize_save_area_alloca that extra work is going to
|
|
|
|
|
need to go on during alloca. */
|
|
|
|
|
current_function_calls_setjmp = 1;
|
|
|
|
|
|
|
|
|
|
/* Set this so all the registers get saved in our frame; we need to be
|
|
|
|
|
able to copy the saved values for any registers from frames we unwind. */
|
|
|
|
|
current_function_has_nonlocal_label = 1;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Construct the trailing part of a __builtin_setjmp call.
|
|
|
|
|
This is used directly by sjlj exception handling code. */
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
expand_builtin_setjmp_receiver (receiver_label)
|
|
|
|
|
rtx receiver_label ATTRIBUTE_UNUSED;
|
|
|
|
|
{
|
|
|
|
|
/* Clobber the FP when we get here, so we have to make sure it's
|
|
|
|
|
marked as used by this function. */
|
|
|
|
|
emit_insn (gen_rtx_USE (VOIDmode, hard_frame_pointer_rtx));
|
|
|
|
|
|
|
|
|
|
/* Mark the static chain as clobbered here so life information
|
|
|
|
|
doesn't get messed up for it. */
|
|
|
|
|
emit_insn (gen_rtx_CLOBBER (VOIDmode, static_chain_rtx));
|
|
|
|
|
|
|
|
|
|
/* Now put in the code to restore the frame pointer, and argument
|
|
|
|
|
pointer, if needed. The code below is from expand_end_bindings
|
|
|
|
|
in stmt.c; see detailed documentation there. */
|
|
|
|
|
#ifdef HAVE_nonlocal_goto
|
|
|
|
|
if (! HAVE_nonlocal_goto)
|
|
|
|
|
#endif
|
|
|
|
|
emit_move_insn (virtual_stack_vars_rtx, hard_frame_pointer_rtx);
|
|
|
|
|
|
|
|
|
|
#if ARG_POINTER_REGNUM != HARD_FRAME_POINTER_REGNUM
|
|
|
|
|
if (fixed_regs[ARG_POINTER_REGNUM])
|
|
|
|
|
{
|
|
|
|
|
#ifdef ELIMINABLE_REGS
|
|
|
|
|
size_t i;
|
|
|
|
|
static const struct elims {const int from, to;} elim_regs[] = ELIMINABLE_REGS;
|
|
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE (elim_regs); i++)
|
|
|
|
|
if (elim_regs[i].from == ARG_POINTER_REGNUM
|
|
|
|
|
&& elim_regs[i].to == HARD_FRAME_POINTER_REGNUM)
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
if (i == ARRAY_SIZE (elim_regs))
|
|
|
|
|
#endif
|
|
|
|
|
{
|
|
|
|
|
/* Now restore our arg pointer from the address at which it
|
|
|
|
|
was saved in our stack frame. */
|
|
|
|
|
emit_move_insn (virtual_incoming_args_rtx,
|
|
|
|
|
copy_to_reg (get_arg_pointer_save_area (cfun)));
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_builtin_setjmp_receiver
|
|
|
|
|
if (HAVE_builtin_setjmp_receiver)
|
|
|
|
|
emit_insn (gen_builtin_setjmp_receiver (receiver_label));
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
#ifdef HAVE_nonlocal_goto_receiver
|
|
|
|
|
if (HAVE_nonlocal_goto_receiver)
|
|
|
|
|
emit_insn (gen_nonlocal_goto_receiver ());
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
{ /* Nothing */ }
|
|
|
|
|
|
|
|
|
|
/* @@@ This is a kludge. Not all machine descriptions define a blockage
|
|
|
|
|
insn, but we must not allow the code we just generated to be reordered
|
|
|
|
|
by scheduling. Specifically, the update of the frame pointer must
|
|
|
|
|
happen immediately, not later. So emit an ASM_INPUT to act as blockage
|
|
|
|
|
insn. */
|
|
|
|
|
emit_insn (gen_rtx_ASM_INPUT (VOIDmode, ""));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* __builtin_setjmp is passed a pointer to an array of five words (not
|
|
|
|
|
all will be used on all machines). It operates similarly to the C
|
|
|
|
|
library function of the same name, but is more efficient. Much of
|
|
|
|
|
the code below (and for longjmp) is copied from the handling of
|
|
|
|
|
non-local gotos.
|
|
|
|
|
|
|
|
|
|
NOTE: This is intended for use by GNAT and the exception handling
|
|
|
|
|
scheme in the compiler and will only work in the method used by
|
|
|
|
|
them. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_setjmp (arglist, target)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
{
|
|
|
|
|
rtx buf_addr, next_lab, cont_lab;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return NULL_RTX;
|
|
|
|
|
|
|
|
|
|
if (target == 0 || GET_CODE (target) != REG
|
|
|
|
|
|| REGNO (target) < FIRST_PSEUDO_REGISTER)
|
|
|
|
|
target = gen_reg_rtx (TYPE_MODE (integer_type_node));
|
|
|
|
|
|
|
|
|
|
buf_addr = expand_expr (TREE_VALUE (arglist), NULL_RTX, VOIDmode, 0);
|
|
|
|
|
|
|
|
|
|
next_lab = gen_label_rtx ();
|
|
|
|
|
cont_lab = gen_label_rtx ();
|
|
|
|
|
|
|
|
|
|
expand_builtin_setjmp_setup (buf_addr, next_lab);
|
|
|
|
|
|
|
|
|
|
/* Set TARGET to zero and branch to the continue label. */
|
|
|
|
|
emit_move_insn (target, const0_rtx);
|
|
|
|
|
emit_jump_insn (gen_jump (cont_lab));
|
|
|
|
|
emit_barrier ();
|
|
|
|
|
emit_label (next_lab);
|
|
|
|
|
|
|
|
|
|
expand_builtin_setjmp_receiver (next_lab);
|
|
|
|
|
|
|
|
|
|
/* Set TARGET to one. */
|
|
|
|
|
emit_move_insn (target, const1_rtx);
|
|
|
|
|
emit_label (cont_lab);
|
|
|
|
|
|
|
|
|
|
/* Tell flow about the strange goings on. Putting `next_lab' on
|
|
|
|
|
`nonlocal_goto_handler_labels' to indicates that function
|
|
|
|
|
calls may traverse the arc back to this label. */
|
|
|
|
|
|
|
|
|
|
current_function_has_nonlocal_label = 1;
|
|
|
|
|
nonlocal_goto_handler_labels
|
|
|
|
|
= gen_rtx_EXPR_LIST (VOIDmode, next_lab, nonlocal_goto_handler_labels);
|
|
|
|
|
|
|
|
|
|
return target;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* __builtin_longjmp is passed a pointer to an array of five words (not
|
|
|
|
|
all will be used on all machines). It operates similarly to the C
|
|
|
|
|
library function of the same name, but is more efficient. Much of
|
|
|
|
|
the code below is copied from the handling of non-local gotos.
|
|
|
|
|
|
|
|
|
|
NOTE: This is intended for use by GNAT and the exception handling
|
|
|
|
|
scheme in the compiler and will only work in the method used by
|
|
|
|
|
them. */
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
expand_builtin_longjmp (buf_addr, value)
|
|
|
|
|
rtx buf_addr, value;
|
|
|
|
|
{
|
|
|
|
|
rtx fp, lab, stack, insn;
|
|
|
|
|
enum machine_mode sa_mode = STACK_SAVEAREA_MODE (SAVE_NONLOCAL);
|
|
|
|
|
|
|
|
|
|
if (setjmp_alias_set == -1)
|
|
|
|
|
setjmp_alias_set = new_alias_set ();
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (buf_addr) != Pmode)
|
|
|
|
|
buf_addr = convert_memory_address (Pmode, buf_addr);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
buf_addr = force_reg (Pmode, buf_addr);
|
|
|
|
|
|
|
|
|
|
/* We used to store value in static_chain_rtx, but that fails if pointers
|
|
|
|
|
are smaller than integers. We instead require that the user must pass
|
|
|
|
|
a second argument of 1, because that is what builtin_setjmp will
|
|
|
|
|
return. This also makes EH slightly more efficient, since we are no
|
|
|
|
|
longer copying around a value that we don't care about. */
|
|
|
|
|
if (value != const1_rtx)
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
current_function_calls_longjmp = 1;
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_builtin_longjmp
|
|
|
|
|
if (HAVE_builtin_longjmp)
|
|
|
|
|
emit_insn (gen_builtin_longjmp (buf_addr));
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
{
|
|
|
|
|
fp = gen_rtx_MEM (Pmode, buf_addr);
|
|
|
|
|
lab = gen_rtx_MEM (Pmode, plus_constant (buf_addr,
|
|
|
|
|
GET_MODE_SIZE (Pmode)));
|
|
|
|
|
|
|
|
|
|
stack = gen_rtx_MEM (sa_mode, plus_constant (buf_addr,
|
|
|
|
|
2 * GET_MODE_SIZE (Pmode)));
|
|
|
|
|
set_mem_alias_set (fp, setjmp_alias_set);
|
|
|
|
|
set_mem_alias_set (lab, setjmp_alias_set);
|
|
|
|
|
set_mem_alias_set (stack, setjmp_alias_set);
|
|
|
|
|
|
|
|
|
|
/* Pick up FP, label, and SP from the block and jump. This code is
|
|
|
|
|
from expand_goto in stmt.c; see there for detailed comments. */
|
|
|
|
|
#if HAVE_nonlocal_goto
|
|
|
|
|
if (HAVE_nonlocal_goto)
|
|
|
|
|
/* We have to pass a value to the nonlocal_goto pattern that will
|
|
|
|
|
get copied into the static_chain pointer, but it does not matter
|
|
|
|
|
what that value is, because builtin_setjmp does not use it. */
|
|
|
|
|
emit_insn (gen_nonlocal_goto (value, lab, stack, fp));
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
{
|
|
|
|
|
lab = copy_to_reg (lab);
|
|
|
|
|
|
|
|
|
|
emit_move_insn (hard_frame_pointer_rtx, fp);
|
|
|
|
|
emit_stack_restore (SAVE_NONLOCAL, stack, NULL_RTX);
|
|
|
|
|
|
|
|
|
|
emit_insn (gen_rtx_USE (VOIDmode, hard_frame_pointer_rtx));
|
|
|
|
|
emit_insn (gen_rtx_USE (VOIDmode, stack_pointer_rtx));
|
|
|
|
|
emit_indirect_jump (lab);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Search backwards and mark the jump insn as a non-local goto.
|
|
|
|
|
Note that this precludes the use of __builtin_longjmp to a
|
|
|
|
|
__builtin_setjmp target in the same function. However, we've
|
|
|
|
|
already cautioned the user that these functions are for
|
|
|
|
|
internal exception handling use only. */
|
|
|
|
|
for (insn = get_last_insn (); insn; insn = PREV_INSN (insn))
|
|
|
|
|
{
|
|
|
|
|
if (GET_CODE (insn) == JUMP_INSN)
|
|
|
|
|
{
|
|
|
|
|
REG_NOTES (insn) = alloc_EXPR_LIST (REG_NON_LOCAL_GOTO, const0_rtx,
|
|
|
|
|
REG_NOTES (insn));
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
else if (GET_CODE (insn) == CALL_INSN)
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to __builtin_prefetch. For a target that does not support
|
|
|
|
|
data prefetch, evaluate the memory address argument in case it has side
|
|
|
|
|
effects. */
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
expand_builtin_prefetch (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
tree arg0, arg1, arg2;
|
|
|
|
|
rtx op0, op1, op2;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, 0))
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
arg0 = TREE_VALUE (arglist);
|
|
|
|
|
/* Arguments 1 and 2 are optional; argument 1 (read/write) defaults to
|
|
|
|
|
zero (read) and argument 2 (locality) defaults to 3 (high degree of
|
|
|
|
|
locality). */
|
|
|
|
|
if (TREE_CHAIN (arglist))
|
|
|
|
|
{
|
|
|
|
|
arg1 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
if (TREE_CHAIN (TREE_CHAIN (arglist)))
|
|
|
|
|
arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
else
|
|
|
|
|
arg2 = build_int_2 (3, 0);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
arg1 = integer_zero_node;
|
|
|
|
|
arg2 = build_int_2 (3, 0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Argument 0 is an address. */
|
|
|
|
|
op0 = expand_expr (arg0, NULL_RTX, Pmode, EXPAND_NORMAL);
|
|
|
|
|
|
|
|
|
|
/* Argument 1 (read/write flag) must be a compile-time constant int. */
|
|
|
|
|
if (TREE_CODE (arg1) != INTEGER_CST)
|
|
|
|
|
{
|
|
|
|
|
error ("second arg to `__builtin_prefetch' must be a constant");
|
|
|
|
|
arg1 = integer_zero_node;
|
|
|
|
|
}
|
|
|
|
|
op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
/* Argument 1 must be either zero or one. */
|
|
|
|
|
if (INTVAL (op1) != 0 && INTVAL (op1) != 1)
|
|
|
|
|
{
|
|
|
|
|
warning ("invalid second arg to __builtin_prefetch; using zero");
|
|
|
|
|
op1 = const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Argument 2 (locality) must be a compile-time constant int. */
|
|
|
|
|
if (TREE_CODE (arg2) != INTEGER_CST)
|
|
|
|
|
{
|
|
|
|
|
error ("third arg to `__builtin_prefetch' must be a constant");
|
|
|
|
|
arg2 = integer_zero_node;
|
|
|
|
|
}
|
|
|
|
|
op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
/* Argument 2 must be 0, 1, 2, or 3. */
|
|
|
|
|
if (INTVAL (op2) < 0 || INTVAL (op2) > 3)
|
|
|
|
|
{
|
|
|
|
|
warning ("invalid third arg to __builtin_prefetch; using zero");
|
|
|
|
|
op2 = const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_prefetch
|
|
|
|
|
if (HAVE_prefetch)
|
|
|
|
|
{
|
|
|
|
|
if (! (*insn_data[(int)CODE_FOR_prefetch].operand[0].predicate)
|
|
|
|
|
(op0,
|
|
|
|
|
insn_data[(int)CODE_FOR_prefetch].operand[0].mode))
|
|
|
|
|
op0 = force_reg (Pmode, op0);
|
|
|
|
|
emit_insn (gen_prefetch (op0, op1, op2));
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
op0 = protect_from_queue (op0, 0);
|
|
|
|
|
/* Don't do anything with direct references to volatile memory, but
|
|
|
|
|
generate code to handle other side effects. */
|
|
|
|
|
if (GET_CODE (op0) != MEM && side_effects_p (op0))
|
|
|
|
|
emit_insn (op0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Get a MEM rtx for expression EXP which is the address of an operand
|
|
|
|
|
to be used to be used in a string instruction (cmpstrsi, movstrsi, ..). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
get_memory_rtx (exp)
|
|
|
|
|
tree exp;
|
|
|
|
|
{
|
|
|
|
|
rtx addr = expand_expr (exp, NULL_RTX, ptr_mode, EXPAND_SUM);
|
|
|
|
|
rtx mem;
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (addr) != Pmode)
|
|
|
|
|
addr = convert_memory_address (Pmode, addr);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
mem = gen_rtx_MEM (BLKmode, memory_address (BLKmode, addr));
|
|
|
|
|
|
|
|
|
|
/* Get an expression we can use to find the attributes to assign to MEM.
|
|
|
|
|
If it is an ADDR_EXPR, use the operand. Otherwise, dereference it if
|
|
|
|
|
we can. First remove any nops. */
|
|
|
|
|
while ((TREE_CODE (exp) == NOP_EXPR || TREE_CODE (exp) == CONVERT_EXPR
|
|
|
|
|
|| TREE_CODE (exp) == NON_LVALUE_EXPR)
|
|
|
|
|
&& POINTER_TYPE_P (TREE_TYPE (TREE_OPERAND (exp, 0))))
|
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (exp) == ADDR_EXPR)
|
|
|
|
|
{
|
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
|
|
|
set_mem_attributes (mem, exp, 0);
|
|
|
|
|
}
|
|
|
|
|
else if (POINTER_TYPE_P (TREE_TYPE (exp)))
|
|
|
|
|
{
|
|
|
|
|
exp = build1 (INDIRECT_REF, TREE_TYPE (TREE_TYPE (exp)), exp);
|
|
|
|
|
/* memcpy, memset and other builtin stringops can alias with anything. */
|
|
|
|
|
set_mem_alias_set (mem, 0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return mem;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Built-in functions to perform an untyped call and return. */
|
|
|
|
|
|
|
|
|
|
/* For each register that may be used for calling a function, this
|
|
|
|
|
gives a mode used to copy the register's value. VOIDmode indicates
|
|
|
|
|
the register is not used for calling a function. If the machine
|
|
|
|
|
has register windows, this gives only the outbound registers.
|
|
|
|
|
INCOMING_REGNO gives the corresponding inbound register. */
|
|
|
|
|
static enum machine_mode apply_args_mode[FIRST_PSEUDO_REGISTER];
|
|
|
|
|
|
|
|
|
|
/* For each register that may be used for returning values, this gives
|
|
|
|
|
a mode used to copy the register's value. VOIDmode indicates the
|
|
|
|
|
register is not used for returning values. If the machine has
|
|
|
|
|
register windows, this gives only the outbound registers.
|
|
|
|
|
INCOMING_REGNO gives the corresponding inbound register. */
|
|
|
|
|
static enum machine_mode apply_result_mode[FIRST_PSEUDO_REGISTER];
|
|
|
|
|
|
|
|
|
|
/* For each register that may be used for calling a function, this
|
|
|
|
|
gives the offset of that register into the block returned by
|
|
|
|
|
__builtin_apply_args. 0 indicates that the register is not
|
|
|
|
|
used for calling a function. */
|
|
|
|
|
static int apply_args_reg_offset[FIRST_PSEUDO_REGISTER];
|
|
|
|
|
|
|
|
|
|
/* Return the offset of register REGNO into the block returned by
|
|
|
|
|
__builtin_apply_args. This is not declared static, since it is
|
|
|
|
|
needed in objc-act.c. */
|
|
|
|
|
|
|
|
|
|
int
|
|
|
|
|
apply_args_register_offset (regno)
|
|
|
|
|
int regno;
|
|
|
|
|
{
|
|
|
|
|
apply_args_size ();
|
|
|
|
|
|
|
|
|
|
/* Arguments are always put in outgoing registers (in the argument
|
|
|
|
|
block) if such make sense. */
|
|
|
|
|
#ifdef OUTGOING_REGNO
|
|
|
|
|
regno = OUTGOING_REGNO(regno);
|
|
|
|
|
#endif
|
|
|
|
|
return apply_args_reg_offset[regno];
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Return the size required for the block returned by __builtin_apply_args,
|
|
|
|
|
and initialize apply_args_mode. */
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
|
apply_args_size ()
|
|
|
|
|
{
|
|
|
|
|
static int size = -1;
|
|
|
|
|
int align;
|
|
|
|
|
unsigned int regno;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
|
|
|
|
|
/* The values computed by this function never change. */
|
|
|
|
|
if (size < 0)
|
|
|
|
|
{
|
|
|
|
|
/* The first value is the incoming arg-pointer. */
|
|
|
|
|
size = GET_MODE_SIZE (Pmode);
|
|
|
|
|
|
|
|
|
|
/* The second value is the structure value address unless this is
|
|
|
|
|
passed as an "invisible" first argument. */
|
|
|
|
|
if (struct_value_rtx)
|
|
|
|
|
size += GET_MODE_SIZE (Pmode);
|
|
|
|
|
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if (FUNCTION_ARG_REGNO_P (regno))
|
|
|
|
|
{
|
|
|
|
|
/* Search for the proper mode for copying this register's
|
|
|
|
|
value. I'm not sure this is right, but it works so far. */
|
|
|
|
|
enum machine_mode best_mode = VOIDmode;
|
|
|
|
|
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_INT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& HARD_REGNO_NREGS (regno, mode) == 1)
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
if (best_mode == VOIDmode)
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_FLOAT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& have_insn_for (SET, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
if (best_mode == VOIDmode)
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_FLOAT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& have_insn_for (SET, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
if (best_mode == VOIDmode)
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_INT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& have_insn_for (SET, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
mode = best_mode;
|
|
|
|
|
if (mode == VOIDmode)
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
|
|
|
|
if (size % align != 0)
|
|
|
|
|
size = CEIL (size, align) * align;
|
|
|
|
|
apply_args_reg_offset[regno] = size;
|
|
|
|
|
size += GET_MODE_SIZE (mode);
|
|
|
|
|
apply_args_mode[regno] = mode;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
apply_args_mode[regno] = VOIDmode;
|
|
|
|
|
apply_args_reg_offset[regno] = 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
return size;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Return the size required for the block returned by __builtin_apply,
|
|
|
|
|
and initialize apply_result_mode. */
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
|
apply_result_size ()
|
|
|
|
|
{
|
|
|
|
|
static int size = -1;
|
|
|
|
|
int align, regno;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
|
|
|
|
|
/* The values computed by this function never change. */
|
|
|
|
|
if (size < 0)
|
|
|
|
|
{
|
|
|
|
|
size = 0;
|
|
|
|
|
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if (FUNCTION_VALUE_REGNO_P (regno))
|
|
|
|
|
{
|
|
|
|
|
/* Search for the proper mode for copying this register's
|
|
|
|
|
value. I'm not sure this is right, but it works so far. */
|
|
|
|
|
enum machine_mode best_mode = VOIDmode;
|
|
|
|
|
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_INT);
|
|
|
|
|
mode != TImode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
if (best_mode == VOIDmode)
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_FLOAT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& have_insn_for (SET, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
if (best_mode == VOIDmode)
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_FLOAT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& have_insn_for (SET, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
if (best_mode == VOIDmode)
|
|
|
|
|
for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_INT);
|
|
|
|
|
mode != VOIDmode;
|
|
|
|
|
mode = GET_MODE_WIDER_MODE (mode))
|
|
|
|
|
if (HARD_REGNO_MODE_OK (regno, mode)
|
|
|
|
|
&& have_insn_for (SET, mode))
|
|
|
|
|
best_mode = mode;
|
|
|
|
|
|
|
|
|
|
mode = best_mode;
|
|
|
|
|
if (mode == VOIDmode)
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
|
|
|
|
if (size % align != 0)
|
|
|
|
|
size = CEIL (size, align) * align;
|
|
|
|
|
size += GET_MODE_SIZE (mode);
|
|
|
|
|
apply_result_mode[regno] = mode;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
apply_result_mode[regno] = VOIDmode;
|
|
|
|
|
|
|
|
|
|
/* Allow targets that use untyped_call and untyped_return to override
|
|
|
|
|
the size so that machine-specific information can be stored here. */
|
|
|
|
|
#ifdef APPLY_RESULT_SIZE
|
|
|
|
|
size = APPLY_RESULT_SIZE;
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
return size;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#if defined (HAVE_untyped_call) || defined (HAVE_untyped_return)
|
|
|
|
|
/* Create a vector describing the result block RESULT. If SAVEP is true,
|
|
|
|
|
the result block is used to save the values; otherwise it is used to
|
|
|
|
|
restore the values. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
result_vector (savep, result)
|
|
|
|
|
int savep;
|
|
|
|
|
rtx result;
|
|
|
|
|
{
|
|
|
|
|
int regno, size, align, nelts;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
rtx reg, mem;
|
|
|
|
|
rtx *savevec = (rtx *) alloca (FIRST_PSEUDO_REGISTER * sizeof (rtx));
|
|
|
|
|
|
|
|
|
|
size = nelts = 0;
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if ((mode = apply_result_mode[regno]) != VOIDmode)
|
|
|
|
|
{
|
|
|
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
|
|
|
|
if (size % align != 0)
|
|
|
|
|
size = CEIL (size, align) * align;
|
|
|
|
|
reg = gen_rtx_REG (mode, savep ? regno : INCOMING_REGNO (regno));
|
|
|
|
|
mem = adjust_address (result, mode, size);
|
|
|
|
|
savevec[nelts++] = (savep
|
|
|
|
|
? gen_rtx_SET (VOIDmode, mem, reg)
|
|
|
|
|
: gen_rtx_SET (VOIDmode, reg, mem));
|
|
|
|
|
size += GET_MODE_SIZE (mode);
|
|
|
|
|
}
|
|
|
|
|
return gen_rtx_PARALLEL (VOIDmode, gen_rtvec_v (nelts, savevec));
|
|
|
|
|
}
|
|
|
|
|
#endif /* HAVE_untyped_call or HAVE_untyped_return */
|
|
|
|
|
|
|
|
|
|
/* Save the state required to perform an untyped call with the same
|
|
|
|
|
arguments as were passed to the current function. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_apply_args_1 ()
|
|
|
|
|
{
|
|
|
|
|
rtx registers;
|
|
|
|
|
int size, align, regno;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
|
|
|
|
|
/* Create a block where the arg-pointer, structure value address,
|
|
|
|
|
and argument registers can be saved. */
|
|
|
|
|
registers = assign_stack_local (BLKmode, apply_args_size (), -1);
|
|
|
|
|
|
|
|
|
|
/* Walk past the arg-pointer and structure value address. */
|
|
|
|
|
size = GET_MODE_SIZE (Pmode);
|
|
|
|
|
if (struct_value_rtx)
|
|
|
|
|
size += GET_MODE_SIZE (Pmode);
|
|
|
|
|
|
|
|
|
|
/* Save each register used in calling a function to the block. */
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if ((mode = apply_args_mode[regno]) != VOIDmode)
|
|
|
|
|
{
|
|
|
|
|
rtx tem;
|
|
|
|
|
|
|
|
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
|
|
|
|
if (size % align != 0)
|
|
|
|
|
size = CEIL (size, align) * align;
|
|
|
|
|
|
|
|
|
|
tem = gen_rtx_REG (mode, INCOMING_REGNO (regno));
|
|
|
|
|
|
|
|
|
|
emit_move_insn (adjust_address (registers, mode, size), tem);
|
|
|
|
|
size += GET_MODE_SIZE (mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Save the arg pointer to the block. */
|
|
|
|
|
emit_move_insn (adjust_address (registers, Pmode, 0),
|
|
|
|
|
copy_to_reg (virtual_incoming_args_rtx));
|
|
|
|
|
size = GET_MODE_SIZE (Pmode);
|
|
|
|
|
|
|
|
|
|
/* Save the structure value address unless this is passed as an
|
|
|
|
|
"invisible" first argument. */
|
|
|
|
|
if (struct_value_incoming_rtx)
|
|
|
|
|
{
|
|
|
|
|
emit_move_insn (adjust_address (registers, Pmode, size),
|
|
|
|
|
copy_to_reg (struct_value_incoming_rtx));
|
|
|
|
|
size += GET_MODE_SIZE (Pmode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Return the address of the block. */
|
|
|
|
|
return copy_addr_to_reg (XEXP (registers, 0));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* __builtin_apply_args returns block of memory allocated on
|
|
|
|
|
the stack into which is stored the arg pointer, structure
|
|
|
|
|
value address, static chain, and all the registers that might
|
|
|
|
|
possibly be used in performing a function call. The code is
|
|
|
|
|
moved to the start of the function so the incoming values are
|
|
|
|
|
saved. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_apply_args ()
|
|
|
|
|
{
|
|
|
|
|
/* Don't do __builtin_apply_args more than once in a function.
|
|
|
|
|
Save the result of the first call and reuse it. */
|
|
|
|
|
if (apply_args_value != 0)
|
|
|
|
|
return apply_args_value;
|
|
|
|
|
{
|
|
|
|
|
/* When this function is called, it means that registers must be
|
|
|
|
|
saved on entry to this function. So we migrate the
|
|
|
|
|
call to the first insn of this function. */
|
|
|
|
|
rtx temp;
|
|
|
|
|
rtx seq;
|
|
|
|
|
|
|
|
|
|
start_sequence ();
|
|
|
|
|
temp = expand_builtin_apply_args_1 ();
|
|
|
|
|
seq = get_insns ();
|
|
|
|
|
end_sequence ();
|
|
|
|
|
|
|
|
|
|
apply_args_value = temp;
|
|
|
|
|
|
|
|
|
|
/* Put the sequence after the NOTE that starts the function.
|
|
|
|
|
If this is inside a SEQUENCE, make the outer-level insn
|
|
|
|
|
chain current, so the code is placed at the start of the
|
|
|
|
|
function. */
|
|
|
|
|
push_topmost_sequence ();
|
|
|
|
|
emit_insns_before (seq, NEXT_INSN (get_insns ()));
|
|
|
|
|
pop_topmost_sequence ();
|
|
|
|
|
return temp;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Perform an untyped call and save the state required to perform an
|
|
|
|
|
untyped return of whatever value was returned by the given function. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_apply (function, arguments, argsize)
|
|
|
|
|
rtx function, arguments, argsize;
|
|
|
|
|
{
|
|
|
|
|
int size, align, regno;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
rtx incoming_args, result, reg, dest, src, call_insn;
|
|
|
|
|
rtx old_stack_level = 0;
|
|
|
|
|
rtx call_fusage = 0;
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (arguments) != Pmode)
|
|
|
|
|
arguments = convert_memory_address (Pmode, arguments);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Create a block where the return registers can be saved. */
|
|
|
|
|
result = assign_stack_local (BLKmode, apply_result_size (), -1);
|
|
|
|
|
|
|
|
|
|
/* Fetch the arg pointer from the ARGUMENTS block. */
|
|
|
|
|
incoming_args = gen_reg_rtx (Pmode);
|
|
|
|
|
emit_move_insn (incoming_args, gen_rtx_MEM (Pmode, arguments));
|
|
|
|
|
#ifndef STACK_GROWS_DOWNWARD
|
|
|
|
|
incoming_args = expand_simple_binop (Pmode, MINUS, incoming_args, argsize,
|
|
|
|
|
incoming_args, 0, OPTAB_LIB_WIDEN);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Perform postincrements before actually calling the function. */
|
|
|
|
|
emit_queue ();
|
|
|
|
|
|
|
|
|
|
/* Push a new argument block and copy the arguments. Do not allow
|
|
|
|
|
the (potential) memcpy call below to interfere with our stack
|
|
|
|
|
manipulations. */
|
|
|
|
|
do_pending_stack_adjust ();
|
|
|
|
|
NO_DEFER_POP;
|
|
|
|
|
|
|
|
|
|
/* Save the stack with nonlocal if available */
|
|
|
|
|
#ifdef HAVE_save_stack_nonlocal
|
|
|
|
|
if (HAVE_save_stack_nonlocal)
|
|
|
|
|
emit_stack_save (SAVE_NONLOCAL, &old_stack_level, NULL_RTX);
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
emit_stack_save (SAVE_BLOCK, &old_stack_level, NULL_RTX);
|
|
|
|
|
|
|
|
|
|
/* Push a block of memory onto the stack to store the memory arguments.
|
|
|
|
|
Save the address in a register, and copy the memory arguments. ??? I
|
|
|
|
|
haven't figured out how the calling convention macros effect this,
|
|
|
|
|
but it's likely that the source and/or destination addresses in
|
|
|
|
|
the block copy will need updating in machine specific ways. */
|
|
|
|
|
dest = allocate_dynamic_stack_space (argsize, 0, BITS_PER_UNIT);
|
|
|
|
|
dest = gen_rtx_MEM (BLKmode, dest);
|
|
|
|
|
set_mem_align (dest, PARM_BOUNDARY);
|
|
|
|
|
src = gen_rtx_MEM (BLKmode, incoming_args);
|
|
|
|
|
set_mem_align (src, PARM_BOUNDARY);
|
|
|
|
|
emit_block_move (dest, src, argsize);
|
|
|
|
|
|
|
|
|
|
/* Refer to the argument block. */
|
|
|
|
|
apply_args_size ();
|
|
|
|
|
arguments = gen_rtx_MEM (BLKmode, arguments);
|
|
|
|
|
set_mem_align (arguments, PARM_BOUNDARY);
|
|
|
|
|
|
|
|
|
|
/* Walk past the arg-pointer and structure value address. */
|
|
|
|
|
size = GET_MODE_SIZE (Pmode);
|
|
|
|
|
if (struct_value_rtx)
|
|
|
|
|
size += GET_MODE_SIZE (Pmode);
|
|
|
|
|
|
|
|
|
|
/* Restore each of the registers previously saved. Make USE insns
|
|
|
|
|
for each of these registers for use in making the call. */
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if ((mode = apply_args_mode[regno]) != VOIDmode)
|
|
|
|
|
{
|
|
|
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
|
|
|
|
if (size % align != 0)
|
|
|
|
|
size = CEIL (size, align) * align;
|
|
|
|
|
reg = gen_rtx_REG (mode, regno);
|
|
|
|
|
emit_move_insn (reg, adjust_address (arguments, mode, size));
|
|
|
|
|
use_reg (&call_fusage, reg);
|
|
|
|
|
size += GET_MODE_SIZE (mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Restore the structure value address unless this is passed as an
|
|
|
|
|
"invisible" first argument. */
|
|
|
|
|
size = GET_MODE_SIZE (Pmode);
|
|
|
|
|
if (struct_value_rtx)
|
|
|
|
|
{
|
|
|
|
|
rtx value = gen_reg_rtx (Pmode);
|
|
|
|
|
emit_move_insn (value, adjust_address (arguments, Pmode, size));
|
|
|
|
|
emit_move_insn (struct_value_rtx, value);
|
|
|
|
|
if (GET_CODE (struct_value_rtx) == REG)
|
|
|
|
|
use_reg (&call_fusage, struct_value_rtx);
|
|
|
|
|
size += GET_MODE_SIZE (Pmode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* All arguments and registers used for the call are set up by now! */
|
|
|
|
|
function = prepare_call_address (function, NULL_TREE, &call_fusage, 0, 0);
|
|
|
|
|
|
|
|
|
|
/* Ensure address is valid. SYMBOL_REF is already valid, so no need,
|
|
|
|
|
and we don't want to load it into a register as an optimization,
|
|
|
|
|
because prepare_call_address already did it if it should be done. */
|
|
|
|
|
if (GET_CODE (function) != SYMBOL_REF)
|
|
|
|
|
function = memory_address (FUNCTION_MODE, function);
|
|
|
|
|
|
|
|
|
|
/* Generate the actual call instruction and save the return value. */
|
|
|
|
|
#ifdef HAVE_untyped_call
|
|
|
|
|
if (HAVE_untyped_call)
|
|
|
|
|
emit_call_insn (gen_untyped_call (gen_rtx_MEM (FUNCTION_MODE, function),
|
|
|
|
|
result, result_vector (1, result)));
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
#ifdef HAVE_call_value
|
|
|
|
|
if (HAVE_call_value)
|
|
|
|
|
{
|
|
|
|
|
rtx valreg = 0;
|
|
|
|
|
|
|
|
|
|
/* Locate the unique return register. It is not possible to
|
|
|
|
|
express a call that sets more than one return register using
|
|
|
|
|
call_value; use untyped_call for that. In fact, untyped_call
|
|
|
|
|
only needs to save the return registers in the given block. */
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if ((mode = apply_result_mode[regno]) != VOIDmode)
|
|
|
|
|
{
|
|
|
|
|
if (valreg)
|
|
|
|
|
abort (); /* HAVE_untyped_call required. */
|
|
|
|
|
valreg = gen_rtx_REG (mode, regno);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
emit_call_insn (GEN_CALL_VALUE (valreg,
|
|
|
|
|
gen_rtx_MEM (FUNCTION_MODE, function),
|
|
|
|
|
const0_rtx, NULL_RTX, const0_rtx));
|
|
|
|
|
|
|
|
|
|
emit_move_insn (adjust_address (result, GET_MODE (valreg), 0), valreg);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
/* Find the CALL insn we just emitted. */
|
|
|
|
|
for (call_insn = get_last_insn ();
|
|
|
|
|
call_insn && GET_CODE (call_insn) != CALL_INSN;
|
|
|
|
|
call_insn = PREV_INSN (call_insn))
|
|
|
|
|
;
|
|
|
|
|
|
|
|
|
|
if (! call_insn)
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
/* Put the register usage information on the CALL. If there is already
|
|
|
|
|
some usage information, put ours at the end. */
|
|
|
|
|
if (CALL_INSN_FUNCTION_USAGE (call_insn))
|
|
|
|
|
{
|
|
|
|
|
rtx link;
|
|
|
|
|
|
|
|
|
|
for (link = CALL_INSN_FUNCTION_USAGE (call_insn); XEXP (link, 1) != 0;
|
|
|
|
|
link = XEXP (link, 1))
|
|
|
|
|
;
|
|
|
|
|
|
|
|
|
|
XEXP (link, 1) = call_fusage;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
CALL_INSN_FUNCTION_USAGE (call_insn) = call_fusage;
|
|
|
|
|
|
|
|
|
|
/* Restore the stack. */
|
|
|
|
|
#ifdef HAVE_save_stack_nonlocal
|
|
|
|
|
if (HAVE_save_stack_nonlocal)
|
|
|
|
|
emit_stack_restore (SAVE_NONLOCAL, old_stack_level, NULL_RTX);
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
emit_stack_restore (SAVE_BLOCK, old_stack_level, NULL_RTX);
|
|
|
|
|
|
|
|
|
|
OK_DEFER_POP;
|
|
|
|
|
|
|
|
|
|
/* Return the address of the result block. */
|
|
|
|
|
return copy_addr_to_reg (XEXP (result, 0));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Perform an untyped return. */
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
expand_builtin_return (result)
|
|
|
|
|
rtx result;
|
|
|
|
|
{
|
|
|
|
|
int size, align, regno;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
rtx reg;
|
|
|
|
|
rtx call_fusage = 0;
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (result) != Pmode)
|
|
|
|
|
result = convert_memory_address (Pmode, result);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
apply_result_size ();
|
|
|
|
|
result = gen_rtx_MEM (BLKmode, result);
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_untyped_return
|
|
|
|
|
if (HAVE_untyped_return)
|
|
|
|
|
{
|
|
|
|
|
emit_jump_insn (gen_untyped_return (result, result_vector (0, result)));
|
|
|
|
|
emit_barrier ();
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Restore the return value and note that each value is used. */
|
|
|
|
|
size = 0;
|
|
|
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
|
|
|
|
if ((mode = apply_result_mode[regno]) != VOIDmode)
|
|
|
|
|
{
|
|
|
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
|
|
|
|
if (size % align != 0)
|
|
|
|
|
size = CEIL (size, align) * align;
|
|
|
|
|
reg = gen_rtx_REG (mode, INCOMING_REGNO (regno));
|
|
|
|
|
emit_move_insn (reg, adjust_address (result, mode, size));
|
|
|
|
|
|
|
|
|
|
push_to_sequence (call_fusage);
|
|
|
|
|
emit_insn (gen_rtx_USE (VOIDmode, reg));
|
|
|
|
|
call_fusage = get_insns ();
|
|
|
|
|
end_sequence ();
|
|
|
|
|
size += GET_MODE_SIZE (mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Put the USE insns before the return. */
|
|
|
|
|
emit_insns (call_fusage);
|
|
|
|
|
|
|
|
|
|
/* Return whatever values was restored by jumping directly to the end
|
|
|
|
|
of the function. */
|
|
|
|
|
expand_null_return ();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Used by expand_builtin_classify_type and fold_builtin_classify_type. */
|
|
|
|
|
|
|
|
|
|
static enum type_class
|
|
|
|
|
type_to_class (type)
|
|
|
|
|
tree type;
|
|
|
|
|
{
|
|
|
|
|
switch (TREE_CODE (type))
|
|
|
|
|
{
|
|
|
|
|
case VOID_TYPE: return void_type_class;
|
|
|
|
|
case INTEGER_TYPE: return integer_type_class;
|
|
|
|
|
case CHAR_TYPE: return char_type_class;
|
|
|
|
|
case ENUMERAL_TYPE: return enumeral_type_class;
|
|
|
|
|
case BOOLEAN_TYPE: return boolean_type_class;
|
|
|
|
|
case POINTER_TYPE: return pointer_type_class;
|
|
|
|
|
case REFERENCE_TYPE: return reference_type_class;
|
|
|
|
|
case OFFSET_TYPE: return offset_type_class;
|
|
|
|
|
case REAL_TYPE: return real_type_class;
|
|
|
|
|
case COMPLEX_TYPE: return complex_type_class;
|
|
|
|
|
case FUNCTION_TYPE: return function_type_class;
|
|
|
|
|
case METHOD_TYPE: return method_type_class;
|
|
|
|
|
case RECORD_TYPE: return record_type_class;
|
|
|
|
|
case UNION_TYPE:
|
|
|
|
|
case QUAL_UNION_TYPE: return union_type_class;
|
|
|
|
|
case ARRAY_TYPE: return (TYPE_STRING_FLAG (type)
|
|
|
|
|
? string_type_class : array_type_class);
|
|
|
|
|
case SET_TYPE: return set_type_class;
|
|
|
|
|
case FILE_TYPE: return file_type_class;
|
|
|
|
|
case LANG_TYPE: return lang_type_class;
|
|
|
|
|
default: return no_type_class;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to __builtin_classify_type with arguments found in
|
|
|
|
|
ARGLIST. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_classify_type (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
if (arglist != 0)
|
|
|
|
|
return GEN_INT (type_to_class (TREE_TYPE (TREE_VALUE (arglist))));
|
|
|
|
|
return GEN_INT (no_type_class);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to __builtin_constant_p. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_constant_p (exp)
|
|
|
|
|
tree exp;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
enum machine_mode value_mode = TYPE_MODE (TREE_TYPE (exp));
|
|
|
|
|
rtx tmp;
|
|
|
|
|
|
|
|
|
|
if (arglist == 0)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
arglist = TREE_VALUE (arglist);
|
|
|
|
|
|
|
|
|
|
/* We have taken care of the easy cases during constant folding. This
|
|
|
|
|
case is not obvious, so emit (constant_p_rtx (ARGLIST)) and let CSE get a
|
|
|
|
|
chance to see if it can deduce whether ARGLIST is constant. */
|
|
|
|
|
|
|
|
|
|
tmp = expand_expr (arglist, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
tmp = gen_rtx_CONSTANT_P_RTX (value_mode, tmp);
|
|
|
|
|
return tmp;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to one of the builtin math functions (sin, cos, or sqrt).
|
|
|
|
|
Return 0 if a normal call should be emitted rather than expanding the
|
|
|
|
|
function in-line. EXP is the expression that is a call to the builtin
|
|
|
|
|
function; if convenient, the result should be placed in TARGET.
|
|
|
|
|
SUBTARGET may be used as the target for computing one of EXP's operands. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_mathfn (exp, target, subtarget)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target, subtarget;
|
|
|
|
|
{
|
|
|
|
|
optab builtin_optab;
|
|
|
|
|
rtx op0, insns;
|
|
|
|
|
tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, REAL_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Stabilize and compute the argument. */
|
|
|
|
|
if (TREE_CODE (TREE_VALUE (arglist)) != VAR_DECL
|
|
|
|
|
&& TREE_CODE (TREE_VALUE (arglist)) != PARM_DECL)
|
|
|
|
|
{
|
|
|
|
|
exp = copy_node (exp);
|
|
|
|
|
TREE_OPERAND (exp, 1) = arglist;
|
|
|
|
|
/* Wrap the computation of the argument in a SAVE_EXPR. That
|
|
|
|
|
way, if we need to expand the argument again (as in the
|
|
|
|
|
flag_errno_math case below where we cannot directly set
|
|
|
|
|
errno), we will not perform side-effects more than once.
|
|
|
|
|
Note that here we're mutating the original EXP as well as the
|
|
|
|
|
copy; that's the right thing to do in case the original EXP
|
|
|
|
|
is expanded later. */
|
|
|
|
|
TREE_VALUE (arglist) = save_expr (TREE_VALUE (arglist));
|
|
|
|
|
arglist = copy_node (arglist);
|
|
|
|
|
}
|
|
|
|
|
op0 = expand_expr (TREE_VALUE (arglist), subtarget, VOIDmode, 0);
|
|
|
|
|
|
|
|
|
|
/* Make a suitable register to place result in. */
|
|
|
|
|
target = gen_reg_rtx (TYPE_MODE (TREE_TYPE (exp)));
|
|
|
|
|
|
|
|
|
|
emit_queue ();
|
|
|
|
|
start_sequence ();
|
|
|
|
|
|
|
|
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
|
|
|
|
{
|
|
|
|
|
case BUILT_IN_SIN:
|
|
|
|
|
case BUILT_IN_SINF:
|
|
|
|
|
case BUILT_IN_SINL:
|
|
|
|
|
builtin_optab = sin_optab; break;
|
|
|
|
|
case BUILT_IN_COS:
|
|
|
|
|
case BUILT_IN_COSF:
|
|
|
|
|
case BUILT_IN_COSL:
|
|
|
|
|
builtin_optab = cos_optab; break;
|
|
|
|
|
case BUILT_IN_SQRT:
|
|
|
|
|
case BUILT_IN_SQRTF:
|
|
|
|
|
case BUILT_IN_SQRTL:
|
|
|
|
|
builtin_optab = sqrt_optab; break;
|
|
|
|
|
default:
|
|
|
|
|
abort ();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Compute into TARGET.
|
|
|
|
|
Set TARGET to wherever the result comes back. */
|
|
|
|
|
target = expand_unop (TYPE_MODE (TREE_TYPE (TREE_VALUE (arglist))),
|
|
|
|
|
builtin_optab, op0, target, 0);
|
|
|
|
|
|
|
|
|
|
/* If we were unable to expand via the builtin, stop the
|
|
|
|
|
sequence (without outputting the insns) and return 0, causing
|
|
|
|
|
a call to the library function. */
|
|
|
|
|
if (target == 0)
|
|
|
|
|
{
|
|
|
|
|
end_sequence ();
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If errno must be maintained and if we are not allowing unsafe
|
|
|
|
|
math optimizations, check the result. */
|
|
|
|
|
|
|
|
|
|
if (flag_errno_math && ! flag_unsafe_math_optimizations)
|
|
|
|
|
{
|
|
|
|
|
rtx lab1;
|
|
|
|
|
|
|
|
|
|
/* Don't define the builtin FP instructions
|
|
|
|
|
if your machine is not IEEE. */
|
|
|
|
|
if (TARGET_FLOAT_FORMAT != IEEE_FLOAT_FORMAT)
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
lab1 = gen_label_rtx ();
|
|
|
|
|
|
|
|
|
|
/* Test the result; if it is NaN, set errno=EDOM because
|
|
|
|
|
the argument was not in the domain. */
|
|
|
|
|
emit_cmp_and_jump_insns (target, target, EQ, 0, GET_MODE (target),
|
|
|
|
|
0, lab1);
|
|
|
|
|
|
|
|
|
|
#ifdef TARGET_EDOM
|
|
|
|
|
{
|
|
|
|
|
#ifdef GEN_ERRNO_RTX
|
|
|
|
|
rtx errno_rtx = GEN_ERRNO_RTX;
|
|
|
|
|
#else
|
|
|
|
|
rtx errno_rtx
|
|
|
|
|
= gen_rtx_MEM (word_mode, gen_rtx_SYMBOL_REF (Pmode, "errno"));
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
emit_move_insn (errno_rtx, GEN_INT (TARGET_EDOM));
|
|
|
|
|
}
|
|
|
|
|
#else
|
|
|
|
|
/* We can't set errno=EDOM directly; let the library call do it.
|
|
|
|
|
Pop the arguments right away in case the call gets deleted. */
|
|
|
|
|
NO_DEFER_POP;
|
|
|
|
|
expand_call (exp, target, 0);
|
|
|
|
|
OK_DEFER_POP;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
emit_label (lab1);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Output the entire sequence. */
|
|
|
|
|
insns = get_insns ();
|
|
|
|
|
end_sequence ();
|
|
|
|
|
emit_insns (insns);
|
|
|
|
|
|
|
|
|
|
return target;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP which is a call to the strlen builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call, otherwise
|
|
|
|
|
try to get the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strlen (exp, target)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
enum machine_mode value_mode = TYPE_MODE (TREE_TYPE (exp));
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
rtx pat;
|
|
|
|
|
tree src = TREE_VALUE (arglist);
|
|
|
|
|
|
|
|
|
|
int align
|
|
|
|
|
= get_pointer_alignment (src, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
|
|
|
|
|
|
|
|
|
|
rtx result, src_reg, char_rtx, before_strlen;
|
|
|
|
|
enum machine_mode insn_mode = value_mode, char_mode;
|
|
|
|
|
enum insn_code icode = CODE_FOR_nothing;
|
|
|
|
|
|
|
|
|
|
/* If SRC is not a pointer type, don't do this operation inline. */
|
|
|
|
|
if (align == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Bail out if we can't compute strlen in the right mode. */
|
|
|
|
|
while (insn_mode != VOIDmode)
|
|
|
|
|
{
|
|
|
|
|
icode = strlen_optab->handlers[(int) insn_mode].insn_code;
|
|
|
|
|
if (icode != CODE_FOR_nothing)
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
insn_mode = GET_MODE_WIDER_MODE (insn_mode);
|
|
|
|
|
}
|
|
|
|
|
if (insn_mode == VOIDmode)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Make a place to write the result of the instruction. */
|
|
|
|
|
result = target;
|
|
|
|
|
if (! (result != 0
|
|
|
|
|
&& GET_CODE (result) == REG
|
|
|
|
|
&& GET_MODE (result) == insn_mode
|
|
|
|
|
&& REGNO (result) >= FIRST_PSEUDO_REGISTER))
|
|
|
|
|
result = gen_reg_rtx (insn_mode);
|
|
|
|
|
|
|
|
|
|
/* Make a place to hold the source address. We will not expand
|
|
|
|
|
the actual source until we are sure that the expansion will
|
|
|
|
|
not fail -- there are trees that cannot be expanded twice. */
|
|
|
|
|
src_reg = gen_reg_rtx (Pmode);
|
|
|
|
|
|
|
|
|
|
/* Mark the beginning of the strlen sequence so we can emit the
|
|
|
|
|
source operand later. */
|
|
|
|
|
before_strlen = get_last_insn();
|
|
|
|
|
|
|
|
|
|
char_rtx = const0_rtx;
|
|
|
|
|
char_mode = insn_data[(int) icode].operand[2].mode;
|
|
|
|
|
if (! (*insn_data[(int) icode].operand[2].predicate) (char_rtx,
|
|
|
|
|
char_mode))
|
|
|
|
|
char_rtx = copy_to_mode_reg (char_mode, char_rtx);
|
|
|
|
|
|
|
|
|
|
pat = GEN_FCN (icode) (result, gen_rtx_MEM (BLKmode, src_reg),
|
|
|
|
|
char_rtx, GEN_INT (align));
|
|
|
|
|
if (! pat)
|
|
|
|
|
return 0;
|
|
|
|
|
emit_insn (pat);
|
|
|
|
|
|
|
|
|
|
/* Now that we are assured of success, expand the source. */
|
|
|
|
|
start_sequence ();
|
|
|
|
|
pat = memory_address (BLKmode,
|
|
|
|
|
expand_expr (src, src_reg, ptr_mode, EXPAND_SUM));
|
|
|
|
|
if (pat != src_reg)
|
|
|
|
|
emit_move_insn (src_reg, pat);
|
|
|
|
|
pat = gen_sequence ();
|
|
|
|
|
end_sequence ();
|
|
|
|
|
|
|
|
|
|
if (before_strlen)
|
|
|
|
|
emit_insn_after (pat, before_strlen);
|
|
|
|
|
else
|
|
|
|
|
emit_insn_before (pat, get_insns ());
|
|
|
|
|
|
|
|
|
|
/* Return the value in the proper mode for this function. */
|
|
|
|
|
if (GET_MODE (result) == value_mode)
|
|
|
|
|
target = result;
|
|
|
|
|
else if (target != 0)
|
|
|
|
|
convert_move (target, result, 0);
|
|
|
|
|
else
|
|
|
|
|
target = convert_to_mode (value_mode, result, 0);
|
|
|
|
|
|
|
|
|
|
return target;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the strstr builtin. Return 0 if we failed the
|
|
|
|
|
caller should emit a normal call, otherwise try to get the result
|
|
|
|
|
in TARGET, if convenient (and in mode MODE if that's convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strstr (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree s1 = TREE_VALUE (arglist), s2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
tree fn;
|
|
|
|
|
const char *p1, *p2;
|
|
|
|
|
|
|
|
|
|
p2 = c_getstr (s2);
|
|
|
|
|
if (p2 == NULL)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (s1);
|
|
|
|
|
if (p1 != NULL)
|
|
|
|
|
{
|
|
|
|
|
const char *r = strstr (p1, p2);
|
|
|
|
|
|
|
|
|
|
if (r == NULL)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
/* Return an offset into the constant string argument. */
|
|
|
|
|
return expand_expr (fold (build (PLUS_EXPR, TREE_TYPE (s1),
|
|
|
|
|
s1, ssize_int (r - p1))),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (p2[0] == '\0')
|
|
|
|
|
return expand_expr (s1, target, mode, EXPAND_NORMAL);
|
|
|
|
|
|
|
|
|
|
if (p2[1] != '\0')
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
fn = built_in_decls[BUILT_IN_STRCHR];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* New argument list transforming strstr(s1, s2) to
|
|
|
|
|
strchr(s1, s2[0]). */
|
|
|
|
|
arglist =
|
|
|
|
|
build_tree_list (NULL_TREE, build_int_2 (p2[0], 0));
|
|
|
|
|
arglist = tree_cons (NULL_TREE, s1, arglist);
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the strchr builtin. Return 0 if we failed the
|
|
|
|
|
caller should emit a normal call, otherwise try to get the result
|
|
|
|
|
in TARGET, if convenient (and in mode MODE if that's convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strchr (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree s1 = TREE_VALUE (arglist), s2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
const char *p1;
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (s2) != INTEGER_CST)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (s1);
|
|
|
|
|
if (p1 != NULL)
|
|
|
|
|
{
|
|
|
|
|
char c;
|
|
|
|
|
const char *r;
|
|
|
|
|
|
|
|
|
|
if (target_char_cast (s2, &c))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
r = strchr (p1, c);
|
|
|
|
|
|
|
|
|
|
if (r == NULL)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
/* Return an offset into the constant string argument. */
|
|
|
|
|
return expand_expr (fold (build (PLUS_EXPR, TREE_TYPE (s1),
|
|
|
|
|
s1, ssize_int (r - p1))),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* FIXME: Should use here strchrM optab so that ports can optimize
|
|
|
|
|
this. */
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the strrchr builtin. Return 0 if we failed the
|
|
|
|
|
caller should emit a normal call, otherwise try to get the result
|
|
|
|
|
in TARGET, if convenient (and in mode MODE if that's convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strrchr (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree s1 = TREE_VALUE (arglist), s2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
tree fn;
|
|
|
|
|
const char *p1;
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (s2) != INTEGER_CST)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (s1);
|
|
|
|
|
if (p1 != NULL)
|
|
|
|
|
{
|
|
|
|
|
char c;
|
|
|
|
|
const char *r;
|
|
|
|
|
|
|
|
|
|
if (target_char_cast (s2, &c))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
r = strrchr (p1, c);
|
|
|
|
|
|
|
|
|
|
if (r == NULL)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
/* Return an offset into the constant string argument. */
|
|
|
|
|
return expand_expr (fold (build (PLUS_EXPR, TREE_TYPE (s1),
|
|
|
|
|
s1, ssize_int (r - p1))),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (! integer_zerop (s2))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
fn = built_in_decls[BUILT_IN_STRCHR];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Transform strrchr(s1, '\0') to strchr(s1, '\0'). */
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the strpbrk builtin. Return 0 if we failed the
|
|
|
|
|
caller should emit a normal call, otherwise try to get the result
|
|
|
|
|
in TARGET, if convenient (and in mode MODE if that's convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strpbrk (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree s1 = TREE_VALUE (arglist), s2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
tree fn;
|
|
|
|
|
const char *p1, *p2;
|
|
|
|
|
|
|
|
|
|
p2 = c_getstr (s2);
|
|
|
|
|
if (p2 == NULL)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (s1);
|
|
|
|
|
if (p1 != NULL)
|
|
|
|
|
{
|
|
|
|
|
const char *r = strpbrk (p1, p2);
|
|
|
|
|
|
|
|
|
|
if (r == NULL)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
/* Return an offset into the constant string argument. */
|
|
|
|
|
return expand_expr (fold (build (PLUS_EXPR, TREE_TYPE (s1),
|
|
|
|
|
s1, ssize_int (r - p1))),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (p2[0] == '\0')
|
|
|
|
|
{
|
|
|
|
|
/* strpbrk(x, "") == NULL.
|
|
|
|
|
Evaluate and ignore the arguments in case they had
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (s1, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (p2[1] != '\0')
|
|
|
|
|
return 0; /* Really call strpbrk. */
|
|
|
|
|
|
|
|
|
|
fn = built_in_decls[BUILT_IN_STRCHR];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* New argument list transforming strpbrk(s1, s2) to
|
|
|
|
|
strchr(s1, s2[0]). */
|
|
|
|
|
arglist =
|
|
|
|
|
build_tree_list (NULL_TREE, build_int_2 (p2[0], 0));
|
|
|
|
|
arglist = tree_cons (NULL_TREE, s1, arglist);
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Callback routine for store_by_pieces. Read GET_MODE_BITSIZE (MODE)
|
|
|
|
|
bytes from constant string DATA + OFFSET and return it as target
|
|
|
|
|
constant. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
builtin_memcpy_read_str (data, offset, mode)
|
|
|
|
|
PTR data;
|
|
|
|
|
HOST_WIDE_INT offset;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
const char *str = (const char *) data;
|
|
|
|
|
|
|
|
|
|
if (offset < 0
|
|
|
|
|
|| ((unsigned HOST_WIDE_INT) offset + GET_MODE_SIZE (mode)
|
|
|
|
|
> strlen (str) + 1))
|
|
|
|
|
abort (); /* Attempt to read past the end of constant string. */
|
|
|
|
|
|
|
|
|
|
return c_readstr (str + offset, mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the memcpy builtin, with arguments in ARGLIST.
|
|
|
|
|
Return 0 if we failed, the caller should emit a normal call, otherwise
|
|
|
|
|
try to get the result in TARGET, if convenient (and in mode MODE if
|
|
|
|
|
that's convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_memcpy (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist,
|
|
|
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree dest = TREE_VALUE (arglist);
|
|
|
|
|
tree src = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
tree len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
const char *src_str;
|
|
|
|
|
|
|
|
|
|
unsigned int src_align = get_pointer_alignment (src, BIGGEST_ALIGNMENT);
|
|
|
|
|
unsigned int dest_align
|
|
|
|
|
= get_pointer_alignment (dest, BIGGEST_ALIGNMENT);
|
|
|
|
|
rtx dest_mem, src_mem, dest_addr, len_rtx;
|
|
|
|
|
|
|
|
|
|
/* If DEST is not a pointer type, call the normal function. */
|
|
|
|
|
if (dest_align == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* If the LEN parameter is zero, return DEST. */
|
|
|
|
|
if (host_integerp (len, 1) && tree_low_cst (len, 1) == 0)
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore SRC in case it has side-effects. */
|
|
|
|
|
expand_expr (src, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return expand_expr (dest, target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If either SRC is not a pointer type, don't do this
|
|
|
|
|
operation in-line. */
|
|
|
|
|
if (src_align == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
dest_mem = get_memory_rtx (dest);
|
|
|
|
|
set_mem_align (dest_mem, dest_align);
|
|
|
|
|
len_rtx = expand_expr (len, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
src_str = c_getstr (src);
|
|
|
|
|
|
|
|
|
|
/* If SRC is a string constant and block move would be done
|
|
|
|
|
by pieces, we can avoid loading the string from memory
|
|
|
|
|
and only stored the computed constants. */
|
|
|
|
|
if (src_str
|
|
|
|
|
&& GET_CODE (len_rtx) == CONST_INT
|
|
|
|
|
&& (unsigned HOST_WIDE_INT) INTVAL (len_rtx) <= strlen (src_str) + 1
|
|
|
|
|
&& can_store_by_pieces (INTVAL (len_rtx), builtin_memcpy_read_str,
|
|
|
|
|
(PTR) src_str, dest_align))
|
|
|
|
|
{
|
|
|
|
|
store_by_pieces (dest_mem, INTVAL (len_rtx),
|
|
|
|
|
builtin_memcpy_read_str,
|
|
|
|
|
(PTR) src_str, dest_align);
|
|
|
|
|
return force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
src_mem = get_memory_rtx (src);
|
|
|
|
|
set_mem_align (src_mem, src_align);
|
|
|
|
|
|
|
|
|
|
/* Copy word part most expediently. */
|
|
|
|
|
dest_addr = emit_block_move (dest_mem, src_mem, len_rtx);
|
|
|
|
|
|
|
|
|
|
if (dest_addr == 0)
|
|
|
|
|
dest_addr = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
|
|
|
|
|
|
|
|
|
return dest_addr;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strcpy builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call, otherwise try to get
|
|
|
|
|
the result in TARGET, if convenient (and in mode MODE if that's
|
|
|
|
|
convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strcpy (exp, target, mode)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
tree fn, len;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
fn = built_in_decls[BUILT_IN_MEMCPY];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
len = c_strlen (TREE_VALUE (TREE_CHAIN (arglist)));
|
|
|
|
|
if (len == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
len = size_binop (PLUS_EXPR, len, ssize_int (1));
|
|
|
|
|
chainon (arglist, build_tree_list (NULL_TREE, len));
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Callback routine for store_by_pieces. Read GET_MODE_BITSIZE (MODE)
|
|
|
|
|
bytes from constant string DATA + OFFSET and return it as target
|
|
|
|
|
constant. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
builtin_strncpy_read_str (data, offset, mode)
|
|
|
|
|
PTR data;
|
|
|
|
|
HOST_WIDE_INT offset;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
const char *str = (const char *) data;
|
|
|
|
|
|
|
|
|
|
if ((unsigned HOST_WIDE_INT) offset > strlen (str))
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
return c_readstr (str + offset, mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strncpy builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strncpy (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist,
|
|
|
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree slen = c_strlen (TREE_VALUE (TREE_CHAIN (arglist)));
|
|
|
|
|
tree len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
tree fn;
|
|
|
|
|
|
|
|
|
|
/* We must be passed a constant len parameter. */
|
|
|
|
|
if (TREE_CODE (len) != INTEGER_CST)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* If the len parameter is zero, return the dst parameter. */
|
|
|
|
|
if (integer_zerop (len))
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore the src argument in case it has
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (TREE_VALUE (TREE_CHAIN (arglist)), const0_rtx,
|
|
|
|
|
VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
/* Return the dst parameter. */
|
|
|
|
|
return expand_expr (TREE_VALUE (arglist), target, mode,
|
|
|
|
|
EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Now, we must be passed a constant src ptr parameter. */
|
|
|
|
|
if (slen == 0 || TREE_CODE (slen) != INTEGER_CST)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
slen = size_binop (PLUS_EXPR, slen, ssize_int (1));
|
|
|
|
|
|
|
|
|
|
/* We're required to pad with trailing zeros if the requested
|
|
|
|
|
len is greater than strlen(s2)+1. In that case try to
|
|
|
|
|
use store_by_pieces, if it fails, punt. */
|
|
|
|
|
if (tree_int_cst_lt (slen, len))
|
|
|
|
|
{
|
|
|
|
|
tree dest = TREE_VALUE (arglist);
|
|
|
|
|
unsigned int dest_align
|
|
|
|
|
= get_pointer_alignment (dest, BIGGEST_ALIGNMENT);
|
|
|
|
|
const char *p = c_getstr (TREE_VALUE (TREE_CHAIN (arglist)));
|
|
|
|
|
rtx dest_mem;
|
|
|
|
|
|
|
|
|
|
if (!p || dest_align == 0 || !host_integerp (len, 1)
|
|
|
|
|
|| !can_store_by_pieces (tree_low_cst (len, 1),
|
|
|
|
|
builtin_strncpy_read_str,
|
|
|
|
|
(PTR) p, dest_align))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
dest_mem = get_memory_rtx (dest);
|
|
|
|
|
store_by_pieces (dest_mem, tree_low_cst (len, 1),
|
|
|
|
|
builtin_strncpy_read_str,
|
|
|
|
|
(PTR) p, dest_align);
|
|
|
|
|
return force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* OK transform into builtin memcpy. */
|
|
|
|
|
fn = built_in_decls[BUILT_IN_MEMCPY];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Callback routine for store_by_pieces. Read GET_MODE_BITSIZE (MODE)
|
|
|
|
|
bytes from constant string DATA + OFFSET and return it as target
|
|
|
|
|
constant. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
builtin_memset_read_str (data, offset, mode)
|
|
|
|
|
PTR data;
|
|
|
|
|
HOST_WIDE_INT offset ATTRIBUTE_UNUSED;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
const char *c = (const char *) data;
|
|
|
|
|
char *p = alloca (GET_MODE_SIZE (mode));
|
|
|
|
|
|
|
|
|
|
memset (p, *c, GET_MODE_SIZE (mode));
|
|
|
|
|
|
|
|
|
|
return c_readstr (p, mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the memset builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call, otherwise try to get
|
|
|
|
|
the result in TARGET, if convenient (and in mode MODE if that's
|
|
|
|
|
convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_memset (exp, target, mode)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist,
|
|
|
|
|
POINTER_TYPE, INTEGER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree dest = TREE_VALUE (arglist);
|
|
|
|
|
tree val = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
tree len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
char c;
|
|
|
|
|
|
|
|
|
|
unsigned int dest_align
|
|
|
|
|
= get_pointer_alignment (dest, BIGGEST_ALIGNMENT);
|
|
|
|
|
rtx dest_mem, dest_addr, len_rtx;
|
|
|
|
|
|
|
|
|
|
/* If DEST is not a pointer type, don't do this
|
|
|
|
|
operation in-line. */
|
|
|
|
|
if (dest_align == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* If the LEN parameter is zero, return DEST. */
|
|
|
|
|
if (host_integerp (len, 1) && tree_low_cst (len, 1) == 0)
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore VAL in case it has side-effects. */
|
|
|
|
|
expand_expr (val, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return expand_expr (dest, target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (val) != INTEGER_CST)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
if (target_char_cast (val, &c))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
if (c)
|
|
|
|
|
{
|
|
|
|
|
if (!host_integerp (len, 1))
|
|
|
|
|
return 0;
|
|
|
|
|
if (!can_store_by_pieces (tree_low_cst (len, 1),
|
|
|
|
|
builtin_memset_read_str, (PTR) &c,
|
|
|
|
|
dest_align))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
dest_mem = get_memory_rtx (dest);
|
|
|
|
|
store_by_pieces (dest_mem, tree_low_cst (len, 1),
|
|
|
|
|
builtin_memset_read_str,
|
|
|
|
|
(PTR) &c, dest_align);
|
|
|
|
|
return force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
len_rtx = expand_expr (len, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
|
|
|
|
|
dest_mem = get_memory_rtx (dest);
|
|
|
|
|
set_mem_align (dest_mem, dest_align);
|
|
|
|
|
dest_addr = clear_storage (dest_mem, len_rtx);
|
|
|
|
|
|
|
|
|
|
if (dest_addr == 0)
|
|
|
|
|
dest_addr = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
|
|
|
|
|
|
|
|
|
return dest_addr;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the bzero builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_bzero (exp)
|
|
|
|
|
tree exp;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
tree dest, size, newarglist;
|
|
|
|
|
rtx result;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return NULL_RTX;
|
|
|
|
|
|
|
|
|
|
dest = TREE_VALUE (arglist);
|
|
|
|
|
size = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
|
|
|
|
|
/* New argument list transforming bzero(ptr x, int y) to
|
|
|
|
|
memset(ptr x, int 0, size_t y). This is done this way
|
|
|
|
|
so that if it isn't expanded inline, we fallback to
|
|
|
|
|
calling bzero instead of memset. */
|
|
|
|
|
|
|
|
|
|
newarglist = build_tree_list (NULL_TREE, convert (sizetype, size));
|
|
|
|
|
newarglist = tree_cons (NULL_TREE, integer_zero_node, newarglist);
|
|
|
|
|
newarglist = tree_cons (NULL_TREE, dest, newarglist);
|
|
|
|
|
|
|
|
|
|
TREE_OPERAND (exp, 1) = newarglist;
|
|
|
|
|
result = expand_builtin_memset (exp, const0_rtx, VOIDmode);
|
|
|
|
|
|
|
|
|
|
/* Always restore the original arguments. */
|
|
|
|
|
TREE_OPERAND (exp, 1) = arglist;
|
|
|
|
|
|
|
|
|
|
return result;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the memcmp or the strcmp builtin.
|
|
|
|
|
ARGLIST is the argument list for this call. Return 0 if we failed and the
|
|
|
|
|
caller should emit a normal call, otherwise try to get the result in
|
|
|
|
|
TARGET, if convenient (and in mode MODE, if that's convenient). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_memcmp (exp, arglist, target, mode)
|
|
|
|
|
tree exp ATTRIBUTE_UNUSED;
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
tree arg1, arg2, len;
|
|
|
|
|
const char *p1, *p2;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist,
|
|
|
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
arg1 = TREE_VALUE (arglist);
|
|
|
|
|
arg2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
|
|
|
|
|
/* If the len parameter is zero, return zero. */
|
|
|
|
|
if (host_integerp (len, 1) && tree_low_cst (len, 1) == 0)
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore arg1 and arg2 in case they have
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (arg1, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
expand_expr (arg2, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (arg1);
|
|
|
|
|
p2 = c_getstr (arg2);
|
|
|
|
|
|
|
|
|
|
/* If all arguments are constant, and the value of len is not greater
|
|
|
|
|
than the lengths of arg1 and arg2, evaluate at compile-time. */
|
|
|
|
|
if (host_integerp (len, 1) && p1 && p2
|
|
|
|
|
&& compare_tree_int (len, strlen (p1) + 1) <= 0
|
|
|
|
|
&& compare_tree_int (len, strlen (p2) + 1) <= 0)
|
|
|
|
|
{
|
|
|
|
|
const int r = memcmp (p1, p2, tree_low_cst (len, 1));
|
|
|
|
|
|
|
|
|
|
return (r < 0 ? constm1_rtx : (r > 0 ? const1_rtx : const0_rtx));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If len parameter is one, return an expression corresponding to
|
|
|
|
|
(*(const unsigned char*)arg1 - (const unsigned char*)arg2). */
|
|
|
|
|
if (host_integerp (len, 1) && tree_low_cst (len, 1) == 1)
|
|
|
|
|
{
|
|
|
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
|
|
|
|
tree cst_uchar_ptr_node = build_pointer_type (cst_uchar_node);
|
|
|
|
|
tree ind1 =
|
|
|
|
|
fold (build1 (CONVERT_EXPR, integer_type_node,
|
|
|
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
|
|
|
|
build1 (NOP_EXPR, cst_uchar_ptr_node, arg1))));
|
|
|
|
|
tree ind2 =
|
|
|
|
|
fold (build1 (CONVERT_EXPR, integer_type_node,
|
|
|
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
|
|
|
|
build1 (NOP_EXPR, cst_uchar_ptr_node, arg2))));
|
|
|
|
|
tree result = fold (build (MINUS_EXPR, integer_type_node, ind1, ind2));
|
|
|
|
|
return expand_expr (result, target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_cmpstrsi
|
|
|
|
|
{
|
|
|
|
|
rtx arg1_rtx, arg2_rtx, arg3_rtx;
|
|
|
|
|
rtx result;
|
|
|
|
|
rtx insn;
|
|
|
|
|
|
|
|
|
|
int arg1_align
|
|
|
|
|
= get_pointer_alignment (arg1, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
|
|
|
|
|
int arg2_align
|
|
|
|
|
= get_pointer_alignment (arg2, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
|
|
|
|
|
enum machine_mode insn_mode
|
|
|
|
|
= insn_data[(int) CODE_FOR_cmpstrsi].operand[0].mode;
|
|
|
|
|
|
|
|
|
|
/* If we don't have POINTER_TYPE, call the function. */
|
|
|
|
|
if (arg1_align == 0 || arg2_align == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Make a place to write the result of the instruction. */
|
|
|
|
|
result = target;
|
|
|
|
|
if (! (result != 0
|
|
|
|
|
&& GET_CODE (result) == REG && GET_MODE (result) == insn_mode
|
|
|
|
|
&& REGNO (result) >= FIRST_PSEUDO_REGISTER))
|
|
|
|
|
result = gen_reg_rtx (insn_mode);
|
|
|
|
|
|
|
|
|
|
arg1_rtx = get_memory_rtx (arg1);
|
|
|
|
|
arg2_rtx = get_memory_rtx (arg2);
|
|
|
|
|
arg3_rtx = expand_expr (len, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
if (!HAVE_cmpstrsi)
|
|
|
|
|
insn = NULL_RTX;
|
|
|
|
|
else
|
|
|
|
|
insn = gen_cmpstrsi (result, arg1_rtx, arg2_rtx, arg3_rtx,
|
|
|
|
|
GEN_INT (MIN (arg1_align, arg2_align)));
|
|
|
|
|
|
|
|
|
|
if (insn)
|
|
|
|
|
emit_insn (insn);
|
|
|
|
|
else
|
|
|
|
|
emit_library_call_value (memcmp_libfunc, result, LCT_PURE_MAKE_BLOCK,
|
|
|
|
|
TYPE_MODE (integer_type_node), 3,
|
|
|
|
|
XEXP (arg1_rtx, 0), Pmode,
|
|
|
|
|
XEXP (arg2_rtx, 0), Pmode,
|
|
|
|
|
convert_to_mode (TYPE_MODE (sizetype), arg3_rtx,
|
|
|
|
|
TREE_UNSIGNED (sizetype)),
|
|
|
|
|
TYPE_MODE (sizetype));
|
|
|
|
|
|
|
|
|
|
/* Return the value in the proper mode for this function. */
|
|
|
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
|
|
|
|
if (GET_MODE (result) == mode)
|
|
|
|
|
return result;
|
|
|
|
|
else if (target != 0)
|
|
|
|
|
{
|
|
|
|
|
convert_move (target, result, 0);
|
|
|
|
|
return target;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
return convert_to_mode (mode, result, 0);
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strcmp builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call, otherwise try to get
|
|
|
|
|
the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strcmp (exp, target, mode)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
tree arg1, arg2, len, len2, fn;
|
|
|
|
|
const char *p1, *p2;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
arg1 = TREE_VALUE (arglist);
|
|
|
|
|
arg2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (arg1);
|
|
|
|
|
p2 = c_getstr (arg2);
|
|
|
|
|
|
|
|
|
|
if (p1 && p2)
|
|
|
|
|
{
|
|
|
|
|
const int i = strcmp (p1, p2);
|
|
|
|
|
return (i < 0 ? constm1_rtx : (i > 0 ? const1_rtx : const0_rtx));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If either arg is "", return an expression corresponding to
|
|
|
|
|
(*(const unsigned char*)arg1 - (const unsigned char*)arg2). */
|
|
|
|
|
if ((p1 && *p1 == '\0') || (p2 && *p2 == '\0'))
|
|
|
|
|
{
|
|
|
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
|
|
|
|
tree cst_uchar_ptr_node = build_pointer_type (cst_uchar_node);
|
|
|
|
|
tree ind1 =
|
|
|
|
|
fold (build1 (CONVERT_EXPR, integer_type_node,
|
|
|
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
|
|
|
|
build1 (NOP_EXPR, cst_uchar_ptr_node, arg1))));
|
|
|
|
|
tree ind2 =
|
|
|
|
|
fold (build1 (CONVERT_EXPR, integer_type_node,
|
|
|
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
|
|
|
|
build1 (NOP_EXPR, cst_uchar_ptr_node, arg2))));
|
|
|
|
|
tree result = fold (build (MINUS_EXPR, integer_type_node, ind1, ind2));
|
|
|
|
|
return expand_expr (result, target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
len = c_strlen (arg1);
|
|
|
|
|
len2 = c_strlen (arg2);
|
|
|
|
|
|
|
|
|
|
if (len)
|
|
|
|
|
len = size_binop (PLUS_EXPR, ssize_int (1), len);
|
|
|
|
|
|
|
|
|
|
if (len2)
|
|
|
|
|
len2 = size_binop (PLUS_EXPR, ssize_int (1), len2);
|
|
|
|
|
|
|
|
|
|
/* If we don't have a constant length for the first, use the length
|
|
|
|
|
of the second, if we know it. We don't require a constant for
|
|
|
|
|
this case; some cost analysis could be done if both are available
|
|
|
|
|
but neither is constant. For now, assume they're equally cheap
|
|
|
|
|
unless one has side effects.
|
|
|
|
|
|
|
|
|
|
If both strings have constant lengths, use the smaller. This
|
|
|
|
|
could arise if optimization results in strcpy being called with
|
|
|
|
|
two fixed strings, or if the code was machine-generated. We should
|
|
|
|
|
add some code to the `memcmp' handler below to deal with such
|
|
|
|
|
situations, someday. */
|
|
|
|
|
|
|
|
|
|
if (!len || TREE_CODE (len) != INTEGER_CST)
|
|
|
|
|
{
|
|
|
|
|
if (len2 && !TREE_SIDE_EFFECTS (len2))
|
|
|
|
|
len = len2;
|
|
|
|
|
else if (len == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
else if (len2 && TREE_CODE (len2) == INTEGER_CST
|
|
|
|
|
&& tree_int_cst_lt (len2, len))
|
|
|
|
|
len = len2;
|
|
|
|
|
|
|
|
|
|
/* If both arguments have side effects, we cannot optimize. */
|
|
|
|
|
if (TREE_SIDE_EFFECTS (len))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
fn = built_in_decls[BUILT_IN_MEMCMP];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
chainon (arglist, build_tree_list (NULL_TREE, len));
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strncmp builtin. Return 0
|
|
|
|
|
if we failed the caller should emit a normal call, otherwise try to get
|
|
|
|
|
the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strncmp (exp, target, mode)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
tree fn, newarglist, len = 0;
|
|
|
|
|
tree arg1, arg2, arg3;
|
|
|
|
|
const char *p1, *p2;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist,
|
|
|
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
arg1 = TREE_VALUE (arglist);
|
|
|
|
|
arg2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
arg3 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
|
|
|
|
|
/* If the len parameter is zero, return zero. */
|
|
|
|
|
if (host_integerp (arg3, 1) && tree_low_cst (arg3, 1) == 0)
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore arg1 and arg2 in case they have
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (arg1, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
expand_expr (arg2, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
p1 = c_getstr (arg1);
|
|
|
|
|
p2 = c_getstr (arg2);
|
|
|
|
|
|
|
|
|
|
/* If all arguments are constant, evaluate at compile-time. */
|
|
|
|
|
if (host_integerp (arg3, 1) && p1 && p2)
|
|
|
|
|
{
|
|
|
|
|
const int r = strncmp (p1, p2, tree_low_cst (arg3, 1));
|
|
|
|
|
return (r < 0 ? constm1_rtx : (r > 0 ? const1_rtx : const0_rtx));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If len == 1 or (either string parameter is "" and (len >= 1)),
|
|
|
|
|
return (*(const u_char*)arg1 - *(const u_char*)arg2). */
|
|
|
|
|
if (host_integerp (arg3, 1)
|
|
|
|
|
&& (tree_low_cst (arg3, 1) == 1
|
|
|
|
|
|| (tree_low_cst (arg3, 1) > 1
|
|
|
|
|
&& ((p1 && *p1 == '\0') || (p2 && *p2 == '\0')))))
|
|
|
|
|
{
|
|
|
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
|
|
|
|
tree cst_uchar_ptr_node = build_pointer_type (cst_uchar_node);
|
|
|
|
|
tree ind1 =
|
|
|
|
|
fold (build1 (CONVERT_EXPR, integer_type_node,
|
|
|
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
|
|
|
|
build1 (NOP_EXPR, cst_uchar_ptr_node, arg1))));
|
|
|
|
|
tree ind2 =
|
|
|
|
|
fold (build1 (CONVERT_EXPR, integer_type_node,
|
|
|
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
|
|
|
|
build1 (NOP_EXPR, cst_uchar_ptr_node, arg2))));
|
|
|
|
|
tree result = fold (build (MINUS_EXPR, integer_type_node, ind1, ind2));
|
|
|
|
|
return expand_expr (result, target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If c_strlen can determine an expression for one of the string
|
|
|
|
|
lengths, and it doesn't have side effects, then call
|
|
|
|
|
expand_builtin_memcmp() using length MIN(strlen(string)+1, arg3). */
|
|
|
|
|
|
|
|
|
|
/* Perhaps one of the strings is really constant, if so prefer
|
|
|
|
|
that constant length over the other string's length. */
|
|
|
|
|
if (p1)
|
|
|
|
|
len = c_strlen (arg1);
|
|
|
|
|
else if (p2)
|
|
|
|
|
len = c_strlen (arg2);
|
|
|
|
|
|
|
|
|
|
/* If we still don't have a len, try either string arg as long
|
|
|
|
|
as they don't have side effects. */
|
|
|
|
|
if (!len && !TREE_SIDE_EFFECTS (arg1))
|
|
|
|
|
len = c_strlen (arg1);
|
|
|
|
|
if (!len && !TREE_SIDE_EFFECTS (arg2))
|
|
|
|
|
len = c_strlen (arg2);
|
|
|
|
|
/* If we still don't have a length, punt. */
|
|
|
|
|
if (!len)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
fn = built_in_decls[BUILT_IN_MEMCMP];
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Add one to the string length. */
|
|
|
|
|
len = fold (size_binop (PLUS_EXPR, len, ssize_int (1)));
|
|
|
|
|
|
|
|
|
|
/* The actual new length parameter is MIN(len,arg3). */
|
|
|
|
|
len = fold (build (MIN_EXPR, TREE_TYPE (len), len, arg3));
|
|
|
|
|
|
|
|
|
|
newarglist = build_tree_list (NULL_TREE, len);
|
|
|
|
|
newarglist = tree_cons (NULL_TREE, arg2, newarglist);
|
|
|
|
|
newarglist = tree_cons (NULL_TREE, arg1, newarglist);
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, newarglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strcat builtin.
|
|
|
|
|
Return 0 if we failed the caller should emit a normal call,
|
|
|
|
|
otherwise try to get the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strcat (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree dst = TREE_VALUE (arglist),
|
|
|
|
|
src = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
const char *p = c_getstr (src);
|
|
|
|
|
|
|
|
|
|
/* If the string length is zero, return the dst parameter. */
|
|
|
|
|
if (p && *p == '\0')
|
|
|
|
|
return expand_expr (dst, target, mode, EXPAND_NORMAL);
|
|
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strncat builtin.
|
|
|
|
|
Return 0 if we failed the caller should emit a normal call,
|
|
|
|
|
otherwise try to get the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strncat (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist,
|
|
|
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree dst = TREE_VALUE (arglist),
|
|
|
|
|
src = TREE_VALUE (TREE_CHAIN (arglist)),
|
|
|
|
|
len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
|
|
|
|
|
const char *p = c_getstr (src);
|
|
|
|
|
|
|
|
|
|
/* If the requested length is zero, or the src parameter string
|
|
|
|
|
length is zero, return the dst parameter. */
|
|
|
|
|
if (integer_zerop (len) || (p && *p == '\0'))
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore the src and len parameters in case
|
|
|
|
|
they have side-effects. */
|
|
|
|
|
expand_expr (src, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
expand_expr (len, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return expand_expr (dst, target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If the requested len is greater than or equal to the string
|
|
|
|
|
length, call strcat. */
|
|
|
|
|
if (TREE_CODE (len) == INTEGER_CST && p
|
|
|
|
|
&& compare_tree_int (len, strlen (p)) >= 0)
|
|
|
|
|
{
|
|
|
|
|
tree newarglist
|
|
|
|
|
= tree_cons (NULL_TREE, dst, build_tree_list (NULL_TREE, src));
|
|
|
|
|
tree fn = built_in_decls[BUILT_IN_STRCAT];
|
|
|
|
|
|
|
|
|
|
/* If the replacement _DECL isn't initialized, don't do the
|
|
|
|
|
transformation. */
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, newarglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strspn builtin.
|
|
|
|
|
Return 0 if we failed the caller should emit a normal call,
|
|
|
|
|
otherwise try to get the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strspn (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree s1 = TREE_VALUE (arglist), s2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
const char *p1 = c_getstr (s1), *p2 = c_getstr (s2);
|
|
|
|
|
|
|
|
|
|
/* If both arguments are constants, evaluate at compile-time. */
|
|
|
|
|
if (p1 && p2)
|
|
|
|
|
{
|
|
|
|
|
const size_t r = strspn (p1, p2);
|
|
|
|
|
return expand_expr (size_int (r), target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If either argument is "", return 0. */
|
|
|
|
|
if ((p1 && *p1 == '\0') || (p2 && *p2 == '\0'))
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore both arguments in case either one has
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (s1, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
expand_expr (s2, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand expression EXP, which is a call to the strcspn builtin.
|
|
|
|
|
Return 0 if we failed the caller should emit a normal call,
|
|
|
|
|
otherwise try to get the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_strcspn (arglist, target, mode)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
{
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree s1 = TREE_VALUE (arglist), s2 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
const char *p1 = c_getstr (s1), *p2 = c_getstr (s2);
|
|
|
|
|
|
|
|
|
|
/* If both arguments are constants, evaluate at compile-time. */
|
|
|
|
|
if (p1 && p2)
|
|
|
|
|
{
|
|
|
|
|
const size_t r = strcspn (p1, p2);
|
|
|
|
|
return expand_expr (size_int (r), target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If the first argument is "", return 0. */
|
|
|
|
|
if (p1 && *p1 == '\0')
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore argument s2 in case it has
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (s2, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If the second argument is "", return __builtin_strlen(s1). */
|
|
|
|
|
if (p2 && *p2 == '\0')
|
|
|
|
|
{
|
|
|
|
|
tree newarglist = build_tree_list (NULL_TREE, s1),
|
|
|
|
|
fn = built_in_decls[BUILT_IN_STRLEN];
|
|
|
|
|
|
|
|
|
|
/* If the replacement _DECL isn't initialized, don't do the
|
|
|
|
|
transformation. */
|
|
|
|
|
if (!fn)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, newarglist),
|
|
|
|
|
target, mode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to __builtin_saveregs, generating the result in TARGET,
|
|
|
|
|
if that's convenient. */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
expand_builtin_saveregs ()
|
|
|
|
|
{
|
|
|
|
|
rtx val, seq;
|
|
|
|
|
|
|
|
|
|
/* Don't do __builtin_saveregs more than once in a function.
|
|
|
|
|
Save the result of the first call and reuse it. */
|
|
|
|
|
if (saveregs_value != 0)
|
|
|
|
|
return saveregs_value;
|
|
|
|
|
|
|
|
|
|
/* When this function is called, it means that registers must be
|
|
|
|
|
saved on entry to this function. So we migrate the call to the
|
|
|
|
|
first insn of this function. */
|
|
|
|
|
|
|
|
|
|
start_sequence ();
|
|
|
|
|
|
|
|
|
|
#ifdef EXPAND_BUILTIN_SAVEREGS
|
|
|
|
|
/* Do whatever the machine needs done in this case. */
|
|
|
|
|
val = EXPAND_BUILTIN_SAVEREGS ();
|
|
|
|
|
#else
|
|
|
|
|
/* ??? We used to try and build up a call to the out of line function,
|
|
|
|
|
guessing about what registers needed saving etc. This became much
|
|
|
|
|
harder with __builtin_va_start, since we don't have a tree for a
|
|
|
|
|
call to __builtin_saveregs to fall back on. There was exactly one
|
|
|
|
|
port (i860) that used this code, and I'm unconvinced it could actually
|
|
|
|
|
handle the general case. So we no longer try to handle anything
|
|
|
|
|
weird and make the backend absorb the evil. */
|
|
|
|
|
|
|
|
|
|
error ("__builtin_saveregs not supported by this target");
|
|
|
|
|
val = const0_rtx;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
seq = get_insns ();
|
|
|
|
|
end_sequence ();
|
|
|
|
|
|
|
|
|
|
saveregs_value = val;
|
|
|
|
|
|
|
|
|
|
/* Put the sequence after the NOTE that starts the function. If this
|
|
|
|
|
is inside a SEQUENCE, make the outer-level insn chain current, so
|
|
|
|
|
the code is placed at the start of the function. */
|
|
|
|
|
push_topmost_sequence ();
|
|
|
|
|
emit_insns_after (seq, get_insns ());
|
|
|
|
|
pop_topmost_sequence ();
|
|
|
|
|
|
|
|
|
|
return val;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* __builtin_args_info (N) returns word N of the arg space info
|
|
|
|
|
for the current function. The number and meanings of words
|
|
|
|
|
is controlled by the definition of CUMULATIVE_ARGS. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_args_info (exp)
|
|
|
|
|
tree exp;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
int nwords = sizeof (CUMULATIVE_ARGS) / sizeof (int);
|
|
|
|
|
int *word_ptr = (int *) ¤t_function_args_info;
|
|
|
|
|
#if 0
|
|
|
|
|
/* These are used by the code below that is if 0'ed away */
|
|
|
|
|
int i;
|
|
|
|
|
tree type, elts, result;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
if (sizeof (CUMULATIVE_ARGS) % sizeof (int) != 0)
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
if (arglist != 0)
|
|
|
|
|
{
|
|
|
|
|
if (!host_integerp (TREE_VALUE (arglist), 0))
|
|
|
|
|
error ("argument of `__builtin_args_info' must be constant");
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
HOST_WIDE_INT wordnum = tree_low_cst (TREE_VALUE (arglist), 0);
|
|
|
|
|
|
|
|
|
|
if (wordnum < 0 || wordnum >= nwords)
|
|
|
|
|
error ("argument of `__builtin_args_info' out of range");
|
|
|
|
|
else
|
|
|
|
|
return GEN_INT (word_ptr[wordnum]);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
error ("missing argument in `__builtin_args_info'");
|
|
|
|
|
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
#if 0
|
|
|
|
|
for (i = 0; i < nwords; i++)
|
|
|
|
|
elts = tree_cons (NULL_TREE, build_int_2 (word_ptr[i], 0));
|
|
|
|
|
|
|
|
|
|
type = build_array_type (integer_type_node,
|
|
|
|
|
build_index_type (build_int_2 (nwords, 0)));
|
|
|
|
|
result = build (CONSTRUCTOR, type, NULL_TREE, nreverse (elts));
|
|
|
|
|
TREE_CONSTANT (result) = 1;
|
|
|
|
|
TREE_STATIC (result) = 1;
|
|
|
|
|
result = build1 (INDIRECT_REF, build_pointer_type (type), result);
|
|
|
|
|
TREE_CONSTANT (result) = 1;
|
|
|
|
|
return expand_expr (result, NULL_RTX, VOIDmode, 0);
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand ARGLIST, from a call to __builtin_next_arg. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_next_arg (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
tree fntype = TREE_TYPE (current_function_decl);
|
|
|
|
|
|
|
|
|
|
if ((TYPE_ARG_TYPES (fntype) == 0
|
|
|
|
|
|| (TREE_VALUE (tree_last (TYPE_ARG_TYPES (fntype)))
|
|
|
|
|
== void_type_node))
|
|
|
|
|
&& ! current_function_varargs)
|
|
|
|
|
{
|
|
|
|
|
error ("`va_start' used in function with fixed args");
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (arglist)
|
|
|
|
|
{
|
|
|
|
|
tree last_parm = tree_last (DECL_ARGUMENTS (current_function_decl));
|
|
|
|
|
tree arg = TREE_VALUE (arglist);
|
|
|
|
|
|
|
|
|
|
/* Strip off all nops for the sake of the comparison. This
|
|
|
|
|
is not quite the same as STRIP_NOPS. It does more.
|
|
|
|
|
We must also strip off INDIRECT_EXPR for C++ reference
|
|
|
|
|
parameters. */
|
|
|
|
|
while (TREE_CODE (arg) == NOP_EXPR
|
|
|
|
|
|| TREE_CODE (arg) == CONVERT_EXPR
|
|
|
|
|
|| TREE_CODE (arg) == NON_LVALUE_EXPR
|
|
|
|
|
|| TREE_CODE (arg) == INDIRECT_REF)
|
|
|
|
|
arg = TREE_OPERAND (arg, 0);
|
|
|
|
|
if (arg != last_parm)
|
|
|
|
|
warning ("second parameter of `va_start' not last named argument");
|
|
|
|
|
}
|
|
|
|
|
else if (! current_function_varargs)
|
|
|
|
|
/* Evidently an out of date version of <stdarg.h>; can't validate
|
|
|
|
|
va_start's second argument, but can still work as intended. */
|
|
|
|
|
warning ("`__builtin_next_arg' called without an argument");
|
|
|
|
|
|
|
|
|
|
return expand_binop (Pmode, add_optab,
|
|
|
|
|
current_function_internal_arg_pointer,
|
|
|
|
|
current_function_arg_offset_rtx,
|
|
|
|
|
NULL_RTX, 0, OPTAB_LIB_WIDEN);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Make it easier for the backends by protecting the valist argument
|
|
|
|
|
from multiple evaluations. */
|
|
|
|
|
|
|
|
|
|
static tree
|
|
|
|
|
stabilize_va_list (valist, needs_lvalue)
|
|
|
|
|
tree valist;
|
|
|
|
|
int needs_lvalue;
|
|
|
|
|
{
|
|
|
|
|
if (TREE_CODE (va_list_type_node) == ARRAY_TYPE)
|
|
|
|
|
{
|
|
|
|
|
if (TREE_SIDE_EFFECTS (valist))
|
|
|
|
|
valist = save_expr (valist);
|
|
|
|
|
|
|
|
|
|
/* For this case, the backends will be expecting a pointer to
|
|
|
|
|
TREE_TYPE (va_list_type_node), but it's possible we've
|
|
|
|
|
actually been given an array (an actual va_list_type_node).
|
|
|
|
|
So fix it. */
|
|
|
|
|
if (TREE_CODE (TREE_TYPE (valist)) == ARRAY_TYPE)
|
|
|
|
|
{
|
|
|
|
|
tree p1 = build_pointer_type (TREE_TYPE (va_list_type_node));
|
|
|
|
|
tree p2 = build_pointer_type (va_list_type_node);
|
|
|
|
|
|
|
|
|
|
valist = build1 (ADDR_EXPR, p2, valist);
|
|
|
|
|
valist = fold (build1 (NOP_EXPR, p1, valist));
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
tree pt;
|
|
|
|
|
|
|
|
|
|
if (! needs_lvalue)
|
|
|
|
|
{
|
|
|
|
|
if (! TREE_SIDE_EFFECTS (valist))
|
|
|
|
|
return valist;
|
|
|
|
|
|
|
|
|
|
pt = build_pointer_type (va_list_type_node);
|
|
|
|
|
valist = fold (build1 (ADDR_EXPR, pt, valist));
|
|
|
|
|
TREE_SIDE_EFFECTS (valist) = 1;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (TREE_SIDE_EFFECTS (valist))
|
|
|
|
|
valist = save_expr (valist);
|
|
|
|
|
valist = fold (build1 (INDIRECT_REF, TREE_TYPE (TREE_TYPE (valist)),
|
|
|
|
|
valist));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return valist;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* The "standard" implementation of va_start: just assign `nextarg' to
|
|
|
|
|
the variable. */
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
std_expand_builtin_va_start (stdarg_p, valist, nextarg)
|
|
|
|
|
int stdarg_p;
|
|
|
|
|
tree valist;
|
|
|
|
|
rtx nextarg;
|
|
|
|
|
{
|
|
|
|
|
tree t;
|
|
|
|
|
|
|
|
|
|
if (! stdarg_p)
|
|
|
|
|
{
|
|
|
|
|
/* The dummy named parameter is declared as a 'word' sized
|
|
|
|
|
object, but if a 'word' is smaller than an 'int', it would
|
|
|
|
|
have been promoted to int when it was added to the arglist. */
|
|
|
|
|
int align = PARM_BOUNDARY / BITS_PER_UNIT;
|
|
|
|
|
int size = MAX (UNITS_PER_WORD,
|
|
|
|
|
GET_MODE_SIZE (TYPE_MODE (integer_type_node)));
|
|
|
|
|
int offset = ((size + align - 1) / align) * align;
|
|
|
|
|
nextarg = plus_constant (nextarg, -offset);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
t = build (MODIFY_EXPR, TREE_TYPE (valist), valist,
|
|
|
|
|
make_tree (ptr_type_node, nextarg));
|
|
|
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
|
|
|
|
|
|
|
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand ARGLIST, which from a call to __builtin_stdarg_va_start or
|
|
|
|
|
__builtin_varargs_va_start, depending on STDARG_P. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_va_start (stdarg_p, arglist)
|
|
|
|
|
int stdarg_p;
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
rtx nextarg;
|
|
|
|
|
tree chain = arglist, valist;
|
|
|
|
|
|
|
|
|
|
if (stdarg_p)
|
|
|
|
|
nextarg = expand_builtin_next_arg (chain = TREE_CHAIN (arglist));
|
|
|
|
|
else
|
|
|
|
|
nextarg = expand_builtin_next_arg (NULL_TREE);
|
|
|
|
|
|
|
|
|
|
if (TREE_CHAIN (chain))
|
|
|
|
|
error ("too many arguments to function `va_start'");
|
|
|
|
|
|
|
|
|
|
valist = stabilize_va_list (TREE_VALUE (arglist), 1);
|
|
|
|
|
|
|
|
|
|
#ifdef EXPAND_BUILTIN_VA_START
|
|
|
|
|
EXPAND_BUILTIN_VA_START (stdarg_p, valist, nextarg);
|
|
|
|
|
#else
|
|
|
|
|
std_expand_builtin_va_start (stdarg_p, valist, nextarg);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* The "standard" implementation of va_arg: read the value from the
|
|
|
|
|
current (padded) address and increment by the (padded) size. */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
std_expand_builtin_va_arg (valist, type)
|
|
|
|
|
tree valist, type;
|
|
|
|
|
{
|
2002-05-09 20:02:13 +00:00
|
|
|
|
tree addr_tree, t, type_size = NULL;
|
|
|
|
|
tree align, alignm1;
|
|
|
|
|
tree rounded_size;
|
2002-02-01 18:16:02 +00:00
|
|
|
|
rtx addr;
|
|
|
|
|
|
|
|
|
|
/* Compute the rounded size of the type. */
|
2002-05-09 20:02:13 +00:00
|
|
|
|
align = size_int (PARM_BOUNDARY / BITS_PER_UNIT);
|
|
|
|
|
alignm1 = size_int (PARM_BOUNDARY / BITS_PER_UNIT - 1);
|
|
|
|
|
if (type == error_mark_node
|
|
|
|
|
|| (type_size = TYPE_SIZE_UNIT (TYPE_MAIN_VARIANT (type))) == NULL
|
|
|
|
|
|| TREE_OVERFLOW (type_size))
|
|
|
|
|
rounded_size = size_zero_node;
|
|
|
|
|
else
|
|
|
|
|
rounded_size = fold (build (MULT_EXPR, sizetype,
|
|
|
|
|
fold (build (TRUNC_DIV_EXPR, sizetype,
|
|
|
|
|
fold (build (PLUS_EXPR, sizetype,
|
|
|
|
|
type_size, alignm1)),
|
|
|
|
|
align)),
|
|
|
|
|
align));
|
2002-02-01 18:16:02 +00:00
|
|
|
|
|
|
|
|
|
/* Get AP. */
|
|
|
|
|
addr_tree = valist;
|
2002-05-09 20:02:13 +00:00
|
|
|
|
if (PAD_VARARGS_DOWN && ! integer_zerop (rounded_size))
|
2002-02-01 18:16:02 +00:00
|
|
|
|
{
|
|
|
|
|
/* Small args are padded downward. */
|
2002-05-09 20:02:13 +00:00
|
|
|
|
addr_tree = fold (build (PLUS_EXPR, TREE_TYPE (addr_tree), addr_tree,
|
|
|
|
|
fold (build (COND_EXPR, sizetype,
|
|
|
|
|
fold (build (GT_EXPR, sizetype,
|
|
|
|
|
rounded_size,
|
|
|
|
|
align)),
|
|
|
|
|
size_zero_node,
|
|
|
|
|
fold (build (MINUS_EXPR, sizetype,
|
|
|
|
|
rounded_size,
|
|
|
|
|
type_size))))));
|
2002-02-01 18:16:02 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
addr = expand_expr (addr_tree, NULL_RTX, Pmode, EXPAND_NORMAL);
|
|
|
|
|
addr = copy_to_reg (addr);
|
|
|
|
|
|
|
|
|
|
/* Compute new value for AP. */
|
2002-05-09 20:02:13 +00:00
|
|
|
|
if (! integer_zerop (rounded_size))
|
|
|
|
|
{
|
|
|
|
|
t = build (MODIFY_EXPR, TREE_TYPE (valist), valist,
|
|
|
|
|
build (PLUS_EXPR, TREE_TYPE (valist), valist,
|
|
|
|
|
rounded_size));
|
|
|
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
|
|
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
}
|
2002-02-01 18:16:02 +00:00
|
|
|
|
|
|
|
|
|
return addr;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand __builtin_va_arg, which is not really a builtin function, but
|
|
|
|
|
a very special sort of operator. */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
expand_builtin_va_arg (valist, type)
|
|
|
|
|
tree valist, type;
|
|
|
|
|
{
|
|
|
|
|
rtx addr, result;
|
|
|
|
|
tree promoted_type, want_va_type, have_va_type;
|
|
|
|
|
|
|
|
|
|
/* Verify that valist is of the proper type. */
|
|
|
|
|
|
|
|
|
|
want_va_type = va_list_type_node;
|
|
|
|
|
have_va_type = TREE_TYPE (valist);
|
|
|
|
|
if (TREE_CODE (want_va_type) == ARRAY_TYPE)
|
|
|
|
|
{
|
|
|
|
|
/* If va_list is an array type, the argument may have decayed
|
|
|
|
|
to a pointer type, e.g. by being passed to another function.
|
|
|
|
|
In that case, unwrap both types so that we can compare the
|
|
|
|
|
underlying records. */
|
|
|
|
|
if (TREE_CODE (have_va_type) == ARRAY_TYPE
|
|
|
|
|
|| TREE_CODE (have_va_type) == POINTER_TYPE)
|
|
|
|
|
{
|
|
|
|
|
want_va_type = TREE_TYPE (want_va_type);
|
|
|
|
|
have_va_type = TREE_TYPE (have_va_type);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
if (TYPE_MAIN_VARIANT (want_va_type) != TYPE_MAIN_VARIANT (have_va_type))
|
|
|
|
|
{
|
|
|
|
|
error ("first argument to `va_arg' not of type `va_list'");
|
|
|
|
|
addr = const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate a diagnostic for requesting data of a type that cannot
|
|
|
|
|
be passed through `...' due to type promotion at the call site. */
|
|
|
|
|
else if ((promoted_type = (*lang_type_promotes_to) (type)) != NULL_TREE)
|
|
|
|
|
{
|
|
|
|
|
const char *name = "<anonymous type>", *pname = 0;
|
2002-05-09 20:02:13 +00:00
|
|
|
|
static bool gave_help;
|
2002-02-01 18:16:02 +00:00
|
|
|
|
|
|
|
|
|
if (TYPE_NAME (type))
|
|
|
|
|
{
|
|
|
|
|
if (TREE_CODE (TYPE_NAME (type)) == IDENTIFIER_NODE)
|
|
|
|
|
name = IDENTIFIER_POINTER (TYPE_NAME (type));
|
|
|
|
|
else if (TREE_CODE (TYPE_NAME (type)) == TYPE_DECL
|
|
|
|
|
&& DECL_NAME (TYPE_NAME (type)))
|
|
|
|
|
name = IDENTIFIER_POINTER (DECL_NAME (TYPE_NAME (type)));
|
|
|
|
|
}
|
|
|
|
|
if (TYPE_NAME (promoted_type))
|
|
|
|
|
{
|
|
|
|
|
if (TREE_CODE (TYPE_NAME (promoted_type)) == IDENTIFIER_NODE)
|
|
|
|
|
pname = IDENTIFIER_POINTER (TYPE_NAME (promoted_type));
|
|
|
|
|
else if (TREE_CODE (TYPE_NAME (promoted_type)) == TYPE_DECL
|
|
|
|
|
&& DECL_NAME (TYPE_NAME (promoted_type)))
|
|
|
|
|
pname = IDENTIFIER_POINTER (DECL_NAME (TYPE_NAME (promoted_type)));
|
|
|
|
|
}
|
|
|
|
|
|
2002-05-09 20:02:13 +00:00
|
|
|
|
/* Unfortunately, this is merely undefined, rather than a constraint
|
|
|
|
|
violation, so we cannot make this an error. If this call is never
|
|
|
|
|
executed, the program is still strictly conforming. */
|
|
|
|
|
warning ("`%s' is promoted to `%s' when passed through `...'",
|
|
|
|
|
name, pname);
|
2002-02-01 18:16:02 +00:00
|
|
|
|
if (! gave_help)
|
|
|
|
|
{
|
2002-05-09 20:02:13 +00:00
|
|
|
|
gave_help = true;
|
|
|
|
|
warning ("(so you should pass `%s' not `%s' to `va_arg')",
|
|
|
|
|
pname, name);
|
2002-02-01 18:16:02 +00:00
|
|
|
|
}
|
|
|
|
|
|
2002-05-09 20:02:13 +00:00
|
|
|
|
/* We can, however, treat "undefined" any way we please.
|
|
|
|
|
Call abort to encourage the user to fix the program. */
|
|
|
|
|
expand_builtin_trap ();
|
|
|
|
|
|
|
|
|
|
/* This is dead code, but go ahead and finish so that the
|
|
|
|
|
mode of the result comes out right. */
|
2002-02-01 18:16:02 +00:00
|
|
|
|
addr = const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
/* Make it easier for the backends by protecting the valist argument
|
|
|
|
|
from multiple evaluations. */
|
|
|
|
|
valist = stabilize_va_list (valist, 0);
|
|
|
|
|
|
|
|
|
|
#ifdef EXPAND_BUILTIN_VA_ARG
|
|
|
|
|
addr = EXPAND_BUILTIN_VA_ARG (valist, type);
|
|
|
|
|
#else
|
|
|
|
|
addr = std_expand_builtin_va_arg (valist, type);
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (addr) != Pmode)
|
|
|
|
|
addr = convert_memory_address (Pmode, addr);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
result = gen_rtx_MEM (TYPE_MODE (type), addr);
|
|
|
|
|
set_mem_alias_set (result, get_varargs_alias_set ());
|
|
|
|
|
|
|
|
|
|
return result;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand ARGLIST, from a call to __builtin_va_end. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_va_end (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
tree valist = TREE_VALUE (arglist);
|
|
|
|
|
|
|
|
|
|
#ifdef EXPAND_BUILTIN_VA_END
|
|
|
|
|
valist = stabilize_va_list (valist, 0);
|
|
|
|
|
EXPAND_BUILTIN_VA_END(arglist);
|
|
|
|
|
#else
|
|
|
|
|
/* Evaluate for side effects, if needed. I hate macros that don't
|
|
|
|
|
do that. */
|
|
|
|
|
if (TREE_SIDE_EFFECTS (valist))
|
|
|
|
|
expand_expr (valist, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand ARGLIST, from a call to __builtin_va_copy. We do this as a
|
|
|
|
|
builtin rather than just as an assignment in stdarg.h because of the
|
|
|
|
|
nastiness of array-type va_list types. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_va_copy (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
tree dst, src, t;
|
|
|
|
|
|
|
|
|
|
dst = TREE_VALUE (arglist);
|
|
|
|
|
src = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
|
|
|
|
|
dst = stabilize_va_list (dst, 1);
|
|
|
|
|
src = stabilize_va_list (src, 0);
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (va_list_type_node) != ARRAY_TYPE)
|
|
|
|
|
{
|
|
|
|
|
t = build (MODIFY_EXPR, va_list_type_node, dst, src);
|
|
|
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
|
|
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
rtx dstb, srcb, size;
|
|
|
|
|
|
|
|
|
|
/* Evaluate to pointers. */
|
|
|
|
|
dstb = expand_expr (dst, NULL_RTX, Pmode, EXPAND_NORMAL);
|
|
|
|
|
srcb = expand_expr (src, NULL_RTX, Pmode, EXPAND_NORMAL);
|
|
|
|
|
size = expand_expr (TYPE_SIZE_UNIT (va_list_type_node), NULL_RTX,
|
|
|
|
|
VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (dstb) != Pmode)
|
|
|
|
|
dstb = convert_memory_address (Pmode, dstb);
|
|
|
|
|
|
|
|
|
|
if (GET_MODE (srcb) != Pmode)
|
|
|
|
|
srcb = convert_memory_address (Pmode, srcb);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* "Dereference" to BLKmode memories. */
|
|
|
|
|
dstb = gen_rtx_MEM (BLKmode, dstb);
|
|
|
|
|
set_mem_alias_set (dstb, get_alias_set (TREE_TYPE (TREE_TYPE (dst))));
|
|
|
|
|
set_mem_align (dstb, TYPE_ALIGN (va_list_type_node));
|
|
|
|
|
srcb = gen_rtx_MEM (BLKmode, srcb);
|
|
|
|
|
set_mem_alias_set (srcb, get_alias_set (TREE_TYPE (TREE_TYPE (src))));
|
|
|
|
|
set_mem_align (srcb, TYPE_ALIGN (va_list_type_node));
|
|
|
|
|
|
|
|
|
|
/* Copy. */
|
|
|
|
|
emit_block_move (dstb, srcb, size);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to one of the builtin functions __builtin_frame_address or
|
|
|
|
|
__builtin_return_address. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_frame_address (exp)
|
|
|
|
|
tree exp;
|
|
|
|
|
{
|
|
|
|
|
tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
|
|
|
|
|
/* The argument must be a nonnegative integer constant.
|
|
|
|
|
It counts the number of frames to scan up the stack.
|
|
|
|
|
The value is the return address saved in that frame. */
|
|
|
|
|
if (arglist == 0)
|
|
|
|
|
/* Warning about missing arg was already issued. */
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
else if (! host_integerp (TREE_VALUE (arglist), 1))
|
|
|
|
|
{
|
|
|
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_FRAME_ADDRESS)
|
|
|
|
|
error ("invalid arg to `__builtin_frame_address'");
|
|
|
|
|
else
|
|
|
|
|
error ("invalid arg to `__builtin_return_address'");
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
rtx tem
|
|
|
|
|
= expand_builtin_return_addr (DECL_FUNCTION_CODE (fndecl),
|
|
|
|
|
tree_low_cst (TREE_VALUE (arglist), 1),
|
|
|
|
|
hard_frame_pointer_rtx);
|
|
|
|
|
|
|
|
|
|
/* Some ports cannot access arbitrary stack frames. */
|
|
|
|
|
if (tem == NULL)
|
|
|
|
|
{
|
|
|
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_FRAME_ADDRESS)
|
|
|
|
|
warning ("unsupported arg to `__builtin_frame_address'");
|
|
|
|
|
else
|
|
|
|
|
warning ("unsupported arg to `__builtin_return_address'");
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* For __builtin_frame_address, return what we've got. */
|
|
|
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_FRAME_ADDRESS)
|
|
|
|
|
return tem;
|
|
|
|
|
|
|
|
|
|
if (GET_CODE (tem) != REG
|
|
|
|
|
&& ! CONSTANT_P (tem))
|
|
|
|
|
tem = copy_to_mode_reg (Pmode, tem);
|
|
|
|
|
return tem;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the alloca builtin, with arguments ARGLIST. Return 0 if
|
|
|
|
|
we failed and the caller should emit a normal call, otherwise try to get
|
|
|
|
|
the result in TARGET, if convenient. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_alloca (arglist, target)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
{
|
|
|
|
|
rtx op0;
|
|
|
|
|
rtx result;
|
|
|
|
|
|
|
|
|
|
if (!validate_arglist (arglist, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Compute the argument. */
|
|
|
|
|
op0 = expand_expr (TREE_VALUE (arglist), NULL_RTX, VOIDmode, 0);
|
|
|
|
|
|
|
|
|
|
/* Allocate the desired space. */
|
|
|
|
|
result = allocate_dynamic_stack_space (op0, target, BITS_PER_UNIT);
|
|
|
|
|
|
|
|
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
|
|
|
|
if (GET_MODE (result) != ptr_mode)
|
|
|
|
|
result = convert_memory_address (ptr_mode, result);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
return result;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to the ffs builtin. The arguments are in ARGLIST.
|
|
|
|
|
Return 0 if a normal call should be emitted rather than expanding the
|
|
|
|
|
function in-line. If convenient, the result should be placed in TARGET.
|
|
|
|
|
SUBTARGET may be used as the target for computing one of EXP's operands. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_ffs (arglist, target, subtarget)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target, subtarget;
|
|
|
|
|
{
|
|
|
|
|
rtx op0;
|
|
|
|
|
if (!validate_arglist (arglist, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Compute the argument. */
|
|
|
|
|
op0 = expand_expr (TREE_VALUE (arglist), subtarget, VOIDmode, 0);
|
|
|
|
|
/* Compute ffs, into TARGET if possible.
|
|
|
|
|
Set TARGET to wherever the result comes back. */
|
|
|
|
|
target = expand_unop (TYPE_MODE (TREE_TYPE (TREE_VALUE (arglist))),
|
|
|
|
|
ffs_optab, op0, target, 1);
|
|
|
|
|
if (target == 0)
|
|
|
|
|
abort ();
|
|
|
|
|
return target;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If the string passed to fputs is a constant and is one character
|
|
|
|
|
long, we attempt to transform this call into __builtin_fputc(). */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_fputs (arglist, ignore, unlocked)
|
|
|
|
|
tree arglist;
|
|
|
|
|
int ignore;
|
|
|
|
|
int unlocked;
|
|
|
|
|
{
|
|
|
|
|
tree len, fn;
|
|
|
|
|
tree fn_fputc = unlocked ? built_in_decls[BUILT_IN_FPUTC_UNLOCKED]
|
|
|
|
|
: built_in_decls[BUILT_IN_FPUTC];
|
|
|
|
|
tree fn_fwrite = unlocked ? built_in_decls[BUILT_IN_FWRITE_UNLOCKED]
|
|
|
|
|
: built_in_decls[BUILT_IN_FWRITE];
|
|
|
|
|
|
|
|
|
|
/* If the return value is used, or the replacement _DECL isn't
|
|
|
|
|
initialized, don't do the transformation. */
|
|
|
|
|
if (!ignore || !fn_fputc || !fn_fwrite)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Verify the arguments in the original call. */
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Get the length of the string passed to fputs. If the length
|
|
|
|
|
can't be determined, punt. */
|
|
|
|
|
if (!(len = c_strlen (TREE_VALUE (arglist)))
|
|
|
|
|
|| TREE_CODE (len) != INTEGER_CST)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
switch (compare_tree_int (len, 1))
|
|
|
|
|
{
|
|
|
|
|
case -1: /* length is 0, delete the call entirely . */
|
|
|
|
|
{
|
|
|
|
|
/* Evaluate and ignore the argument in case it has
|
|
|
|
|
side-effects. */
|
|
|
|
|
expand_expr (TREE_VALUE (TREE_CHAIN (arglist)), const0_rtx,
|
|
|
|
|
VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
case 0: /* length is 1, call fputc. */
|
|
|
|
|
{
|
|
|
|
|
const char *p = c_getstr (TREE_VALUE (arglist));
|
|
|
|
|
|
|
|
|
|
if (p != NULL)
|
|
|
|
|
{
|
|
|
|
|
/* New argument list transforming fputs(string, stream) to
|
|
|
|
|
fputc(string[0], stream). */
|
|
|
|
|
arglist =
|
|
|
|
|
build_tree_list (NULL_TREE, TREE_VALUE (TREE_CHAIN (arglist)));
|
|
|
|
|
arglist =
|
|
|
|
|
tree_cons (NULL_TREE, build_int_2 (p[0], 0), arglist);
|
|
|
|
|
fn = fn_fputc;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
/* FALLTHROUGH */
|
|
|
|
|
case 1: /* length is greater than 1, call fwrite. */
|
|
|
|
|
{
|
|
|
|
|
tree string_arg = TREE_VALUE (arglist);
|
|
|
|
|
|
|
|
|
|
/* New argument list transforming fputs(string, stream) to
|
|
|
|
|
fwrite(string, 1, len, stream). */
|
|
|
|
|
arglist = build_tree_list (NULL_TREE, TREE_VALUE (TREE_CHAIN (arglist)));
|
|
|
|
|
arglist = tree_cons (NULL_TREE, len, arglist);
|
|
|
|
|
arglist = tree_cons (NULL_TREE, size_one_node, arglist);
|
|
|
|
|
arglist = tree_cons (NULL_TREE, string_arg, arglist);
|
|
|
|
|
fn = fn_fwrite;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
default:
|
|
|
|
|
abort ();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return expand_expr (build_function_call_expr (fn, arglist),
|
|
|
|
|
(ignore ? const0_rtx : NULL_RTX),
|
|
|
|
|
VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand a call to __builtin_expect. We return our argument and emit a
|
|
|
|
|
NOTE_INSN_EXPECTED_VALUE note. This is the expansion of __builtin_expect in
|
|
|
|
|
a non-jump context. */
|
|
|
|
|
|
|
|
|
|
static rtx
|
|
|
|
|
expand_builtin_expect (arglist, target)
|
|
|
|
|
tree arglist;
|
|
|
|
|
rtx target;
|
|
|
|
|
{
|
|
|
|
|
tree exp, c;
|
|
|
|
|
rtx note, rtx_c;
|
|
|
|
|
|
|
|
|
|
if (arglist == NULL_TREE
|
|
|
|
|
|| TREE_CHAIN (arglist) == NULL_TREE)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
exp = TREE_VALUE (arglist);
|
|
|
|
|
c = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
|
|
|
|
|
if (TREE_CODE (c) != INTEGER_CST)
|
|
|
|
|
{
|
|
|
|
|
error ("second arg to `__builtin_expect' must be a constant");
|
|
|
|
|
c = integer_zero_node;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
target = expand_expr (exp, target, VOIDmode, EXPAND_NORMAL);
|
|
|
|
|
|
|
|
|
|
/* Don't bother with expected value notes for integral constants. */
|
|
|
|
|
if (GET_CODE (target) != CONST_INT)
|
|
|
|
|
{
|
|
|
|
|
/* We do need to force this into a register so that we can be
|
|
|
|
|
moderately sure to be able to correctly interpret the branch
|
|
|
|
|
condition later. */
|
|
|
|
|
target = force_reg (GET_MODE (target), target);
|
|
|
|
|
|
|
|
|
|
rtx_c = expand_expr (c, NULL_RTX, GET_MODE (target), EXPAND_NORMAL);
|
|
|
|
|
|
|
|
|
|
note = emit_note (NULL, NOTE_INSN_EXPECTED_VALUE);
|
|
|
|
|
NOTE_EXPECTED_VALUE (note) = gen_rtx_EQ (VOIDmode, target, rtx_c);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return target;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Like expand_builtin_expect, except do this in a jump context. This is
|
|
|
|
|
called from do_jump if the conditional is a __builtin_expect. Return either
|
|
|
|
|
a SEQUENCE of insns to emit the jump or NULL if we cannot optimize
|
|
|
|
|
__builtin_expect. We need to optimize this at jump time so that machines
|
|
|
|
|
like the PowerPC don't turn the test into a SCC operation, and then jump
|
|
|
|
|
based on the test being 0/1. */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
expand_builtin_expect_jump (exp, if_false_label, if_true_label)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx if_false_label;
|
|
|
|
|
rtx if_true_label;
|
|
|
|
|
{
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
tree arg0 = TREE_VALUE (arglist);
|
|
|
|
|
tree arg1 = TREE_VALUE (TREE_CHAIN (arglist));
|
|
|
|
|
rtx ret = NULL_RTX;
|
|
|
|
|
|
|
|
|
|
/* Only handle __builtin_expect (test, 0) and
|
|
|
|
|
__builtin_expect (test, 1). */
|
|
|
|
|
if (TREE_CODE (TREE_TYPE (arg1)) == INTEGER_TYPE
|
|
|
|
|
&& (integer_zerop (arg1) || integer_onep (arg1)))
|
|
|
|
|
{
|
|
|
|
|
int j;
|
|
|
|
|
int num_jumps = 0;
|
|
|
|
|
|
|
|
|
|
/* If we fail to locate an appropriate conditional jump, we'll
|
|
|
|
|
fall back to normal evaluation. Ensure that the expression
|
|
|
|
|
can be re-evaluated. */
|
|
|
|
|
switch (unsafe_for_reeval (arg0))
|
|
|
|
|
{
|
|
|
|
|
case 0: /* Safe. */
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case 1: /* Mildly unsafe. */
|
|
|
|
|
arg0 = unsave_expr (arg0);
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case 2: /* Wildly unsafe. */
|
|
|
|
|
return NULL_RTX;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Expand the jump insns. */
|
|
|
|
|
start_sequence ();
|
|
|
|
|
do_jump (arg0, if_false_label, if_true_label);
|
|
|
|
|
ret = gen_sequence ();
|
|
|
|
|
end_sequence ();
|
|
|
|
|
|
|
|
|
|
/* Now that the __builtin_expect has been validated, go through and add
|
|
|
|
|
the expect's to each of the conditional jumps. If we run into an
|
|
|
|
|
error, just give up and generate the 'safe' code of doing a SCC
|
|
|
|
|
operation and then doing a branch on that. */
|
|
|
|
|
for (j = 0; j < XVECLEN (ret, 0); j++)
|
|
|
|
|
{
|
|
|
|
|
rtx insn = XVECEXP (ret, 0, j);
|
|
|
|
|
rtx pattern;
|
|
|
|
|
|
|
|
|
|
if (GET_CODE (insn) == JUMP_INSN && any_condjump_p (insn)
|
|
|
|
|
&& (pattern = pc_set (insn)) != NULL_RTX)
|
|
|
|
|
{
|
|
|
|
|
rtx ifelse = SET_SRC (pattern);
|
|
|
|
|
rtx label;
|
|
|
|
|
int taken;
|
|
|
|
|
|
|
|
|
|
if (GET_CODE (ifelse) != IF_THEN_ELSE)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
if (GET_CODE (XEXP (ifelse, 1)) == LABEL_REF)
|
|
|
|
|
{
|
|
|
|
|
taken = 1;
|
|
|
|
|
label = XEXP (XEXP (ifelse, 1), 0);
|
|
|
|
|
}
|
|
|
|
|
/* An inverted jump reverses the probabilities. */
|
|
|
|
|
else if (GET_CODE (XEXP (ifelse, 2)) == LABEL_REF)
|
|
|
|
|
{
|
|
|
|
|
taken = 0;
|
|
|
|
|
label = XEXP (XEXP (ifelse, 2), 0);
|
|
|
|
|
}
|
|
|
|
|
/* We shouldn't have to worry about conditional returns during
|
|
|
|
|
the expansion stage, but handle it gracefully anyway. */
|
|
|
|
|
else if (GET_CODE (XEXP (ifelse, 1)) == RETURN)
|
|
|
|
|
{
|
|
|
|
|
taken = 1;
|
|
|
|
|
label = NULL_RTX;
|
|
|
|
|
}
|
|
|
|
|
/* An inverted return reverses the probabilities. */
|
|
|
|
|
else if (GET_CODE (XEXP (ifelse, 2)) == RETURN)
|
|
|
|
|
{
|
|
|
|
|
taken = 0;
|
|
|
|
|
label = NULL_RTX;
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
/* If the test is expected to fail, reverse the
|
|
|
|
|
probabilities. */
|
|
|
|
|
if (integer_zerop (arg1))
|
|
|
|
|
taken = 1 - taken;
|
|
|
|
|
|
|
|
|
|
/* If we are jumping to the false label, reverse the
|
|
|
|
|
probabilities. */
|
|
|
|
|
if (label == NULL_RTX)
|
|
|
|
|
; /* conditional return */
|
|
|
|
|
else if (label == if_false_label)
|
|
|
|
|
taken = 1 - taken;
|
|
|
|
|
else if (label != if_true_label)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
num_jumps++;
|
|
|
|
|
predict_insn_def (insn, PRED_BUILTIN_EXPECT, taken);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If no jumps were modified, fail and do __builtin_expect the normal
|
|
|
|
|
way. */
|
|
|
|
|
if (num_jumps == 0)
|
|
|
|
|
ret = NULL_RTX;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
|
}
|
2002-05-09 20:02:13 +00:00
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
expand_builtin_trap ()
|
|
|
|
|
{
|
|
|
|
|
#ifdef HAVE_trap
|
|
|
|
|
if (HAVE_trap)
|
|
|
|
|
emit_insn (gen_trap ());
|
|
|
|
|
else
|
|
|
|
|
#endif
|
|
|
|
|
emit_library_call (abort_libfunc, LCT_NORETURN, VOIDmode, 0);
|
|
|
|
|
emit_barrier ();
|
|
|
|
|
}
|
2002-02-01 18:16:02 +00:00
|
|
|
|
|
|
|
|
|
/* Expand an expression EXP that calls a built-in function,
|
|
|
|
|
with result going to TARGET if that's convenient
|
|
|
|
|
(and in mode MODE if that's convenient).
|
|
|
|
|
SUBTARGET may be used as the target for computing one of EXP's operands.
|
|
|
|
|
IGNORE is nonzero if the value is to be ignored. */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
expand_builtin (exp, target, subtarget, mode, ignore)
|
|
|
|
|
tree exp;
|
|
|
|
|
rtx target;
|
|
|
|
|
rtx subtarget;
|
|
|
|
|
enum machine_mode mode;
|
|
|
|
|
int ignore;
|
|
|
|
|
{
|
|
|
|
|
tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
|
|
|
|
|
|
|
|
|
if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
|
|
|
|
return (*targetm.expand_builtin) (exp, target, subtarget, mode, ignore);
|
|
|
|
|
|
|
|
|
|
/* When not optimizing, generate calls to library functions for a certain
|
|
|
|
|
set of builtins. */
|
|
|
|
|
if (!optimize && !CALLED_AS_BUILT_IN (fndecl))
|
|
|
|
|
switch (fcode)
|
|
|
|
|
{
|
|
|
|
|
case BUILT_IN_SIN:
|
|
|
|
|
case BUILT_IN_COS:
|
|
|
|
|
case BUILT_IN_SQRT:
|
|
|
|
|
case BUILT_IN_SQRTF:
|
|
|
|
|
case BUILT_IN_SQRTL:
|
|
|
|
|
case BUILT_IN_MEMSET:
|
|
|
|
|
case BUILT_IN_MEMCPY:
|
|
|
|
|
case BUILT_IN_MEMCMP:
|
|
|
|
|
case BUILT_IN_BCMP:
|
|
|
|
|
case BUILT_IN_BZERO:
|
|
|
|
|
case BUILT_IN_INDEX:
|
|
|
|
|
case BUILT_IN_RINDEX:
|
|
|
|
|
case BUILT_IN_STRCHR:
|
|
|
|
|
case BUILT_IN_STRRCHR:
|
|
|
|
|
case BUILT_IN_STRLEN:
|
|
|
|
|
case BUILT_IN_STRCPY:
|
|
|
|
|
case BUILT_IN_STRNCPY:
|
|
|
|
|
case BUILT_IN_STRNCMP:
|
|
|
|
|
case BUILT_IN_STRSTR:
|
|
|
|
|
case BUILT_IN_STRPBRK:
|
|
|
|
|
case BUILT_IN_STRCAT:
|
|
|
|
|
case BUILT_IN_STRNCAT:
|
|
|
|
|
case BUILT_IN_STRSPN:
|
|
|
|
|
case BUILT_IN_STRCSPN:
|
|
|
|
|
case BUILT_IN_STRCMP:
|
|
|
|
|
case BUILT_IN_FFS:
|
|
|
|
|
case BUILT_IN_PUTCHAR:
|
|
|
|
|
case BUILT_IN_PUTS:
|
|
|
|
|
case BUILT_IN_PRINTF:
|
|
|
|
|
case BUILT_IN_FPUTC:
|
|
|
|
|
case BUILT_IN_FPUTS:
|
|
|
|
|
case BUILT_IN_FWRITE:
|
|
|
|
|
case BUILT_IN_PUTCHAR_UNLOCKED:
|
|
|
|
|
case BUILT_IN_PUTS_UNLOCKED:
|
|
|
|
|
case BUILT_IN_PRINTF_UNLOCKED:
|
|
|
|
|
case BUILT_IN_FPUTC_UNLOCKED:
|
|
|
|
|
case BUILT_IN_FPUTS_UNLOCKED:
|
|
|
|
|
case BUILT_IN_FWRITE_UNLOCKED:
|
|
|
|
|
return expand_call (exp, target, ignore);
|
|
|
|
|
|
|
|
|
|
default:
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
switch (fcode)
|
|
|
|
|
{
|
|
|
|
|
case BUILT_IN_ABS:
|
|
|
|
|
case BUILT_IN_LABS:
|
|
|
|
|
case BUILT_IN_LLABS:
|
|
|
|
|
case BUILT_IN_IMAXABS:
|
|
|
|
|
case BUILT_IN_FABS:
|
|
|
|
|
case BUILT_IN_FABSF:
|
|
|
|
|
case BUILT_IN_FABSL:
|
|
|
|
|
/* build_function_call changes these into ABS_EXPR. */
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_CONJ:
|
|
|
|
|
case BUILT_IN_CONJF:
|
|
|
|
|
case BUILT_IN_CONJL:
|
|
|
|
|
case BUILT_IN_CREAL:
|
|
|
|
|
case BUILT_IN_CREALF:
|
|
|
|
|
case BUILT_IN_CREALL:
|
|
|
|
|
case BUILT_IN_CIMAG:
|
|
|
|
|
case BUILT_IN_CIMAGF:
|
|
|
|
|
case BUILT_IN_CIMAGL:
|
|
|
|
|
/* expand_tree_builtin changes these into CONJ_EXPR, REALPART_EXPR
|
|
|
|
|
and IMAGPART_EXPR. */
|
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_SIN:
|
|
|
|
|
case BUILT_IN_SINF:
|
|
|
|
|
case BUILT_IN_SINL:
|
|
|
|
|
case BUILT_IN_COS:
|
|
|
|
|
case BUILT_IN_COSF:
|
|
|
|
|
case BUILT_IN_COSL:
|
|
|
|
|
/* Treat these like sqrt only if unsafe math optimizations are allowed,
|
|
|
|
|
because of possible accuracy problems. */
|
|
|
|
|
if (! flag_unsafe_math_optimizations)
|
|
|
|
|
break;
|
|
|
|
|
case BUILT_IN_SQRT:
|
|
|
|
|
case BUILT_IN_SQRTF:
|
|
|
|
|
case BUILT_IN_SQRTL:
|
|
|
|
|
target = expand_builtin_mathfn (exp, target, subtarget);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_FMOD:
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_APPLY_ARGS:
|
|
|
|
|
return expand_builtin_apply_args ();
|
|
|
|
|
|
|
|
|
|
/* __builtin_apply (FUNCTION, ARGUMENTS, ARGSIZE) invokes
|
|
|
|
|
FUNCTION with a copy of the parameters described by
|
|
|
|
|
ARGUMENTS, and ARGSIZE. It returns a block of memory
|
|
|
|
|
allocated on the stack into which is stored all the registers
|
|
|
|
|
that might possibly be used for returning the result of a
|
|
|
|
|
function. ARGUMENTS is the value returned by
|
|
|
|
|
__builtin_apply_args. ARGSIZE is the number of bytes of
|
|
|
|
|
arguments that must be copied. ??? How should this value be
|
|
|
|
|
computed? We'll also need a safe worst case value for varargs
|
|
|
|
|
functions. */
|
|
|
|
|
case BUILT_IN_APPLY:
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE,
|
|
|
|
|
POINTER_TYPE, INTEGER_TYPE, VOID_TYPE)
|
|
|
|
|
&& !validate_arglist (arglist, REFERENCE_TYPE,
|
|
|
|
|
POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
int i;
|
|
|
|
|
tree t;
|
|
|
|
|
rtx ops[3];
|
|
|
|
|
|
|
|
|
|
for (t = arglist, i = 0; t; t = TREE_CHAIN (t), i++)
|
|
|
|
|
ops[i] = expand_expr (TREE_VALUE (t), NULL_RTX, VOIDmode, 0);
|
|
|
|
|
|
|
|
|
|
return expand_builtin_apply (ops[0], ops[1], ops[2]);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* __builtin_return (RESULT) causes the function to return the
|
|
|
|
|
value described by RESULT. RESULT is address of the block of
|
|
|
|
|
memory returned by __builtin_apply. */
|
|
|
|
|
case BUILT_IN_RETURN:
|
|
|
|
|
if (validate_arglist (arglist, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
expand_builtin_return (expand_expr (TREE_VALUE (arglist),
|
|
|
|
|
NULL_RTX, VOIDmode, 0));
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_SAVEREGS:
|
|
|
|
|
return expand_builtin_saveregs ();
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_ARGS_INFO:
|
|
|
|
|
return expand_builtin_args_info (exp);
|
|
|
|
|
|
|
|
|
|
/* Return the address of the first anonymous stack arg. */
|
|
|
|
|
case BUILT_IN_NEXT_ARG:
|
|
|
|
|
return expand_builtin_next_arg (arglist);
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_CLASSIFY_TYPE:
|
|
|
|
|
return expand_builtin_classify_type (arglist);
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_CONSTANT_P:
|
|
|
|
|
return expand_builtin_constant_p (exp);
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_FRAME_ADDRESS:
|
|
|
|
|
case BUILT_IN_RETURN_ADDRESS:
|
|
|
|
|
return expand_builtin_frame_address (exp);
|
|
|
|
|
|
|
|
|
|
/* Returns the address of the area where the structure is returned.
|
|
|
|
|
0 otherwise. */
|
|
|
|
|
case BUILT_IN_AGGREGATE_INCOMING_ADDRESS:
|
|
|
|
|
if (arglist != 0
|
|
|
|
|
|| ! AGGREGATE_TYPE_P (TREE_TYPE (TREE_TYPE (current_function_decl)))
|
|
|
|
|
|| GET_CODE (DECL_RTL (DECL_RESULT (current_function_decl))) != MEM)
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
else
|
|
|
|
|
return XEXP (DECL_RTL (DECL_RESULT (current_function_decl)), 0);
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_ALLOCA:
|
|
|
|
|
target = expand_builtin_alloca (arglist, target);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_FFS:
|
|
|
|
|
target = expand_builtin_ffs (arglist, target, subtarget);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRLEN:
|
|
|
|
|
target = expand_builtin_strlen (exp, target);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRCPY:
|
|
|
|
|
target = expand_builtin_strcpy (exp, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRNCPY:
|
|
|
|
|
target = expand_builtin_strncpy (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRCAT:
|
|
|
|
|
target = expand_builtin_strcat (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRNCAT:
|
|
|
|
|
target = expand_builtin_strncat (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRSPN:
|
|
|
|
|
target = expand_builtin_strspn (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRCSPN:
|
|
|
|
|
target = expand_builtin_strcspn (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRSTR:
|
|
|
|
|
target = expand_builtin_strstr (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRPBRK:
|
|
|
|
|
target = expand_builtin_strpbrk (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_INDEX:
|
|
|
|
|
case BUILT_IN_STRCHR:
|
|
|
|
|
target = expand_builtin_strchr (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_RINDEX:
|
|
|
|
|
case BUILT_IN_STRRCHR:
|
|
|
|
|
target = expand_builtin_strrchr (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_MEMCPY:
|
|
|
|
|
target = expand_builtin_memcpy (arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_MEMSET:
|
|
|
|
|
target = expand_builtin_memset (exp, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_BZERO:
|
|
|
|
|
target = expand_builtin_bzero (exp);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRCMP:
|
|
|
|
|
target = expand_builtin_strcmp (exp, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRNCMP:
|
|
|
|
|
target = expand_builtin_strncmp (exp, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_BCMP:
|
|
|
|
|
case BUILT_IN_MEMCMP:
|
|
|
|
|
target = expand_builtin_memcmp (exp, arglist, target, mode);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_SETJMP:
|
|
|
|
|
target = expand_builtin_setjmp (arglist, target);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
/* __builtin_longjmp is passed a pointer to an array of five words.
|
|
|
|
|
It's similar to the C library longjmp function but works with
|
|
|
|
|
__builtin_setjmp above. */
|
|
|
|
|
case BUILT_IN_LONGJMP:
|
|
|
|
|
if (!validate_arglist (arglist, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
|
|
|
|
break;
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
rtx buf_addr = expand_expr (TREE_VALUE (arglist), subtarget,
|
|
|
|
|
VOIDmode, 0);
|
|
|
|
|
rtx value = expand_expr (TREE_VALUE (TREE_CHAIN (arglist)),
|
|
|
|
|
NULL_RTX, VOIDmode, 0);
|
|
|
|
|
|
|
|
|
|
if (value != const1_rtx)
|
|
|
|
|
{
|
|
|
|
|
error ("__builtin_longjmp second argument must be 1");
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
expand_builtin_longjmp (buf_addr, value);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_TRAP:
|
2002-05-09 20:02:13 +00:00
|
|
|
|
expand_builtin_trap ();
|
2002-02-01 18:16:02 +00:00
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_PUTCHAR:
|
|
|
|
|
case BUILT_IN_PUTS:
|
|
|
|
|
case BUILT_IN_FPUTC:
|
|
|
|
|
case BUILT_IN_FWRITE:
|
|
|
|
|
case BUILT_IN_PUTCHAR_UNLOCKED:
|
|
|
|
|
case BUILT_IN_PUTS_UNLOCKED:
|
|
|
|
|
case BUILT_IN_FPUTC_UNLOCKED:
|
|
|
|
|
case BUILT_IN_FWRITE_UNLOCKED:
|
|
|
|
|
break;
|
|
|
|
|
case BUILT_IN_FPUTS:
|
|
|
|
|
target = expand_builtin_fputs (arglist, ignore,/*unlocked=*/ 0);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
case BUILT_IN_FPUTS_UNLOCKED:
|
|
|
|
|
target = expand_builtin_fputs (arglist, ignore,/*unlocked=*/ 1);
|
|
|
|
|
if (target)
|
|
|
|
|
return target;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
/* Various hooks for the DWARF 2 __throw routine. */
|
|
|
|
|
case BUILT_IN_UNWIND_INIT:
|
|
|
|
|
expand_builtin_unwind_init ();
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
case BUILT_IN_DWARF_CFA:
|
|
|
|
|
return virtual_cfa_rtx;
|
|
|
|
|
#ifdef DWARF2_UNWIND_INFO
|
|
|
|
|
case BUILT_IN_DWARF_FP_REGNUM:
|
|
|
|
|
return expand_builtin_dwarf_fp_regnum ();
|
|
|
|
|
case BUILT_IN_INIT_DWARF_REG_SIZES:
|
|
|
|
|
expand_builtin_init_dwarf_reg_sizes (TREE_VALUE (arglist));
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
#endif
|
|
|
|
|
case BUILT_IN_FROB_RETURN_ADDR:
|
|
|
|
|
return expand_builtin_frob_return_addr (TREE_VALUE (arglist));
|
|
|
|
|
case BUILT_IN_EXTRACT_RETURN_ADDR:
|
|
|
|
|
return expand_builtin_extract_return_addr (TREE_VALUE (arglist));
|
|
|
|
|
case BUILT_IN_EH_RETURN:
|
|
|
|
|
expand_builtin_eh_return (TREE_VALUE (arglist),
|
|
|
|
|
TREE_VALUE (TREE_CHAIN (arglist)));
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
#ifdef EH_RETURN_DATA_REGNO
|
|
|
|
|
case BUILT_IN_EH_RETURN_DATA_REGNO:
|
|
|
|
|
return expand_builtin_eh_return_data_regno (arglist);
|
|
|
|
|
#endif
|
|
|
|
|
case BUILT_IN_VARARGS_START:
|
|
|
|
|
return expand_builtin_va_start (0, arglist);
|
|
|
|
|
case BUILT_IN_STDARG_START:
|
|
|
|
|
return expand_builtin_va_start (1, arglist);
|
|
|
|
|
case BUILT_IN_VA_END:
|
|
|
|
|
return expand_builtin_va_end (arglist);
|
|
|
|
|
case BUILT_IN_VA_COPY:
|
|
|
|
|
return expand_builtin_va_copy (arglist);
|
|
|
|
|
case BUILT_IN_EXPECT:
|
|
|
|
|
return expand_builtin_expect (arglist, target);
|
|
|
|
|
case BUILT_IN_PREFETCH:
|
|
|
|
|
expand_builtin_prefetch (arglist);
|
|
|
|
|
return const0_rtx;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
default: /* just do library call, if unknown builtin */
|
|
|
|
|
error ("built-in function `%s' not currently supported",
|
|
|
|
|
IDENTIFIER_POINTER (DECL_NAME (fndecl)));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* The switch statement above can drop through to cause the function
|
|
|
|
|
to be called normally. */
|
|
|
|
|
return expand_call (exp, target, ignore);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Fold a call to __builtin_constant_p, if we know it will evaluate to a
|
|
|
|
|
constant. ARGLIST is the argument list of the call. */
|
|
|
|
|
|
|
|
|
|
static tree
|
|
|
|
|
fold_builtin_constant_p (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
if (arglist == 0)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
arglist = TREE_VALUE (arglist);
|
|
|
|
|
|
|
|
|
|
/* We return 1 for a numeric type that's known to be a constant
|
|
|
|
|
value at compile-time or for an aggregate type that's a
|
|
|
|
|
literal constant. */
|
|
|
|
|
STRIP_NOPS (arglist);
|
|
|
|
|
|
|
|
|
|
/* If we know this is a constant, emit the constant of one. */
|
|
|
|
|
if (TREE_CODE_CLASS (TREE_CODE (arglist)) == 'c'
|
|
|
|
|
|| (TREE_CODE (arglist) == CONSTRUCTOR
|
|
|
|
|
&& TREE_CONSTANT (arglist))
|
|
|
|
|
|| (TREE_CODE (arglist) == ADDR_EXPR
|
|
|
|
|
&& TREE_CODE (TREE_OPERAND (arglist, 0)) == STRING_CST))
|
|
|
|
|
return integer_one_node;
|
|
|
|
|
|
|
|
|
|
/* If we aren't going to be running CSE or this expression
|
|
|
|
|
has side effects, show we don't know it to be a constant.
|
|
|
|
|
Likewise if it's a pointer or aggregate type since in those
|
|
|
|
|
case we only want literals, since those are only optimized
|
|
|
|
|
when generating RTL, not later.
|
|
|
|
|
And finally, if we are compiling an initializer, not code, we
|
|
|
|
|
need to return a definite result now; there's not going to be any
|
|
|
|
|
more optimization done. */
|
|
|
|
|
if (TREE_SIDE_EFFECTS (arglist) || cse_not_expected
|
|
|
|
|
|| AGGREGATE_TYPE_P (TREE_TYPE (arglist))
|
|
|
|
|
|| POINTER_TYPE_P (TREE_TYPE (arglist))
|
|
|
|
|
|| cfun == 0)
|
|
|
|
|
return integer_zero_node;
|
|
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Fold a call to __builtin_classify_type. */
|
|
|
|
|
|
|
|
|
|
static tree
|
|
|
|
|
fold_builtin_classify_type (arglist)
|
|
|
|
|
tree arglist;
|
|
|
|
|
{
|
|
|
|
|
if (arglist == 0)
|
|
|
|
|
return build_int_2 (no_type_class, 0);
|
|
|
|
|
|
|
|
|
|
return build_int_2 (type_to_class (TREE_TYPE (TREE_VALUE (arglist))), 0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Used by constant folding to eliminate some builtin calls early. EXP is
|
|
|
|
|
the CALL_EXPR of a call to a builtin function. */
|
|
|
|
|
|
|
|
|
|
tree
|
|
|
|
|
fold_builtin (exp)
|
|
|
|
|
tree exp;
|
|
|
|
|
{
|
|
|
|
|
tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
|
|
|
|
tree arglist = TREE_OPERAND (exp, 1);
|
|
|
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
|
|
|
|
|
|
|
|
|
if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
switch (fcode)
|
|
|
|
|
{
|
|
|
|
|
case BUILT_IN_CONSTANT_P:
|
|
|
|
|
return fold_builtin_constant_p (arglist);
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_CLASSIFY_TYPE:
|
|
|
|
|
return fold_builtin_classify_type (arglist);
|
|
|
|
|
|
|
|
|
|
case BUILT_IN_STRLEN:
|
|
|
|
|
if (validate_arglist (arglist, POINTER_TYPE, VOID_TYPE))
|
|
|
|
|
{
|
|
|
|
|
tree len = c_strlen (TREE_VALUE (arglist));
|
|
|
|
|
if (len != 0)
|
|
|
|
|
return len;
|
|
|
|
|
}
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
default:
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static tree
|
|
|
|
|
build_function_call_expr (fn, arglist)
|
|
|
|
|
tree fn, arglist;
|
|
|
|
|
{
|
|
|
|
|
tree call_expr;
|
|
|
|
|
|
|
|
|
|
call_expr = build1 (ADDR_EXPR, build_pointer_type (TREE_TYPE (fn)), fn);
|
|
|
|
|
call_expr = build (CALL_EXPR, TREE_TYPE (TREE_TYPE (fn)),
|
|
|
|
|
call_expr, arglist);
|
|
|
|
|
TREE_SIDE_EFFECTS (call_expr) = 1;
|
|
|
|
|
return fold (call_expr);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* This function validates the types of a function call argument list
|
|
|
|
|
represented as a tree chain of parameters against a specified list
|
|
|
|
|
of tree_codes. If the last specifier is a 0, that represents an
|
|
|
|
|
ellipses, otherwise the last specifier must be a VOID_TYPE. */
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
|
validate_arglist VPARAMS ((tree arglist, ...))
|
|
|
|
|
{
|
|
|
|
|
enum tree_code code;
|
|
|
|
|
int res = 0;
|
|
|
|
|
|
|
|
|
|
VA_OPEN (ap, arglist);
|
|
|
|
|
VA_FIXEDARG (ap, tree, arglist);
|
|
|
|
|
|
|
|
|
|
do {
|
|
|
|
|
code = va_arg (ap, enum tree_code);
|
|
|
|
|
switch (code)
|
|
|
|
|
{
|
|
|
|
|
case 0:
|
|
|
|
|
/* This signifies an ellipses, any further arguments are all ok. */
|
|
|
|
|
res = 1;
|
|
|
|
|
goto end;
|
|
|
|
|
case VOID_TYPE:
|
|
|
|
|
/* This signifies an endlink, if no arguments remain, return
|
|
|
|
|
true, otherwise return false. */
|
|
|
|
|
res = arglist == 0;
|
|
|
|
|
goto end;
|
|
|
|
|
default:
|
|
|
|
|
/* If no parameters remain or the parameter's code does not
|
|
|
|
|
match the specified code, return false. Otherwise continue
|
|
|
|
|
checking any remaining arguments. */
|
|
|
|
|
if (arglist == 0 || code != TREE_CODE (TREE_TYPE (TREE_VALUE (arglist))))
|
|
|
|
|
goto end;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
arglist = TREE_CHAIN (arglist);
|
|
|
|
|
} while (1);
|
|
|
|
|
|
|
|
|
|
/* We need gotos here since we can only have one VA_CLOSE in a
|
|
|
|
|
function. */
|
|
|
|
|
end: ;
|
|
|
|
|
VA_CLOSE (ap);
|
|
|
|
|
|
|
|
|
|
return res;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Default version of target-specific builtin setup that does nothing. */
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
default_init_builtins ()
|
|
|
|
|
{
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Default target-specific builtin expander that does nothing. */
|
|
|
|
|
|
|
|
|
|
rtx
|
|
|
|
|
default_expand_builtin (exp, target, subtarget, mode, ignore)
|
|
|
|
|
tree exp ATTRIBUTE_UNUSED;
|
|
|
|
|
rtx target ATTRIBUTE_UNUSED;
|
|
|
|
|
rtx subtarget ATTRIBUTE_UNUSED;
|
|
|
|
|
enum machine_mode mode ATTRIBUTE_UNUSED;
|
|
|
|
|
int ignore ATTRIBUTE_UNUSED;
|
|
|
|
|
{
|
|
|
|
|
return NULL_RTX;
|
|
|
|
|
}
|