/* Functions related to invoking methods and overloaded functions.
Copyright (C) 1987, 1992, 1993 Free Software Foundation, Inc.
Contributed by Michael Tiemann (tiemann@cygnus.com)
This file is part of GNU CC.
GNU CC is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2, or (at your option)
GNU CC is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with GNU CC; see the file COPYING. If not, write to
the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. */
/* High-level class interface. */
#define obstack_chunk_alloc xmalloc
#define obstack_chunk_free free
extern int inhibit_warnings
;
extern int flag_assume_nonnull_objects
;
extern tree ctor_label
, dtor_label
;
extern tree
unary_complex_lvalue ();
/* Compute the ease with which a conversion can be performed
between an expected and the given type. */
static int convert_harshness ();
#define EVIL_HARSHNESS(ARG) ((ARG) & 1)
#define ELLIPSIS_HARSHNESS(ARG) ((ARG) & 2)
#define USER_HARSHNESS(ARG) ((ARG) & 4)
#define CONTRAVARIANT_HARSHNESS(ARG) ((ARG) & 8)
#define BASE_DERIVED_HARSHNESS(ARG) ((ARG) & 16)
#define INT_TO_BD_HARSHNESS(ARG) (((ARG) << 5) | 16)
#define INT_FROM_BD_HARSHNESS(ARG) ((ARG) >> 5)
#define INT_TO_EASY_HARSHNESS(ARG) ((ARG) << 5)
#define INT_FROM_EASY_HARSHNESS(ARG) ((ARG) >> 5)
#define ONLY_EASY_HARSHNESS(ARG) (((ARG) & 31) == 0)
#define CONST_HARSHNESS(ARG) ((ARG) & 2048)
/* Ordering function for overload resolution. */
return y
->evil
- x
->evil
;
if (CONST_HARSHNESS (y
->harshness
[0]) ^ CONST_HARSHNESS (x
->harshness
[0]))
return y
->harshness
[0] - x
->harshness
[0];
if (y
->ellipsis
- x
->ellipsis
)
return y
->ellipsis
- x
->ellipsis
;
return y
->user
- x
->user
;
if (y
->b_or_d
- x
->b_or_d
)
return y
->b_or_d
- x
->b_or_d
;
return y
->easy
- x
->easy
;
/* TYPE is the type we wish to convert to. PARM is the parameter
we have to work with. We use a somewhat arbitrary cost function
to measure this conversion. */
convert_harshness (type
, parmtype
, parm
)
register tree type
, parmtype
;
register enum tree_code codel
= TREE_CODE (type
);
register enum tree_code coder
= TREE_CODE (parmtype
);
if (TYPE_MAIN_VARIANT (parmtype
) == TYPE_MAIN_VARIANT (type
))
if (codel
== POINTER_TYPE
&& (coder
== METHOD_TYPE
|| coder
== FUNCTION_TYPE
))
int harshness
, new_harshness
;
/* Get to the METHOD_TYPE or FUNCTION_TYPE that this might be. */
if (coder
!= TREE_CODE (type
))
/* We allow the default conversion between function type
and pointer-to-function type for free. */
/* Compare return types. */
p2
= TREE_TYPE (parmtype
);
new_harshness
= convert_harshness (p1
, p2
, NULL_TREE
);
if (BASE_DERIVED_HARSHNESS (new_harshness
))
/* This only works for pointers. */
if (TREE_CODE (p1
) != POINTER_TYPE
&& TREE_CODE (p1
) != REFERENCE_TYPE
)
if (CONTRAVARIANT_HARSHNESS (new_harshness
))
binfo
= get_binfo (p2
, p1
, 0);
binfo
= get_binfo (p1
, p2
, 0);
if (! BINFO_OFFSET_ZEROP (binfo
))
static int explained
= 0;
if (CONTRAVARIANT_HARSHNESS (new_harshness
))
message_2_types (sorry
, "cannot cast `%d' to `%d' at function call site", p2
, p1
);
message_2_types (sorry
, "cannot cast `%d' to `%d' at function call site", p1
, p2
);
sorry ("(because pointer values change during conversion)");
harshness
|= new_harshness
;
p1
= TYPE_ARG_TYPES (type
);
p2
= TYPE_ARG_TYPES (parmtype
);
while (p1
&& TREE_VALUE (p1
) != void_type_node
&& p2
&& TREE_VALUE (p2
) != void_type_node
)
new_harshness
= convert_harshness (TREE_VALUE (p1
), TREE_VALUE (p2
), NULL_TREE
);
if (EVIL_HARSHNESS (new_harshness
))
if (BASE_DERIVED_HARSHNESS (new_harshness
))
/* This only works for pointers and references. */
if (TREE_CODE (TREE_VALUE (p1
)) != POINTER_TYPE
&& TREE_CODE (TREE_VALUE (p1
)) != REFERENCE_TYPE
)
new_harshness
^= CONTRAVARIANT_HARSHNESS (new_harshness
);
harshness
|= new_harshness
;
/* This trick allows use to accumulate easy type
conversions without messing up the bits that encode
info about more involved things. */
else if (ONLY_EASY_HARSHNESS (new_harshness
))
harshness
+= new_harshness
;
harshness
|= new_harshness
;
return p1
? 1 : (harshness
| ELLIPSIS_HARSHNESS (-1));
return harshness
| (TREE_PURPOSE (p1
) == NULL_TREE
);
else if (codel
== POINTER_TYPE
&& coder
== OFFSET_TYPE
)
/* XXX: Note this is set a few times, but it's never actually
/* Get to the OFFSET_TYPE that this might be. */
if (coder
!= TREE_CODE (type
))
if (TYPE_OFFSET_BASETYPE (type
) == TYPE_OFFSET_BASETYPE (parmtype
))
else if (UNIQUELY_DERIVED_FROM_P (TYPE_OFFSET_BASETYPE (type
),
TYPE_OFFSET_BASETYPE (parmtype
)))
harshness
= INT_TO_BD_HARSHNESS (1);
else if (UNIQUELY_DERIVED_FROM_P (TYPE_OFFSET_BASETYPE (parmtype
),
TYPE_OFFSET_BASETYPE (type
)))
harshness
= CONTRAVARIANT_HARSHNESS (-1);
/* Now test the OFFSET_TYPE's target compatibility. */
parmtype
= TREE_TYPE (parmtype
);
if (coder
== UNKNOWN_TYPE
)
if (codel
== FUNCTION_TYPE
|| (codel
== POINTER_TYPE
&& (TREE_CODE (TREE_TYPE (type
)) == FUNCTION_TYPE
|| TREE_CODE (TREE_TYPE (type
)) == METHOD_TYPE
)))
if (codel
== ENUMERAL_TYPE
|| codel
== INTEGER_TYPE
)
/* Control equivalence of ints an enums. */
if (codel
== ENUMERAL_TYPE
&& flag_int_enum_equivalence
== 0)
/* Enums can be converted to ints, but not vice-versa. */
if (coder
!= ENUMERAL_TYPE
|| TYPE_MAIN_VARIANT (type
) != TYPE_MAIN_VARIANT (parmtype
))
/* else enums and ints (almost) freely interconvert. */
if (coder
== INTEGER_TYPE
|| coder
== ENUMERAL_TYPE
)
int easy
= TREE_UNSIGNED (type
) ^ TREE_UNSIGNED (parmtype
);
if (TYPE_MODE (type
) != TYPE_MODE (parmtype
))
return INT_TO_EASY_HARSHNESS (easy
);
else if (coder
== REAL_TYPE
)
return INT_TO_EASY_HARSHNESS (4);
/* Shun converting between float and double if a choice exists. */
if (TYPE_MODE (type
) != TYPE_MODE (parmtype
))
return INT_TO_EASY_HARSHNESS (2);
else if (coder
== INTEGER_TYPE
|| coder
== ENUMERAL_TYPE
)
return INT_TO_EASY_HARSHNESS (4);
/* convert arrays which have not previously been converted. */
/* Conversions among pointers */
if (codel
== POINTER_TYPE
&& coder
== POINTER_TYPE
)
register tree ttl
= TYPE_MAIN_VARIANT (TREE_TYPE (type
));
register tree ttr
= TYPE_MAIN_VARIANT (TREE_TYPE (parmtype
));
int penalty
= 4 * (ttl
!= ttr
);
/* Anything converts to void *. void * converts to anything.
Since these may be `const void *' (etc.) use VOID_TYPE
instead of void_type_node.
Otherwise, the targets must be the same,
except that we do allow (at some cost) conversion
between signed and unsinged pointer types. */
if ((TREE_CODE (ttl
) == METHOD_TYPE
|| TREE_CODE (ttl
) == FUNCTION_TYPE
)
&& TREE_CODE (ttl
) == TREE_CODE (ttr
))
if (comptypes (ttl
, ttr
, -1))
return INT_TO_EASY_HARSHNESS (penalty
);
if (!(TREE_CODE (ttl
) == VOID_TYPE
|| TREE_CODE (ttr
) == VOID_TYPE
|| (TREE_UNSIGNED (ttl
) ^ TREE_UNSIGNED (ttr
)
&& (ttl
= unsigned_type (ttl
),
ttr
= unsigned_type (ttr
),
|| (comp_target_types (ttl
, ttr
, 0))))
return INT_TO_EASY_HARSHNESS (10);
return INT_TO_BD_HARSHNESS (0);
if (TREE_CODE (ttl
) == RECORD_TYPE
&& TREE_CODE (ttr
) == RECORD_TYPE
)
int b_or_d
= get_base_distance (ttl
, ttr
, 0, 0);
b_or_d
= get_base_distance (ttr
, ttl
, 0, 0);
return CONTRAVARIANT_HARSHNESS (-1);
return INT_TO_BD_HARSHNESS (b_or_d
);
/* If converting from a `class*' to a `void*', make it
less favorable than any inheritance relationship. */
if (TREE_CODE (ttl
) == VOID_TYPE
&& IS_AGGR_TYPE (ttr
))
return INT_TO_BD_HARSHNESS (CLASSTYPE_MAX_DEPTH (ttr
)+1);
return INT_TO_EASY_HARSHNESS (penalty
);
if (codel
== POINTER_TYPE
&& coder
== INTEGER_TYPE
)
/* This is not a bad match, but don't let it beat
integer-enum combinations. */
if (parm
&& integer_zerop (parm
))
return INT_TO_EASY_HARSHNESS (4);
/* C++: one of the types must be a reference type. */
register tree intype
= TYPE_MAIN_VARIANT (parmtype
);
register enum tree_code form
= TREE_CODE (intype
);
if (codel
== REFERENCE_TYPE
|| coder
== REFERENCE_TYPE
)
ttl
= TYPE_MAIN_VARIANT (type
);
if (codel
== REFERENCE_TYPE
)
/* When passing a non-const argument into a const reference,
dig it a little, so a non-const reference is preferred over
if (parm
&& TREE_READONLY (ttl
) && ! TREE_READONLY (parm
))
ttl
= TYPE_MAIN_VARIANT (ttl
);
intype
= TREE_TYPE (intype
);
form
= TREE_CODE (intype
);
if (form
== REFERENCE_TYPE
)
intype
= TYPE_MAIN_VARIANT (TREE_TYPE (intype
));
/* Can reference be built up? */
if (ttl
== intype
&& penalty
== 0) {
/* Because the READONLY bits and VOLATILE bits are not
always in the type, this extra check is necessary. The
problem should be fixed someplace else, and this extra
Also, if type if a reference, the readonly bits could
either be in the outer type (with reference) or on the
inner type (the thing being referenced). (mrs) */
&& ((TREE_READONLY (parm
)
&& ! (TYPE_READONLY (type
)
|| (TREE_CODE (type
) == REFERENCE_TYPE
&& TYPE_READONLY (TREE_TYPE (type
)))))
|| (TREE_SIDE_EFFECTS (parm
)
&& ! (TYPE_VOLATILE (type
)
|| (TREE_CODE (type
) == REFERENCE_TYPE
&& TYPE_VOLATILE (TREE_TYPE (type
)))))))
else if (form
== REFERENCE_TYPE
)
tree tmp
= convert_from_reference (parm
);
intype
= TYPE_MAIN_VARIANT (TREE_TYPE (tmp
));
intype
= TREE_TYPE (intype
);
while (TREE_CODE (intype
) == REFERENCE_TYPE
);
intype
= TYPE_MAIN_VARIANT (intype
);
if (TREE_UNSIGNED (ttl
) ^ TREE_UNSIGNED (intype
))
ttl
= unsigned_type (ttl
);
intype
= unsigned_type (intype
);
/* If the initializer is not an lvalue, then it does not
matter if we make life easier for the programmer
by creating a temporary variable with which to
if (parm
&& (coder
== INTEGER_TYPE
|| coder
== ENUMERAL_TYPE
return (convert_harshness (ttl
, ttr
, NULL_TREE
)
| INT_TO_EASY_HARSHNESS (penalty
));
return INT_TO_EASY_HARSHNESS (penalty
);
return INT_TO_BD_HARSHNESS (0);
/* Pointers to voids always convert for pointers. But
make them less natural than more specific matches. */
if (TREE_CODE (ttl
) == POINTER_TYPE
&& TREE_CODE (ttr
) == POINTER_TYPE
)
if (TREE_TYPE (ttl
) == void_type_node
|| TREE_TYPE (ttr
) == void_type_node
)
return INT_TO_EASY_HARSHNESS (penalty
+1);
if (parm
&& codel
!= REFERENCE_TYPE
)
return (convert_harshness (ttl
, ttr
, NULL_TREE
)
| INT_TO_EASY_HARSHNESS (penalty
));
/* Here it does matter. If this conversion is from
derived to base, allow it. Otherwise, types must
be compatible in the strong sense. */
if (TREE_CODE (ttl
) == RECORD_TYPE
&& TREE_CODE (ttr
) == RECORD_TYPE
)
int b_or_d
= get_base_distance (ttl
, ttr
, 0, 0);
b_or_d
= get_base_distance (ttr
, ttl
, 0, 0);
return CONTRAVARIANT_HARSHNESS (-1);
/* Say that this conversion is relatively painless.
If it turns out that there is a user-defined X(X&)
constructor, then that will be invoked, but that's
preferable to dealing with other user-defined conversions
that may produce surprising results. */
return INT_TO_BD_HARSHNESS (b_or_d
);
if (comp_target_types (ttl
, intype
, 1))
return INT_TO_EASY_HARSHNESS (penalty
);
if (codel
== RECORD_TYPE
&& coder
== RECORD_TYPE
)
int b_or_d
= get_base_distance (type
, parmtype
, 0, 0);
b_or_d
= get_base_distance (parmtype
, type
, 0, 0);
return CONTRAVARIANT_HARSHNESS (-1);
return INT_TO_BD_HARSHNESS (b_or_d
);
/* Algorithm: Start out with no strikes against. For each argument
which requires a (subjective) hard conversion (such as between
floating point and integer), issue a strike. If there are the same
number of formal and actual parameters in the list, there will be at
least on strike, otherwise an exact match would have been found. If
there are not the same number of arguments in the type lists, we are
not dead yet: a `...' means that we can have more parms then were
declared, and if we wind up in the default argument section of the
list those can be used as well. If an exact match could be found for
one of those cases, return it immediately. Otherwise, rank the fields
so that fields with fewer strikes are tried first.
Conversions between builtin and user-defined types are allowed, but
no function involving such a conversion is preferred to one which
does not require such a conversion. Furthermore, such conversions
compute_conversion_costs (function
, tta_in
, cp
, arglen
)
tree ttf_in
= TYPE_ARG_TYPES (TREE_TYPE (function
));
/* Start out with no strikes against. */
int ellipsis_strikes
= 0;
int strike_index
= 0, win
, lose
;
n_compute_conversion_costs
++;
cp
->arg
= tta
? TREE_VALUE (tta
) : NULL_TREE
;
cp
->u
.bad_arg
= 0; /* optimistic! */
bzero (cp
->harshness
, (arglen
+1) * sizeof (short));
if (ttf
== void_list_node
)
if (type_unknown_p (TREE_VALUE (tta
)))
/* Must perform some instantiation here. */
tree rhs
= TREE_VALUE (tta
);
tree lhstype
= TREE_VALUE (ttf
);
/* Keep quiet about possible contravariance violations. */
int old_inhibit_warnings
= inhibit_warnings
;
/* @@ This is to undo what `grokdeclarator' does to
parameter types. It really should go through
something more general. */
TREE_TYPE (tta
) = unknown_type_node
;
rhs
= instantiate_type (lhstype
, rhs
, 0);
inhibit_warnings
= old_inhibit_warnings
;
if (TREE_CODE (rhs
) == ERROR_MARK
)
harshness
= convert_harshness (lhstype
, TREE_TYPE (rhs
), rhs
);
harshness
= convert_harshness (TREE_VALUE (ttf
), TREE_TYPE (TREE_VALUE (tta
)), TREE_VALUE (tta
));
cp
->harshness
[strike_index
] = harshness
;
if (EVIL_HARSHNESS (harshness
)
|| CONTRAVARIANT_HARSHNESS (harshness
))
cp
->u
.bad_arg
= strike_index
;
else if (ELLIPSIS_HARSHNESS (harshness
))
/* This is never set by `convert_harshness'. */
else if (USER_HARSHNESS (harshness
))
else if (BASE_DERIVED_HARSHNESS (harshness
))
b_or_d_strikes
+= INT_FROM_BD_HARSHNESS (harshness
);
easy_strikes
+= INT_FROM_EASY_HARSHNESS (harshness
);
/* ran out of formals, and parmlist is fixed size. */
if (ttf
/* == void_type_node */)
else if (ttf
&& ttf
!= void_list_node
)
/* ran out of actuals, and no defaults. */
if (TREE_PURPOSE (ttf
) == NULL_TREE
)
/* Store index of first default. */
cp
->harshness
[arglen
] = strike_index
+1;
else cp
->harshness
[arglen
] = 0;
/* Argument list lengths work out, so don't need to check them again. */
/* We do not check for derived->base conversions here, since in
no case would they give evil strike counts, unless such conversions
are somehow ambiguous. */
/* See if any user-defined conversions apply.
But make sure that we do not loop. */
static int dont_convert_types
= 0;
win
= 0; /* Only get one chance to win. */
ttf
= TYPE_ARG_TYPES (TREE_TYPE (function
));
if (ttf
== void_list_node
)
lose
= cp
->harshness
[strike_index
];
if (EVIL_HARSHNESS (lose
)
|| CONTRAVARIANT_HARSHNESS (lose
))
tree actual_type
= TREE_TYPE (TREE_VALUE (tta
));
tree formal_type
= TREE_VALUE (ttf
);
if (TREE_CODE (formal_type
) == REFERENCE_TYPE
)
formal_type
= TREE_TYPE (formal_type
);
if (TREE_CODE (actual_type
) == REFERENCE_TYPE
)
actual_type
= TREE_TYPE (actual_type
);
if (formal_type
!= error_mark_node
&& actual_type
!= error_mark_node
)
formal_type
= TYPE_MAIN_VARIANT (formal_type
);
actual_type
= TYPE_MAIN_VARIANT (actual_type
);
if (TYPE_HAS_CONSTRUCTOR (formal_type
))
/* If it has a constructor for this type, try to use it. */
if (convert_to_aggr (formal_type
, TREE_VALUE (tta
), 0, 1)
/* @@ There is no way to save this result yet.
@@ So success is NULL_TREE for now. */
if (TYPE_LANG_SPECIFIC (actual_type
) && TYPE_HAS_CONVERSION (actual_type
))
if (TREE_CODE (formal_type
) == INTEGER_TYPE
&& TYPE_HAS_INT_CONVERSION (actual_type
))
else if (TREE_CODE (formal_type
) == REAL_TYPE
&& TYPE_HAS_REAL_CONVERSION (actual_type
))
tree conv
= build_type_conversion (CALL_EXPR
, TREE_VALUE (ttf
), TREE_VALUE (tta
), 0);
if (conv
== error_mark_node
)
else if (TREE_CODE (TREE_VALUE (ttf
)) == REFERENCE_TYPE
)
conv
= build_type_conversion (CALL_EXPR
, formal_type
, TREE_VALUE (tta
), 0);
if (conv
== error_mark_node
)
cp
->harshness
[strike_index
] = USER_HARSHNESS (-1);
if (cp
->u
.bad_arg
> strike_index
)
cp
->u
.bad_arg
= strike_index
;
evil_strikes
= win
? 2 : 1;
/* Const member functions get a small penalty because defaulting
to const is less useful than defaulting to non-const. */
/* This is bogus, it does not correspond to anything in the ARM.
This code will be fixed when this entire section is rewritten
to conform to the ARM. (mrs) */
if (TREE_CODE (TREE_TYPE (function
)) == METHOD_TYPE
)
if (TYPE_READONLY (TREE_TYPE (TREE_VALUE (ttf_in
))))
cp
->harshness
[0] += INT_TO_EASY_HARSHNESS (1);
/* Calling a non-const member function from a const member function
is probably invalid, but for now we let it only draw a warning.
We indicate that such a mismatch has occurred by setting the
harshness to a maximum value. */
if (TREE_CODE (TREE_TYPE (TREE_VALUE (tta_in
))) == POINTER_TYPE
&& (TYPE_READONLY (TREE_TYPE (TREE_TYPE (TREE_VALUE (tta_in
))))))
cp
->harshness
[0] |= CONST_HARSHNESS (-1);
cp
->ellipsis
= ellipsis_strikes
;
cp
->b_or_d
= b_or_d_strikes
;
/* When one of several possible overloaded functions and/or methods
can be called, choose the best candidate for overloading.
BASETYPE is the context from which we start method resolution
or NULL if we are comparing overloaded functions.
CANDIDATES is the array of candidates we have to choose from.
N_CANDIDATES is the length of CANDIDATES.
PARMS is a TREE_LIST of parameters to the function we'll ultimately
choose. It is modified in place when resolving methods. It is not
modified in place when resolving overloaded functions.
LEN is the length of the parameter list. */
static struct candidate
*
ideal_candidate (basetype
, candidates
, n_candidates
, parms
, len
)
struct candidate
*candidates
;
struct candidate
*cp
= candidates
+ n_candidates
;
qsort (candidates
, /* char *base */
n_candidates
, /* int nel */
sizeof (struct candidate
), /* int width */
rank_for_overload
); /* int (*compar)() */
/* If the best candidate requires user-defined conversions,
and its user-defined conversions are a strict subset
of all other candidates requiring user-defined conversions,
then it is, in fact, the best. */
for (i
= -1; cp
+ i
!= candidates
; i
--)
/* Check that every other candidate requires those conversions
as a strict subset of their conversions. */
if (cp
[i
].user
== cp
[-1].user
)
/* Look at subset relationship more closely. */
for (ttf
= TYPE_ARG_TYPES (TREE_TYPE (cp
[i
].function
)),
ttf0
= TYPE_ARG_TYPES (TREE_TYPE (cp
[-1].function
)),
ttf
= TREE_CHAIN (ttf
), ttf0
= TREE_CHAIN (ttf0
), index
++)
if (USER_HARSHNESS (cp
[i
].harshness
[index
]))
/* If our "best" candidate also needs a conversion,
it must be the same one. */
if (USER_HARSHNESS (cp
[-1].harshness
[index
])
&& TREE_VALUE (ttf
) != TREE_VALUE (ttf0
))
/* The best was the best. */
/* Use other rules for determining "bestness". */
/* If the best two candidates we find require user-defined
conversions, we may need to report and error message. */
if (cp
[-1].user
&& cp
[-2].user
&& (cp
[-1].b_or_d
|| cp
[-2].b_or_d
== 0))
/* If the best two methods found involved user-defined
type conversions, then we must see whether one
of them is exactly what we wanted. If not, then
/* Stash all of our parameters in safe places
so that we can perform type conversions in place. */
TREE_PURPOSE (tta
) = TREE_VALUE (tta
);
int exact_conversions
= 0;
if (DECL_STATIC_FUNCTION_P (cp
[i
].function
))
/* special note, we don't go through len parameters, because we
may only need len-1 parameters because of a call to a static
for (ttf
= TYPE_ARG_TYPES (TREE_TYPE (cp
[i
].function
)), index
= 0;
tta
= TREE_CHAIN (tta
), ttf
= TREE_CHAIN (ttf
), index
++)
if (USER_HARSHNESS (cp
[i
].harshness
[index
]))
tree this_parm
= build_type_conversion (CALL_EXPR
, TREE_VALUE (ttf
), TREE_PURPOSE (tta
), 2);
if (basetype
!= NULL_TREE
)
TREE_VALUE (tta
) = this_parm
;
if (TREE_CODE (this_parm
) != CONVERT_EXPR
&& (TREE_CODE (this_parm
) != NOP_EXPR
|| comp_target_types (TREE_TYPE (this_parm
),
TREE_TYPE (TREE_OPERAND (this_parm
, 0)), 1)))
else if (PROMOTES_TO_AGGR_TYPE (TREE_VALUE (ttf
), REFERENCE_TYPE
))
/* To get here we had to have succeeded via
TREE_VALUE (tta
) = TREE_PURPOSE (tta
);
if (exact_conversions
== cp
[i
].user
)
p1
= TYPE_ARG_TYPES (TREE_TYPE (f1
));
/* Don't complain if next best is from base class. */
tree f2
= cp
[i
].function
;
if (TREE_CODE (TREE_TYPE (f1
)) == METHOD_TYPE
&& TREE_CODE (TREE_TYPE (f2
)) == METHOD_TYPE
&& BASE_DERIVED_HARSHNESS (cp
[i
].harshness
[0])
&& cp
[best
].harshness
[0] < cp
[i
].harshness
[0])
tree p2
= TYPE_ARG_TYPES (TREE_TYPE (f2
));
if (! compparms (TREE_CHAIN (p1
), TREE_CHAIN (p2
), 1))
/* Ensure that there's nothing ambiguous about these
for (index
= 0; index
< len
; index
++)
/* Type conversions must be piecewise equivalent. */
if (USER_HARSHNESS (cp
[best
].harshness
[index
])
!= USER_HARSHNESS (cp
[i
].harshness
[index
]))
/* If there's anything we like better about the
other function, consider it ambiguous. */
if (cp
[i
].harshness
[index
] < cp
[best
].harshness
[index
])
/* If any single one it diffent, then the whole is
if (cp
[i
].harshness
[index
] != cp
[best
].harshness
[index
])
/* If we can't tell the difference between the two, it
/* If we made it to here, it means we're satisfied that
} while (cp
+ i
!= candidates
);
int exact_conversions
= cp
[best
].user
;
if (DECL_STATIC_FUNCTION_P (cp
[best
].function
))
tta
= TREE_CHAIN (parms
);
for (ttf
= TYPE_ARG_TYPES (TREE_TYPE (cp
[best
].function
)), index
= 0;
tta
= TREE_CHAIN (tta
), ttf
= TREE_CHAIN (ttf
), index
++)
if (USER_HARSHNESS (cp
[best
].harshness
[index
]))
/* We must now fill in the slot we left behind.
@@ This could be optimized to use the value previously
@@ computed by build_type_conversion in some cases. */
if (basetype
!= NULL_TREE
)
TREE_VALUE (tta
) = convert (TREE_VALUE (ttf
), TREE_PURPOSE (tta
));
else TREE_VALUE (tta
) = TREE_PURPOSE (tta
);
/* If the best two candidates we find both use default parameters,
we may need to report and error. Don't need to worry if next-best
candidate is forced to use user-defined conversion when best is not. */
&& cp
[-1].harshness
[len
] != 0 && cp
[-2].harshness
[len
] != 0)
tree tt1
= TYPE_ARG_TYPES (TREE_TYPE (cp
[-1].function
));
tree tt2
= TYPE_ARG_TYPES (TREE_TYPE (cp
[-2].function
));
unsigned i
= cp
[-1].harshness
[len
];
if (cp
[-2].harshness
[len
] < i
)
i
= cp
[-2].harshness
[len
];
if (TYPE_MAIN_VARIANT (TREE_VALUE (tt1
))
!= TYPE_MAIN_VARIANT (TREE_VALUE (tt2
)))
/* These lists are not identical, so we can choose our best candidate. */
/* To get here, both lists had the same parameters up to the defaults
which were used. This is an ambiguous request. */
/* Otherwise, return our best candidate. Note that if we get candidates
from independent base classes, we have an ambiguity, even if one
argument list look a little better than another one. */
if (cp
[-1].b_or_d
&& basetype
&& TYPE_USES_MULTIPLE_INHERITANCE (basetype
))
int i
= n_candidates
- 1, best
= i
;
if (TREE_CODE (TREE_TYPE (candidates
[i
].function
)) == FUNCTION_TYPE
)
for (; i
>= 0 && candidates
[i
].user
== 0 && candidates
[i
].evil
== 0; i
--)
if (TREE_CODE (TREE_TYPE (candidates
[i
].function
)) == METHOD_TYPE
)
tree newbase
= DECL_CLASS_CONTEXT (candidates
[i
].function
);
/* newbase could be a base or a parent of base1 */
if (newbase
!= base1
&& ! UNIQUELY_DERIVED_FROM_P (newbase
, base1
)
&& ! UNIQUELY_DERIVED_FROM_P (base1
, newbase
))
error_with_aggr_type (basetype
, "ambiguous request for function from distinct base classes of type `%s'");
error (" first candidate is `%s'",
fndecl_as_string (0, candidates
[best
].function
, 1));
error (" second candidate is `%s'",
fndecl_as_string (0, candidates
[i
].function
, 1));
/* Don't accept a candidate as being ideal if it's indistinguishable
from another candidate. */
if (rank_for_overload (cp
-1, cp
-2) == 0)
/* If the types are distinguishably different (like
`long' vs. `unsigned long'), that's ok. But if they are arbitrarily
different, such as `int (*)(void)' vs. `void (*)(int)',
tree p1
= TYPE_ARG_TYPES (TREE_TYPE (cp
[-1].function
));
tree p2
= TYPE_ARG_TYPES (TREE_TYPE (cp
[-2].function
));
if (TREE_CODE (TREE_VALUE (p1
)) == POINTER_TYPE
&& TREE_CODE (TREE_TYPE (TREE_VALUE (p1
))) == FUNCTION_TYPE
&& TREE_VALUE (p1
) != TREE_VALUE (p2
))
/* In the case where there is no ideal candidate, restore
TREE_VALUE slots of PARMS from TREE_PURPOSE slots. */
TREE_VALUE (parms
) = TREE_PURPOSE (parms
);
parms
= TREE_CHAIN (parms
);
/* Assume that if the class referred to is not in the
current class hierarchy, that it may be remote.
PARENT is assumed to be of aggregate type here. */
if (TYPE_OVERLOADS_METHOD_CALL_EXPR (parent
) == 0)
if (current_class_type
== NULL_TREE
)
if (parent
== current_class_type
)
if (UNIQUELY_DERIVED_FROM_P (parent
, current_class_type
))
build_vfield_ref (datum
, type
)
int old_assume_nonnull_objects
= flag_assume_nonnull_objects
;
if (datum
== error_mark_node
)
/* Vtable references are always made from non-null objects. */
flag_assume_nonnull_objects
= 1;
if (TREE_CODE (TREE_TYPE (datum
)) == REFERENCE_TYPE
)
datum
= convert_from_reference (datum
);
if (! TYPE_USES_COMPLEX_INHERITANCE (type
))
rval
= build (COMPONENT_REF
, TREE_TYPE (CLASSTYPE_VFIELD (type
)),
datum
, CLASSTYPE_VFIELD (type
));
rval
= build_component_ref (datum
, DECL_NAME (CLASSTYPE_VFIELD (type
)), 0, 0);
flag_assume_nonnull_objects
= old_assume_nonnull_objects
;
/* Build a call to a member of an object. I.e., one that overloads
operator ()(), or is a pointer-to-function or pointer-to-method. */
build_field_call (basetype_path
, instance_ptr
, name
, parms
, err_name
)
tree instance_ptr
, name
, parms
;
if (instance_ptr
== current_class_decl
)
/* Check to see if we really have a reference to an instance variable
with `operator()()' overloaded. */
field
= IDENTIFIER_CLASS_VALUE (name
);
error ("`this' has no member named `%s'", err_name
);
if (TREE_CODE (field
) == FIELD_DECL
)
/* If it's a field, try overloading operator (),
or calling if the field is a pointer-to-function. */
instance
= build_component_ref_1 (C_C_D
, field
, 0);
if (instance
== error_mark_node
)
if (TYPE_LANG_SPECIFIC (TREE_TYPE (instance
))
&& TYPE_OVERLOADS_CALL_EXPR (TREE_TYPE (instance
)))
return build_opfncall (CALL_EXPR
, LOOKUP_NORMAL
, instance
, parms
, NULL_TREE
);
if (TREE_CODE (TREE_TYPE (instance
)) == POINTER_TYPE
)
if (TREE_CODE (TREE_TYPE (TREE_TYPE (instance
))) == FUNCTION_TYPE
)
return build_function_call (instance
, parms
);
else if (TREE_CODE (TREE_TYPE (TREE_TYPE (instance
))) == METHOD_TYPE
)
return build_function_call (instance
, tree_cons (NULL_TREE
, current_class_decl
, parms
));
/* Check to see if this is not really a reference to an instance variable
with `operator()()' overloaded. */
field
= lookup_field (basetype_path
, name
, 1, 0);
/* This can happen if the reference was ambiguous
or for visibility violations. */
if (field
== error_mark_node
)
tree ftype
= TREE_TYPE (field
);
if (TYPE_LANG_SPECIFIC (ftype
) && TYPE_OVERLOADS_CALL_EXPR (ftype
))
/* Make the next search for this field very short. */
basetype
= DECL_FIELD_CONTEXT (field
);
instance_ptr
= convert_pointer_to (basetype
, instance_ptr
);
instance
= build_indirect_ref (instance_ptr
, NULL
);
return build_opfncall (CALL_EXPR
, LOOKUP_NORMAL
,
build_component_ref_1 (instance
, field
, 0),
if (TREE_CODE (ftype
) == POINTER_TYPE
)
if (TREE_CODE (TREE_TYPE (ftype
)) == FUNCTION_TYPE
|| TREE_CODE (TREE_TYPE (ftype
)) == METHOD_TYPE
)
/* This is a member which is a pointer to function. */
= build_component_ref_1 (build_indirect_ref (instance_ptr
,
if (ref
== error_mark_node
)
return build_function_call (ref
, parms
);
else if (TREE_CODE (ftype
) == METHOD_TYPE
)
error ("invalid call via pointer-to-member function");
find_scoped_type (type
, inner_name
, inner_types
)
tree type
, inner_name
, inner_types
;
tree tags
= CLASSTYPE_TAGS (type
);
/* The TREE_PURPOSE of an enum tag (which becomes a member of the
enclosing class) is set to the name for the enum type. So, if
inner_name is `bar', and we strike `baz' for `enum bar { baz }',
then this test will be true. */
if (TREE_PURPOSE (tags
) == inner_name
)
if (inner_types
== NULL_TREE
)
return DECL_NESTED_TYPENAME (TYPE_NAME (TREE_VALUE (tags
)));
return resolve_scope_to_name (TREE_VALUE (tags
), inner_types
);
tags
= TREE_CHAIN (tags
);
/* Look for a TYPE_DECL. */
for (tags
= TYPE_FIELDS (type
); tags
; tags
= TREE_CHAIN (tags
))
if (TREE_CODE (tags
) == TYPE_DECL
&& DECL_NAME (tags
) == inner_name
)
if (inner_types
== NULL_TREE
)
return DECL_NESTED_TYPENAME (tags
);
return resolve_scope_to_name (TREE_TYPE (tags
), inner_types
);
/* Resolve an expression NAME1::NAME2::...::NAMEn to
the name that names the above nested type. INNER_TYPES
is a chain of nested type names (held together by SCOPE_REFs);
OUTER_TYPE is the type we know to enclose INNER_TYPES.
Returns NULL_TREE if there is an error. */
resolve_scope_to_name (outer_type
, inner_types
)
tree outer_type
, inner_types
;
if (outer_type
== NULL_TREE
&& current_class_type
!= NULL_TREE
)
/* We first try to look for a nesting in our current class context,
then try any enclosing classes. */
tree type
= current_class_type
;
while (type
&& (TREE_CODE (type
) == RECORD_TYPE
|| TREE_CODE (type
) == UNION_TYPE
))
tree rval
= resolve_scope_to_name (type
, inner_types
);
type
= DECL_CONTEXT (TYPE_NAME (type
));
if (TREE_CODE (inner_types
) == SCOPE_REF
)
inner_name
= TREE_OPERAND (inner_types
, 0);
inner_types
= TREE_OPERAND (inner_types
, 1);
inner_name
= inner_types
;
if (outer_type
== NULL_TREE
)
/* If we have something that's already a type by itself,
if (IDENTIFIER_HAS_TYPE_VALUE (inner_name
))
return resolve_scope_to_name (IDENTIFIER_TYPE_VALUE (inner_name
),
if (! IS_AGGR_TYPE (outer_type
))
/* Look for member classes or enums. */
tmp
= find_scoped_type (outer_type
, inner_name
, inner_types
);
/* If it's not a type in this class, then go down into the
base classes and search there. */
if (! tmp
&& TYPE_BINFO (outer_type
))
tree binfos
= TYPE_BINFO_BASETYPES (outer_type
);
int i
, n_baselinks
= binfos
? TREE_VEC_LENGTH (binfos
) : 0;
for (i
= 0; i
< n_baselinks
; i
++)
tree base_binfo
= TREE_VEC_ELT (binfos
, i
);
tmp
= find_scoped_type (BINFO_TYPE (base_binfo
),
inner_name
, inner_types
);
/* Build a method call of the form `EXP->SCOPES::NAME (PARMS)'.
This is how virtual function calls are avoided. */
build_scoped_method_call (exp
, scopes
, name
, parms
)
/* Because this syntactic form does not allow
a pointer to a base class to be `stolen',
we need not protect the derived->base conversion
@@ But we do have to check visibility privileges later. */
tree basename
= resolve_scope_to_name (NULL_TREE
, scopes
);
tree basetype
, binfo
, decl
;
tree type
= TREE_TYPE (exp
);
if (type
== error_mark_node
|| ! is_aggr_typedef (basename
, 1))
if (! IS_AGGR_TYPE (type
))
error ("base object of scoped method call is not of aggregate type");
basetype
= IDENTIFIER_TYPE_VALUE (basename
);
if (binfo
= binfo_or_else (basetype
, type
))
if (binfo
== error_mark_node
)
if (TREE_CODE (exp
) == INDIRECT_REF
)
decl
= build_indirect_ref (convert_pointer_to (binfo
,
build_unary_op (ADDR_EXPR
, exp
, 0)), NULL
);
decl
= build_scoped_ref (exp
, scopes
);
/* Call to a destructor. */
if (TREE_CODE (name
) == BIT_NOT_EXPR
)
/* Explicit call to destructor. */
name
= TREE_OPERAND (name
, 0);
if (! is_aggr_typedef (name
, 1))
if (TREE_TYPE (decl
) != IDENTIFIER_TYPE_VALUE (name
))
error_with_aggr_type (TREE_TYPE (decl
),
"qualified type `%s' does not match destructor type `%s'",
IDENTIFIER_POINTER (name
));
if (! TYPE_HAS_DESTRUCTOR (TREE_TYPE (decl
)))
error_with_aggr_type (TREE_TYPE (decl
), "type `%s' has no destructor");
return build_delete (TREE_TYPE (decl
), decl
, integer_two_node
,
LOOKUP_NORMAL
|LOOKUP_NONVIRTUAL
|LOOKUP_DESTRUCTOR
,
return build_method_call (decl
, name
, parms
, NULL_TREE
,
LOOKUP_NORMAL
|LOOKUP_NONVIRTUAL
);
/* Build something of the form ptr->method (args)
or object.method (args). This can also build
calls to constructors, and find friends.
Member functions always take their class variable
INSTANCE is a class instance.
NAME is the NAME field of the struct, union, or class
whose type is that of INSTANCE.
PARMS help to figure out what that NAME really refers to.
BASETYPE_PATH, if non-NULL, tells which basetypes of INSTANCE
we should be traversed before starting our search. We need
this information to get protected accesses correct.
FLAGS is the logical disjunction of zero or more LOOKUP_
flags. See cp-tree.h for more info.
If this is all OK, calls build_function_call with the resolved
This function must also handle being called to perform
initialization, promotion/coercion of arguments, and
instantiation of default parameters.
Note that NAME may refer to an instance variable name. If
`operator()()' is defined for the type of that field, then we return
build_method_call (instance
, name
, parms
, basetype_path
, flags
)
tree instance
, name
, parms
, basetype_path
;
register tree function
, fntype
, value_type
;
register tree basetype
, save_basetype
;
register tree baselink
, result
, method_name
, parmtypes
, parm
;
enum visibility_type visibility
;
/* Range of cases for vtable optimization. */
enum vtable_needs
{ not_needed
, maybe_needed
, unneeded
, needed
};
enum vtable_needs need_vtbl
= not_needed
;
tree instance_ptr
= NULL_TREE
;
int all_virtual
= flag_all_virtual
;
int static_call_context
= 0;
tree saw_private
= NULL_TREE
;
tree saw_protected
= NULL_TREE
;
/* Keep track of `const' and `volatile' objects. */
if (instance
== error_mark_node
|| name
== error_mark_node
|| parms
== error_mark_node
|| (instance
!= NULL_TREE
&& TREE_TYPE (instance
) == error_mark_node
))
/* This is the logic that magically deletes the second argument to
operator delete, if it is not needed. */
if (name
== ansi_opname
[(int) DELETE_EXPR
] && list_length (parms
)==2)
tree save_last
= TREE_CHAIN (parms
);
/* get rid of unneeded argument */
TREE_CHAIN (parms
) = NULL_TREE
;
result
= build_method_call (instance
, name
, parms
, basetype_path
,
(LOOKUP_SPECULATIVELY
|flags
)
/* If it works, return it. */
if (result
&& result
!= error_mark_node
)
return build_method_call (instance
, name
, parms
, basetype_path
, flags
);
/* If it doesn't work, two argument delete must work */
TREE_CHAIN (parms
) = save_last
;
/* C++ 2.1 does not allow this, but ANSI probably will. */
if (TREE_CODE (name
) == BIT_NOT_EXPR
)
error ("invalid call to destructor, use qualified name `%s::~%s'",
IDENTIFIER_POINTER (name
), IDENTIFIER_POINTER (name
));
if (TREE_CODE (name
) == BIT_NOT_EXPR
)
flags
|= LOOKUP_DESTRUCTOR
;
name
= TREE_OPERAND (name
, 0);
if (! is_aggr_typedef (name
, 1))
error ("destructors take no parameters");
basetype
= IDENTIFIER_TYPE_VALUE (name
);
if (! TYPE_HAS_DESTRUCTOR (basetype
))
#if 0 /* ARM says tp->~T() without T::~T() is valid. */
error_with_aggr_type (basetype
, "type `%s' has no destructor");
/* A destructive destructor wouldn't be a bad idea, but let's
return build_c_cast (void_type_node
, instance
);
instance
= default_conversion (instance
);
if (TREE_CODE (TREE_TYPE (instance
)) == POINTER_TYPE
)
instance_ptr
= build_unary_op (ADDR_EXPR
, instance
, 0);
return build_delete (build_pointer_type (basetype
),
instance_ptr
, integer_two_node
,
LOOKUP_NORMAL
|LOOKUP_DESTRUCTOR
, 0, 0);
/* Initialize name for error reporting. */
if (IDENTIFIER_TYPENAME_P (name
))
err_name
= "type conversion operator";
else if (IDENTIFIER_OPNAME_P (name
))
char *p
= operator_name_string (name
);
err_name
= (char *)alloca (strlen (p
) + 10);
sprintf (err_name
, "operator %s", p
);
else if (TREE_CODE (name
) == SCOPE_REF
)
err_name
= IDENTIFIER_POINTER (TREE_OPERAND (name
, 1));
err_name
= IDENTIFIER_POINTER (name
);
if (IDENTIFIER_OPNAME_P (name
))
GNU_xref_call (current_function_decl
, IDENTIFIER_POINTER (name
));
GNU_xref_call (current_function_decl
, err_name
);
if (instance
== NULL_TREE
)
/* Check cases where this is really a call to raise
if (current_class_type
&& TREE_CODE (name
) == IDENTIFIER_NODE
)
basetype
= purpose_member (name
, CLASSTYPE_TAGS (current_class_type
));
basetype
= TREE_VALUE (basetype
);
else if (TREE_CODE (name
) == SCOPE_REF
&& TREE_CODE (TREE_OPERAND (name
, 0)) == IDENTIFIER_NODE
)
if (! is_aggr_typedef (TREE_OPERAND (name
, 0), 1))
basetype
= purpose_member (TREE_OPERAND (name
, 1),
CLASSTYPE_TAGS (IDENTIFIER_TYPE_VALUE (TREE_OPERAND (name
, 0))));
basetype
= TREE_VALUE (basetype
);
if (basetype
!= NULL_TREE
)
/* call to a constructor... */
else if (IDENTIFIER_HAS_TYPE_VALUE (name
))
basetype
= IDENTIFIER_TYPE_VALUE (name
);
name
= constructor_name (basetype
);
tree typedef_name
= lookup_name (name
, 1);
if (typedef_name
&& TREE_CODE (typedef_name
) == TYPE_DECL
)
/* Canonicalize the typedef name. */
basetype
= TREE_TYPE (typedef_name
);
name
= TYPE_IDENTIFIER (basetype
);
error ("no constructor named `%s' in visible scope",
IDENTIFIER_POINTER (name
));
if (! IS_AGGR_TYPE (basetype
))
if ((flags
& LOOKUP_COMPLAIN
) && TREE_CODE (basetype
) != ERROR_MARK
)
error ("request for member `%s' in something not a structure or union", err_name
);
else if (instance
== C_C_D
|| instance
== current_class_decl
)
/* When doing initialization, we side-effect the TREE_TYPE of
C_C_D, hence we cannot set up BASETYPE from CURRENT_CLASS_TYPE. */
basetype
= TREE_TYPE (C_C_D
);
/* Anything manifestly `this' in constructors and destructors
has a known type, so virtual function tables are not needed. */
if (TYPE_VIRTUAL_P (basetype
)
&& !(flags
& LOOKUP_NONVIRTUAL
))
need_vtbl
= (dtor_label
|| ctor_label
)
? unneeded
: maybe_needed
;
instance_ptr
= current_class_decl
;
result
= build_field_call (TYPE_BINFO (current_class_type
),
instance_ptr
, name
, parms
, err_name
);
else if (TREE_CODE (instance
) == RESULT_DECL
)
basetype
= TREE_TYPE (instance
);
/* Should we ever have to make a virtual function reference
from a RESULT_DECL, know that it must be of fixed type
within the scope of this function. */
if (!(flags
& LOOKUP_NONVIRTUAL
) && TYPE_VIRTUAL_P (basetype
))
need_vtbl
= maybe_needed
;
instance_ptr
= build1 (ADDR_EXPR
, TYPE_POINTER_TO (basetype
), instance
);
else if (instance
== current_exception_object
)
instance_ptr
= build1 (ADDR_EXPR
, TYPE_POINTER_TO (current_exception_type
),
TREE_OPERAND (current_exception_object
, 0));
mark_addressable (TREE_OPERAND (current_exception_object
, 0));
result
= build_field_call (TYPE_BINFO (current_exception_type
),
instance_ptr
, name
, parms
, err_name
);
error ("exception member `%s' cannot be invoked", err_name
);
/* The MAIN_VARIANT of the type that `instance_ptr' winds up being. */
static_call_context
= (TREE_CODE (instance
) == NOP_EXPR
&& TREE_OPERAND (instance
, 0) == error_mark_node
);
/* the base type of an instance variable is pointer to class */
basetype
= TREE_TYPE (instance
);
if (TREE_CODE (basetype
) == REFERENCE_TYPE
)
basetype
= TYPE_MAIN_VARIANT (TREE_TYPE (basetype
));
if (! IS_AGGR_TYPE (basetype
))
/* Call to convert not needed because we are remaining
instance_ptr
= build1 (NOP_EXPR
, TYPE_POINTER_TO (basetype
), instance
);
inst_ptr_basetype
= basetype
;
if (TREE_CODE (basetype
) == POINTER_TYPE
)
basetype
= TREE_TYPE (basetype
);
if (! IS_AGGR_TYPE (basetype
))
&& (instance_ptr
= build_unary_op (ADDR_EXPR
, instance
, 0)))
|| (instance_ptr
= unary_complex_lvalue (ADDR_EXPR
, instance
)))
if (instance_ptr
== error_mark_node
)
else if (TREE_CODE (instance
) == NOP_EXPR
|| TREE_CODE (instance
) == CONSTRUCTOR
)
/* A cast is not an lvalue. Initialize a fresh temp
with the value we are casting from, and proceed with
that temporary. We can't cast to a reference type,
so that simplifies the initialization to something
tree temp
= get_temp_name (TREE_TYPE (instance
), 0);
if (IS_AGGR_TYPE (TREE_TYPE (instance
)))
expand_aggr_init (temp
, instance
, 0);
store_init_value (temp
, instance
);
instance_ptr
= build_unary_op (ADDR_EXPR
, instance
, 0);
if (TREE_CODE (instance
) != CALL_EXPR
)
if (TYPE_NEEDS_CONSTRUCTOR (basetype
))
instance
= build_cplus_new (basetype
, instance
, 0);
instance
= get_temp_name (basetype
, 0);
TREE_ADDRESSABLE (instance
) = 1;
instance_ptr
= build_unary_op (ADDR_EXPR
, instance
, 0);
/* @@ Should we call comp_target_types here? */
inst_ptr_basetype
= TREE_TYPE (TREE_TYPE (instance_ptr
));
if (TYPE_MAIN_VARIANT (basetype
) == TYPE_MAIN_VARIANT (inst_ptr_basetype
))
basetype
= inst_ptr_basetype
;
instance_ptr
= convert (TYPE_POINTER_TO (basetype
), instance_ptr
);
if (instance_ptr
== error_mark_node
)
inst_ptr_basetype
= TREE_TYPE (TREE_TYPE (instance_ptr
));
if (basetype_path
== NULL_TREE
)
basetype_path
= TYPE_BINFO (inst_ptr_basetype
);
result
= build_field_call (basetype_path
, instance_ptr
, name
, parms
, err_name
);
if (!(flags
& LOOKUP_NONVIRTUAL
) && TYPE_VIRTUAL_P (basetype
))
if (TREE_SIDE_EFFECTS (instance_ptr
))
/* This action is needed because the instance is needed
for providing the base of the virtual function table.
Without using a SAVE_EXPR, the function we are building
may be called twice, or side effects on the instance
variable (such as a post-increment), may happen twice. */
instance_ptr
= save_expr (instance_ptr
);
instance
= build_indirect_ref (instance_ptr
, NULL
);
else if (TREE_CODE (TREE_TYPE (instance
)) == POINTER_TYPE
)
/* This happens when called for operator new (). */
instance
= build_indirect_ref (instance
, NULL
);
need_vtbl
= maybe_needed
;
if (TYPE_SIZE (basetype
) == 0)
/* This is worth complaining about, I think. */
error_with_aggr_type (basetype
, "cannot lookup method in incomplete type `%s'");
save_basetype
= basetype
;
&& (! strncmp (IDENTIFIER_POINTER (name
), OPERATOR_METHOD_FORMAT
,
|| instance_ptr
== NULL_TREE
|| (TYPE_OVERLOADS_METHOD_CALL_EXPR (basetype
) == 0)))
for (parmtypes
= NULL_TREE
, parm
= parms
; parm
; parm
= TREE_CHAIN (parm
))
tree t
= TREE_TYPE (TREE_VALUE (parm
));
if (TREE_CODE (t
) == OFFSET_TYPE
)
/* Convert OFFSET_TYPE entities to their normal selves. */
TREE_VALUE (parm
) = resolve_offset_ref (TREE_VALUE (parm
));
t
= TREE_TYPE (TREE_VALUE (parm
));
if (TREE_CODE (t
) == ARRAY_TYPE
)
/* Perform the conversion from ARRAY_TYPE to POINTER_TYPE in place.
This eliminates needless calls to `compute_conversion_costs'. */
TREE_VALUE (parm
) = default_conversion (TREE_VALUE (parm
));
t
= TREE_TYPE (TREE_VALUE (parm
));
if (t
== error_mark_node
)
last
= build_tree_list (NULL_TREE
, t
);
parmtypes
= chainon (parmtypes
, last
);
constp
= TREE_READONLY (instance
);
volatilep
= TREE_THIS_VOLATILE (instance
);
parms
= tree_cons (NULL_TREE
, instance_ptr
, parms
);
/* Raw constructors are always in charge. */
if (TYPE_USES_VIRTUAL_BASECLASSES (basetype
)
&& ! (flags
& LOOKUP_HAS_IN_CHARGE
))
flags
|= LOOKUP_HAS_IN_CHARGE
;
parms
= tree_cons (NULL_TREE
, integer_one_node
, parms
);
parmtypes
= tree_cons (NULL_TREE
, integer_type_node
, parmtypes
);
if (flag_this_is_variable
> 0)
parms
= tree_cons (NULL_TREE
, build1 (NOP_EXPR
, TYPE_POINTER_TO (basetype
), integer_zero_node
), parms
);
instance_ptr
= build_new (NULL_TREE
, basetype
, void_type_node
, 0);
if (instance_ptr
== error_mark_node
)
instance_ptr
= save_expr (instance_ptr
);
TREE_CALLS_NEW (instance_ptr
) = 1;
instance
= build_indirect_ref (instance_ptr
, NULL
);
/* If it's a default argument initialized from a ctor, what we get
from instance_ptr will match the arglist for the FUNCTION_DECL
if (parms
&& TREE_CODE (TREE_VALUE (parms
)) == CALL_EXPR
&& TREE_OPERAND (TREE_VALUE (parms
), 1)
&& TREE_CALLS_NEW (TREE_VALUE (TREE_OPERAND (TREE_VALUE (parms
), 1))))
parms
= build_tree_list (NULL_TREE
, instance_ptr
);
parms
= tree_cons (NULL_TREE
, instance_ptr
, parms
);
parmtypes
= tree_cons (NULL_TREE
,
build_pointer_type (build_type_variant (basetype
, constp
, volatilep
)),
/* Look up function name in the structure type definition. */
if ((IDENTIFIER_HAS_TYPE_VALUE (name
)
&& IS_AGGR_TYPE (IDENTIFIER_TYPE_VALUE (name
)))
|| name
== constructor_name (basetype
))
if (IDENTIFIER_TYPE_VALUE (name
) == basetype
|| name
== constructor_name (basetype
))
tmp
= TYPE_BINFO (basetype
);
tmp
= get_binfo (IDENTIFIER_TYPE_VALUE (name
), basetype
, 0);
name_kind
= "constructor";
if (TYPE_USES_VIRTUAL_BASECLASSES (basetype
)
&& ! (flags
& LOOKUP_HAS_IN_CHARGE
))
/* Constructors called for initialization
only are never in charge. */
flags
|= LOOKUP_HAS_IN_CHARGE
;
tmplist
= tree_cons (NULL_TREE
, integer_zero_node
,
TREE_CHAIN (parms
) = tmplist
;
tmplist
= tree_cons (NULL_TREE
, integer_type_node
, TREE_CHAIN (parmtypes
));
TREE_CHAIN (parmtypes
) = tmplist
;
basetype
= BINFO_TYPE (tmp
);
else name_kind
= "method";
if (basetype_path
== NULL_TREE
)
basetype_path
= TYPE_BINFO (basetype
);
result
= lookup_fnfields (basetype_path
, name
,
(flags
& LOOKUP_COMPLAIN
));
if (result
== error_mark_node
)
/* Now, go look for this method name. We do not find destructors here.
Putting `void_list_node' on the end of the parmtypes
fakes out `build_decl_overload' into doing the right thing. */
TREE_CHAIN (last
) = void_list_node
;
method_name
= build_decl_overload (name
, parmtypes
,
1 + (name
== constructor_name (save_basetype
)));
TREE_CHAIN (last
) = NULL_TREE
;
for (pass
= 0; pass
< 2; pass
++)
struct candidate
*candidates
;
/* This increments every time we go up the type hierarchy.
The idea is to prefer a function of the derived class if possible. */
= (struct candidate
*) alloca ((ever_seen
+1)
* sizeof (struct candidate
));
len
= list_length (parms
);
/* First see if a global function has a shot at it. */
if (flags
& LOOKUP_GLOBAL
)
tree parm
= TREE_VALUE (parms
);
if (TREE_CODE (TREE_TYPE (parm
)) == REFERENCE_TYPE
)
else if (TREE_CODE (TREE_TYPE (parm
)) == POINTER_TYPE
)
parm
= build_indirect_ref (parm
, "friendifying parms (compiler error)");
new_type
= build_reference_type (TREE_TYPE (parm
));
/* It is possible that this should go down a layer. */
new_type
= build_type_variant (new_type
,
TREE_THIS_VOLATILE (parm
));
parm
= convert (new_type
, parm
);
friend_parms
= tree_cons (NULL_TREE
, parm
, TREE_CHAIN (parms
));
= (unsigned short *)alloca ((len
+1) * sizeof (short));
result
= build_overload_call (name
, friend_parms
, 0, cp
);
/* If it turns out to be the one we were actually looking for
(it was probably a friend function), the return the
if (TREE_CODE (result
) == CALL_EXPR
)
/* non-standard uses: set the field to 0 to indicate
we are using a non-member function. */
if (cp
->harshness
[len
] == 0
&& cp
->harshness
[len
] == 0
&& cp
->user
== 0 && cp
->b_or_d
== 0
/* We have a hit (of sorts). If the parameter list is
"error_mark_node", or some variant thereof, it won't
match any methods. Since we have verified that the is
some method vaguely matching this one (in name at least),
Don't stop for friends, however. */
tree basetypes
= TREE_PURPOSE (baselink
);
function
= TREE_VALUE (baselink
);
if (TREE_CODE (basetypes
) == TREE_LIST
)
basetypes
= TREE_VALUE (basetypes
);
basetype
= BINFO_TYPE (basetypes
);
/* Cast the instance variable to the appropriate type. */
TREE_VALUE (parmtypes
) = TYPE_POINTER_TO (basetype
);
if (DESTRUCTOR_NAME_P (DECL_ASSEMBLER_NAME (function
)))
function
= DECL_CHAIN (function
);
for (; function
; function
= DECL_CHAIN (function
))
n_inner_fields_searched
++;
/* Not looking for friends here. */
if (TREE_CODE (TREE_TYPE (function
)) == FUNCTION_TYPE
&& ! DECL_STATIC_FUNCTION_P (function
))
&& DECL_ASSEMBLER_NAME (function
) == method_name
)
if (flags
& LOOKUP_PROTECT
)
visibility
= compute_visibility (basetypes
, function
);
if (visibility
== visibility_protected
&& flags
& LOOKUP_PROTECTED_OK
)
visibility
= visibility_public
;
if ((flags
& LOOKUP_PROTECT
) == 0
|| visibility
== visibility_public
)
else if (visibility
== visibility_private
)
else if (visibility
== visibility_protected
)
saw_protected
= function
;
/* If we fail on the exact match, we have
tree these_parms
= parms
;
n_inner_fields_searched
++;
= (unsigned short *)alloca ((len
+1) * sizeof (short));
if (DECL_STATIC_FUNCTION_P (function
))
these_parms
= TREE_CHAIN (these_parms
);
compute_conversion_costs (function
, these_parms
, cp
, len
);
if (flags
& LOOKUP_PROTECT
)
enum visibility_type this_v
;
this_v
= compute_visibility (basetypes
, function
);
if (this_v
== visibility_protected
&& (flags
& LOOKUP_PROTECTED_OK
))
this_v
= visibility_public
;
if (this_v
!= visibility_public
)
if (this_v
== visibility_private
)
saw_protected
= function
;
/* No "two-level" conversions. */
if (flags
& LOOKUP_NO_CONVERSION
&& cp
->user
!= 0)
/* If we used default parameters, we must
check to see whether anyone else might
use them also, and report a possible
if (! TYPE_USES_MULTIPLE_INHERITANCE (save_basetype
)
&& cp
->harshness
[len
] == 0
&& CONST_HARSHNESS (cp
->harshness
[0]) == 0
&& cp
->user
== 0 && cp
->b_or_d
== 0
if (! DECL_STATIC_FUNCTION_P (function
))
TREE_VALUE (parms
) = cp
->arg
;
goto found_and_maybe_warn
;
/* Now we have run through one link's member functions.
arrange to head-insert this link's links. */
baselink
= next_baselink (baselink
);
/* No exact match could be found. Now try to find match
using default conversions. */
if ((flags
& LOOKUP_GLOBAL
) && IDENTIFIER_GLOBAL_VALUE (name
))
if (TREE_CODE (IDENTIFIER_GLOBAL_VALUE (name
)) == FUNCTION_DECL
)
else if (TREE_CODE (IDENTIFIER_GLOBAL_VALUE (name
)) == TREE_LIST
)
ever_seen
+= list_length (IDENTIFIER_GLOBAL_VALUE (name
));
if ((flags
& (LOOKUP_SPECULATIVELY
|LOOKUP_COMPLAIN
))
if (flags
& LOOKUP_GLOBAL
)
error ("no global or non-hidden member function `%s' defined", err_name
);
error_with_aggr_type (save_basetype
, "no non-hidden member function `%s::%s' defined", err_name
);
if (cp
- candidates
!= 0)
/* Rank from worst to best. Then cp will point to best one.
Private fields have their bits flipped. For unsigned
numbers, this should make them look very large.
If the best alternate has a (signed) negative value,
then all we ever saw were private members. */
cp
= ideal_candidate (save_basetype
, candidates
,
cp
- candidates
, parms
, len
);
if (cp
== (struct candidate
*)0)
error ("ambiguous type conversion requested for %s `%s'",
else if (cp
[-1].evil
== 2)
error ("ambiguous type conversion requested for %s `%s'",
/* The global function was the best, so use it. */
/* We must convert the instance pointer into a reference type.
Global overloaded functions can only either take
aggregate objects (which come for free from references)
or reference data types anyway. */
TREE_VALUE (parms
) = copy_node (instance_ptr
);
TREE_TYPE (TREE_VALUE (parms
)) = build_reference_type (TREE_TYPE (TREE_TYPE (instance_ptr
)));
return build_function_call (cp
->function
, parms
);
if (! DECL_STATIC_FUNCTION_P (function
))
TREE_VALUE (parms
) = cp
->arg
;
goto found_and_maybe_warn
;
if ((flags
& ~LOOKUP_GLOBAL
) & (LOOKUP_COMPLAIN
|LOOKUP_SPECULATIVELY
))
if ((flags
& (LOOKUP_SPECULATIVELY
|LOOKUP_COMPLAIN
))
if (DECL_STATIC_FUNCTION_P (cp
->function
))
parms
= TREE_CHAIN (parms
);
if (((HOST_WIDE_INT
)saw_protected
|(HOST_WIDE_INT
)saw_private
) == 0)
if (flags
& LOOKUP_SPECULATIVELY
)
if (static_call_context
&& TREE_CODE (TREE_TYPE (cp
->function
)) == METHOD_TYPE
)
error_with_aggr_type (TREE_TYPE (TREE_TYPE (instance_ptr
)),
"object missing in call to `%s::%s'",
report_type_mismatch (cp
, parms
, name_kind
, err_name
);
msg
= "%s `%%s' (and the like) are private or protected";
msg
= "the %s `%%s' is private";
msg
= "the %s `%%s' is protected";
sprintf (buf
, msg
, name_kind
);
error_with_decl (seen
, buf
);
error ("within this context");
if ((flags
& (LOOKUP_SPECULATIVELY
|LOOKUP_COMPLAIN
))
if (TREE_CODE (save_basetype
) == RECORD_TYPE
)
buf
= (char *)alloca (30 + strlen (err_name
));
strcpy (buf
, "%s has no method named `%s'");
error (buf
, tag_name
, err_name
);
if (CONST_HARSHNESS (cp
->harshness
[0]))
if (flags
& LOOKUP_COMPLAIN
)
error_with_decl (cp
->function
, "non-const member function `%s'");
error ("called for const object at this point in file");
/* Not good enough for a match. */
/* Silently return error_mark_node. */
if (visibility
== visibility_private
)
if (flags
& LOOKUP_COMPLAIN
)
error_with_file_and_line (DECL_SOURCE_FILE (function
),
DECL_SOURCE_LINE (function
),
: "%s `%s' is from private base class",
lang_printable_name (function
));
error ("within this context");
else if (visibility
== visibility_protected
)
if (flags
& LOOKUP_COMPLAIN
)
error_with_file_and_line (DECL_SOURCE_FILE (function
),
DECL_SOURCE_LINE (function
),
TREE_PROTECTED (function
)
: "%s `%s' has protected visibility from this point",
lang_printable_name (function
));
error ("within this context");
/* From here on down, BASETYPE is the type that INSTANCE_PTR's
type (if it exists) is a pointer to. */
function
= DECL_MAIN_VARIANT (function
);
/* Declare external function if necessary. */
assemble_external (function
);
fntype
= TREE_TYPE (function
);
if (TREE_CODE (fntype
) == POINTER_TYPE
)
fntype
= TREE_TYPE (fntype
);
basetype
= DECL_CLASS_CONTEXT (function
);
/* If we are referencing a virtual function from an object
of effectively static type, then there is no need
to go through the virtual function table. */
if (need_vtbl
== maybe_needed
)
int fixed_type
= resolves_to_fixed_type_p (instance
, 0);
&& DECL_VINDEX (function
)
&& may_be_remote (basetype
))
else if (DECL_VINDEX (function
))
need_vtbl
= fixed_type
? unneeded
: needed
;
if (TREE_CODE (fntype
) == METHOD_TYPE
&& static_call_context
&& !DECL_CONSTRUCTOR_P (function
))
/* Let's be nice to the user for now, and give reasonable
instance_ptr
= current_class_decl
;
if (basetype
!= current_class_type
)
tree binfo
= get_binfo (basetype
, current_class_type
, 1);
error_not_base_type (function
, current_class_type
);
else if (basetype
== error_mark_node
)
else if (! TREE_STATIC (function
))
error_with_aggr_type (basetype
, "cannot call member function `%s::%s' without object",
value_type
= TREE_TYPE (fntype
) ? TREE_TYPE (fntype
) : void_type_node
;
if (TYPE_SIZE (value_type
) == 0)
if (flags
& LOOKUP_COMPLAIN
)
incomplete_type_error (0, value_type
);
/* We do not pass FUNCTION into `convert_arguments', because by
now everything should be ok. If not, then we have a serious error. */
if (DECL_STATIC_FUNCTION_P (function
))
parms
= convert_arguments (NULL_TREE
, TYPE_ARG_TYPES (fntype
),
TREE_CHAIN (parms
), NULL_TREE
, LOOKUP_NORMAL
);
else if (need_vtbl
== unneeded
)
int sub_flags
= DECL_CONSTRUCTOR_P (function
) ? flags
: LOOKUP_NORMAL
;
basetype
= TREE_TYPE (instance
);
if (TYPE_METHOD_BASETYPE (TREE_TYPE (function
)) != TYPE_MAIN_VARIANT (basetype
)
&& TYPE_USES_COMPLEX_INHERITANCE (basetype
))
basetype
= DECL_CLASS_CONTEXT (function
);
instance_ptr
= convert_pointer_to (basetype
, instance_ptr
);
instance
= build_indirect_ref (instance_ptr
, NULL
);
parms
= tree_cons (NULL_TREE
, instance_ptr
,
convert_arguments (NULL_TREE
, TREE_CHAIN (TYPE_ARG_TYPES (fntype
)), TREE_CHAIN (parms
), NULL_TREE
, sub_flags
));
if ((flags
& LOOKUP_NONVIRTUAL
) == 0)
basetype
= DECL_CONTEXT (function
);
/* First parm could be integer_zerop with casts like
((Object*)0)->Object::IsA() */
if (!integer_zerop (TREE_VALUE (parms
)))
instance_ptr
= convert_pointer_to (build_type_variant (basetype
, constp
, volatilep
),
if (TREE_CODE (instance_ptr
) == COND_EXPR
)
instance_ptr
= save_expr (instance_ptr
);
instance
= build_indirect_ref (instance_ptr
, NULL
);
else if (TREE_CODE (instance_ptr
) == NOP_EXPR
&& TREE_CODE (TREE_OPERAND (instance_ptr
, 0)) == ADDR_EXPR
&& TREE_OPERAND (TREE_OPERAND (instance_ptr
, 0), 0) == instance
)
/* The call to `convert_pointer_to' may return error_mark_node. */
else if (TREE_CODE (instance_ptr
) == ERROR_MARK
)
else if (instance
== NULL_TREE
|| TREE_CODE (instance
) != INDIRECT_REF
|| TREE_OPERAND (instance
, 0) != instance_ptr
)
instance
= build_indirect_ref (instance_ptr
, NULL
);
parms
= tree_cons (NULL_TREE
, instance_ptr
,
convert_arguments (NULL_TREE
, TREE_CHAIN (TYPE_ARG_TYPES (fntype
)), TREE_CHAIN (parms
), NULL_TREE
, LOOKUP_NORMAL
));
/* Constructors do not overload method calls. */
else if (TYPE_OVERLOADS_METHOD_CALL_EXPR (basetype
)
&& name
!= TYPE_IDENTIFIER (basetype
)
&& (TREE_CODE (function
) != FUNCTION_DECL
|| strncmp (IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (function
)),
&& (may_be_remote (basetype
)
|| (C_C_D
? TREE_TYPE (instance
) != current_class_type
: 1))
/* This change by Larry Ketcham. */
&& (may_be_remote (basetype
) || instance
!= C_C_D
)
parms
= TREE_CHAIN (parms
);
if (!all_virtual
&& TREE_CODE (function
) == FUNCTION_DECL
)
fn_as_int
= build_unary_op (ADDR_EXPR
, function
, 0);
fn_as_int
= convert (TREE_TYPE (default_conversion (function
)), DECL_VINDEX (function
));
fn_as_int
= convert (integer_type_node
, fn_as_int
);
result
= build_opfncall (METHOD_CALL_EXPR
, LOOKUP_NORMAL
, instance
, fn_as_int
, parms
);
compiler_error ("could not overload `operator->()(...)'");
else if (result
== error_mark_node
)
/* Do this if we want the result of operator->() to inherit
the type of the function it is subbing for. */
TREE_TYPE (result
) = value_type
;
function
= build_vfn_ref (&TREE_VALUE (parms
), instance
, DECL_VINDEX (function
));
TREE_TYPE (function
) = build_pointer_type (fntype
);
if (TREE_CODE (function
) == FUNCTION_DECL
)
GNU_xref_call (current_function_decl
,
IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (function
)));
if (TREE_CODE (function
) == FUNCTION_DECL
)
if (DECL_INLINE (function
))
function
= build1 (ADDR_EXPR
, build_pointer_type (fntype
), function
);
assemble_external (function
);
TREE_USED (function
) = 1;
function
= default_conversion (function
);
function
= default_conversion (function
);
build_nt (CALL_EXPR
, function
, parms
, NULL_TREE
);
TREE_TYPE (result
) = value_type
;
TREE_SIDE_EFFECTS (result
) = 1;
= TYPE_RAISES_EXCEPTIONS (fntype
) || (parms
&& TREE_RAISES (parms
));
/* Similar to `build_method_call', but for overloaded non-member functions.
The name of this function comes through NAME. The name depends
Note that this function must handle simple `C' promotions,
as well as variable numbers of arguments (...), and
default arguments to boot.
If the overloading is successful, we return a tree node which
contains the call to the function.
If overloading produces candidates which are probable, but not definite,
we hold these candidates. If FINAL_CP is non-zero, then we are free
to assume that final_cp points to enough storage for all candidates that
this function might generate. The `harshness' array is preallocated for
the first candidate, but not for subsequent ones.
Note that the DECL_RTL of FUNCTION must be made to agree with this
build_overload_call_real (fnname
, parms
, complain
, final_cp
, buildxxx
)
struct candidate
*final_cp
;
/* must check for overloading here */
tree overload_name
, functions
, function
, parm
;
tree parmtypes
= NULL_TREE
, last
= NULL_TREE
;
int parmlength
= list_length (parms
);
struct candidate
*candidates
, *cp
;
final_cp
[0].function
= 0;
for (parm
= parms
; parm
; parm
= TREE_CHAIN (parm
))
register tree t
= TREE_TYPE (TREE_VALUE (parm
));
if (t
== error_mark_node
)
if (TREE_CODE (t
) == ARRAY_TYPE
|| TREE_CODE (t
) == OFFSET_TYPE
)
/* Perform the conversion from ARRAY_TYPE to POINTER_TYPE in place.
Also convert OFFSET_TYPE entities to their normal selves.
This eliminates needless calls to `compute_conversion_costs'. */
TREE_VALUE (parm
) = default_conversion (TREE_VALUE (parm
));
t
= TREE_TYPE (TREE_VALUE (parm
));
last
= build_tree_list (NULL_TREE
, t
);
parmtypes
= chainon (parmtypes
, last
);
TREE_CHAIN (last
) = void_list_node
;
parmtypes
= void_list_node
;
overload_name
= build_decl_overload (fnname
, parmtypes
, 0);
/* Now check to see whether or not we can win.
Note that if we are called from `build_method_call',
then we cannot have a mis-match, because we would have
already found such a winning case. */
if (IDENTIFIER_GLOBAL_VALUE (overload_name
))
if (TREE_CODE (IDENTIFIER_GLOBAL_VALUE (overload_name
)) != TREE_LIST
)
return build_function_call (DECL_MAIN_VARIANT (IDENTIFIER_GLOBAL_VALUE (overload_name
)), parms
);
functions
= IDENTIFIER_GLOBAL_VALUE (fnname
);
if (functions
== NULL_TREE
)
error ("only member functions apply");
if (TREE_CODE (functions
) == FUNCTION_DECL
)
functions
= DECL_MAIN_VARIANT (functions
);
/* We are just curious whether this is a viable alternative or not. */
compute_conversion_costs (functions
, parms
, final_cp
, parmlength
);
return build_function_call (functions
, parms
);
if (TREE_VALUE (functions
) == NULL_TREE
)
error ("function `%s' declared overloaded, but no instances of that function declared",
IDENTIFIER_POINTER (TREE_PURPOSE (functions
)));
if (TREE_CODE (TREE_VALUE (functions
)) == TREE_LIST
)
/* The list-of-lists should only occur for class things. */
my_friendly_assert (functions
== IDENTIFIER_CLASS_VALUE (fnname
), 168);
for (outer
= functions
; outer
; outer
= TREE_CHAIN (outer
))
length
+= decl_list_length (TREE_VALUE (TREE_VALUE (outer
)));
length
+= list_length (TREE_TYPE (TREE_VALUE (outer
)));
length
= list_length (functions
);
candidates
= (struct candidate
*)alloca ((length
+1) * sizeof (struct candidate
));
my_friendly_assert (TREE_CODE (TREE_VALUE (functions
)) != TREE_LIST
, 169);
/* OUTER is the list of FUNCTION_DECLS, in a TREE_LIST. */
for (outer
= functions
; outer
; outer
= TREE_CHAIN (outer
))
function
= TREE_VALUE (outer
);
if (TREE_CODE (function
) != FUNCTION_DECL
&& ! (TREE_CODE (function
) == TEMPLATE_DECL
&& ! DECL_TEMPLATE_IS_CLASS (function
)
&& TREE_CODE (DECL_TEMPLATE_RESULT (function
)) == FUNCTION_DECL
))
enum tree_code code
= TREE_CODE (function
);
if (code
== TEMPLATE_DECL
)
code
= TREE_CODE (DECL_TEMPLATE_RESULT (function
));
error_with_decl (function
, "enumeral value `%s' conflicts with function of same name");
else if (code
== VAR_DECL
)
if (TREE_STATIC (function
))
error_with_decl (function
, "variable `%s' conflicts with function of same name");
error_with_decl (function
, "constant field `%s' conflicts with function of same name");
else if (code
== TYPE_DECL
)
else my_friendly_abort (2);
error ("at this point in file");
if (TREE_CODE (function
) == TEMPLATE_DECL
)
int ntparms
= TREE_VEC_LENGTH (DECL_TEMPLATE_PARMS (function
));
tree
*targs
= (tree
*) alloca (sizeof (tree
) * ntparms
);
i
= type_unification (DECL_TEMPLATE_PARMS (function
), targs
,
TYPE_ARG_TYPES (TREE_TYPE (function
)),
function
= instantiate_template (function
, targs
);
if (TREE_CODE (function
) == TEMPLATE_DECL
)
/* Unconverted template -- failed match. */
cp
->evil
= 1, cp
->function
= function
, cp
->u
.bad_arg
= -4;
function
= DECL_MAIN_VARIANT (function
);
/* Can't use alloca here, since result might be
passed to calling function. */
= (unsigned short *)oballoc ((parmlength
+1) * sizeof (short));
compute_conversion_costs (function
, parms
, cp
, parmlength
);
/* Should really add another field... */
cp
->easy
= cp
->easy
* 128 + template_cost
;
&& cp
[0].user
== 0 && cp
[0].b_or_d
== 0
final_cp
[0].easy
= cp
[0].easy
;
tree rval
= error_mark_node
;
struct candidate
*best_cp
= ideal_candidate (NULL_TREE
, candidates
,
cp
- candidates
, parms
, parmlength
);
if (best_cp
== (struct candidate
*)0)
error ("call of overloaded `%s' is ambiguous", IDENTIFIER_POINTER (fnname
));
rval
= best_cp
->function
;
error ("type conversion ambiguous");
return buildxxx
? build_function_call_maybe (rval
, parms
)
: build_function_call (rval
, parms
);
/* Initialize name for error reporting. */
if (TREE_CODE (functions
) == TREE_LIST
)
name
= TREE_PURPOSE (functions
);
else if (TREE_CODE (functions
) == ADDR_EXPR
)
/* Since the implicit `operator new' and `operator delete' functions
are set up to have IDENTIFIER_GLOBAL_VALUEs that are unary ADDR_EXPRs
by default_conversion(), we must compensate for that here by
using the name of the ADDR_EXPR's operand. */
name
= DECL_NAME (TREE_OPERAND (functions
, 0));
name
= DECL_NAME (functions
);
if (IDENTIFIER_OPNAME_P (name
))
char *opname
= operator_name_string (name
);
err_name
= (char *)alloca (strlen (opname
) + 12);
sprintf (err_name
, "operator %s", opname
);
err_name
= IDENTIFIER_POINTER (name
);
report_type_mismatch (cp
, parms
, "function", err_name
);
build_overload_call (fnname
, parms
, complain
, final_cp
)
struct candidate
*final_cp
;
return build_overload_call_real (fnname
, parms
, complain
, final_cp
, 0);
build_overload_call_maybe (fnname
, parms
, complain
, final_cp
)
struct candidate
*final_cp
;
return build_overload_call_real (fnname
, parms
, complain
, final_cp
, 1);