https://github.com/JuliaLang/julia
Tip revision: 7790d6f06411be1fd5aec7cb6fffdb38c89c0c2a authored by Kristoffer Carlsson on 13 February 2024, 20:41:13 UTC
release-1.10: set VERSION to 1.10.1 (#53306)
release-1.10: set VERSION to 1.10.1 (#53306)
Tip revision: 7790d6f
typemap.c
// This file is a part of Julia. License is MIT: https://julialang.org/license
#include <stdlib.h>
#include <string.h>
#include "julia.h"
#include "julia_internal.h"
#ifndef _OS_WINDOWS_
#include <unistd.h>
#endif
#include "julia_assert.h"
#define MAX_METHLIST_COUNT 6 // this helps configure the sysimg size and speed.
#ifdef __cplusplus
extern "C" {
#endif
// compute whether the specificity of this type is equivalent to Any in the sort order
static int jl_is_any(jl_value_t *t1)
{
while (jl_is_typevar(t1))
t1 = ((jl_tvar_t*)t1)->ub;
return t1 == (jl_value_t*)jl_any_type;
}
static jl_value_t *jl_type_extract_name(jl_value_t *t1 JL_PROPAGATES_ROOT) JL_NOTSAFEPOINT
{
if (jl_is_unionall(t1))
t1 = jl_unwrap_unionall(t1);
if (jl_is_vararg(t1)) {
return jl_type_extract_name(jl_unwrap_vararg(t1));
}
else if (jl_is_typevar(t1)) {
return jl_type_extract_name(((jl_tvar_t*)t1)->ub);
}
else if (t1 == jl_bottom_type || t1 == (jl_value_t*)jl_typeofbottom_type || t1 == (jl_value_t*)jl_typeofbottom_type->super) {
return (jl_value_t*)jl_typeofbottom_type->name; // put Union{} and typeof(Union{}) and Type{Union{}} together for convenience
}
else if (jl_is_datatype(t1)) {
jl_datatype_t *dt = (jl_datatype_t*)t1;
if (!jl_is_kind(t1))
return (jl_value_t*)dt->name;
return NULL;
}
else if (jl_is_uniontype(t1)) {
jl_uniontype_t *u1 = (jl_uniontype_t*)t1;
jl_value_t *tn1 = jl_type_extract_name(u1->a);
jl_value_t *tn2 = jl_type_extract_name(u1->b);
if (tn1 == tn2)
return tn1;
// TODO: if invariant is false, instead find the nearest common ancestor
return NULL;
}
return NULL;
}
// return false if the name extracted above is an over-approximation
// (such that intersection also needs to consider the subtypes)
static int jl_type_extract_name_precise(jl_value_t *t1, int invariant)
{
if (jl_is_unionall(t1))
t1 = jl_unwrap_unionall(t1);
if (jl_is_vararg(t1)) {
return jl_type_extract_name_precise(jl_unwrap_vararg(t1), invariant);
}
else if (jl_is_typevar(t1)) {
return jl_type_extract_name_precise(((jl_tvar_t*)t1)->ub, 0);
}
else if (t1 == jl_bottom_type || t1 == (jl_value_t*)jl_typeofbottom_type || t1 == (jl_value_t*)jl_typeofbottom_type->super) {
return 1;
}
else if (jl_is_datatype(t1)) {
jl_datatype_t *dt = (jl_datatype_t*)t1;
if ((invariant || !dt->name->abstract) && !jl_is_kind(t1))
return 1;
return 0;
}
else if (jl_is_uniontype(t1)) {
jl_uniontype_t *u1 = (jl_uniontype_t*)t1;
if (!jl_type_extract_name_precise(u1->a, invariant))
return 0;
if (!jl_type_extract_name_precise(u1->b, invariant))
return 0;
jl_value_t *tn1 = jl_type_extract_name(u1->a);
jl_value_t *tn2 = jl_type_extract_name(u1->b);
if (tn1 == tn2)
return 1;
return 0;
}
return 1;
}
// return whether Type{Union{}} is a subtype of Type{t1} (which may have free typevars)
static int jl_parameter_includes_bottom(jl_value_t *t1)
{
if (jl_is_typevar(t1) || t1 == jl_bottom_type)
return 1;
else if (jl_is_uniontype(t1)) {
jl_uniontype_t *u1 = (jl_uniontype_t*)t1;
return jl_parameter_includes_bottom(u1->a) && jl_parameter_includes_bottom(u1->b);
}
return 0;
}
// ----- Type Signature Subtype Testing ----- //
static int sig_match_by_type_leaf(jl_value_t **types, jl_tupletype_t *sig, size_t n)
{
size_t i;
for (i = 0; i < n; i++) {
jl_value_t *decl = jl_tparam(sig, i);
jl_value_t *a = types[i];
if (jl_is_type_type(a)) // decl is not Type, because it wouldn't be leafsig
a = jl_typeof(jl_tparam0(a));
if (!jl_types_equal(a, decl))
return 0;
}
return 1;
}
static int sig_match_by_type_simple(jl_value_t **types, size_t n, jl_tupletype_t *sig, size_t lensig, int va)
{
size_t i;
if (va) lensig -= 1;
for (i = 0; i < lensig; i++) {
jl_value_t *decl = jl_tparam(sig, i);
jl_value_t *a = types[i];
jl_value_t *unw = jl_is_unionall(decl) ? ((jl_unionall_t*)decl)->body : decl;
if (jl_is_vararg(a))
return 0;
if (jl_is_type_type(unw)) {
jl_value_t *tp0 = jl_tparam0(unw);
if (jl_is_type_type(a)) {
if (jl_is_typevar(tp0)) {
// in the case of Type{_}, the types don't have to match exactly.
// this is cached as `Type{T} where T`.
if (((jl_tvar_t*)tp0)->ub != (jl_value_t*)jl_any_type &&
!jl_subtype(jl_tparam0(a), ((jl_tvar_t*)tp0)->ub))
return 0;
}
else if (!jl_types_equal(jl_tparam0(a), tp0)) {
return 0;
}
}
else if (!jl_is_kind(a) || !jl_is_typevar(tp0) || ((jl_tvar_t*)tp0)->ub != (jl_value_t*)jl_any_type) {
// manually unroll jl_subtype(a, decl)
// where `a` can be a subtype and decl is Type{T}
return 0;
}
}
else if (decl == (jl_value_t*)jl_any_type) {
}
else {
if (jl_is_type_type(a)) // decl is not Type, because it would be caught above
a = jl_typeof(jl_tparam0(a));
if (!jl_types_equal(a, decl))
return 0;
}
}
if (va) {
jl_value_t *decl = jl_unwrap_unionall(jl_tparam(sig, i));
if (jl_vararg_kind(decl) == JL_VARARG_INT) {
if (n - i != jl_unbox_long(jl_tparam1(decl)))
return 0;
}
jl_value_t *t = jl_unwrap_vararg(decl);
if (jl_is_typevar(t))
t = ((jl_tvar_t*)t)->ub;
for (; i < n; i++) {
jl_value_t *ti = types[i];
if (i == n - 1 && jl_is_vararg(ti))
ti = jl_unwrap_vararg(ti);
if (!jl_subtype(ti, t))
return 0;
}
return 1;
}
return 1;
}
static inline int sig_match_leaf(jl_value_t *arg1, jl_value_t **args, jl_value_t **sig, size_t n)
{
// NOTE: This function is a huge performance hot spot!!
size_t i;
if (jl_typeof(arg1) != sig[0])
return 0;
for (i = 1; i < n; i++) {
jl_value_t *decl = sig[i];
jl_value_t *a = args[i - 1];
if (jl_typeof(a) != decl) {
/*
we are only matching concrete types here, and those types are
hash-consed, so pointer comparison should work.
*/
return 0;
}
}
return 1;
}
static inline int sig_match_simple(jl_value_t *arg1, jl_value_t **args, size_t n, jl_value_t **sig,
int va, size_t lensig)
{
// NOTE: This function is a performance hot spot!!
size_t i;
if (va)
lensig -= 1;
for (i = 0; i < lensig; i++) {
jl_value_t *decl = sig[i];
jl_value_t *a = (i == 0 ? arg1 : args[i - 1]);
if (jl_typeof(a) == decl || decl == (jl_value_t*)jl_any_type) {
/*
we are only matching concrete types here, and those types are
hash-consed, so pointer comparison should work.
*/
continue;
}
jl_value_t *unw = jl_is_unionall(decl) ? ((jl_unionall_t*)decl)->body : decl;
if (jl_is_type_type(unw) && jl_is_type(a)) {
jl_value_t *tp0 = jl_tparam0(unw);
if (jl_is_typevar(tp0)) {
// in the case of Type{_}, the types don't have to match exactly.
// this is cached as `Type{T} where T`.
if (((jl_tvar_t*)tp0)->ub != (jl_value_t*)jl_any_type &&
!jl_subtype(a, ((jl_tvar_t*)tp0)->ub))
return 0;
}
else {
if (a != tp0) {
jl_datatype_t *da = (jl_datatype_t*)a;
jl_datatype_t *dt = (jl_datatype_t*)tp0;
while (jl_is_unionall(da))
da = (jl_datatype_t*)((jl_unionall_t*)da)->body;
while (jl_is_unionall(dt))
dt = (jl_datatype_t*)((jl_unionall_t*)dt)->body;
if (jl_is_datatype(da) && jl_is_datatype(dt) && da->name != dt->name)
return 0;
if (!jl_types_equal(a, tp0))
return 0;
}
}
}
else {
return 0;
}
}
if (va) {
jl_value_t *decl = sig[i];
if (jl_vararg_kind(decl) == JL_VARARG_INT) {
if (n - i != jl_unbox_long(jl_tparam1(decl)))
return 0;
}
jl_value_t *t = jl_unwrap_vararg(decl);
for (; i < n; i++) {
jl_value_t *a = (i == 0 ? arg1 : args[i - 1]);
if (!jl_isa(a, t))
return 0;
}
return 1;
}
return 1;
}
// ----- MethodCache helper functions ----- //
// predicate to fast-test if this type is a leaf type that can exist in the cache
// and does not need a more expensive linear scan to find all intersections
// we try not to put non-leaf types or DataType/UnionAll/Union in the
// argument cache, since they should have a lower priority and so will go in some
// later list
static int is_cache_leaf(jl_value_t *ty, int tparam)
{
if (ty == jl_bottom_type)
return 1;
return (jl_is_concrete_type(ty) && (tparam || !jl_is_kind(ty)));
}
static _Atomic(jl_value_t*) *mtcache_hash_lookup_bp(jl_array_t *cache JL_PROPAGATES_ROOT, jl_value_t *ty) JL_NOTSAFEPOINT
{
if (cache == (jl_array_t*)jl_an_empty_vec_any)
return NULL;
_Atomic(jl_value_t*) *pml = jl_table_peek_bp(cache, ty);
JL_GC_PROMISE_ROOTED(pml); // clang-sa doesn't trust our JL_PROPAGATES_ROOT claim
return pml;
}
static void mtcache_hash_insert(_Atomic(jl_array_t*) *cache, jl_value_t *parent, jl_value_t *key, jl_typemap_t *val)
{
int inserted = 0;
jl_array_t *a = jl_atomic_load_relaxed(cache);
if (a == (jl_array_t*)jl_an_empty_vec_any) {
a = jl_alloc_vec_any(16);
jl_atomic_store_release(cache, a);
if (parent)
jl_gc_wb(parent, a);
}
a = jl_eqtable_put(a, key, val, &inserted);
assert(inserted);
if (a != jl_atomic_load_relaxed(cache)) {
jl_atomic_store_release(cache, a);
if (parent)
jl_gc_wb(parent, a);
}
}
static jl_typemap_t *mtcache_hash_lookup(jl_array_t *cache JL_PROPAGATES_ROOT, jl_value_t *ty) JL_NOTSAFEPOINT
{
if (cache == (jl_array_t*)jl_an_empty_vec_any)
return (jl_typemap_t*)jl_nothing;
jl_typemap_t *ml = (jl_typemap_t*)jl_eqtable_get(cache, ty, jl_nothing);
return ml;
}
// ----- Sorted Type Signature Lookup Matching ----- //
static int jl_typemap_array_visitor(jl_array_t *a, jl_typemap_visitor_fptr fptr, void *closure)
{
size_t i, l = jl_array_len(a);
_Atomic(jl_typemap_t*) *data = (_Atomic(jl_typemap_t*)*)jl_array_data(a);
for (i = 1; i < l; i += 2) {
jl_value_t *d = jl_atomic_load_relaxed(&data[i]);
JL_GC_PROMISE_ROOTED(d);
if (d == NULL)
continue;
if (jl_is_array(d)) {
if (!jl_typemap_array_visitor((jl_array_t*)d, fptr, closure))
return 0;
}
else {
if (!jl_typemap_visitor(d, fptr, closure))
return 0;
}
}
return 1;
}
// calls fptr on each jl_typemap_entry_t in cache in sort order, until fptr return false
static int jl_typemap_node_visitor(jl_typemap_entry_t *ml, jl_typemap_visitor_fptr fptr, void *closure)
{
while (ml != (void*)jl_nothing) {
if (!fptr(ml, closure))
return 0;
ml = jl_atomic_load_relaxed(&ml->next);
}
return 1;
}
int jl_typemap_visitor(jl_typemap_t *cache, jl_typemap_visitor_fptr fptr, void *closure)
{
if (jl_typeof(cache) == (jl_value_t*)jl_typemap_level_type) {
jl_typemap_level_t *node = (jl_typemap_level_t*)cache;
jl_array_t *a;
JL_GC_PUSH1(&a);
a = jl_atomic_load_relaxed(&node->targ);
if (a != (jl_array_t*)jl_an_empty_vec_any)
if (!jl_typemap_array_visitor(a, fptr, closure))
goto exit;
a = jl_atomic_load_relaxed(&node->arg1);
if (a != (jl_array_t*)jl_an_empty_vec_any)
if (!jl_typemap_array_visitor(a, fptr, closure))
goto exit;
a = jl_atomic_load_relaxed(&node->tname);
if (a != (jl_array_t*)jl_an_empty_vec_any)
if (!jl_typemap_array_visitor(a, fptr, closure))
goto exit;
a = jl_atomic_load_relaxed(&node->name1);
if (a != (jl_array_t*)jl_an_empty_vec_any)
if (!jl_typemap_array_visitor(a, fptr, closure))
goto exit;
if (!jl_typemap_node_visitor(jl_atomic_load_relaxed(&node->linear), fptr, closure))
goto exit;
if (!jl_typemap_visitor(jl_atomic_load_relaxed(&node->any), fptr, closure))
goto exit;
JL_GC_POP();
return 1;
exit:
JL_GC_POP();
return 0;
}
else {
return jl_typemap_node_visitor((jl_typemap_entry_t*)cache, fptr, closure);
}
}
static unsigned jl_supertype_height(jl_datatype_t *dt) JL_NOTSAFEPOINT
{
unsigned height = 1;
while (dt != jl_any_type) {
height++;
dt = dt->super;
}
return height;
}
// return true if a and b might intersect in the type domain (over just their type-names)
static int tname_intersection_dt(jl_datatype_t *a, jl_typename_t *bname, unsigned ha) JL_NOTSAFEPOINT
{
if (a == jl_any_type)
return 1;
jl_datatype_t *b = (jl_datatype_t*)jl_unwrap_unionall(bname->wrapper);
unsigned hb = 1;
while (b != jl_any_type) {
if (a->name == b->name)
return 1;
hb++;
b = b->super;
}
while (ha > hb) {
a = a->super;
ha--;
}
return a->name == bname;
}
static int tname_intersection(jl_value_t *a, jl_typename_t *bname, int8_t tparam) JL_NOTSAFEPOINT
{
if (a == (jl_value_t*)jl_any_type)
return 1;
a = jl_unwrap_unionall(a);
assert(!jl_is_vararg(a));
if (jl_is_uniontype(a))
return tname_intersection(((jl_uniontype_t*)a)->a, bname, tparam) ||
tname_intersection(((jl_uniontype_t*)a)->b, bname, tparam);
if (jl_is_typevar(a))
return tname_intersection(((jl_tvar_t*)a)->ub, bname, tparam);
if (jl_is_datatype(a)) {
if (tparam) {
if (!jl_is_type_type(a))
return 0;
a = jl_unwrap_unionall(jl_tparam0(a));
if (!jl_is_datatype(a))
return tname_intersection(a, bname, 0);
}
return tname_intersection_dt((jl_datatype_t*)a, bname, jl_supertype_height((jl_datatype_t*)a));
}
return 0;
}
static int concrete_intersects(jl_value_t *t, jl_value_t *ty, int8_t tparam)
{
if (ty == (jl_value_t*)jl_any_type) // easy case: Any always matches
return 1;
if (tparam & 1)
return jl_isa(t, ty); // (Type{t} <: ty), where is_leaf_type(t) => isa(t, ty)
else
return t == ty || jl_subtype(t, ty);
}
// tparam bit 0 is ::Type{T} (vs. T)
// tparam bit 1 is typename(T) (vs. T)
static int jl_typemap_intersection_array_visitor(jl_array_t *a, jl_value_t *ty, int8_t tparam,
int8_t offs, struct typemap_intersection_env *closure)
{
JL_GC_PUSH1(&a);
size_t i, l = jl_array_len(a);
_Atomic(jl_typemap_t*) *data = (_Atomic(jl_typemap_t*)*)jl_array_data(a);
unsigned height = 0;
jl_datatype_t *tydt = jl_any_type;
if (tparam & 2) {
// try to extract a description of ty for intersections, but since we
jl_value_t *ttype = jl_unwrap_unionall(ty);
if (tparam & 1)
// extract T from Type{T} (if possible)
ttype = jl_is_type_type(ttype) ? jl_tparam0(ttype) : NULL;
if (ttype && jl_is_datatype(ttype)) {
tydt = (jl_datatype_t*)ttype;
}
else if (ttype) {
ttype = jl_type_extract_name(ttype);
tydt = ttype ? (jl_datatype_t*)jl_unwrap_unionall(((jl_typename_t*)ttype)->wrapper) : NULL;
}
if (tydt == jl_any_type)
ty = (jl_value_t*)jl_any_type;
else if (tydt == NULL)
tydt = jl_any_type;
else
height = jl_supertype_height(tydt);
}
for (i = 0; i < l; i += 2) {
jl_value_t *t = jl_atomic_load_relaxed(&data[i]);
JL_GC_PROMISE_ROOTED(t);
if (t == jl_nothing || t == NULL)
continue;
if (tparam & 2) {
jl_typemap_t *ml = jl_atomic_load_relaxed(&data[i + 1]);
JL_GC_PROMISE_ROOTED(ml);
if (tydt == jl_any_type ?
tname_intersection(ty, (jl_typename_t*)t, tparam & 1) :
tname_intersection_dt(tydt, (jl_typename_t*)t, height)) {
if ((tparam & 1) && t == (jl_value_t*)jl_typeofbottom_type->name) // skip Type{Union{}} and Type{typeof(Union{})}, since the caller should have already handled those
continue;
if (jl_is_array(ml)) {
if (!jl_typemap_intersection_array_visitor((jl_array_t*)ml, ty, tparam & ~2, offs, closure))
goto exit;
}
else {
if (!jl_typemap_intersection_visitor(ml, offs + 1, closure))
goto exit;
}
}
}
else {
// `t` is a leaftype, so intersection test becomes subtype (after excluding kinds)
if (concrete_intersects(t, ty, tparam)) {
jl_typemap_t *ml = jl_atomic_load_relaxed(&data[i + 1]);
JL_GC_PROMISE_ROOTED(ml);
// NOTE: ml might be NULL if we're racing with the thread that's inserting the item
if (ml != NULL && !jl_typemap_intersection_visitor(ml, offs + 1, closure))
goto exit;
}
}
}
JL_GC_POP();
return 1;
exit:
JL_GC_POP();
return 0;
}
// calls fptr on each jl_typemap_entry_t in cache in sort order
// for which type ∩ ml->type != Union{}, until fptr return false
static int jl_typemap_intersection_node_visitor(jl_typemap_entry_t *ml, struct typemap_intersection_env *closure)
{
// slow-path scan everything in ml
// mark this `register` because (for branch prediction)
// that can be absolutely critical for speed
register jl_typemap_intersection_visitor_fptr fptr = closure->fptr;
for (; ml != (void*)jl_nothing; ml = jl_atomic_load_relaxed(&ml->next)) {
if (closure->max_valid < ml->min_world)
continue;
if (closure->min_valid > ml->max_world)
continue;
jl_svec_t **penv = NULL;
if (closure->env) {
closure->env = jl_emptysvec;
penv = &closure->env;
}
closure->ti = jl_type_intersection_env_s(closure->type, (jl_value_t*)ml->sig, penv, &closure->issubty);
if (closure->ti != (jl_value_t*)jl_bottom_type) {
// In some corner cases type intersection is conservative and returns something
// for intersect(A, B) even though A is a dispatch tuple and !(A <: B).
// For dispatch purposes in such a case we know there's no match. This check
// fixes issue #30394.
if (closure->issubty || !jl_is_dispatch_tupletype(closure->type))
if (!fptr(ml, closure))
return 0;
}
}
return 1;
}
int jl_has_intersect_type_not_kind(jl_value_t *t);
int jl_has_intersect_kind_not_type(jl_value_t *t);
// if TypeVar tv is used covariantly, it cannot be Union{}
int has_covariant_var(jl_datatype_t *ttypes, jl_tvar_t *tv)
{
size_t i, l = jl_nparams(ttypes);
for (i = 0; i < l; i++)
if (jl_tparam(ttypes, i) == (jl_value_t*)tv)
return 1;
return 0;
}
void typemap_slurp_search(jl_typemap_entry_t *ml, struct typemap_intersection_env *closure)
{
// n.b. we could consider mt->max_args here too, so this optimization
// usually works even if the user forgets the `slurp...` argument, but
// there is discussion that parameter may be going away? (and it is
// already not accurately up-to-date for all tables currently anyways)
if (closure->search_slurp && ml->va) {
jl_value_t *sig = jl_unwrap_unionall((jl_value_t*)ml->sig);
size_t nargs = jl_nparams(sig);
if (nargs > 1 && nargs - 1 == closure->search_slurp) {
jl_vararg_t *va = (jl_vararg_t*)jl_tparam(sig, nargs - 1);
assert(jl_is_vararg((jl_value_t*)va));
if (va->T == (jl_value_t*)jl_any_type && va->N == NULL) {
// instruct typemap it can set exclude_typeofbottom on parameter nargs
// since we found the necessary slurp argument
closure->search_slurp = 0;
}
}
}
}
int jl_typemap_intersection_visitor(jl_typemap_t *map, int offs,
struct typemap_intersection_env *closure)
{
jl_value_t *ttypes = jl_unwrap_unionall(closure->type);
assert(jl_is_datatype(ttypes));
//TODO: fast-path for leaf-type tuples?
//if (ttypes->isdispatchtuple) {
// register jl_typemap_intersection_visitor_fptr fptr = closure->fptr;
// struct jl_typemap_assoc search = {(jl_value_t*)closure->type, world, closure->env, 0, ~(size_t)0};
// jl_typemap_entry_t *ml = jl_typemap_assoc_by_type(map, search, offs, /*subtype*/1);
// if (ml) {
// closure->env = search->env;
// if (!fptr(ml, closure))
// return 0;
// }
// return 1;
//}
if (jl_typeof(map) == (jl_value_t *)jl_typemap_level_type) {
jl_typemap_level_t *cache = (jl_typemap_level_t*)map;
jl_value_t *ty;
size_t l = jl_nparams(ttypes);
if (closure->va && l <= offs + 1) {
ty = closure->va;
}
else if (l > offs) {
ty = jl_tparam(ttypes, offs);
}
else {
ty = NULL;
}
if (ty == (jl_value_t*)jl_typeofbottom_type)
ty = (jl_value_t*)jl_assume(jl_typeofbottom_type)->super;
if (ty) {
while (jl_is_typevar(ty))
ty = ((jl_tvar_t*)ty)->ub;
// approxify the tparam until we have a valid type
if (jl_has_free_typevars(ty))
ty = jl_rewrap_unionall(ty, closure->type);
JL_GC_PUSH1(&ty);
jl_array_t *targ = jl_atomic_load_relaxed(&cache->targ);
jl_array_t *tname = jl_atomic_load_relaxed(&cache->tname);
int maybe_type = 0;
int maybe_kind = 0;
int exclude_typeofbottom = 0;
jl_value_t *typetype = NULL;
jl_value_t *name = NULL;
// pre-check: optimized pre-intersection test to see if `ty` could intersect with any Type or Kind
if (targ != (jl_array_t*)jl_an_empty_vec_any || tname != (jl_array_t*)jl_an_empty_vec_any) {
maybe_kind = jl_has_intersect_kind_not_type(ty);
maybe_type = maybe_kind || jl_has_intersect_type_not_kind(ty);
if (maybe_type && !maybe_kind) {
typetype = jl_unwrap_unionall(ty);
typetype = jl_is_type_type(typetype) ? jl_tparam0(typetype) : NULL;
name = typetype ? jl_type_extract_name(typetype) : NULL;
if (!typetype)
exclude_typeofbottom = !jl_subtype((jl_value_t*)jl_typeofbottom_type, ty);
else if (jl_is_typevar(typetype))
exclude_typeofbottom = has_covariant_var((jl_datatype_t*)ttypes, (jl_tvar_t*)typetype);
else
exclude_typeofbottom = !jl_parameter_includes_bottom(typetype);
}
}
// First check for intersections with methods defined on Type{T}, where T was a concrete type
if (targ != (jl_array_t*)jl_an_empty_vec_any && maybe_type &&
(!typetype || jl_has_free_typevars(typetype) || is_cache_leaf(typetype, 1))) { // otherwise cannot contain this particular kind, so don't bother with checking
if (!exclude_typeofbottom) {
// detect Type{Union{}}, Type{Type{Union{}}}, and Type{typeof(Union{}} and do those early here
// otherwise the possibility of encountering `Type{Union{}}` in this intersection may
// be forcing us to do some extra work here whenever we see a typevar, even though
// the likelihood of that value actually occurring is frequently likely to be
// zero (or result in an ambiguous match)
targ = jl_atomic_load_relaxed(&cache->targ); // may be GC'd during type-intersection
jl_value_t *ml = mtcache_hash_lookup(targ, (jl_value_t*)jl_typeofbottom_type->name);
if (ml != jl_nothing) {
size_t search_slurp = closure->search_slurp;
closure->search_slurp = offs + 1;
if (!jl_typemap_intersection_visitor((jl_typemap_t*)ml, offs+1, closure)) {
closure->search_slurp = search_slurp;
JL_GC_POP();
return 0;
}
if (closure->search_slurp == 0)
exclude_typeofbottom = 1;
closure->search_slurp = search_slurp;
}
}
if (name != (jl_value_t*)jl_typeofbottom_type->name) {
targ = jl_atomic_load_relaxed(&cache->targ); // may be GC'd earlier
if (exclude_typeofbottom && name && jl_type_extract_name_precise(typetype, 1)) {
// attempt semi-direct lookup of types via their names
// consider the type name first
jl_value_t *ml = mtcache_hash_lookup(targ, (jl_value_t*)name);
if (jl_is_array(ml)) {
if (typetype && !jl_has_free_typevars(typetype)) {
// direct lookup of leaf types
if (is_cache_leaf(typetype, 1)) {
ml = mtcache_hash_lookup((jl_array_t*)ml, typetype);
if (ml != jl_nothing) {
if (!jl_typemap_intersection_visitor((jl_typemap_t*)ml, offs+1, closure)) { JL_GC_POP(); return 0; }
}
}
}
else {
if (!jl_typemap_intersection_array_visitor((jl_array_t*)ml, ty, 1, offs, closure)) { JL_GC_POP(); return 0; }
}
}
else if (ml != jl_nothing) {
if (!jl_typemap_intersection_visitor((jl_typemap_t*)ml, offs+1, closure)) { JL_GC_POP(); return 0; }
}
}
else {
// else an array scan is required to consider all the possible subtypes
if (!jl_typemap_intersection_array_visitor(targ, exclude_typeofbottom && !maybe_kind ? ty : (jl_value_t*)jl_any_type, 3, offs, closure)) { JL_GC_POP(); return 0; }
}
}
}
jl_array_t *cachearg1 = jl_atomic_load_relaxed(&cache->arg1);
if (cachearg1 != (jl_array_t*)jl_an_empty_vec_any) {
if (is_cache_leaf(ty, 0)) {
jl_typename_t *name = ty == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)ty)->name;
// direct lookup of leaf types
jl_value_t *ml = mtcache_hash_lookup(cachearg1, (jl_value_t*)name);
if (jl_is_array(ml))
ml = mtcache_hash_lookup((jl_array_t*)ml, ty);
if (ml != jl_nothing) {
if (!jl_typemap_intersection_visitor(ml, offs+1, closure)) { JL_GC_POP(); return 0; }
}
}
else {
jl_value_t *name = jl_type_extract_name(ty);
if (name && jl_type_extract_name_precise(ty, 0)) {
// direct lookup of leaf types
jl_value_t *ml = mtcache_hash_lookup(cachearg1, name);
if (jl_is_array(ml)) {
if (!jl_typemap_intersection_array_visitor((jl_array_t*)ml, ty, 0, offs, closure)) { JL_GC_POP(); return 0; }
}
else {
if (!jl_typemap_intersection_visitor((jl_typemap_t*)ml, offs+1, closure)) { JL_GC_POP(); return 0; }
}
}
else {
// else an array scan is required to check subtypes
if (!jl_typemap_intersection_array_visitor(cachearg1, ty, 2, offs, closure)) { JL_GC_POP(); return 0; }
}
}
}
// Next check for intersections with methods defined on Type{T}, where T was not concrete (it might even have been a TypeVar), but had an extractable TypeName
if (tname != (jl_array_t*)jl_an_empty_vec_any && maybe_type) {
if (!exclude_typeofbottom || (!typetype && jl_isa((jl_value_t*)jl_typeofbottom_type, ty))) {
// detect Type{Union{}}, Type{Type{Union{}}}, and Type{typeof(Union{}} and do those early here
// otherwise the possibility of encountering `Type{Union{}}` in this intersection may
// be forcing us to do some extra work here whenever we see a typevar, even though
// the likelihood of that value actually occurring is frequently likely to be
// zero (or result in an ambiguous match)
tname = jl_atomic_load_relaxed(&cache->tname); // may be GC'd earlier
jl_value_t *ml = mtcache_hash_lookup(tname, (jl_value_t*)jl_typeofbottom_type->name);
if (ml != jl_nothing) {
size_t search_slurp = closure->search_slurp;
closure->search_slurp = offs + 1;
if (!jl_typemap_intersection_visitor((jl_typemap_t*)ml, offs+1, closure)) {
closure->search_slurp = search_slurp;
JL_GC_POP();
return 0;
}
if (closure->search_slurp == 0)
exclude_typeofbottom = 1;
closure->search_slurp = search_slurp;
}
}
if (exclude_typeofbottom && name && jl_type_extract_name_precise(typetype, 1)) {
// semi-direct lookup of types
// just consider the type and its direct super types
jl_datatype_t *super = (jl_datatype_t*)jl_unwrap_unionall(((jl_typename_t*)name)->wrapper);
if (super->name == jl_typeofbottom_type->name)
super = super->super; // this was handled above
while (1) {
tname = jl_atomic_load_relaxed(&cache->tname); // reload after callback
jl_typemap_t *ml = mtcache_hash_lookup(tname, (jl_value_t*)super->name);
if (ml != jl_nothing) {
if (!jl_typemap_intersection_visitor(ml, offs+1, closure)) { JL_GC_POP(); return 0; }
}
if (super == jl_any_type)
break;
super = super->super;
}
}
else {
// else an array scan is required to check subtypes of typetype too
tname = jl_atomic_load_relaxed(&cache->tname); // may be GC'd earlier
if (!jl_typemap_intersection_array_visitor(tname, exclude_typeofbottom && !maybe_kind ? ty : (jl_value_t*)jl_any_type, 3, offs, closure)) { JL_GC_POP(); return 0; }
}
}
jl_array_t *name1 = jl_atomic_load_relaxed(&cache->name1);
if (name1 != (jl_array_t*)jl_an_empty_vec_any) {
jl_value_t *name = jl_type_extract_name(ty);
if (name && jl_type_extract_name_precise(ty, 0)) {
jl_datatype_t *super = (jl_datatype_t*)jl_unwrap_unionall(((jl_typename_t*)name)->wrapper);
// direct lookup of concrete types
while (1) {
name1 = jl_atomic_load_relaxed(&cache->name1); // reload after callback
jl_typemap_t *ml = mtcache_hash_lookup(name1, (jl_value_t*)super->name);
if (ml != jl_nothing) {
if (!jl_typemap_intersection_visitor(ml, offs+1, closure)) { JL_GC_POP(); return 0; }
}
if (super == jl_any_type)
break;
super = super->super;
}
}
else {
// else an array scan is required to check subtypes
if (!jl_typemap_intersection_array_visitor(name1, ty, 2, offs, closure)) { JL_GC_POP(); return 0; }
}
}
JL_GC_POP();
}
if (!jl_typemap_intersection_node_visitor(jl_atomic_load_relaxed(&cache->linear), closure))
return 0;
return jl_typemap_intersection_visitor(jl_atomic_load_relaxed(&cache->any), offs+1, closure);
}
else {
return jl_typemap_intersection_node_visitor(
(jl_typemap_entry_t*)map, closure);
}
}
/*
Method caches are divided into three parts: one for signatures where
the first argument is a singleton kind (Type{Foo}), one indexed by the
UID of the first argument's type in normal cases, and a fallback
table of everything else.
Note that the "primary key" is the type of the first *argument*, since
there tends to be lots of variation there. The type of the 0th argument
(the function) is always the same for most functions.
*/
static jl_typemap_entry_t *jl_typemap_entry_assoc_by_type(
jl_typemap_entry_t *ml,
struct jl_typemap_assoc *search)
{
jl_value_t *types = search->types;
JL_GC_PROMISE_ROOTED(types);
jl_value_t *unw = jl_unwrap_unionall((jl_value_t*)types);
int isua = jl_is_unionall(types);
size_t n = jl_nparams(unw);
int typesisva = n == 0 ? 0 : jl_is_vararg(jl_tparam(unw, n-1));
for (; ml != (void*)jl_nothing; ml = jl_atomic_load_relaxed(&ml->next)) {
if (search->max_valid < ml->min_world)
continue;
if (search->min_valid > ml->max_world)
continue;
size_t lensig = jl_nparams(jl_unwrap_unionall((jl_value_t*)ml->sig));
if (lensig == n || (ml->va && lensig <= n+1)) {
int resetenv = 0, ismatch = 1;
if (ml->simplesig != (void*)jl_nothing && !isua) {
size_t lensimplesig = jl_nparams(ml->simplesig);
int isva = lensimplesig > 0 && jl_is_vararg(jl_tparam(ml->simplesig, lensimplesig - 1));
if (lensig == n || (isva && lensimplesig <= n + 1))
ismatch = sig_match_by_type_simple(jl_svec_data(((jl_datatype_t*)types)->parameters), n,
ml->simplesig, lensimplesig, isva);
else
ismatch = 0;
}
if (ismatch == 0)
; // nothing
else if (ml->isleafsig && !typesisva && !isua)
ismatch = sig_match_by_type_leaf(jl_svec_data(((jl_datatype_t*)types)->parameters),
ml->sig, lensig);
else if (ml->issimplesig && !typesisva && !isua)
ismatch = sig_match_by_type_simple(jl_svec_data(((jl_datatype_t*)types)->parameters), n,
ml->sig, lensig, ml->va);
else {
ismatch = jl_subtype_matching(types, (jl_value_t*)ml->sig, search->env ? &search->env : NULL);
if (ismatch && search->env)
resetenv = 1;
}
if (ismatch) {
size_t i, l;
for (i = 0, l = jl_svec_len(ml->guardsigs); i < l; i++) {
// see corresponding code in jl_typemap_entry_assoc_exact
if (jl_subtype(types, jl_svecref(ml->guardsigs, i))) {
ismatch = 0;
break;
}
}
if (ismatch) {
if (search->world < ml->min_world) {
// ignore method table entries that are part of a later world
if (search->max_valid >= ml->min_world)
search->max_valid = ml->min_world - 1;
}
else if (search->world > ml->max_world) {
// ignore method table entries that have been replaced in the current world
if (search->min_valid <= ml->max_world)
search->min_valid = ml->max_world + 1;
}
else {
// intersect the env valid range with method's valid range
if (search->min_valid < ml->min_world)
search->min_valid = ml->min_world;
if (search->max_valid > ml->max_world)
search->max_valid = ml->max_world;
return ml;
}
}
}
if (resetenv)
search->env = jl_emptysvec;
}
}
return NULL;
}
static jl_typemap_entry_t *jl_typemap_entry_lookup_by_type(
jl_typemap_entry_t *ml, struct jl_typemap_assoc *search)
{
for (; ml != (void*)jl_nothing; ml = jl_atomic_load_relaxed(&ml->next)) {
if (search->world < ml->min_world || search->world > ml->max_world)
continue;
// unroll the first few cases here, to the extent that is possible to do fast and easily
jl_value_t *types = search->types;
JL_GC_PROMISE_ROOTED(types);
jl_value_t *a = jl_unwrap_unionall(types);
jl_value_t *b = jl_unwrap_unionall((jl_value_t*)ml->sig);
size_t na = jl_nparams(a);
size_t nb = jl_nparams(b);
int va_a = na > 0 && jl_is_vararg(jl_tparam(a, na - 1));
int va_b = nb > 0 && jl_is_vararg(jl_tparam(b, nb - 1));
if (!va_a && !va_b) {
if (na != nb)
continue;
}
if (na - va_a > 0 && nb - va_b > 0) {
if (jl_obviously_unequal(jl_tparam(a, 0), jl_tparam(b, 0)))
continue;
if (na - va_a > 1 && nb - va_b > 1) {
if (jl_obviously_unequal(jl_tparam(a, 1), jl_tparam(b, 1)))
continue;
if (na - va_a > 2 && nb - va_b > 2) {
if (jl_obviously_unequal(jl_tparam(a, 2), jl_tparam(b, 2)))
continue;
}
}
}
if (jl_types_equal(types, (jl_value_t*)ml->sig))
return ml;
}
return NULL;
}
// this is the general entry point for looking up a type in the cache
// as a subtype, or with type_equal
jl_typemap_entry_t *jl_typemap_assoc_by_type(
jl_typemap_t *ml_or_cache,
struct jl_typemap_assoc *search,
int8_t offs, uint8_t subtype)
{
if (jl_typeof(ml_or_cache) == (jl_value_t *)jl_typemap_level_type) {
jl_typemap_level_t *cache = (jl_typemap_level_t*)ml_or_cache;
// called object is the primary key for constructors, otherwise first argument
jl_value_t *ty;
jl_value_t *ttypes = jl_unwrap_unionall((jl_value_t*)search->types);
JL_GC_PROMISE_ROOTED(ttypes);
assert(jl_is_datatype(ttypes));
size_t l = jl_nparams(ttypes);
int isva = 0;
// compute the type at offset `offs` into `types`, which may be a Vararg
if (l <= offs + 1) {
ty = jl_tparam(ttypes, l - 1);
if (jl_is_vararg(ty)) {
ty = jl_unwrap_vararg(ty);
isva = 1;
}
else if (l <= offs) {
ty = NULL;
}
}
else if (l > offs) {
ty = jl_tparam(ttypes, offs);
}
else {
ty = NULL;
}
if (ty == (jl_value_t*)jl_typeofbottom_type)
ty = (jl_value_t*)jl_assume(jl_typeofbottom_type)->super;
// If there is a type at offs, look in the optimized leaf type caches
if (ty && !subtype) {
if (jl_is_any(ty))
return jl_typemap_assoc_by_type(jl_atomic_load_relaxed(&cache->any), search, offs + 1, subtype);
if (isva) // in lookup mode, want to match Vararg exactly, not as a subtype
ty = NULL;
}
if (ty) {
// now look at the optimized leaftype caches
if (jl_is_type_type(ty)) {
jl_value_t *a0 = jl_tparam0(ty);
if (is_cache_leaf(a0, 1)) {
jl_array_t *targ = jl_atomic_load_relaxed(&cache->targ);
if (targ != (jl_array_t*)jl_an_empty_vec_any) {
jl_typename_t *name = a0 == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)a0)->name;
jl_value_t *ml = mtcache_hash_lookup(targ, (jl_value_t*)name);
if (jl_is_array(ml))
ml = mtcache_hash_lookup((jl_array_t*)ml, a0);
if (ml != jl_nothing) {
jl_typemap_entry_t *li = jl_typemap_assoc_by_type((jl_typemap_t*)ml, search, offs + 1, subtype);
if (li) return li;
}
}
if (!subtype) return NULL;
}
}
if (is_cache_leaf(ty, 0)) {
jl_array_t *cachearg1 = jl_atomic_load_relaxed(&cache->arg1);
if (cachearg1 != (jl_array_t*)jl_an_empty_vec_any) {
jl_typename_t *name = ty == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)ty)->name;
jl_value_t *ml = mtcache_hash_lookup(cachearg1, (jl_value_t*)name);
if (jl_is_array(ml))
ml = mtcache_hash_lookup((jl_array_t*)ml, ty);
if (ml != jl_nothing) {
jl_typemap_entry_t *li = jl_typemap_assoc_by_type((jl_typemap_t*)ml, search, offs + 1, subtype);
if (li) return li;
}
}
if (!subtype) return NULL;
}
}
if (ty || subtype) {
// now look at the optimized TypeName caches
jl_array_t *tname = jl_atomic_load_relaxed(&cache->tname);
if (tname != (jl_array_t*)jl_an_empty_vec_any) {
jl_value_t *a0 = ty && jl_is_type_type(ty) ? jl_type_extract_name(jl_tparam0(ty)) : NULL;
if (a0) { // TODO: if we start analyzing Union types in jl_type_extract_name, then a0 might be over-approximated here, leading us to miss possible subtypes
jl_datatype_t *super = (jl_datatype_t*)jl_unwrap_unionall(((jl_typename_t*)a0)->wrapper);
while (1) {
tname = jl_atomic_load_relaxed(&cache->tname); // reload after tree descent (which may hit safepoints)
jl_typemap_t *ml = mtcache_hash_lookup(tname, (jl_value_t*)super->name);
if (ml != (void*)jl_nothing) {
jl_typemap_entry_t *li = jl_typemap_assoc_by_type(ml, search, offs + 1, subtype);
if (li) return li;
}
if (super == jl_any_type || !subtype)
break;
super = super->super;
}
}
else {
if (!ty || !jl_has_empty_intersection((jl_value_t*)jl_type_type, ty)) {
// couldn't figure out unique `a0` initial point, so scan all for matches
size_t i, l = jl_array_len(tname);
_Atomic(jl_typemap_t*) *data = (_Atomic(jl_typemap_t*)*)jl_array_ptr_data(tname);
JL_GC_PUSH1(&tname);
for (i = 1; i < l; i += 2) {
jl_typemap_t *ml = jl_atomic_load_relaxed(&data[i]);
if (ml == NULL || ml == jl_nothing)
continue;
jl_typemap_entry_t *li = jl_typemap_assoc_by_type(ml, search, offs + 1, subtype);
if (li) {
JL_GC_POP();
return li;
}
}
JL_GC_POP();
}
}
}
jl_array_t *name1 = jl_atomic_load_relaxed(&cache->name1);
if (name1 != (jl_array_t*)jl_an_empty_vec_any) {
if (ty) {
jl_value_t *a0 = jl_type_extract_name(ty);
if (a0) { // TODO: if we start analyzing Union types in jl_type_extract_name, then a0 might be over-approximated here, leading us to miss possible subtypes
jl_datatype_t *super = (jl_datatype_t*)jl_unwrap_unionall(((jl_typename_t*)a0)->wrapper);
while (1) {
name1 = jl_atomic_load_relaxed(&cache->name1); // reload after tree descent (which may hit safepoints)
jl_typemap_t *ml = mtcache_hash_lookup(name1, (jl_value_t*)super->name);
if (ml != (void*)jl_nothing) {
jl_typemap_entry_t *li =
jl_typemap_assoc_by_type(ml, search, offs + 1, subtype);
if (li) return li;
}
if (super == jl_any_type || !subtype)
break;
super = super->super;
}
}
}
else {
// doing subtype, but couldn't figure out unique `ty`, so scan all for supertypes
size_t i, l = jl_array_len(name1);
_Atomic(jl_typemap_t*) *data = (_Atomic(jl_typemap_t*)*)jl_array_ptr_data(name1);
JL_GC_PUSH1(&name1);
for (i = 1; i < l; i += 2) {
jl_typemap_t *ml = jl_atomic_load_relaxed(&data[i]);
if (ml == NULL || ml == jl_nothing)
continue;
jl_typemap_entry_t *li = jl_typemap_assoc_by_type(ml, search, offs + 1, subtype);
if (li) {
JL_GC_POP();
return li;
}
}
JL_GC_POP();
}
}
}
// Always check the list (since offs doesn't always start at 0)
if (subtype) {
jl_typemap_entry_t *li = jl_typemap_entry_assoc_by_type(jl_atomic_load_relaxed(&cache->linear), search);
if (li) return li;
return jl_typemap_assoc_by_type(jl_atomic_load_relaxed(&cache->any), search, offs + 1, subtype);
}
else {
return jl_typemap_entry_lookup_by_type(jl_atomic_load_relaxed(&cache->linear), search);
}
}
else {
jl_typemap_entry_t *leaf = (jl_typemap_entry_t*)ml_or_cache;
return subtype ?
jl_typemap_entry_assoc_by_type(leaf, search) :
jl_typemap_entry_lookup_by_type(leaf, search);
}
}
jl_typemap_entry_t *jl_typemap_entry_assoc_exact(jl_typemap_entry_t *ml, jl_value_t *arg1, jl_value_t **args, size_t n, size_t world)
{
// some manually-unrolled common special cases
while (ml->simplesig == (void*)jl_nothing && ml->guardsigs == jl_emptysvec && ml->isleafsig) {
// use a tight loop for as long as possible
if (world >= ml->min_world && world <= ml->max_world) {
if (n == jl_nparams(ml->sig) && jl_typeof(arg1) == jl_tparam(ml->sig, 0)) {
if (n == 1)
return ml;
if (n == 2) {
if (jl_typeof(args[0]) == jl_tparam(ml->sig, 1))
return ml;
}
else if (n == 3) {
if (jl_typeof(args[0]) == jl_tparam(ml->sig, 1) &&
jl_typeof(args[1]) == jl_tparam(ml->sig, 2))
return ml;
}
else {
if (sig_match_leaf(arg1, args, jl_svec_data(ml->sig->parameters), n))
return ml;
}
}
}
ml = jl_atomic_load_relaxed(&ml->next);
if (ml == (void*)jl_nothing)
return NULL;
}
for (; ml != (void*)jl_nothing; ml = jl_atomic_load_relaxed(&ml->next)) {
if (world < ml->min_world || world > ml->max_world)
continue; // ignore replaced methods
size_t lensig = jl_nparams(ml->sig);
if (lensig == n || (ml->va && lensig <= n+1)) {
if (ml->simplesig != (void*)jl_nothing) {
size_t lensimplesig = jl_nparams(ml->simplesig);
int isva = lensimplesig > 0 && jl_is_vararg(jl_tparam(ml->simplesig, lensimplesig - 1));
if (lensig == n || (isva && lensimplesig <= n + 1)) {
if (!sig_match_simple(arg1, args, n, jl_svec_data(ml->simplesig->parameters), isva, lensimplesig))
continue;
}
else {
continue;
}
}
if (ml->isleafsig) {
if (!sig_match_leaf(arg1, args, jl_svec_data(ml->sig->parameters), n))
continue;
}
else if (ml->issimplesig) {
if (!sig_match_simple(arg1, args, n, jl_svec_data(ml->sig->parameters), ml->va, lensig))
continue;
}
else {
if (!jl_tuple1_isa(arg1, args, n, ml->sig))
continue;
}
size_t i, l;
if (ml->guardsigs != jl_emptysvec) {
for (i = 0, l = jl_svec_len(ml->guardsigs); i < l; i++) {
// checking guard entries require a more
// expensive subtype check, since guard entries added for @nospecialize might be
// abstract. this fixed issue #12967.
if (jl_tuple1_isa(arg1, args, n, (jl_tupletype_t*)jl_svecref(ml->guardsigs, i))) {
goto nomatch;
}
}
}
return ml;
nomatch:
continue;
}
}
return NULL;
}
jl_typemap_entry_t *jl_typemap_level_assoc_exact(jl_typemap_level_t *cache, jl_value_t *arg1, jl_value_t **args, size_t n, int8_t offs, size_t world)
{
if (n > offs) {
jl_value_t *a1 = (offs == 0 ? arg1 : args[offs - 1]);
jl_value_t *ty = jl_typeof(a1);
assert(jl_is_datatype(ty));
jl_array_t *targ = jl_atomic_load_relaxed(&cache->targ);
if (targ != (jl_array_t*)jl_an_empty_vec_any && is_cache_leaf(a1, 1)) {
jl_typename_t *name = a1 == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)a1)->name;
jl_value_t *ml_or_cache = mtcache_hash_lookup(targ, (jl_value_t*)name);
if (jl_is_array(ml_or_cache))
ml_or_cache = mtcache_hash_lookup((jl_array_t*)ml_or_cache, a1);
jl_typemap_entry_t *ml = jl_typemap_assoc_exact(ml_or_cache, arg1, args, n, offs+1, world);
if (ml) return ml;
}
jl_array_t *cachearg1 = jl_atomic_load_relaxed(&cache->arg1);
if (cachearg1 != (jl_array_t*)jl_an_empty_vec_any && is_cache_leaf(ty, 0)) {
jl_typename_t *name = ty == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)ty)->name;
jl_value_t *ml_or_cache = mtcache_hash_lookup(cachearg1, (jl_value_t*)name);
if (jl_is_array(ml_or_cache))
ml_or_cache = mtcache_hash_lookup((jl_array_t*)ml_or_cache, ty);
jl_typemap_entry_t *ml = jl_typemap_assoc_exact((jl_typemap_t*)ml_or_cache, arg1, args, n, offs+1, world);
if (ml) return ml;
}
jl_array_t *tname = jl_atomic_load_relaxed(&cache->tname);
if (jl_is_kind(ty) && tname != (jl_array_t*)jl_an_empty_vec_any) {
jl_value_t *name = jl_type_extract_name(a1);
if (name) {
if (ty != (jl_value_t*)jl_datatype_type)
a1 = jl_unwrap_unionall(((jl_typename_t*)name)->wrapper);
while (1) {
tname = jl_atomic_load_relaxed(&cache->tname); // reload after tree descent (which may hit safepoints)
jl_typemap_t *ml_or_cache = mtcache_hash_lookup(
tname, (jl_value_t*)((jl_datatype_t*)a1)->name);
jl_typemap_entry_t *ml = jl_typemap_assoc_exact(ml_or_cache, arg1, args, n, offs+1, world);
if (ml) return ml;
if (a1 == (jl_value_t*)jl_any_type)
break;
a1 = (jl_value_t*)((jl_datatype_t*)a1)->super;
}
}
else {
// couldn't figure out unique `name` initial point, so must scan all for matches
size_t i, l = jl_array_len(tname);
_Atomic(jl_typemap_t*) *data = (_Atomic(jl_typemap_t*)*)jl_array_ptr_data(tname);
JL_GC_PUSH1(&tname);
for (i = 1; i < l; i += 2) {
jl_typemap_t *ml_or_cache = jl_atomic_load_relaxed(&data[i]);
if (ml_or_cache == NULL || ml_or_cache == jl_nothing)
continue;
jl_typemap_entry_t *ml = jl_typemap_assoc_exact(ml_or_cache, arg1, args, n, offs + 1, world);
if (ml) {
JL_GC_POP();
return ml;
}
}
JL_GC_POP();
}
}
jl_array_t *name1 = jl_atomic_load_relaxed(&cache->name1);
if (name1 != (jl_array_t*)jl_an_empty_vec_any) {
while (1) {
name1 = jl_atomic_load_relaxed(&cache->name1); // reload after tree descent (which may hit safepoints)
jl_typemap_t *ml_or_cache = mtcache_hash_lookup(
name1, (jl_value_t*)((jl_datatype_t*)ty)->name);
jl_typemap_entry_t *ml = jl_typemap_assoc_exact(ml_or_cache, arg1, args, n, offs+1, world);
if (ml) return ml;
if (ty == (jl_value_t*)jl_any_type)
break;
ty = (jl_value_t*)((jl_datatype_t*)ty)->super;
}
}
}
jl_typemap_entry_t *linear = jl_atomic_load_relaxed(&cache->linear);
if (linear != (jl_typemap_entry_t*)jl_nothing) {
jl_typemap_entry_t *ml = jl_typemap_entry_assoc_exact(linear, arg1, args, n, world);
if (ml) return ml;
}
jl_typemap_t *cacheany = jl_atomic_load_relaxed(&cache->any);
if (cacheany != (jl_typemap_t*)jl_nothing)
return jl_typemap_assoc_exact(cacheany, arg1, args, n, offs+1, world);
return NULL;
}
// ----- Method List Insertion Management ----- //
static unsigned jl_typemap_list_count_locked(jl_typemap_entry_t *ml) JL_NOTSAFEPOINT
{
unsigned count = 0;
while (ml != (void*)jl_nothing) {
count++;
ml = jl_atomic_load_relaxed(&ml->next);
}
return count;
}
static void jl_typemap_level_insert_(jl_typemap_t *map, jl_typemap_level_t *cache, jl_typemap_entry_t *newrec, int8_t offs);
static jl_typemap_level_t *jl_new_typemap_level(void)
{
jl_task_t *ct = jl_current_task;
jl_typemap_level_t *cache =
(jl_typemap_level_t*)jl_gc_alloc(ct->ptls, sizeof(jl_typemap_level_t),
jl_typemap_level_type);
jl_atomic_store_relaxed(&cache->arg1, (jl_array_t*)jl_an_empty_vec_any);
jl_atomic_store_relaxed(&cache->targ, (jl_array_t*)jl_an_empty_vec_any);
jl_atomic_store_relaxed(&cache->name1, (jl_array_t*)jl_an_empty_vec_any);
jl_atomic_store_relaxed(&cache->tname, (jl_array_t*)jl_an_empty_vec_any);
jl_atomic_store_relaxed(&cache->linear, (jl_typemap_entry_t*)jl_nothing);
jl_atomic_store_relaxed(&cache->any, jl_nothing);
return cache;
}
static void jl_typemap_array_insert_(
jl_typemap_t *map, _Atomic(jl_array_t*) *pcache, jl_value_t *key, jl_typemap_entry_t *newrec,
jl_value_t *parent, int8_t tparam, int8_t offs, jl_value_t *doublesplit);
static jl_value_t *jl_method_convert_list_to_cache(
jl_typemap_t *map, jl_typemap_entry_t *ml, int8_t tparam, int8_t offs, int8_t doublesplit)
{
jl_value_t *cache = doublesplit ? jl_an_empty_vec_any : (jl_value_t*)jl_new_typemap_level();
jl_typemap_entry_t *next = NULL;
JL_GC_PUSH3(&cache, &next, &ml);
while (ml != (void*)jl_nothing) {
next = jl_atomic_load_relaxed(&ml->next);
jl_atomic_store_relaxed(&ml->next, (jl_typemap_entry_t*)jl_nothing);
// n.b. this is being done concurrently with lookups!
// TODO: is it safe to be doing this concurrently with lookups?
if (doublesplit) {
jl_value_t *key = jl_unwrap_unionall((jl_value_t*)ml->sig);
size_t len = jl_nparams(key);
if (offs < len-1)
key = jl_tparam(key, offs);
else
key = jl_tparam(key, len-1);
if (jl_is_vararg(key))
key = jl_unwrap_vararg(key);
if (key == (jl_value_t*)jl_typeofbottom_type)
key = (jl_value_t*)jl_assume(jl_typeofbottom_type)->super;
if (tparam) {
assert(jl_is_type_type(key));
key = jl_tparam0(key);
}
jl_typemap_array_insert_(map, (_Atomic(jl_array_t*)*)&cache, key, ml, NULL, 0, offs, NULL);
}
else
jl_typemap_level_insert_(map, (jl_typemap_level_t*)cache, ml, offs);
ml = next;
}
JL_GC_POP();
return cache;
}
static void jl_typemap_list_insert_(
jl_typemap_t *map, _Atomic(jl_typemap_entry_t*) *pml, jl_value_t *parent,
jl_typemap_entry_t *newrec)
{
jl_typemap_entry_t *l = jl_atomic_load_relaxed(pml);
while ((jl_value_t*)l != jl_nothing) {
if (newrec->isleafsig || !l->isleafsig)
if (newrec->issimplesig || !l->issimplesig)
break;
pml = &l->next;
parent = (jl_value_t*)l;
l = jl_atomic_load_relaxed(&l->next);
}
jl_atomic_store_relaxed(&newrec->next, l);
jl_gc_wb(newrec, l);
jl_atomic_store_release(pml, newrec);
jl_gc_wb(parent, newrec);
}
// n.b. tparam value only needed if doublesplit is set (for jl_method_convert_list_to_cache)
static void jl_typemap_insert_generic(
jl_typemap_t *map, _Atomic(jl_value_t*) *pml, jl_value_t *parent,
jl_typemap_entry_t *newrec, int8_t tparam, int8_t offs, jl_value_t *doublesplit)
{
jl_value_t *ml = jl_atomic_load_relaxed(pml);
if (jl_is_array(ml)) {
assert(doublesplit);
jl_typemap_array_insert_(map, (_Atomic(jl_array_t*)*)pml, doublesplit, newrec, parent, 0, offs, NULL);
return;
}
if (jl_typeof(ml) == (jl_value_t*)jl_typemap_level_type) {
assert(!doublesplit);
jl_typemap_level_insert_(map, (jl_typemap_level_t*)ml, newrec, offs);
return;
}
unsigned count = jl_typemap_list_count_locked((jl_typemap_entry_t*)ml);
if (count > MAX_METHLIST_COUNT) {
ml = jl_method_convert_list_to_cache(
map, (jl_typemap_entry_t*)ml, tparam, offs, doublesplit != NULL);
jl_atomic_store_release(pml, ml);
jl_gc_wb(parent, ml);
if (doublesplit)
jl_typemap_array_insert_(map, (_Atomic(jl_array_t*)*)pml, doublesplit, newrec, parent, 0, offs, NULL);
else
jl_typemap_level_insert_(map, (jl_typemap_level_t*)ml, newrec, offs);
return;
}
jl_typemap_list_insert_(map, (_Atomic(jl_typemap_entry_t*)*)pml,
parent, newrec);
}
static void jl_typemap_array_insert_(
jl_typemap_t *map, _Atomic(jl_array_t*) *pcache, jl_value_t *key, jl_typemap_entry_t *newrec,
jl_value_t *parent, int8_t tparam, int8_t offs, jl_value_t *doublesplit)
{
jl_array_t *cache = jl_atomic_load_relaxed(pcache);
_Atomic(jl_value_t*) *pml = mtcache_hash_lookup_bp(cache, key);
if (pml == NULL)
mtcache_hash_insert(pcache, parent, key, (jl_typemap_t*)newrec);
else
jl_typemap_insert_generic(map, pml, (jl_value_t*)cache, newrec, tparam, offs + (doublesplit ? 0 : 1), doublesplit);
}
static void jl_typemap_level_insert_(
jl_typemap_t *map, jl_typemap_level_t *cache, jl_typemap_entry_t *newrec, int8_t offs)
{
jl_value_t *ttypes = jl_unwrap_unionall((jl_value_t*)newrec->sig);
size_t l = jl_nparams(ttypes);
// compute the type at offset `offs` into `sig`, which may be a Vararg
jl_value_t *t1;
int isva = 0;
if (l <= offs + 1) {
t1 = jl_tparam(ttypes, l - 1);
if (jl_is_vararg(t1)) {
isva = 1;
t1 = jl_unwrap_vararg(t1);
}
else if (l <= offs) {
t1 = NULL;
}
}
else if (l > offs) {
t1 = jl_tparam(ttypes, offs);
}
else {
t1 = NULL;
}
if (t1 == (jl_value_t*)jl_typeofbottom_type)
t1 = (jl_value_t*)jl_assume(jl_typeofbottom_type)->super;
// If the type at `offs` is Any, put it in the Any list
if (t1 && jl_is_any(t1)) {
jl_typemap_insert_generic(map, &cache->any, (jl_value_t*)cache, newrec, 0, offs+1, NULL);
return;
}
// Don't put Varargs in the optimized caches (too hard to handle in lookup and bp)
if (t1 && !isva) {
// try to put in leaf type caches
if (jl_is_type_type(t1)) {
// if the argument is Type{...}, this method has specializations for singleton kinds
// and we use the table indexed for that purpose.
jl_value_t *a0 = jl_tparam0(t1);
if (is_cache_leaf(a0, 1)) {
jl_typename_t *name = a0 == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)a0)->name;
jl_typemap_array_insert_(map, &cache->targ, (jl_value_t*)name, newrec, (jl_value_t*)cache, 1, offs, jl_is_datatype(name->wrapper) ? NULL : a0);
return;
}
}
if (is_cache_leaf(t1, 0)) {
jl_typename_t *name = t1 == jl_bottom_type ? jl_typeofbottom_type->name : ((jl_datatype_t*)t1)->name;
jl_typemap_array_insert_(map, &cache->arg1, (jl_value_t*)name, newrec, (jl_value_t*)cache, 0, offs, jl_is_datatype(name->wrapper) ? NULL : t1);
return;
}
// try to put in TypeName caches
jl_value_t *a0;
t1 = jl_unwrap_unionall(t1);
if (jl_is_type_type(t1)) {
a0 = jl_type_extract_name(jl_tparam0(t1));
jl_datatype_t *super = a0 ? (jl_datatype_t*)jl_unwrap_unionall(((jl_typename_t*)a0)->wrapper) : jl_any_type;
jl_typemap_array_insert_(map, &cache->tname, (jl_value_t*)super->name, newrec, (jl_value_t*)cache, 1, offs, NULL);
return;
}
a0 = jl_type_extract_name(t1);
if (a0 && a0 != (jl_value_t*)jl_any_type->name) {
jl_typemap_array_insert_(map, &cache->name1, a0, newrec, (jl_value_t*)cache, 0, offs, NULL);
return;
}
}
jl_typemap_list_insert_(map, &cache->linear, (jl_value_t*)cache, newrec);
}
jl_typemap_entry_t *jl_typemap_alloc(
jl_tupletype_t *type, jl_tupletype_t *simpletype, jl_svec_t *guardsigs,
jl_value_t *newvalue, size_t min_world, size_t max_world)
{
jl_task_t *ct = jl_current_task;
assert(min_world > 0 && max_world > 0);
if (!simpletype)
simpletype = (jl_tupletype_t*)jl_nothing;
jl_value_t *ttype = jl_unwrap_unionall((jl_value_t*)type);
assert(jl_is_tuple_type(ttype));
// compute the complexity of this type signature
int isva = jl_is_va_tuple((jl_datatype_t*)ttype);
int issimplesig = !jl_is_unionall(type); // a TypeVar environment needs a complex matching test
int isleafsig = issimplesig && !isva; // entirely leaf types don't need to be sorted
size_t i, l;
for (i = 0, l = jl_nparams(ttype); i < l && issimplesig; i++) {
jl_value_t *decl = jl_tparam(ttype, i);
if (jl_is_kind(decl))
isleafsig = 0; // Type{} may have a higher priority than a kind
else if (jl_is_type_type(decl))
isleafsig = 0; // Type{} may need special processing to compute the match
else if (jl_is_vararg(decl))
isleafsig = 0; // makes iteration easier when the endpoints are the same
else if (decl == (jl_value_t*)jl_any_type)
isleafsig = 0; // Any needs to go in the general cache
else if (!jl_is_concrete_type(decl)) // anything else needs to go through the general subtyping test
isleafsig = issimplesig = 0;
}
jl_typemap_entry_t *newrec =
(jl_typemap_entry_t*)jl_gc_alloc(ct->ptls, sizeof(jl_typemap_entry_t),
jl_typemap_entry_type);
newrec->sig = type;
newrec->simplesig = simpletype;
newrec->func.value = newvalue;
newrec->guardsigs = guardsigs;
jl_atomic_store_relaxed(&newrec->next, (jl_typemap_entry_t*)jl_nothing);
newrec->min_world = min_world;
newrec->max_world = max_world;
newrec->va = isva;
newrec->issimplesig = issimplesig;
newrec->isleafsig = isleafsig;
return newrec;
}
void jl_typemap_insert(_Atomic(jl_typemap_t *) *pcache, jl_value_t *parent,
jl_typemap_entry_t *newrec, int8_t offs)
{
jl_typemap_t *cache = jl_atomic_load_relaxed(pcache);
jl_typemap_insert_generic(cache, pcache, parent, newrec, 0, offs, NULL);
}
#ifdef __cplusplus
}
#endif