diff options
Diffstat (limited to 'contrib/gcc/ggc-common.c')
-rw-r--r-- | contrib/gcc/ggc-common.c | 716 |
1 files changed, 166 insertions, 550 deletions
diff --git a/contrib/gcc/ggc-common.c b/contrib/gcc/ggc-common.c index b0676b2..528b3f2 100644 --- a/contrib/gcc/ggc-common.c +++ b/contrib/gcc/ggc-common.c @@ -26,33 +26,26 @@ Software Foundation, 59 Temple Place - Suite 330, Boston, MA #include "rtl.h" #include "tree.h" #include "tm_p.h" -#include "hash.h" #include "hashtab.h" #include "varray.h" #include "ggc.h" +#include "langhooks.h" +#include "params.h" +#ifdef HAVE_SYS_RESOURCE_H +# include <sys/resource.h> +#endif +#ifdef ENABLE_VALGRIND_CHECKING +#include <valgrind.h> +#else +/* Avoid #ifdef:s when we can help it. */ +#define VALGRIND_DISCARD(x) +#endif /* Statistics about the allocation. */ static ggc_statistics *ggc_stats; -/* The FALSE_LABEL_STACK, declared in except.h, has language-dependent - semantics. If a front-end needs to mark the false label stack, it - should set this pointer to a non-NULL value. Otherwise, no marking - will be done. */ -void (*lang_mark_false_label_stack) PARAMS ((struct label_node *)); - -/* Trees that have been marked, but whose children still need marking. */ -varray_type ggc_pending_trees; - -static void ggc_mark_rtx_children_1 PARAMS ((rtx)); -static void ggc_mark_rtx_ptr PARAMS ((void *)); -static void ggc_mark_tree_ptr PARAMS ((void *)); -static void ggc_mark_rtx_varray_ptr PARAMS ((void *)); -static void ggc_mark_tree_varray_ptr PARAMS ((void *)); -static void ggc_mark_tree_hash_table_ptr PARAMS ((void *)); static int ggc_htab_delete PARAMS ((void **, void *)); -static void ggc_mark_trees PARAMS ((void)); -static bool ggc_mark_tree_hash_table_entry PARAMS ((struct hash_entry *, - hash_table_key)); +static double ggc_rlimit_bound PARAMS ((double)); /* Maintain global roots that are preserved during GC. */ @@ -70,7 +63,7 @@ struct ggc_root static struct ggc_root *roots; /* Add BASE as a new garbage collection root. It is an array of - length NELT with each element SIZE bytes long. CB is a + length NELT with each element SIZE bytes long. CB is a function that will be called with a pointer to each element of the array; it is the intention that CB call the appropriate routine to mark gc-able memory for that element. */ @@ -92,132 +85,6 @@ ggc_add_root (base, nelt, size, cb) roots = x; } -/* Register an array of rtx as a GC root. */ - -void -ggc_add_rtx_root (base, nelt) - rtx *base; - int nelt; -{ - ggc_add_root (base, nelt, sizeof (rtx), ggc_mark_rtx_ptr); -} - -/* Register an array of trees as a GC root. */ - -void -ggc_add_tree_root (base, nelt) - tree *base; - int nelt; -{ - ggc_add_root (base, nelt, sizeof (tree), ggc_mark_tree_ptr); -} - -/* Register a varray of rtxs as a GC root. */ - -void -ggc_add_rtx_varray_root (base, nelt) - varray_type *base; - int nelt; -{ - ggc_add_root (base, nelt, sizeof (varray_type), - ggc_mark_rtx_varray_ptr); -} - -/* Register a varray of trees as a GC root. */ - -void -ggc_add_tree_varray_root (base, nelt) - varray_type *base; - int nelt; -{ - ggc_add_root (base, nelt, sizeof (varray_type), - ggc_mark_tree_varray_ptr); -} - -/* Register a hash table of trees as a GC root. */ - -void -ggc_add_tree_hash_table_root (base, nelt) - struct hash_table **base; - int nelt; -{ - ggc_add_root (base, nelt, sizeof (struct hash_table *), - ggc_mark_tree_hash_table_ptr); -} - -/* Remove the previously registered GC root at BASE. */ - -void -ggc_del_root (base) - void *base; -{ - struct ggc_root *x, **p; - - p = &roots, x = roots; - while (x) - { - if (x->base == base) - { - *p = x->next; - free (x); - return; - } - p = &x->next; - x = x->next; - } - - abort (); -} - -/* Add a hash table to be scanned when all roots have been processed. We - delete any entry in the table that has not been marked. */ - -struct d_htab_root -{ - struct d_htab_root *next; - htab_t htab; - ggc_htab_marked_p marked_p; - ggc_htab_mark mark; -}; - -static struct d_htab_root *d_htab_roots; - -/* Add X, an htab, to a list of htabs that contain objects which are allocated - from GC memory. Once all other roots are marked, we check each object in - the htab to see if it has already been marked. If not, it is deleted. - - MARKED_P, if specified, is a function that returns 1 if the entry is to - be considered as "marked". If not present, the data structure pointed to - by the htab slot is tested. This function should be supplied if some - other object (such as something pointed to by that object) should be tested - in which case the function tests whether that object (or objects) are - marked (using ggc_marked_p) and returns nonzero if it is. - - MARK, if specified, is a function that is passed the contents of a slot - that has been determined to have been "marked" (via the above function) - and marks any other objects pointed to by that object. For example, - we might have a hash table of memory attribute blocks, which are pointed - to by a MEM RTL but have a pointer to a DECL. MARKED_P in that case will - not be specified because we want to know if the attribute block is pointed - to by the MEM, but MARK must be specified because if the block has been - marked, we need to mark the DECL. */ - -void -ggc_add_deletable_htab (x, marked_p, mark) - PTR x; - ggc_htab_marked_p marked_p; - ggc_htab_mark mark; -{ - struct d_htab_root *r - = (struct d_htab_root *) xmalloc (sizeof (struct d_htab_root)); - - r->next = d_htab_roots; - r->htab = (htab_t) x; - r->marked_p = marked_p ? marked_p : ggc_marked_p; - r->mark = mark; - d_htab_roots = r; -} - /* Process a slot of an htab by deleting it if it has not been marked. */ static int @@ -225,12 +92,12 @@ ggc_htab_delete (slot, info) void **slot; void *info; { - struct d_htab_root *r = (struct d_htab_root *) info; + const struct ggc_cache_tab *r = (const struct ggc_cache_tab *) info; if (! (*r->marked_p) (*slot)) - htab_clear_slot (r->htab, slot); - else if (r->mark) - (*r->mark) (*slot); + htab_clear_slot (*r->base, slot); + else + (*r->cb) (*slot); return 1; } @@ -241,9 +108,20 @@ void ggc_mark_roots () { struct ggc_root *x; - struct d_htab_root *y; - - VARRAY_TREE_INIT (ggc_pending_trees, 4096, "ggc_pending_trees"); + const struct ggc_root_tab *const *rt; + const struct ggc_root_tab *rti; + const struct ggc_cache_tab *const *ct; + const struct ggc_cache_tab *cti; + size_t i; + + for (rt = gt_ggc_deletable_rtab; *rt; rt++) + for (rti = *rt; rti->base != NULL; rti++) + memset (rti->base, 0, rti->stride); + + for (rt = gt_ggc_rtab; *rt; rt++) + for (rti = *rt; rti->base != NULL; rti++) + for (i = 0; i < rti->nelt; i++) + (*rti->cb)(*(void **)((char *)rti->base + rti->stride * i)); for (x = roots; x != NULL; x = x->next) { @@ -256,417 +134,77 @@ ggc_mark_roots () (*cb)(elt); } - /* Mark all the queued up trees, and their children. */ - ggc_mark_trees (); - VARRAY_FREE (ggc_pending_trees); - /* Now scan all hash tables that have objects which are to be deleted if - they are not already marked. Since these may mark more trees, we need - to reinitialize that varray. */ - VARRAY_TREE_INIT (ggc_pending_trees, 1024, "ggc_pending_trees"); - - for (y = d_htab_roots; y != NULL; y = y->next) - htab_traverse (y->htab, ggc_htab_delete, (PTR) y); - ggc_mark_trees (); - VARRAY_FREE (ggc_pending_trees); + they are not already marked. */ + for (ct = gt_ggc_cache_rtab; *ct; ct++) + for (cti = *ct; cti->base != NULL; cti++) + if (*cti->base) + htab_traverse (*cti->base, ggc_htab_delete, (PTR) cti); } -/* R had not been previously marked, but has now been marked via - ggc_set_mark. Now recurse and process the children. */ - -void -ggc_mark_rtx_children (r) - rtx r; -{ - rtx i, last; - - /* Special case the instruction chain. This is a data structure whose - chain length is potentially unbounded, and which contain references - within the chain (e.g. label_ref and insn_list). If do nothing here, - we risk blowing the stack recursing through a long chain of insns. - - Combat this by marking all of the instructions in the chain before - marking the contents of those instructions. */ - - switch (GET_CODE (r)) - { - case INSN: - case JUMP_INSN: - case CALL_INSN: - case NOTE: - case CODE_LABEL: - case BARRIER: - for (i = NEXT_INSN (r); ; i = NEXT_INSN (i)) - if (! ggc_test_and_set_mark (i)) - break; - last = i; - - for (i = NEXT_INSN (r); i != last; i = NEXT_INSN (i)) - ggc_mark_rtx_children_1 (i); - - default: - break; - } - - ggc_mark_rtx_children_1 (r); -} - -static void -ggc_mark_rtx_children_1 (r) - rtx r; +/* Allocate a block of memory, then clear it. */ +void * +ggc_alloc_cleared (size) + size_t size; { - const char *fmt; - int i; - rtx next_rtx; - - do - { - enum rtx_code code = GET_CODE (r); - /* This gets set to a child rtx to eliminate tail recursion. */ - next_rtx = NULL; - - /* Collect statistics, if appropriate. */ - if (ggc_stats) - { - ++ggc_stats->num_rtxs[(int) code]; - ggc_stats->size_rtxs[(int) code] += ggc_get_size (r); - } - - /* ??? If (some of) these are really pass-dependent info, do we - have any right poking our noses in? */ - switch (code) - { - case MEM: - ggc_mark (MEM_ATTRS (r)); - break; - case JUMP_INSN: - ggc_mark_rtx (JUMP_LABEL (r)); - break; - case CODE_LABEL: - ggc_mark_rtx (LABEL_REFS (r)); - break; - case LABEL_REF: - ggc_mark_rtx (LABEL_NEXTREF (r)); - ggc_mark_rtx (CONTAINING_INSN (r)); - break; - case ADDRESSOF: - ggc_mark_tree (ADDRESSOF_DECL (r)); - break; - case CONST_DOUBLE: - ggc_mark_rtx (CONST_DOUBLE_CHAIN (r)); - break; - case NOTE: - switch (NOTE_LINE_NUMBER (r)) - { - case NOTE_INSN_RANGE_BEG: - case NOTE_INSN_RANGE_END: - case NOTE_INSN_LIVE: - case NOTE_INSN_EXPECTED_VALUE: - ggc_mark_rtx (NOTE_RANGE_INFO (r)); - break; - - case NOTE_INSN_BLOCK_BEG: - case NOTE_INSN_BLOCK_END: - ggc_mark_tree (NOTE_BLOCK (r)); - break; - - default: - break; - } - break; - - default: - break; - } - - for (fmt = GET_RTX_FORMAT (GET_CODE (r)), i = 0; *fmt ; ++fmt, ++i) - { - rtx exp; - switch (*fmt) - { - case 'e': case 'u': - exp = XEXP (r, i); - if (ggc_test_and_set_mark (exp)) - { - if (next_rtx == NULL) - next_rtx = exp; - else - ggc_mark_rtx_children (exp); - } - break; - case 'V': case 'E': - ggc_mark_rtvec (XVEC (r, i)); - break; - } - } - } - while ((r = next_rtx) != NULL); + void *buf = ggc_alloc (size); + memset (buf, 0, size); + return buf; } -/* V had not been previously marked, but has now been marked via - ggc_set_mark. Now recurse and process the children. */ - -void -ggc_mark_rtvec_children (v) - rtvec v; +/* Resize a block of memory, possibly re-allocating it. */ +void * +ggc_realloc (x, size) + void *x; + size_t size; { - int i; + void *r; + size_t old_size; - i = GET_NUM_ELEM (v); - while (--i >= 0) - ggc_mark_rtx (RTVEC_ELT (v, i)); -} - -/* Recursively set marks on all of the children of the - GCC_PENDING_TREES. */ + if (x == NULL) + return ggc_alloc (size); -static void -ggc_mark_trees () -{ - while (ggc_pending_trees->elements_used) + old_size = ggc_get_size (x); + if (size <= old_size) { - tree t; - enum tree_code code; - - t = VARRAY_TOP_TREE (ggc_pending_trees); - VARRAY_POP (ggc_pending_trees); - code = TREE_CODE (t); - - /* Collect statistics, if appropriate. */ - if (ggc_stats) - { - ++ggc_stats->num_trees[(int) code]; - ggc_stats->size_trees[(int) code] += ggc_get_size (t); - } - - /* Bits from common. */ - ggc_mark_tree (TREE_TYPE (t)); - ggc_mark_tree (TREE_CHAIN (t)); - - /* Some nodes require special handling. */ - switch (code) - { - case TREE_LIST: - ggc_mark_tree (TREE_PURPOSE (t)); - ggc_mark_tree (TREE_VALUE (t)); - continue; - - case TREE_VEC: - { - int i = TREE_VEC_LENGTH (t); - - while (--i >= 0) - ggc_mark_tree (TREE_VEC_ELT (t, i)); - continue; - } - - case COMPLEX_CST: - ggc_mark_tree (TREE_REALPART (t)); - ggc_mark_tree (TREE_IMAGPART (t)); - break; - - case PARM_DECL: - ggc_mark_rtx (DECL_INCOMING_RTL (t)); - break; - - case FIELD_DECL: - ggc_mark_tree (DECL_FIELD_BIT_OFFSET (t)); - break; - - case IDENTIFIER_NODE: - lang_mark_tree (t); - continue; - - default: - break; - } - - /* But in general we can handle them by class. */ - switch (TREE_CODE_CLASS (code)) - { - case 'd': /* A decl node. */ - ggc_mark_tree (DECL_SIZE (t)); - ggc_mark_tree (DECL_SIZE_UNIT (t)); - ggc_mark_tree (DECL_NAME (t)); - ggc_mark_tree (DECL_CONTEXT (t)); - ggc_mark_tree (DECL_ARGUMENTS (t)); - ggc_mark_tree (DECL_RESULT_FLD (t)); - ggc_mark_tree (DECL_INITIAL (t)); - ggc_mark_tree (DECL_ABSTRACT_ORIGIN (t)); - ggc_mark_tree (DECL_SECTION_NAME (t)); - ggc_mark_tree (DECL_ATTRIBUTES (t)); - if (DECL_RTL_SET_P (t)) - ggc_mark_rtx (DECL_RTL (t)); - ggc_mark_rtx (DECL_LIVE_RANGE_RTL (t)); - ggc_mark_tree (DECL_VINDEX (t)); - if (DECL_ASSEMBLER_NAME_SET_P (t)) - ggc_mark_tree (DECL_ASSEMBLER_NAME (t)); - if (TREE_CODE (t) == FUNCTION_DECL) - { - ggc_mark_tree (DECL_SAVED_TREE (t)); - ggc_mark_tree (DECL_INLINED_FNS (t)); - if (DECL_SAVED_INSNS (t)) - ggc_mark_struct_function (DECL_SAVED_INSNS (t)); - } - lang_mark_tree (t); - break; - - case 't': /* A type node. */ - ggc_mark_tree (TYPE_SIZE (t)); - ggc_mark_tree (TYPE_SIZE_UNIT (t)); - ggc_mark_tree (TYPE_ATTRIBUTES (t)); - ggc_mark_tree (TYPE_VALUES (t)); - ggc_mark_tree (TYPE_POINTER_TO (t)); - ggc_mark_tree (TYPE_REFERENCE_TO (t)); - ggc_mark_tree (TYPE_NAME (t)); - ggc_mark_tree (TYPE_MIN_VALUE (t)); - ggc_mark_tree (TYPE_MAX_VALUE (t)); - ggc_mark_tree (TYPE_NEXT_VARIANT (t)); - ggc_mark_tree (TYPE_MAIN_VARIANT (t)); - ggc_mark_tree (TYPE_BINFO (t)); - ggc_mark_tree (TYPE_CONTEXT (t)); - lang_mark_tree (t); - break; - - case 'b': /* A lexical block. */ - ggc_mark_tree (BLOCK_VARS (t)); - ggc_mark_tree (BLOCK_SUBBLOCKS (t)); - ggc_mark_tree (BLOCK_SUPERCONTEXT (t)); - ggc_mark_tree (BLOCK_ABSTRACT_ORIGIN (t)); - break; - - case 'c': /* A constant. */ - ggc_mark_rtx (TREE_CST_RTL (t)); - break; - - case 'r': case '<': case '1': - case '2': case 'e': case 's': /* Expressions. */ - { - int i = TREE_CODE_LENGTH (TREE_CODE (t)); - int first_rtl = first_rtl_op (TREE_CODE (t)); - - while (--i >= 0) - { - if (i >= first_rtl) - ggc_mark_rtx ((rtx) TREE_OPERAND (t, i)); - else - ggc_mark_tree (TREE_OPERAND (t, i)); - } - break; - } - - case 'x': - lang_mark_tree (t); - break; - } + /* Mark the unwanted memory as unaccessible. We also need to make + the "new" size accessible, since ggc_get_size returns the size of + the pool, not the size of the individually allocated object, the + size which was previously made accessible. Unfortunately, we + don't know that previously allocated size. Without that + knowledge we have to lose some initialization-tracking for the + old parts of the object. An alternative is to mark the whole + old_size as reachable, but that would lose tracking of writes + after the end of the object (by small offsets). Discard the + handle to avoid handle leak. */ + VALGRIND_DISCARD (VALGRIND_MAKE_NOACCESS ((char *) x + size, + old_size - size)); + VALGRIND_DISCARD (VALGRIND_MAKE_READABLE (x, size)); + return x; } -} - -/* Mark all the elements of the varray V, which contains rtxs. */ - -void -ggc_mark_rtx_varray (v) - varray_type v; -{ - int i; - - if (v) - for (i = v->num_elements - 1; i >= 0; --i) - ggc_mark_rtx (VARRAY_RTX (v, i)); -} - -/* Mark all the elements of the varray V, which contains trees. */ - -void -ggc_mark_tree_varray (v) - varray_type v; -{ - int i; - - if (v) - for (i = v->num_elements - 1; i >= 0; --i) - ggc_mark_tree (VARRAY_TREE (v, i)); -} - -/* Mark the hash table-entry HE. Its key field is really a tree. */ - -static bool -ggc_mark_tree_hash_table_entry (he, k) - struct hash_entry *he; - hash_table_key k ATTRIBUTE_UNUSED; -{ - ggc_mark_tree ((tree) he->key); - return true; -} - -/* Mark all the elements of the hash-table H, which contains trees. */ - -void -ggc_mark_tree_hash_table (ht) - struct hash_table *ht; -{ - hash_traverse (ht, ggc_mark_tree_hash_table_entry, /*info=*/0); -} - -/* Type-correct function to pass to ggc_add_root. It just forwards - *ELT (which is an rtx) to ggc_mark_rtx. */ - -static void -ggc_mark_rtx_ptr (elt) - void *elt; -{ - ggc_mark_rtx (*(rtx *) elt); -} -/* Type-correct function to pass to ggc_add_root. It just forwards - *ELT (which is a tree) to ggc_mark_tree. */ + r = ggc_alloc (size); -static void -ggc_mark_tree_ptr (elt) - void *elt; -{ - ggc_mark_tree (*(tree *) elt); -} - -/* Type-correct function to pass to ggc_add_root. It just forwards - ELT (which is really a varray_type *) to ggc_mark_rtx_varray. */ + /* Since ggc_get_size returns the size of the pool, not the size of the + individually allocated object, we'd access parts of the old object + that were marked invalid with the memcpy below. We lose a bit of the + initialization-tracking since some of it may be uninitialized. */ + VALGRIND_DISCARD (VALGRIND_MAKE_READABLE (x, old_size)); -static void -ggc_mark_rtx_varray_ptr (elt) - void *elt; -{ - ggc_mark_rtx_varray (*(varray_type *) elt); -} - -/* Type-correct function to pass to ggc_add_root. It just forwards - ELT (which is really a varray_type *) to ggc_mark_tree_varray. */ - -static void -ggc_mark_tree_varray_ptr (elt) - void *elt; -{ - ggc_mark_tree_varray (*(varray_type *) elt); -} + memcpy (r, x, old_size); -/* Type-correct function to pass to ggc_add_root. It just forwards - ELT (which is really a struct hash_table **) to - ggc_mark_tree_hash_table. */ + /* The old object is not supposed to be used anymore. */ + VALGRIND_DISCARD (VALGRIND_MAKE_NOACCESS (x, old_size)); -static void -ggc_mark_tree_hash_table_ptr (elt) - void *elt; -{ - ggc_mark_tree_hash_table (*(struct hash_table **) elt); + return r; } -/* Allocate a block of memory, then clear it. */ +/* Like ggc_alloc_cleared, but performs a multiplication. */ void * -ggc_alloc_cleared (size) - size_t size; +ggc_calloc (s1, s2) + size_t s1, s2; { - void *buf = ggc_alloc (size); - memset (buf, 0, size); - return buf; + return ggc_alloc_cleared (s1 * s2); } /* Print statistics that are independent of the collector in use. */ @@ -704,17 +242,17 @@ ggc_print_common_statistics (stream, stats) } /* Print the statistics for trees. */ - fprintf (stream, "\n%-17s%10s %16s %10s\n", "Tree", + fprintf (stream, "\n%-17s%10s %16s %10s\n", "Tree", "Number", "Bytes", "% Total"); for (code = 0; code < MAX_TREE_CODES; ++code) - if (ggc_stats->num_trees[code]) + if (ggc_stats->num_trees[code]) { fprintf (stream, "%-17s%10u%16ld%c %10.3f\n", tree_code_name[code], ggc_stats->num_trees[code], SCALE (ggc_stats->size_trees[code]), LABEL (ggc_stats->size_trees[code]), - (100 * ((double) ggc_stats->size_trees[code]) + (100 * ((double) ggc_stats->size_trees[code]) / ggc_stats->total_size_trees)); } fprintf (stream, @@ -724,17 +262,17 @@ ggc_print_common_statistics (stream, stats) LABEL (ggc_stats->total_size_trees)); /* Print the statistics for RTL. */ - fprintf (stream, "\n%-17s%10s %16s %10s\n", "RTX", + fprintf (stream, "\n%-17s%10s %16s %10s\n", "RTX", "Number", "Bytes", "% Total"); for (code = 0; code < NUM_RTX_CODE; ++code) - if (ggc_stats->num_rtxs[code]) + if (ggc_stats->num_rtxs[code]) { fprintf (stream, "%-17s%10u%16ld%c %10.3f\n", rtx_name[code], ggc_stats->num_rtxs[code], SCALE (ggc_stats->size_rtxs[code]), LABEL (ggc_stats->size_rtxs[code]), - (100 * ((double) ggc_stats->size_rtxs[code]) + (100 * ((double) ggc_stats->size_rtxs[code]) / ggc_stats->total_size_rtxs)); } fprintf (stream, @@ -746,3 +284,81 @@ ggc_print_common_statistics (stream, stats) /* Don't gather statistics any more. */ ggc_stats = NULL; } + +/* Modify the bound based on rlimits. Keep the smallest number found. */ +static double +ggc_rlimit_bound (limit) + double limit; +{ +#if defined(HAVE_GETRLIMIT) + struct rlimit rlim; +# ifdef RLIMIT_RSS + if (getrlimit (RLIMIT_RSS, &rlim) == 0 + && rlim.rlim_cur != (rlim_t) RLIM_INFINITY + && rlim.rlim_cur < limit) + limit = rlim.rlim_cur; +# endif +# ifdef RLIMIT_DATA + if (getrlimit (RLIMIT_DATA, &rlim) == 0 + && rlim.rlim_cur != (rlim_t) RLIM_INFINITY + && rlim.rlim_cur < limit) + limit = rlim.rlim_cur; +# endif +# ifdef RLIMIT_AS + if (getrlimit (RLIMIT_AS, &rlim) == 0 + && rlim.rlim_cur != (rlim_t) RLIM_INFINITY + && rlim.rlim_cur < limit) + limit = rlim.rlim_cur; +# endif +#endif /* HAVE_GETRLIMIT */ + + return limit; +} + +/* Heuristic to set a default for GGC_MIN_EXPAND. */ +int +ggc_min_expand_heuristic() +{ + double min_expand = physmem_total(); + + /* Adjust for rlimits. */ + min_expand = ggc_rlimit_bound (min_expand); + + /* The heuristic is a percentage equal to 30% + 70%*(RAM/1GB), yielding + a lower bound of 30% and an upper bound of 100% (when RAM >= 1GB). */ + min_expand /= 1024*1024*1024; + min_expand *= 70; + min_expand = MIN (min_expand, 70); + min_expand += 30; + + return min_expand; +} + +/* Heuristic to set a default for GGC_MIN_HEAPSIZE. */ +int +ggc_min_heapsize_heuristic() +{ + double min_heap_kbytes = physmem_total(); + + /* Adjust for rlimits. */ + min_heap_kbytes = ggc_rlimit_bound (min_heap_kbytes); + + min_heap_kbytes /= 1024; /* convert to Kbytes. */ + + /* The heuristic is RAM/8, with a lower bound of 4M and an upper + bound of 128M (when RAM >= 1GB). */ + min_heap_kbytes /= 8; + min_heap_kbytes = MAX (min_heap_kbytes, 4 * 1024); + min_heap_kbytes = MIN (min_heap_kbytes, 128 * 1024); + + return min_heap_kbytes; +} + +void +init_ggc_heuristics () +{ +#ifndef ENABLE_GC_ALWAYS_COLLECT + set_param_value ("ggc-min-expand", ggc_min_expand_heuristic()); + set_param_value ("ggc-min-heapsize", ggc_min_heapsize_heuristic()); +#endif +} |