| // boehm.cc - interface between libjava and Boehm GC. |
| |
| /* Copyright (C) 1998, 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007 |
| Free Software Foundation |
| |
| This file is part of libgcj. |
| |
| This software is copyrighted work licensed under the terms of the |
| Libgcj License. Please consult the file "LIBGCJ_LICENSE" for |
| details. */ |
| |
| #include <config.h> |
| |
| #include <stdio.h> |
| #include <limits.h> |
| |
| #include <jvm.h> |
| #include <gcj/cni.h> |
| |
| #include <java/lang/Class.h> |
| #include <java/lang/reflect/Modifier.h> |
| #include <java-interp.h> |
| |
| // More nastiness: the GC wants to define TRUE and FALSE. We don't |
| // need the Java definitions (themselves a hack), so we undefine them. |
| #undef TRUE |
| #undef FALSE |
| |
| // We include two autoconf headers. Avoid multiple definition warnings. |
| #undef PACKAGE_NAME |
| #undef PACKAGE_STRING |
| #undef PACKAGE_TARNAME |
| #undef PACKAGE_VERSION |
| |
| #ifdef HAVE_DLFCN_H |
| #undef _GNU_SOURCE |
| #define _GNU_SOURCE |
| #include <dlfcn.h> |
| #endif |
| |
| extern "C" |
| { |
| #include <gc_config.h> |
| |
| // Set GC_DEBUG before including gc.h! |
| #ifdef LIBGCJ_GC_DEBUG |
| # define GC_DEBUG |
| #endif |
| |
| #include <gc_mark.h> |
| #include <gc_gcj.h> |
| #include <javaxfc.h> // GC_finalize_all declaration. |
| |
| #ifdef THREAD_LOCAL_ALLOC |
| # define GC_REDIRECT_TO_LOCAL |
| # include <gc_local_alloc.h> |
| #endif |
| |
| // From boehm's misc.c |
| void GC_enable(); |
| void GC_disable(); |
| }; |
| |
| #define MAYBE_MARK(Obj, Top, Limit, Source) \ |
| Top=GC_MARK_AND_PUSH((GC_PTR) Obj, Top, Limit, (GC_PTR *) Source) |
| |
| // `kind' index used when allocating Java arrays. |
| static int array_kind_x; |
| |
| // Freelist used for Java arrays. |
| static void **array_free_list; |
| |
| static int _Jv_GC_has_static_roots (const char *filename, void *, size_t); |
| |
| |
| |
| // This is called by the GC during the mark phase. It marks a Java |
| // object. We use `void *' arguments and return, and not what the |
| // Boehm GC wants, to avoid pollution in our headers. |
| void * |
| _Jv_MarkObj (void *addr, void *msp, void *msl, void *env) |
| { |
| struct GC_ms_entry *mark_stack_ptr = (struct GC_ms_entry *)msp; |
| struct GC_ms_entry *mark_stack_limit = (struct GC_ms_entry *)msl; |
| |
| if (env == (void *)1) /* Object allocated with debug allocator. */ |
| addr = (GC_PTR)GC_USR_PTR_FROM_BASE(addr); |
| jobject obj = (jobject) addr; |
| |
| _Jv_VTable *dt = *(_Jv_VTable **) addr; |
| // The object might not yet have its vtable set, or it might |
| // really be an object on the freelist. In either case, the vtable slot |
| // will either be 0, or it will point to a cleared object. |
| // This assumes Java objects have size at least 3 words, |
| // including the header. But this should remain true, since this |
| // should only be used with debugging allocation or with large objects. |
| if (__builtin_expect (! dt || !(dt -> get_finalizer()), false)) |
| return mark_stack_ptr; |
| jclass klass = dt->clas; |
| GC_PTR p; |
| |
| p = (GC_PTR) dt; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, obj); |
| |
| # ifndef JV_HASH_SYNCHRONIZATION |
| // Every object has a sync_info pointer. |
| p = (GC_PTR) obj->sync_info; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, obj); |
| # endif |
| |
| if (__builtin_expect (klass == &java::lang::Class::class$, false)) |
| { |
| // Currently we allocate some of the memory referenced from class objects |
| // as pointerfree memory, and then mark it more intelligently here. |
| // We ensure that the ClassClass mark descriptor forces invocation of |
| // this procedure. |
| // Correctness of this is subtle, but it looks OK to me for now. For the incremental |
| // collector, we need to make sure that the class object is written whenever |
| // any of the subobjects are altered and may need rescanning. This may be tricky |
| // during construction, and this may not be the right way to do this with |
| // incremental collection. |
| // If we overflow the mark stack, we will rescan the class object, so we should |
| // be OK. The same applies if we redo the mark phase because win32 unmapped part |
| // of our root set. - HB |
| jclass c = (jclass) addr; |
| |
| p = (GC_PTR) c->name; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| p = (GC_PTR) c->superclass; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| p = (GC_PTR) c->constants.tags; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| p = (GC_PTR) c->constants.data; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| // If the class is an array, then the methods field holds a |
| // pointer to the element class. If the class is primitive, |
| // then the methods field holds a pointer to the array class. |
| p = (GC_PTR) c->methods; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| p = (GC_PTR) c->fields; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| // The vtable might be allocated even for compiled code. |
| p = (GC_PTR) c->vtable; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| p = (GC_PTR) c->interfaces; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| p = (GC_PTR) c->loader; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| // The dispatch tables can be allocated at runtime. |
| p = (GC_PTR) c->ancestors; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| p = (GC_PTR) c->idt; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| p = (GC_PTR) c->arrayclass; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| p = (GC_PTR) c->protectionDomain; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| p = (GC_PTR) c->hack_signers; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| p = (GC_PTR) c->aux_info; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| p = (GC_PTR) c->reflection_data; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| |
| // The class chain must be marked for runtime-allocated Classes |
| // loaded by the bootstrap ClassLoader. |
| p = (GC_PTR) c->next_or_version; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c); |
| } |
| else |
| { |
| // NOTE: each class only holds information about the class |
| // itself. So we must do the marking for the entire inheritance |
| // tree in order to mark all fields. FIXME: what about |
| // interfaces? We skip Object here, because Object only has a |
| // sync_info, and we handled that earlier. |
| // Note: occasionally `klass' can be null. For instance, this |
| // can happen if a GC occurs between the point where an object |
| // is allocated and where the vtbl slot is set. |
| while (klass && klass != &java::lang::Object::class$) |
| { |
| jfieldID field = JvGetFirstInstanceField (klass); |
| jint max = JvNumInstanceFields (klass); |
| |
| for (int i = 0; i < max; ++i) |
| { |
| if (JvFieldIsRef (field)) |
| { |
| jobject val = JvGetObjectField (obj, field); |
| p = (GC_PTR) val; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, obj); |
| } |
| field = field->getNextField (); |
| } |
| klass = klass->getSuperclass(); |
| } |
| } |
| |
| return mark_stack_ptr; |
| } |
| |
| // This is called by the GC during the mark phase. It marks a Java |
| // array (of objects). We use `void *' arguments and return, and not |
| // what the Boehm GC wants, to avoid pollution in our headers. |
| void * |
| _Jv_MarkArray (void *addr, void *msp, void *msl, void *env) |
| { |
| struct GC_ms_entry *mark_stack_ptr = (struct GC_ms_entry *)msp; |
| struct GC_ms_entry *mark_stack_limit = (struct GC_ms_entry *)msl; |
| |
| if (env == (void *)1) /* Object allocated with debug allocator. */ |
| addr = (void *)GC_USR_PTR_FROM_BASE(addr); |
| jobjectArray array = (jobjectArray) addr; |
| |
| _Jv_VTable *dt = *(_Jv_VTable **) addr; |
| // Assumes size >= 3 words. That's currently true since arrays have |
| // a vtable, sync pointer, and size. If the sync pointer goes away, |
| // we may need to round up the size. |
| if (__builtin_expect (! dt || !(dt -> get_finalizer()), false)) |
| return mark_stack_ptr; |
| GC_PTR p; |
| |
| p = (GC_PTR) dt; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, array); |
| |
| # ifndef JV_HASH_SYNCHRONIZATION |
| // Every object has a sync_info pointer. |
| p = (GC_PTR) array->sync_info; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, array); |
| # endif |
| |
| for (int i = 0; i < JvGetArrayLength (array); ++i) |
| { |
| jobject obj = elements (array)[i]; |
| p = (GC_PTR) obj; |
| MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, array); |
| } |
| |
| return mark_stack_ptr; |
| } |
| |
| // Generate a GC marking descriptor for a class. |
| // |
| // We assume that the gcj mark proc has index 0. This is a dubious assumption, |
| // since another one could be registered first. But the compiler also |
| // knows this, so in that case everything else will break, too. |
| #define GCJ_DEFAULT_DESCR GC_MAKE_PROC(GC_GCJ_RESERVED_MARK_PROC_INDEX,0) |
| |
| void * |
| _Jv_BuildGCDescr(jclass self) |
| { |
| jlong desc = 0; |
| jint bits_per_word = CHAR_BIT * sizeof (void *); |
| |
| // Note: for now we only consider a bitmap mark descriptor. We |
| // could also handle the case where the first N fields of a type are |
| // references. However, this is not very likely to be used by many |
| // classes, and it is easier to compute things this way. |
| |
| // The vtable pointer. |
| desc |= 1ULL << (bits_per_word - 1); |
| #ifndef JV_HASH_SYNCHRONIZATION |
| // The sync_info field. |
| desc |= 1ULL << (bits_per_word - 2); |
| #endif |
| |
| for (jclass klass = self; klass != NULL; klass = klass->getSuperclass()) |
| { |
| jfieldID field = JvGetFirstInstanceField(klass); |
| int count = JvNumInstanceFields(klass); |
| |
| for (int i = 0; i < count; ++i) |
| { |
| if (field->isRef()) |
| { |
| unsigned int off = field->getOffset(); |
| // If we run into a weird situation, we bail. |
| if (off % sizeof (void *) != 0) |
| return (void *) (GCJ_DEFAULT_DESCR); |
| off /= sizeof (void *); |
| // If we find a field outside the range of our bitmap, |
| // fall back to procedure marker. The bottom 2 bits are |
| // reserved. |
| if (off >= (unsigned) bits_per_word - 2) |
| return (void *) (GCJ_DEFAULT_DESCR); |
| desc |= 1ULL << (bits_per_word - off - 1); |
| } |
| |
| field = field->getNextField(); |
| } |
| } |
| |
| // For bitmap mark type, bottom bits are 01. |
| desc |= 1; |
| // Bogus warning avoidance (on many platforms). |
| return (void *) (unsigned long) desc; |
| } |
| |
| // Allocate some space that is known to be pointer-free. |
| void * |
| _Jv_AllocBytes (jsize size) |
| { |
| void *r = GC_MALLOC_ATOMIC (size); |
| // We have to explicitly zero memory here, as the GC doesn't |
| // guarantee that PTRFREE allocations are zeroed. Note that we |
| // don't have to do this for other allocation types because we set |
| // the `ok_init' flag in the type descriptor. |
| memset (r, 0, size); |
| return r; |
| } |
| |
| #ifdef LIBGCJ_GC_DEBUG |
| |
| void * |
| _Jv_AllocObj (jsize size, jclass klass) |
| { |
| return GC_GCJ_MALLOC (size, klass->vtable); |
| } |
| |
| void * |
| _Jv_AllocPtrFreeObj (jsize size, jclass klass) |
| { |
| #ifdef JV_HASH_SYNCHRONIZATION |
| void * obj = GC_MALLOC_ATOMIC(size); |
| *((_Jv_VTable **) obj) = klass->vtable; |
| #else |
| void * obj = GC_GCJ_MALLOC(size, klass->vtable); |
| #endif |
| return obj; |
| } |
| |
| #endif /* LIBGCJ_GC_DEBUG */ |
| // In the non-debug case, the above two functions are defined |
| // as inline functions in boehm-gc.h. In the debug case we |
| // really want to take advantage of the definitions in gc_gcj.h. |
| |
| // Allocate space for a new Java array. |
| // Used only for arrays of objects. |
| void * |
| _Jv_AllocArray (jsize size, jclass klass) |
| { |
| void *obj; |
| |
| #ifdef LIBGCJ_GC_DEBUG |
| // There isn't much to lose by scanning this conservatively. |
| // If we didn't, the mark proc would have to understand that |
| // it needed to skip the header. |
| obj = GC_MALLOC(size); |
| #else |
| const jsize min_heap_addr = 16*1024; |
| // A heuristic. If size is less than this value, the size |
| // stored in the array can't possibly be misinterpreted as |
| // a pointer. Thus we lose nothing by scanning the object |
| // completely conservatively, since no misidentification can |
| // take place. |
| |
| if (size < min_heap_addr) |
| obj = GC_MALLOC(size); |
| else |
| obj = GC_generic_malloc (size, array_kind_x); |
| #endif |
| *((_Jv_VTable **) obj) = klass->vtable; |
| return obj; |
| } |
| |
| /* Allocate space for a new non-Java object, which does not have the usual |
| Java object header but may contain pointers to other GC'ed objects. */ |
| void * |
| _Jv_AllocRawObj (jsize size) |
| { |
| return (void *) GC_MALLOC (size ? size : 1); |
| } |
| |
| #ifdef INTERPRETER |
| typedef _Jv_ClosureList *closure_list_pointer; |
| |
| /* Release closures in a _Jv_ClosureList. */ |
| static void |
| finalize_closure_list (GC_PTR obj, GC_PTR) |
| { |
| _Jv_ClosureList **clpp = (_Jv_ClosureList **)obj; |
| _Jv_ClosureList::releaseClosures (clpp); |
| } |
| |
| /* Allocate a double-indirect pointer to a _Jv_ClosureList that will |
| get garbage-collected after this double-indirect pointer becomes |
| unreachable by any other objects, including finalizable ones. */ |
| _Jv_ClosureList ** |
| _Jv_ClosureListFinalizer () |
| { |
| _Jv_ClosureList **clpp; |
| clpp = (_Jv_ClosureList **)_Jv_AllocBytes (sizeof (*clpp)); |
| GC_REGISTER_FINALIZER_UNREACHABLE (clpp, finalize_closure_list, |
| NULL, NULL, NULL); |
| return clpp; |
| } |
| #endif // INTERPRETER |
| |
| static void |
| call_finalizer (GC_PTR obj, GC_PTR client_data) |
| { |
| _Jv_FinalizerFunc *fn = (_Jv_FinalizerFunc *) client_data; |
| jobject jobj = (jobject) obj; |
| |
| (*fn) (jobj); |
| } |
| |
| void |
| _Jv_RegisterFinalizer (void *object, _Jv_FinalizerFunc *meth) |
| { |
| GC_REGISTER_FINALIZER_NO_ORDER (object, call_finalizer, (GC_PTR) meth, |
| NULL, NULL); |
| } |
| |
| void |
| _Jv_RunFinalizers (void) |
| { |
| GC_invoke_finalizers (); |
| } |
| |
| void |
| _Jv_RunAllFinalizers (void) |
| { |
| GC_finalize_all (); |
| } |
| |
| void |
| _Jv_RunGC (void) |
| { |
| GC_gcollect (); |
| } |
| |
| long |
| _Jv_GCTotalMemory (void) |
| { |
| return GC_get_heap_size (); |
| } |
| |
| long |
| _Jv_GCFreeMemory (void) |
| { |
| return GC_get_free_bytes (); |
| } |
| |
| void |
| _Jv_GCSetInitialHeapSize (size_t size) |
| { |
| size_t current = GC_get_heap_size (); |
| if (size > current) |
| GC_expand_hp (size - current); |
| } |
| |
| void |
| _Jv_GCSetMaximumHeapSize (size_t size) |
| { |
| GC_set_max_heap_size ((GC_word) size); |
| } |
| |
| int |
| _Jv_SetGCFreeSpaceDivisor (int div) |
| { |
| return (int)GC_set_free_space_divisor ((GC_word)div); |
| } |
| |
| void |
| _Jv_DisableGC (void) |
| { |
| GC_disable(); |
| } |
| |
| void |
| _Jv_EnableGC (void) |
| { |
| GC_enable(); |
| } |
| |
| static void * handle_out_of_memory(size_t) |
| { |
| _Jv_ThrowNoMemory(); |
| } |
| |
| static void |
| gcj_describe_type_fn(void *obj, char *out_buf) |
| { |
| _Jv_VTable *dt = *(_Jv_VTable **) obj; |
| |
| if (! dt /* Shouldn't happen */) |
| { |
| strcpy(out_buf, "GCJ (bad)"); |
| return; |
| } |
| jclass klass = dt->clas; |
| if (!klass /* shouldn't happen */) |
| { |
| strcpy(out_buf, "GCJ (bad)"); |
| return; |
| } |
| jstring name = klass -> getName(); |
| size_t len = name -> length(); |
| if (len >= GC_TYPE_DESCR_LEN) len = GC_TYPE_DESCR_LEN - 1; |
| JvGetStringUTFRegion (name, 0, len, out_buf); |
| out_buf[len] = '\0'; |
| } |
| |
| void |
| _Jv_InitGC (void) |
| { |
| int proc; |
| static bool gc_initialized; |
| |
| if (gc_initialized) |
| return; |
| |
| gc_initialized = 1; |
| |
| // Ignore pointers that do not point to the start of an object. |
| GC_all_interior_pointers = 0; |
| |
| #if defined (HAVE_DLFCN_H) && defined (HAVE_DLADDR) |
| // Tell the collector to ask us before scanning DSOs. |
| GC_register_has_static_roots_callback (_Jv_GC_has_static_roots); |
| #endif |
| |
| // Configure the collector to use the bitmap marking descriptors that we |
| // stash in the class vtable. |
| // We always use mark proc descriptor 0, since the compiler knows |
| // about it. |
| GC_init_gcj_malloc (0, (void *) _Jv_MarkObj); |
| |
| // Cause an out of memory error to be thrown from the allocators, |
| // instead of returning 0. This is cheaper than checking on allocation. |
| GC_oom_fn = handle_out_of_memory; |
| |
| GC_java_finalization = 1; |
| |
| // We use a different mark procedure for object arrays. This code |
| // configures a different object `kind' for object array allocation and |
| // marking. |
| array_free_list = GC_new_free_list(); |
| proc = GC_new_proc((GC_mark_proc)_Jv_MarkArray); |
| array_kind_x = GC_new_kind(array_free_list, GC_MAKE_PROC (proc, 0), 0, 1); |
| |
| // Arrange to have the GC print Java class names in backtraces, etc. |
| GC_register_describe_type_fn(GC_gcj_kind, gcj_describe_type_fn); |
| GC_register_describe_type_fn(GC_gcj_debug_kind, gcj_describe_type_fn); |
| } |
| |
| #ifdef JV_HASH_SYNCHRONIZATION |
| // Allocate an object with a fake vtable pointer, which causes only |
| // the first field (beyond the fake vtable pointer) to be traced. |
| // Eventually this should probably be generalized. |
| |
| static _Jv_VTable trace_one_vtable = { |
| 0, // class pointer |
| (void *)(2 * sizeof(void *)), |
| // descriptor; scan 2 words incl. vtable ptr. |
| // Least significant bits must be zero to |
| // identify this as a length descriptor |
| {0} // First method |
| }; |
| |
| void * |
| _Jv_AllocTraceOne (jsize size /* includes vtable slot */) |
| { |
| return GC_GCJ_MALLOC (size, &trace_one_vtable); |
| } |
| |
| // Ditto for two words. |
| // the first field (beyond the fake vtable pointer) to be traced. |
| // Eventually this should probably be generalized. |
| |
| static _Jv_VTable trace_two_vtable = |
| { |
| 0, // class pointer |
| (void *)(3 * sizeof(void *)), |
| // descriptor; scan 3 words incl. vtable ptr. |
| {0} // First method |
| }; |
| |
| void * |
| _Jv_AllocTraceTwo (jsize size /* includes vtable slot */) |
| { |
| return GC_GCJ_MALLOC (size, &trace_two_vtable); |
| } |
| |
| #endif /* JV_HASH_SYNCHRONIZATION */ |
| |
| void |
| _Jv_GCInitializeFinalizers (void (*notifier) (void)) |
| { |
| GC_finalize_on_demand = 1; |
| GC_finalizer_notifier = notifier; |
| } |
| |
| void |
| _Jv_GCRegisterDisappearingLink (jobject *objp) |
| { |
| // This test helps to ensure that we meet a precondition of |
| // GC_general_register_disappearing_link, viz. "Obj must be a |
| // pointer to the first word of an object we allocated." |
| if (GC_base(*objp)) |
| GC_general_register_disappearing_link ((GC_PTR *) objp, (GC_PTR) *objp); |
| } |
| |
| jboolean |
| _Jv_GCCanReclaimSoftReference (jobject) |
| { |
| // For now, always reclaim soft references. FIXME. |
| return true; |
| } |
| |
| |
| |
| #if defined (HAVE_DLFCN_H) && defined (HAVE_DLADDR) |
| |
| // We keep a store of the filenames of DSOs that need to be |
| // conservatively scanned by the garbage collector. During collection |
| // the gc calls _Jv_GC_has_static_roots() to see if the data segment |
| // of a DSO should be scanned. |
| typedef struct filename_node |
| { |
| char *name; |
| struct filename_node *link; |
| } filename_node; |
| |
| #define FILENAME_STORE_SIZE 17 |
| static filename_node *filename_store[FILENAME_STORE_SIZE]; |
| |
| // Find a filename in filename_store. |
| static filename_node ** |
| find_file (const char *filename) |
| { |
| int index = strlen (filename) % FILENAME_STORE_SIZE; |
| filename_node **node = &filename_store[index]; |
| |
| while (*node) |
| { |
| if (strcmp ((*node)->name, filename) == 0) |
| return node; |
| node = &(*node)->link; |
| } |
| |
| return node; |
| } |
| |
| // Print the store of filenames of DSOs that need collection. |
| void |
| _Jv_print_gc_store (void) |
| { |
| for (int i = 0; i < FILENAME_STORE_SIZE; i++) |
| { |
| filename_node *node = filename_store[i]; |
| while (node) |
| { |
| fprintf (stderr, "%s\n", node->name); |
| node = node->link; |
| } |
| } |
| } |
| |
| // Create a new node in the store of libraries to collect. |
| static filename_node * |
| new_node (const char *filename) |
| { |
| filename_node *node = (filename_node*)_Jv_Malloc (sizeof (filename_node)); |
| node->name = (char *)_Jv_Malloc (strlen (filename) + 1); |
| node->link = NULL; |
| strcpy (node->name, filename); |
| |
| return node; |
| } |
| |
| // Nonzero if the gc should scan this lib. |
| static int |
| _Jv_GC_has_static_roots (const char *filename, void *, size_t) |
| { |
| if (filename == NULL || strlen (filename) == 0) |
| // No filename; better safe than sorry. |
| return 1; |
| |
| filename_node **node = find_file (filename); |
| if (*node) |
| return 1; |
| |
| return 0; |
| } |
| |
| #endif |
| |
| // Register the DSO that contains p for collection. |
| void |
| _Jv_RegisterLibForGc (const void *p __attribute__ ((__unused__))) |
| { |
| #if defined (HAVE_DLFCN_H) && defined (HAVE_DLADDR) |
| Dl_info info; |
| |
| if (dladdr (const_cast<void *>(p), &info) != 0) |
| { |
| filename_node **node = find_file (info.dli_fname); |
| if (! *node) |
| *node = new_node (info.dli_fname); |
| } |
| #endif |
| } |
| |
| void |
| _Jv_SuspendThread (_Jv_Thread_t *thread) |
| { |
| #if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \ |
| && !defined(GC_WIN32_THREADS) && !defined(GC_DARWIN_THREADS) |
| GC_suspend_thread (_Jv_GetPlatformThreadID (thread)); |
| #endif |
| } |
| |
| void |
| _Jv_ResumeThread (_Jv_Thread_t *thread) |
| { |
| #if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \ |
| && !defined(GC_WIN32_THREADS) && !defined(GC_DARWIN_THREADS) |
| GC_resume_thread (_Jv_GetPlatformThreadID (thread)); |
| #endif |
| } |
| |
| int |
| _Jv_IsThreadSuspended (_Jv_Thread_t *thread) |
| { |
| #if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \ |
| && !defined(GC_WIN32_THREADS) && !defined(GC_DARWIN_THREADS) |
| return GC_is_thread_suspended (_Jv_GetPlatformThreadID (thread)); |
| #else |
| return 0; |
| #endif |
| } |
| |
| void |
| _Jv_GCAttachThread () |
| { |
| // The registration interface is only defined on posixy systems and |
| // only actually works if pthread_getattr_np is defined. |
| // FIXME: until gc7 it is simpler to disable this on solaris. |
| #if defined(HAVE_PTHREAD_GETATTR_NP) && !defined(GC_SOLARIS_THREADS) |
| GC_register_my_thread (); |
| #endif |
| } |
| |
| void |
| _Jv_GCDetachThread () |
| { |
| #if defined(HAVE_PTHREAD_GETATTR_NP) && !defined(GC_SOLARIS_THREADS) |
| GC_unregister_my_thread (); |
| #endif |
| } |