summaryrefslogtreecommitdiff
path: root/libmudflap/mf-hooks3.c
diff options
context:
space:
mode:
Diffstat (limited to 'libmudflap/mf-hooks3.c')
-rw-r--r--libmudflap/mf-hooks3.c573
1 files changed, 573 insertions, 0 deletions
diff --git a/libmudflap/mf-hooks3.c b/libmudflap/mf-hooks3.c
new file mode 100644
index 00000000000..838f3810db5
--- /dev/null
+++ b/libmudflap/mf-hooks3.c
@@ -0,0 +1,573 @@
+/* Mudflap: narrow-pointer bounds-checking by tree rewriting.
+ Copyright (C) 2002, 2003 Free Software Foundation, Inc.
+ Contributed by Frank Ch. Eigler <fche@redhat.com>
+ and Graydon Hoare <graydon@redhat.com>
+
+This file is part of GCC.
+
+GCC is free software; you can redistribute it and/or modify it under
+the terms of the GNU General Public License as published by the Free
+Software Foundation; either version 2, or (at your option) any later
+version.
+
+In addition to the permissions in the GNU General Public License, the
+Free Software Foundation gives you unlimited permission to link the
+compiled version of this file into combinations with other programs,
+and to distribute those combinations without any restriction coming
+from the use of this file. (The General Public License restrictions
+do apply in other respects; for example, they cover modification of
+the file, and distribution when not linked into a combine
+executable.)
+
+GCC is distributed in the hope that it will be useful, but WITHOUT ANY
+WARRANTY; without even the implied warranty of MERCHANTABILITY or
+FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+for more details.
+
+You should have received a copy of the GNU General Public License
+along with GCC; see the file COPYING. If not, write to the Free
+Software Foundation, 59 Temple Place - Suite 330, Boston, MA
+02111-1307, USA. */
+
+
+#include "config.h"
+
+#ifndef HAVE_SOCKLEN_T
+#define socklen_t int
+#endif
+
+/* These attempt to coax various unix flavours to declare all our
+ needed tidbits in the system headers. */
+#if !defined(__FreeBSD__) && !defined(__APPLE__)
+#define _POSIX_SOURCE
+#endif /* Some BSDs break <sys/socket.h> if this is defined. */
+#define _GNU_SOURCE
+#define _XOPEN_SOURCE
+#define _BSD_TYPES
+#define __EXTENSIONS__
+#define _ALL_SOURCE
+#define _LARGE_FILE_API
+#define _XOPEN_SOURCE_EXTENDED 1
+
+#include <string.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <sys/types.h>
+#include <sys/mman.h>
+#include <unistd.h>
+#include <assert.h>
+#include <errno.h>
+#include <limits.h>
+#include <sched.h>
+
+#include "mf-runtime.h"
+#include "mf-impl.h"
+
+#ifdef _MUDFLAP
+#error "Do not compile this file with -fmudflap!"
+#endif
+
+
+/* Multithreading support hooks. */
+
+
+#ifdef WRAP_pthreadstuff
+
+
+#ifndef LIBMUDFLAPTH
+#error "pthreadstuff is to be included only in libmudflapth"
+#endif
+
+
+
+/* Describe a thread (dead or alive). */
+struct pthread_info
+{
+ short used_p; /* Is this slot in use? */
+ short dead_p; /* Is this thread dead? */
+ pthread_t self; /* The thread id. */
+
+ /* If libmudflapth allocated the stack, store its base/size. */
+ void *stack;
+ size_t stack_size;
+
+ int *thread_errno;
+ enum __mf_state_enum state;
+};
+
+
+/* Describe the startup information for a new user thread. */
+struct pthread_start_info
+{
+ /* The user's thread entry point and argument. */
+ void * (*user_fn)(void *);
+ void *user_arg;
+
+ /* Set by user thread when this startup struct may be disposed of. */
+ struct pthread_info *thread_info;
+};
+
+
+
+
+/* To avoid dynamic memory allocation, use static array to store these
+ thread description structs. The second (_idx) array is used as a
+ simple caching hash table, mapping PTHREAD_HASH(thread) to its
+ index in __mf_pthread_info[]. */
+
+#define LIBMUDFLAPTH_THREADS_MAX 1024
+static struct pthread_info __mf_pthread_info[LIBMUDFLAPTH_THREADS_MAX];
+static unsigned __mf_pthread_info_idx[LIBMUDFLAPTH_THREADS_MAX];
+#define PTHREAD_HASH(p) ((unsigned) (p) % LIBMUDFLAPTH_THREADS_MAX)
+
+
+/* Find any old empty entry in __mf_pthread_info; mark it used and
+ return it. Return NULL if there are no more available slots. */
+struct pthread_info*
+__mf_allocate_blank_threadinfo (unsigned* idx)
+{
+ static unsigned probe = LIBMUDFLAPTH_THREADS_MAX-1;
+ unsigned probe_at_start = probe;
+ static pthread_mutex_t mutex =
+#ifdef PTHREAD_ERRORCHECK_MUTEX_INITIALIZER_NP
+ PTHREAD_ERRORCHECK_MUTEX_INITIALIZER_NP;
+#else
+ PTHREAD_MUTEX_INITIALIZER;
+#endif
+ int rc;
+
+ rc = pthread_mutex_lock (& mutex);
+ assert (rc == 0);
+
+ /* Look for a blank spot starting one past the last one we found. */
+ do
+ {
+ probe = (probe + 1) % LIBMUDFLAPTH_THREADS_MAX;
+ struct pthread_info* pi = & __mf_pthread_info [probe];
+ if (! pi->used_p)
+ {
+ /* memset (pi, 0, sizeof (*pi)); */
+ pi->used_p = 1;
+ if (idx != NULL) *idx = probe;
+ /* VERBOSE_TRACE ("allocated threadinfo slot %u\n", probe); */
+ rc = pthread_mutex_unlock (& mutex);
+ assert (rc == 0);
+ return pi;
+ }
+ }
+ while (probe != probe_at_start);
+
+ rc = pthread_mutex_unlock (& mutex);
+ assert (rc == 0);
+ return NULL;
+}
+
+
+/* Find and return the pthread_info struct for the current thread.
+ There might already be one in __mf_pthread_info for this thread, in
+ which case return it. There may not be one (if this is a main
+ thread, an auxiliary -lpthread manager, or an actual user thread
+ making an early call into libmudflap. In these cases, create a new
+ entry. If not it's not the main thread, put it into reentrant
+ initial state.
+*/
+static struct pthread_info*
+__mf_find_threadinfo ()
+{
+ pthread_t it = pthread_self ();
+ unsigned *hash = & __mf_pthread_info_idx [PTHREAD_HASH (it)];
+ struct pthread_info *result = NULL;
+ static pthread_t last;
+ static int main_thread_seen_p;
+
+ /* Check out the lookup cache; failing that, do a linear search
+ around the table. */
+ {
+ struct pthread_info* pi = & __mf_pthread_info [*hash];
+ unsigned i;
+
+ if (pi->used_p && pi->self == it)
+ result = pi;
+ else for (i = 0; i < LIBMUDFLAPTH_THREADS_MAX; i++)
+ {
+ struct pthread_info* pi2 = & __mf_pthread_info [i];
+ if (pi2->used_p && pi2->self == it)
+ {
+ *hash = i;
+ result = pi2;
+ break;
+ }
+ }
+ }
+
+ if (result == NULL)
+ {
+ /* Create a __mf_pthread_info record for the main thread. It's
+ different from the auto-recognized worker bees because for
+ example we can assume that it's a fully stack/errno-equipped
+ thread. */
+
+ /* This must be the main thread, until now unseen in libmudflap. */
+ unsigned *hash = & __mf_pthread_info_idx [PTHREAD_HASH (it)];
+ struct pthread_info* pi = __mf_allocate_blank_threadinfo (hash);
+ assert (pi != NULL);
+ assert (pi->used_p);
+ result = pi;
+ result->self = it;
+
+ if (! main_thread_seen_p)
+ {
+ result->state = active;
+ /* NB: leave result->thread_errno unset, as main thread's errno
+ has already been registered in __mf_init. */
+ /* NB: leave stack-related fields unset, to avoid
+ deallocation. */
+ main_thread_seen_p = 1;
+ VERBOSE_TRACE ("identified self as main thread\n");
+ }
+ else
+ {
+ result->state = reentrant;
+ /* NB: leave result->thread_errno unset, as worker thread's
+ errno is unlikely to be used, and user threads fill them
+ in during __mf_pthread_spawn(). */
+ /* NB: leave stack-related fields unset, leaving pthread_create
+ to fill them in for user threads, leaving them empty for
+ other threads. */
+ VERBOSE_TRACE ("identified self as new aux or user thread\n");
+ }
+ }
+
+ if (last != it)
+ {
+ VERBOSE_TRACE ("found threadinfo for %u, slot %u\n",
+ (unsigned) it,
+ (unsigned) *hash);
+ last = it;
+ }
+
+ assert (result != NULL);
+ assert (result->self == it);
+
+ return result;
+}
+
+
+
+/* Return a pointer to the per-thread __mf_state variable. */
+enum __mf_state_enum *
+__mf_state_perthread ()
+{
+ assert (! __mf_starting_p);
+ return & (__mf_find_threadinfo()->state);
+}
+
+
+static void
+__mf_pthread_cleanup (void *arg)
+{
+ struct pthread_info *pi = arg;
+
+ /* XXX: This unregistration is not safe on platforms where distinct
+ threads share errno (or at least its virtual address). */
+ if (pi->thread_errno != NULL)
+ __mf_unregister (pi->thread_errno, sizeof (int));
+
+ /* XXX: Only detached threads should designate themselves as dead
+ here. Non-detached threads are marked dead after their
+ personalized pthread_join() call. */
+ pi->state = reentrant;
+ pi->dead_p = 1;
+
+ VERBOSE_TRACE ("thread pi %p exiting\n", pi);
+}
+
+
+static void *
+__mf_pthread_spawner (void *arg)
+{
+ struct pthread_info *pi = __mf_find_threadinfo ();
+ void *result = NULL;
+
+ /* Turn off reentrancy indications. */
+ assert (pi->state == reentrant);
+ pi->state = active;
+
+ VERBOSE_TRACE ("new user thread\n");
+
+ if (__mf_opts.heur_std_data)
+ {
+ pi->thread_errno = & errno;
+ __mf_register (pi->thread_errno, sizeof (int),
+ __MF_TYPE_GUESS, "errno area (thread)");
+ /* NB: we could use __MF_TYPE_STATIC above, but we guess that
+ the thread errno is coming out of some dynamically allocated
+ pool that we already know of as __MF_TYPE_HEAP. */
+ }
+
+ /* We considered using pthread_key_t objects instead of these
+ cleanup stacks, but they were less cooperative with the
+ interposed malloc hooks in libmudflap. */
+ pthread_cleanup_push (& __mf_pthread_cleanup, pi);
+
+ /* Call user thread */
+ {
+ /* Extract given entry point and argument. */
+ struct pthread_start_info *psi = arg;
+ void * (*user_fn)(void *) = psi->user_fn;
+ void *user_arg = psi->user_arg;
+
+ /* Signal the main thread to resume. */
+ psi->thread_info = pi;
+
+ result = (*user_fn)(user_arg);
+ }
+
+ pthread_cleanup_pop (1 /* execute */);
+
+ /* NB: there is a slight race here. The pthread_info field will now
+ say this thread is dead, but it may still be running .. right
+ here. We try to check for this possibility using the
+ pthread_kill test below. */
+
+ return result;
+}
+
+
+#if PIC
+/* A special bootstrap variant. */
+int
+__mf_0fn_pthread_create (pthread_t *thr, const pthread_attr_t *attr,
+ void * (*start) (void *), void *arg)
+{
+ return -1;
+}
+#endif
+
+
+#undef pthread_create
+WRAPPER(int, pthread_create, pthread_t *thr, const pthread_attr_t *attr,
+ void * (*start) (void *), void *arg)
+{
+ DECLARE(int, munmap, void *p, size_t l);
+ DECLARE(void *, mmap, void *p, size_t l, int prot, int flags, int fd, off_t of);
+ DECLARE(int, pthread_create, pthread_t *thr, const pthread_attr_t *attr,
+ void * (*start) (void *), void *arg);
+ int result;
+ pthread_attr_t override_attr;
+ void *override_stack;
+ size_t override_stacksize;
+ unsigned i;
+
+ TRACE ("pthread_create\n");
+
+ /* Garbage-collect dead threads' stacks. */
+ LOCKTH ();
+ for (i = 0; i < LIBMUDFLAPTH_THREADS_MAX; i++)
+ {
+ struct pthread_info *pi = & __mf_pthread_info [i];
+ if (! pi->used_p)
+ continue;
+ if (! pi->dead_p)
+ continue;
+
+ /* VERBOSE_TRACE ("thread %u pi %p stack cleanup deferred (%u)\n",
+ (unsigned) pi->self, pi, pi->dead_p); */
+
+ /* Delay actual deallocation by a few cycles, try to discourage the
+ race mentioned at the end of __mf_pthread_spawner(). */
+ if (pi->dead_p)
+ pi->dead_p ++;
+ if (pi->dead_p >= 10 /* XXX */)
+ {
+ if (pi->stack)
+ CALL_REAL (munmap, pi->stack, pi->stack_size);
+
+ VERBOSE_TRACE ("slot %u freed, stack %p\n", i, pi->stack);
+ memset (pi, 0, sizeof (*pi));
+
+ /* One round of garbage collection is enough. */
+ break;
+ }
+ }
+ UNLOCKTH ();
+
+ /* Let's allocate a stack for this thread, if one is not already
+ supplied by the caller. We don't want to let e.g. the
+ linuxthreads manager thread do this allocation. */
+ if (attr != NULL)
+ override_attr = *attr;
+ else
+ pthread_attr_init (& override_attr);
+
+ /* Get supplied attributes, if any. */
+ /* XXX: consider using POSIX2K attr_getstack() */
+ if (pthread_attr_getstackaddr (& override_attr, & override_stack) != 0 ||
+ pthread_attr_getstacksize (& override_attr, & override_stacksize) != 0)
+ {
+ override_stack = NULL;
+ override_stacksize = 0;
+ }
+
+ /* Do we need to allocate the new thread's stack? */
+ if (__mf_opts.thread_stack && override_stack == NULL)
+ {
+ uintptr_t alignment = 256; /* power of two */
+
+ /* Perturb the initial stack addresses slightly, to encourage
+ threads to have nonconflicting entries in the lookup cache
+ for their tracked stack objects. */
+ static unsigned perturb = 0;
+ const unsigned perturb_delta = 32;
+ const unsigned perturb_count = 16;
+ perturb += perturb_delta;
+ if (perturb > perturb_delta*perturb_count) perturb = 0;
+
+ /* Use glibc x86 defaults */
+/* Should have been defined in <limits.h> */
+#ifndef PTHREAD_STACK_MIN
+#define PTHREAD_STACK_MIN 65536
+#endif
+ override_stacksize = max (PTHREAD_STACK_MIN, __mf_opts.thread_stack * 1024);
+
+
+#if defined(MAP_ANONYMOUS)
+#define MF_MAP_ANON MAP_ANONYMOUS
+#elif defined(MAP_ANON)
+#define MF_MAP_ANON MAP_ANON
+#else
+#error "Cannot mmap anonymous memory."
+#endif
+
+ override_stack = CALL_REAL (mmap, NULL, override_stacksize,
+ PROT_READ|PROT_WRITE,
+ MAP_PRIVATE|MF_MAP_ANON,
+ 0, 0);
+ if (override_stack == 0 || override_stack == MAP_FAILED)
+ {
+ errno = EAGAIN;
+ return -1;
+ }
+
+ VERBOSE_TRACE ("thread stack alloc %p size %lu\n",
+ override_stack, (unsigned long) override_stacksize);
+
+ /* The stackaddr pthreads attribute is a candidate stack pointer.
+ It must point near the top or the bottom of this buffer, depending
+ on whether stack grows downward or upward, and suitably aligned.
+ On the x86, it grows down, so we set stackaddr near the top. */
+ override_stack = (void *)
+ (((uintptr_t) override_stack + override_stacksize - alignment - perturb)
+ & (~(uintptr_t)(alignment-1)));
+
+ /* XXX: consider using POSIX2K attr_setstack() */
+ if (pthread_attr_setstackaddr (& override_attr, override_stack) != 0 ||
+ pthread_attr_setstacksize (& override_attr,
+ override_stacksize - alignment - perturb) != 0)
+ {
+ /* This should not happen. */
+ CALL_REAL (munmap, override_stack, override_stacksize);
+ errno = EAGAIN;
+ return -1;
+ }
+ }
+
+ /* Actually start the child thread. */
+ {
+ struct pthread_start_info psi;
+ struct pthread_info *pi = NULL;
+
+ /* Fill in startup-control fields. */
+ psi.user_fn = start;
+ psi.user_arg = arg;
+ psi.thread_info = NULL;
+
+ /* Actually create the thread. */
+ __mf_state = reentrant;
+ result = CALL_REAL (pthread_create, thr, & override_attr,
+ & __mf_pthread_spawner, (void *) & psi);
+ __mf_state = active;
+ /* We also hook pthread_join/pthread_exit to get into reentrant
+ mode during thread shutdown/cleanup. */
+
+ /* Wait until child thread has progressed far enough into its
+ __mf_pthread_spawner() call. */
+ while (1) /* XXX: timeout? */
+ {
+ volatile struct pthread_start_info *psip = & psi;
+ pi = psip->thread_info;
+ if (pi != NULL)
+ break;
+ sched_yield ();
+ }
+
+ /* Fill in remaining fields in pthread_info. */
+ pi->stack = override_stack;
+ pi->stack_size = override_stacksize;
+ /* XXX: this might be too late for future heuristics that attempt
+ to use thread stack bounds. We may need to put the new thread
+ to sleep. */
+ }
+
+
+ /* May need to clean up if we created a pthread_attr_t of our own. */
+ if (attr == NULL)
+ pthread_attr_destroy (& override_attr); /* NB: this shouldn't deallocate stack */
+
+ return result;
+}
+
+
+
+#if PIC
+/* A special bootstrap variant. */
+int
+__mf_0fn_pthread_join (pthread_t thr, void **rc)
+{
+ return -1;
+}
+#endif
+
+
+#undef pthread_join
+WRAPPER(int, pthread_join, pthread_t thr, void **rc)
+{
+ DECLARE(int, pthread_join, pthread_t thr, void **rc);
+ int result;
+
+ TRACE ("pthread_join\n");
+ __mf_state = reentrant;
+ result = CALL_REAL (pthread_join, thr, rc);
+ __mf_state = active;
+
+ return result;
+}
+
+
+#if PIC
+/* A special bootstrap variant. */
+void
+__mf_0fn_pthread_exit (void *rc)
+{
+}
+#endif
+
+
+#undef pthread_exit
+WRAPPER(void, pthread_exit, void *rc)
+{
+ DECLARE(void, pthread_exit, void *rc);
+
+ TRACE ("pthread_exit\n");
+ /* __mf_state = reentrant; */
+ CALL_REAL (pthread_exit, rc);
+ /* NOTREACHED */
+}
+
+
+
+
+
+
+
+#endif /* pthreadstuff */