summaryrefslogtreecommitdiff
path: root/APACHE_1_3_42/src/main/alloc.c
diff options
context:
space:
mode:
Diffstat (limited to 'APACHE_1_3_42/src/main/alloc.c')
-rw-r--r--APACHE_1_3_42/src/main/alloc.c2880
1 files changed, 2880 insertions, 0 deletions
diff --git a/APACHE_1_3_42/src/main/alloc.c b/APACHE_1_3_42/src/main/alloc.c
new file mode 100644
index 0000000000..9d772c37b3
--- /dev/null
+++ b/APACHE_1_3_42/src/main/alloc.c
@@ -0,0 +1,2880 @@
+/* Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+/*
+ * Resource allocation code... the code here is responsible for making
+ * sure that nothing leaks.
+ *
+ * rst --- 4/95 --- 6/95
+ */
+
+#include "httpd.h"
+#include "multithread.h"
+#include "http_log.h"
+
+#include <stdarg.h>
+
+#ifdef OS2
+#define INCL_DOS
+#include <os2.h>
+#endif
+
+/* debugging support, define this to enable code which helps detect re-use
+ * of freed memory and other such nonsense.
+ *
+ * The theory is simple. The FILL_BYTE (0xa5) is written over all malloc'd
+ * memory as we receive it, and is written over everything that we free up
+ * during a clear_pool. We check that blocks on the free list always
+ * have the FILL_BYTE in them, and we check during palloc() that the bytes
+ * still have FILL_BYTE in them. If you ever see garbage URLs or whatnot
+ * containing lots of 0xa5s then you know something used data that's been
+ * freed or uninitialized.
+ */
+/* #define ALLOC_DEBUG */
+
+/* debugging support, if defined all allocations will be done with
+ * malloc and free()d appropriately at the end. This is intended to be
+ * used with something like Electric Fence or Purify to help detect
+ * memory problems. Note that if you're using efence then you should also
+ * add in ALLOC_DEBUG. But don't add in ALLOC_DEBUG if you're using Purify
+ * because ALLOC_DEBUG would hide all the uninitialized read errors that
+ * Purify can diagnose.
+ */
+/* #define ALLOC_USE_MALLOC */
+
+/* Pool debugging support. This is intended to detect cases where the
+ * wrong pool is used when assigning data to an object in another pool.
+ * In particular, it causes the table_{set,add,merge}n routines to check
+ * that their arguments are safe for the table they're being placed in.
+ * It currently only works with the unix multiprocess model, but could
+ * be extended to others.
+ */
+/* #define POOL_DEBUG */
+
+/* Provide diagnostic information about make_table() calls which are
+ * possibly too small. This requires a recent gcc which supports
+ * __builtin_return_address(). The error_log output will be a
+ * message such as:
+ * table_push: table created by 0x804d874 hit limit of 10
+ * Use "l *0x804d874" to find the source that corresponds to. It
+ * indicates that a table allocated by a call at that address has
+ * possibly too small an initial table size guess.
+ */
+/* #define MAKE_TABLE_PROFILE */
+
+/* Provide some statistics on the cost of allocations. It requires a
+ * bit of an understanding of how alloc.c works.
+ */
+/* #define ALLOC_STATS */
+
+#ifdef POOL_DEBUG
+#ifdef ALLOC_USE_MALLOC
+# error "sorry, no support for ALLOC_USE_MALLOC and POOL_DEBUG at the same time"
+#endif
+#ifdef MULTITHREAD
+# error "sorry, no support for MULTITHREAD and POOL_DEBUG at the same time"
+#endif
+#endif
+
+#ifdef ALLOC_USE_MALLOC
+#undef BLOCK_MINFREE
+#undef BLOCK_MINALLOC
+#define BLOCK_MINFREE 0
+#define BLOCK_MINALLOC 0
+#endif
+
+/*****************************************************************
+ *
+ * Managing free storage blocks...
+ */
+
+union align {
+ /* Types which are likely to have the longest RELEVANT alignment
+ * restrictions...
+ */
+
+ char *cp;
+ void (*f) (void);
+ long l;
+ FILE *fp;
+ double d;
+};
+
+#define CLICK_SZ (sizeof(union align))
+
+union block_hdr {
+ union align a;
+
+ /* Actual header... */
+
+ struct {
+ char *endp;
+ union block_hdr *next;
+ char *first_avail;
+#ifdef POOL_DEBUG
+ union block_hdr *global_next;
+ struct pool *owning_pool;
+#endif
+ } h;
+};
+
+static union block_hdr *block_freelist = NULL;
+static mutex *alloc_mutex = NULL;
+static mutex *spawn_mutex = NULL;
+#ifdef POOL_DEBUG
+static char *known_stack_point;
+static int stack_direction;
+static union block_hdr *global_block_list;
+#define FREE_POOL ((struct pool *)(-1))
+#endif
+#ifdef ALLOC_STATS
+static unsigned long long num_free_blocks_calls;
+static unsigned long long num_blocks_freed;
+static unsigned max_blocks_in_one_free;
+static unsigned num_malloc_calls;
+static unsigned num_malloc_bytes;
+#endif
+
+#ifdef ALLOC_DEBUG
+#define FILL_BYTE ((char)(0xa5))
+
+#define debug_fill(ptr,size) ((void)memset((ptr), FILL_BYTE, (size)))
+
+static ap_inline void debug_verify_filled(const char *ptr,
+ const char *endp, const char *error_msg)
+{
+ for (; ptr < endp; ++ptr) {
+ if (*ptr != FILL_BYTE) {
+ fputs(error_msg, stderr);
+ abort();
+ exit(1);
+ }
+ }
+}
+
+#else
+#define debug_fill(a,b)
+#define debug_verify_filled(a,b,c)
+#endif
+
+
+/* Get a completely new block from the system pool. Note that we rely on
+ malloc() to provide aligned memory. */
+
+static union block_hdr *malloc_block(int size)
+{
+ union block_hdr *blok;
+ int request_size;
+
+#ifdef ALLOC_DEBUG
+ /* make some room at the end which we'll fill and expect to be
+ * always filled
+ */
+ size += CLICK_SZ;
+#endif
+#ifdef ALLOC_STATS
+ ++num_malloc_calls;
+ num_malloc_bytes += size + sizeof(union block_hdr);
+#endif
+ request_size = size + sizeof(union block_hdr);
+ blok = (union block_hdr *) malloc(request_size);
+ if (blok == NULL) {
+ fprintf(stderr, "Ouch! malloc(%d) failed in malloc_block()\n",
+ request_size);
+ exit(1);
+ }
+ debug_fill(blok, size + sizeof(union block_hdr));
+ blok->h.next = NULL;
+ blok->h.first_avail = (char *) (blok + 1);
+ blok->h.endp = size + blok->h.first_avail;
+#ifdef ALLOC_DEBUG
+ blok->h.endp -= CLICK_SZ;
+#endif
+#ifdef POOL_DEBUG
+ blok->h.global_next = global_block_list;
+ global_block_list = blok;
+ blok->h.owning_pool = NULL;
+#endif
+
+ return blok;
+}
+
+
+
+#if defined(ALLOC_DEBUG) && !defined(ALLOC_USE_MALLOC)
+static void chk_on_blk_list(union block_hdr *blok, union block_hdr *free_blk)
+{
+ debug_verify_filled(blok->h.endp, blok->h.endp + CLICK_SZ,
+ "Ouch! Someone trounced the padding at the end of a block!\n");
+ while (free_blk) {
+ if (free_blk == blok) {
+ fprintf(stderr, "Ouch! Freeing free block\n");
+ abort();
+ exit(1);
+ }
+ free_blk = free_blk->h.next;
+ }
+}
+#else
+#define chk_on_blk_list(_x, _y)
+#endif
+
+/* Free a chain of blocks --- must be called with alarms blocked. */
+
+static void free_blocks(union block_hdr *blok)
+{
+#ifdef ALLOC_USE_MALLOC
+ union block_hdr *next;
+
+ for (; blok; blok = next) {
+ next = blok->h.next;
+ free(blok);
+ }
+#else
+#ifdef ALLOC_STATS
+ unsigned num_blocks;
+#endif
+ /* First, put new blocks at the head of the free list ---
+ * we'll eventually bash the 'next' pointer of the last block
+ * in the chain to point to the free blocks we already had.
+ */
+
+ union block_hdr *old_free_list;
+
+ if (blok == NULL)
+ return; /* Sanity check --- freeing empty pool? */
+
+ (void) ap_acquire_mutex(alloc_mutex);
+ old_free_list = block_freelist;
+ block_freelist = blok;
+
+ /*
+ * Next, adjust first_avail pointers of each block --- have to do it
+ * sooner or later, and it simplifies the search in new_block to do it
+ * now.
+ */
+
+#ifdef ALLOC_STATS
+ num_blocks = 1;
+#endif
+ while (blok->h.next != NULL) {
+#ifdef ALLOC_STATS
+ ++num_blocks;
+#endif
+ chk_on_blk_list(blok, old_free_list);
+ blok->h.first_avail = (char *) (blok + 1);
+ debug_fill(blok->h.first_avail, blok->h.endp - blok->h.first_avail);
+#ifdef POOL_DEBUG
+ blok->h.owning_pool = FREE_POOL;
+#endif
+ blok = blok->h.next;
+ }
+
+ chk_on_blk_list(blok, old_free_list);
+ blok->h.first_avail = (char *) (blok + 1);
+ debug_fill(blok->h.first_avail, blok->h.endp - blok->h.first_avail);
+#ifdef POOL_DEBUG
+ blok->h.owning_pool = FREE_POOL;
+#endif
+
+ /* Finally, reset next pointer to get the old free blocks back */
+
+ blok->h.next = old_free_list;
+
+#ifdef ALLOC_STATS
+ if (num_blocks > max_blocks_in_one_free) {
+ max_blocks_in_one_free = num_blocks;
+ }
+ ++num_free_blocks_calls;
+ num_blocks_freed += num_blocks;
+#endif
+
+ (void) ap_release_mutex(alloc_mutex);
+#endif
+}
+
+
+/* Get a new block, from our own free list if possible, from the system
+ * if necessary. Must be called with alarms blocked.
+ */
+
+static union block_hdr *new_block(int min_size)
+{
+ union block_hdr **lastptr = &block_freelist;
+ union block_hdr *blok = block_freelist;
+
+ /* First, see if we have anything of the required size
+ * on the free list...
+ */
+
+ while (blok != NULL) {
+ if (min_size + BLOCK_MINFREE <= blok->h.endp - blok->h.first_avail) {
+ *lastptr = blok->h.next;
+ blok->h.next = NULL;
+ debug_verify_filled(blok->h.first_avail, blok->h.endp,
+ "Ouch! Someone trounced a block on the free list!\n");
+ return blok;
+ }
+ else {
+ lastptr = &blok->h.next;
+ blok = blok->h.next;
+ }
+ }
+
+ /* Nope. */
+
+ min_size += BLOCK_MINFREE;
+ blok = malloc_block((min_size > BLOCK_MINALLOC) ? min_size : BLOCK_MINALLOC);
+ return blok;
+}
+
+
+/* Accounting */
+
+static long bytes_in_block_list(union block_hdr *blok)
+{
+ long size = 0;
+
+ while (blok) {
+ size += blok->h.endp - (char *) (blok + 1);
+ blok = blok->h.next;
+ }
+
+ return size;
+}
+
+
+/*****************************************************************
+ *
+ * Pool internals and management...
+ * NB that subprocesses are not handled by the generic cleanup code,
+ * basically because we don't want cleanups for multiple subprocesses
+ * to result in multiple three-second pauses.
+ */
+
+struct process_chain;
+struct cleanup;
+
+static void run_cleanups(struct cleanup *);
+static void free_proc_chain(struct process_chain *);
+
+struct pool {
+ union block_hdr *first;
+ union block_hdr *last;
+ struct cleanup *cleanups;
+ struct process_chain *subprocesses;
+ struct pool *sub_pools;
+ struct pool *sub_next;
+ struct pool *sub_prev;
+ struct pool *parent;
+ char *free_first_avail;
+#ifdef ALLOC_USE_MALLOC
+ void *allocation_list;
+#endif
+#ifdef POOL_DEBUG
+ struct pool *joined;
+#endif
+};
+
+static pool *permanent_pool;
+
+/* Each pool structure is allocated in the start of its own first block,
+ * so we need to know how many bytes that is (once properly aligned...).
+ * This also means that when a pool's sub-pool is destroyed, the storage
+ * associated with it is *completely* gone, so we have to make sure it
+ * gets taken off the parent's sub-pool list...
+ */
+
+#define POOL_HDR_CLICKS (1 + ((sizeof(struct pool) - 1) / CLICK_SZ))
+#define POOL_HDR_BYTES (POOL_HDR_CLICKS * CLICK_SZ)
+
+API_EXPORT(struct pool *) ap_make_sub_pool(struct pool *p)
+{
+ union block_hdr *blok;
+ pool *new_pool;
+
+ ap_block_alarms();
+
+ (void) ap_acquire_mutex(alloc_mutex);
+
+ blok = new_block(POOL_HDR_BYTES);
+ new_pool = (pool *) blok->h.first_avail;
+ blok->h.first_avail += POOL_HDR_BYTES;
+#ifdef POOL_DEBUG
+ blok->h.owning_pool = new_pool;
+#endif
+
+ memset((char *) new_pool, '\0', sizeof(struct pool));
+ new_pool->free_first_avail = blok->h.first_avail;
+ new_pool->first = new_pool->last = blok;
+
+ if (p) {
+ new_pool->parent = p;
+ new_pool->sub_next = p->sub_pools;
+ if (new_pool->sub_next)
+ new_pool->sub_next->sub_prev = new_pool;
+ p->sub_pools = new_pool;
+ }
+
+ (void) ap_release_mutex(alloc_mutex);
+ ap_unblock_alarms();
+
+ return new_pool;
+}
+
+#ifdef POOL_DEBUG
+static void stack_var_init(char *s)
+{
+ char t;
+
+ if (s < &t) {
+ stack_direction = 1; /* stack grows up */
+ }
+ else {
+ stack_direction = -1; /* stack grows down */
+ }
+}
+#endif
+
+#ifdef ALLOC_STATS
+static void dump_stats(void)
+{
+ fprintf(stderr,
+ "alloc_stats: [%d] #free_blocks %llu #blocks %llu max %u #malloc %u #bytes %u\n",
+ (int)getpid(),
+ num_free_blocks_calls,
+ num_blocks_freed,
+ max_blocks_in_one_free,
+ num_malloc_calls,
+ num_malloc_bytes);
+}
+#endif
+
+API_EXPORT(pool *) ap_init_alloc(void)
+{
+#ifdef POOL_DEBUG
+ char s;
+
+ known_stack_point = &s;
+ stack_var_init(&s);
+#endif
+ alloc_mutex = ap_create_mutex(NULL);
+ spawn_mutex = ap_create_mutex(NULL);
+ permanent_pool = ap_make_sub_pool(NULL);
+#ifdef ALLOC_STATS
+ atexit(dump_stats);
+#endif
+
+ return permanent_pool;
+}
+
+void ap_cleanup_alloc(void)
+{
+ ap_destroy_mutex(alloc_mutex);
+ ap_destroy_mutex(spawn_mutex);
+}
+
+API_EXPORT(void) ap_clear_pool(struct pool *a)
+{
+ ap_block_alarms();
+
+ (void) ap_acquire_mutex(alloc_mutex);
+ while (a->sub_pools)
+ ap_destroy_pool(a->sub_pools);
+ (void) ap_release_mutex(alloc_mutex);
+ /* Don't hold the mutex during cleanups. */
+ run_cleanups(a->cleanups);
+ a->cleanups = NULL;
+ free_proc_chain(a->subprocesses);
+ a->subprocesses = NULL;
+ free_blocks(a->first->h.next);
+ a->first->h.next = NULL;
+
+ a->last = a->first;
+ a->first->h.first_avail = a->free_first_avail;
+ debug_fill(a->first->h.first_avail,
+ a->first->h.endp - a->first->h.first_avail);
+
+#ifdef ALLOC_USE_MALLOC
+ {
+ void *c, *n;
+
+ for (c = a->allocation_list; c; c = n) {
+ n = *(void **)c;
+ free(c);
+ }
+ a->allocation_list = NULL;
+ }
+#endif
+
+ ap_unblock_alarms();
+}
+
+API_EXPORT(void) ap_destroy_pool(pool *a)
+{
+ ap_block_alarms();
+ ap_clear_pool(a);
+
+ (void) ap_acquire_mutex(alloc_mutex);
+ if (a->parent) {
+ if (a->parent->sub_pools == a)
+ a->parent->sub_pools = a->sub_next;
+ if (a->sub_prev)
+ a->sub_prev->sub_next = a->sub_next;
+ if (a->sub_next)
+ a->sub_next->sub_prev = a->sub_prev;
+ }
+ (void) ap_release_mutex(alloc_mutex);
+
+ free_blocks(a->first);
+ ap_unblock_alarms();
+}
+
+API_EXPORT(long) ap_bytes_in_pool(pool *p)
+{
+ return bytes_in_block_list(p->first);
+}
+API_EXPORT(long) ap_bytes_in_free_blocks(void)
+{
+ return bytes_in_block_list(block_freelist);
+}
+
+/*****************************************************************
+ * POOL_DEBUG support
+ */
+#ifdef POOL_DEBUG
+
+/* the unix linker defines this symbol as the last byte + 1 of
+ * the executable... so it includes TEXT, BSS, and DATA
+ */
+extern char _end;
+
+/* is ptr in the range [lo,hi) */
+#define is_ptr_in_range(ptr, lo, hi) \
+ (((unsigned long)(ptr) - (unsigned long)(lo)) \
+ < \
+ (unsigned long)(hi) - (unsigned long)(lo))
+
+/* Find the pool that ts belongs to, return NULL if it doesn't
+ * belong to any pool.
+ */
+API_EXPORT(pool *) ap_find_pool(const void *ts)
+{
+ const char *s = ts;
+ union block_hdr **pb;
+ union block_hdr *b;
+
+ /* short-circuit stuff which is in TEXT, BSS, or DATA */
+ if (is_ptr_in_range(s, 0, &_end)) {
+ return NULL;
+ }
+ /* consider stuff on the stack to also be in the NULL pool...
+ * XXX: there's cases where we don't want to assume this
+ */
+ if ((stack_direction == -1 && is_ptr_in_range(s, &ts, known_stack_point))
+ || (stack_direction == 1 && is_ptr_in_range(s, known_stack_point, &ts))) {
+ abort();
+ return NULL;
+ }
+ ap_block_alarms();
+ /* search the global_block_list */
+ for (pb = &global_block_list; *pb; pb = &b->h.global_next) {
+ b = *pb;
+ if (is_ptr_in_range(s, b, b->h.endp)) {
+ if (b->h.owning_pool == FREE_POOL) {
+ fprintf(stderr,
+ "Ouch! find_pool() called on pointer in a free block\n");
+ abort();
+ exit(1);
+ }
+ if (b != global_block_list) {
+ /* promote b to front of list, this is a hack to speed
+ * up the lookup */
+ *pb = b->h.global_next;
+ b->h.global_next = global_block_list;
+ global_block_list = b;
+ }
+ ap_unblock_alarms();
+ return b->h.owning_pool;
+ }
+ }
+ ap_unblock_alarms();
+ return NULL;
+}
+
+/* return TRUE iff a is an ancestor of b
+ * NULL is considered an ancestor of all pools
+ */
+API_EXPORT(int) ap_pool_is_ancestor(pool *a, pool *b)
+{
+ if (a == NULL) {
+ return 1;
+ }
+ while (a->joined) {
+ a = a->joined;
+ }
+ while (b) {
+ if (a == b) {
+ return 1;
+ }
+ b = b->parent;
+ }
+ return 0;
+}
+
+/* All blocks belonging to sub will be changed to point to p
+ * instead. This is a guarantee by the caller that sub will not
+ * be destroyed before p is.
+ */
+API_EXPORT(void) ap_pool_join(pool *p, pool *sub)
+{
+ union block_hdr *b;
+
+ /* We could handle more general cases... but this is it for now. */
+ if (sub->parent != p) {
+ fprintf(stderr, "pool_join: p is not parent of sub\n");
+ abort();
+ }
+ ap_block_alarms();
+ while (p->joined) {
+ p = p->joined;
+ }
+ sub->joined = p;
+ for (b = global_block_list; b; b = b->h.global_next) {
+ if (b->h.owning_pool == sub) {
+ b->h.owning_pool = p;
+ }
+ }
+ ap_unblock_alarms();
+}
+#endif
+
+/*****************************************************************
+ *
+ * Allocating stuff...
+ */
+
+
+API_EXPORT(void *) ap_palloc(struct pool *a, int reqsize)
+{
+#ifdef ALLOC_USE_MALLOC
+ int size = reqsize + CLICK_SZ;
+ void *ptr;
+
+ ap_block_alarms();
+ ptr = malloc(size);
+ if (ptr == NULL) {
+ fputs("Ouch! Out of memory!\n", stderr);
+ exit(1);
+ }
+ debug_fill(ptr, size); /* might as well get uninitialized protection */
+ *(void **)ptr = a->allocation_list;
+ a->allocation_list = ptr;
+ ap_unblock_alarms();
+ return (char *)ptr + CLICK_SZ;
+#else
+
+ /* Round up requested size to an even number of alignment units (core clicks)
+ */
+
+ int nclicks = 1 + ((reqsize - 1) / CLICK_SZ);
+ int size = nclicks * CLICK_SZ;
+
+ /* First, see if we have space in the block most recently
+ * allocated to this pool
+ */
+
+ union block_hdr *blok = a->last;
+ char *first_avail = blok->h.first_avail;
+ char *new_first_avail;
+
+ if (reqsize <= 0)
+ return NULL;
+
+ new_first_avail = first_avail + size;
+
+ if (new_first_avail <= blok->h.endp) {
+ debug_verify_filled(first_avail, blok->h.endp,
+ "Ouch! Someone trounced past the end of their allocation!\n");
+ blok->h.first_avail = new_first_avail;
+ return (void *) first_avail;
+ }
+
+ /* Nope --- get a new one that's guaranteed to be big enough */
+
+ ap_block_alarms();
+
+ (void) ap_acquire_mutex(alloc_mutex);
+
+ blok = new_block(size);
+ a->last->h.next = blok;
+ a->last = blok;
+#ifdef POOL_DEBUG
+ blok->h.owning_pool = a;
+#endif
+
+ (void) ap_release_mutex(alloc_mutex);
+
+ ap_unblock_alarms();
+
+ first_avail = blok->h.first_avail;
+ blok->h.first_avail += size;
+
+ return (void *) first_avail;
+#endif
+}
+
+API_EXPORT(void *) ap_pcalloc(struct pool *a, int size)
+{
+ void *res = ap_palloc(a, size);
+ memset(res, '\0', size);
+ return res;
+}
+
+API_EXPORT(char *) ap_pstrdup(struct pool *a, const char *s)
+{
+ char *res;
+ size_t len;
+
+ if (s == NULL)
+ return NULL;
+ len = strlen(s) + 1;
+ res = ap_palloc(a, len);
+ memcpy(res, s, len);
+ return res;
+}
+
+API_EXPORT(char *) ap_pstrndup(struct pool *a, const char *s, int n)
+{
+ char *res;
+
+ if (s == NULL)
+ return NULL;
+ res = ap_palloc(a, n + 1);
+ memcpy(res, s, n);
+ res[n] = '\0';
+ return res;
+}
+
+API_EXPORT_NONSTD(char *) ap_pstrcat(pool *a,...)
+{
+ char *cp, *argp, *res;
+
+ /* Pass one --- find length of required string */
+
+ int len = 0;
+ va_list adummy;
+
+ va_start(adummy, a);
+
+ while ((cp = va_arg(adummy, char *)) != NULL)
+ len += strlen(cp);
+
+ va_end(adummy);
+
+ /* Allocate the required string */
+
+ res = (char *) ap_palloc(a, len + 1);
+ cp = res;
+ *cp = '\0';
+
+ /* Pass two --- copy the argument strings into the result space */
+
+ va_start(adummy, a);
+
+ while ((argp = va_arg(adummy, char *)) != NULL) {
+ strcpy(cp, argp);
+ cp += strlen(argp);
+ }
+
+ va_end(adummy);
+
+ /* Return the result string */
+
+ return res;
+}
+
+/* ap_psprintf is implemented by writing directly into the current
+ * block of the pool, starting right at first_avail. If there's
+ * insufficient room, then a new block is allocated and the earlier
+ * output is copied over. The new block isn't linked into the pool
+ * until all the output is done.
+ *
+ * Note that this is completely safe because nothing else can
+ * allocate in this pool while ap_psprintf is running. alarms are
+ * blocked, and the only thing outside of alloc.c that's invoked
+ * is ap_vformatter -- which was purposefully written to be
+ * self-contained with no callouts.
+ */
+
+struct psprintf_data {
+ ap_vformatter_buff vbuff;
+#ifdef ALLOC_USE_MALLOC
+ char *base;
+#else
+ union block_hdr *blok;
+ int got_a_new_block;
+#endif
+};
+
+#define AP_PSPRINTF_MIN_SIZE 32 /* Minimum size of allowable avail block */
+
+static int psprintf_flush(ap_vformatter_buff *vbuff)
+{
+ struct psprintf_data *ps = (struct psprintf_data *)vbuff;
+#ifdef ALLOC_USE_MALLOC
+ int cur_len, size;
+ char *ptr;
+
+ cur_len = (char *)ps->vbuff.curpos - ps->base;
+ size = cur_len << 1;
+ if (size < AP_PSPRINTF_MIN_SIZE)
+ size = AP_PSPRINTF_MIN_SIZE;
+ ptr = realloc(ps->base, size);
+ if (ptr == NULL) {
+ fputs("Ouch! Out of memory!\n", stderr);
+ exit(1);
+ }
+ ps->base = ptr;
+ ps->vbuff.curpos = ptr + cur_len;
+ ps->vbuff.endpos = ptr + size - 1;
+ return 0;
+#else
+ union block_hdr *blok;
+ union block_hdr *nblok;
+ size_t cur_len, size;
+ char *strp;
+
+ blok = ps->blok;
+ strp = ps->vbuff.curpos;
+ cur_len = strp - blok->h.first_avail;
+ size = cur_len << 1;
+ if (size < AP_PSPRINTF_MIN_SIZE)
+ size = AP_PSPRINTF_MIN_SIZE;
+
+ /* must try another blok */
+ (void) ap_acquire_mutex(alloc_mutex);
+ nblok = new_block(size);
+ (void) ap_release_mutex(alloc_mutex);
+ memcpy(nblok->h.first_avail, blok->h.first_avail, cur_len);
+ ps->vbuff.curpos = nblok->h.first_avail + cur_len;
+ /* save a byte for the NUL terminator */
+ ps->vbuff.endpos = nblok->h.endp - 1;
+
+ /* did we allocate the current blok? if so free it up */
+ if (ps->got_a_new_block) {
+ debug_fill(blok->h.first_avail, blok->h.endp - blok->h.first_avail);
+ (void) ap_acquire_mutex(alloc_mutex);
+ blok->h.next = block_freelist;
+ block_freelist = blok;
+ (void) ap_release_mutex(alloc_mutex);
+ }
+ ps->blok = nblok;
+ ps->got_a_new_block = 1;
+ /* note that we've deliberately not linked the new block onto
+ * the pool yet... because we may need to flush again later, and
+ * we'd have to spend more effort trying to unlink the block.
+ */
+ return 0;
+#endif
+}
+
+API_EXPORT(char *) ap_pvsprintf(pool *p, const char *fmt, va_list ap)
+{
+#ifdef ALLOC_USE_MALLOC
+ struct psprintf_data ps;
+ void *ptr;
+
+ ap_block_alarms();
+ ps.base = malloc(512);
+ if (ps.base == NULL) {
+ fputs("Ouch! Out of memory!\n", stderr);
+ exit(1);
+ }
+ /* need room at beginning for allocation_list */
+ ps.vbuff.curpos = ps.base + CLICK_SZ;
+ ps.vbuff.endpos = ps.base + 511;
+ ap_vformatter(psprintf_flush, &ps.vbuff, fmt, ap);
+ *ps.vbuff.curpos++ = '\0';
+ ptr = ps.base;
+ /* shrink */
+ ptr = realloc(ptr, (char *)ps.vbuff.curpos - (char *)ptr);
+ if (ptr == NULL) {
+ fputs("Ouch! Out of memory!\n", stderr);
+ exit(1);
+ }
+ *(void **)ptr = p->allocation_list;
+ p->allocation_list = ptr;
+ ap_unblock_alarms();
+ return (char *)ptr + CLICK_SZ;
+#else
+ struct psprintf_data ps;
+ char *strp;
+ int size;
+
+ ap_block_alarms();
+ ps.blok = p->last;
+ ps.vbuff.curpos = ps.blok->h.first_avail;
+ ps.vbuff.endpos = ps.blok->h.endp - 1; /* save one for NUL */
+ ps.got_a_new_block = 0;
+
+ if (ps.blok->h.first_avail == ps.blok->h.endp)
+ psprintf_flush(&ps.vbuff); /* ensure room for NUL */
+ ap_vformatter(psprintf_flush, &ps.vbuff, fmt, ap);
+
+ strp = ps.vbuff.curpos;
+ *strp++ = '\0';
+
+ size = strp - ps.blok->h.first_avail;
+ size = (1 + ((size - 1) / CLICK_SZ)) * CLICK_SZ;
+ strp = ps.blok->h.first_avail; /* save away result pointer */
+ ps.blok->h.first_avail += size;
+
+ /* have to link the block in if it's a new one */
+ if (ps.got_a_new_block) {
+ p->last->h.next = ps.blok;
+ p->last = ps.blok;
+#ifdef POOL_DEBUG
+ ps.blok->h.owning_pool = p;
+#endif
+ }
+ ap_unblock_alarms();
+
+ return strp;
+#endif
+}
+
+API_EXPORT_NONSTD(char *) ap_psprintf(pool *p, const char *fmt, ...)
+{
+ va_list ap;
+ char *res;
+
+ va_start(ap, fmt);
+ res = ap_pvsprintf(p, fmt, ap);
+ va_end(ap);
+ return res;
+}
+
+/*****************************************************************
+ *
+ * The 'array' functions...
+ */
+
+static void make_array_core(array_header *res, pool *p, int nelts, int elt_size)
+{
+ if (nelts < 1)
+ nelts = 1; /* Assure sanity if someone asks for
+ * array of zero elts.
+ */
+
+ res->elts = ap_pcalloc(p, nelts * elt_size);
+
+ res->pool = p;
+ res->elt_size = elt_size;
+ res->nelts = 0; /* No active elements yet... */
+ res->nalloc = nelts; /* ...but this many allocated */
+}
+
+API_EXPORT(array_header *) ap_make_array(pool *p, int nelts, int elt_size)
+{
+ array_header *res = (array_header *) ap_palloc(p, sizeof(array_header));
+
+ make_array_core(res, p, nelts, elt_size);
+ return res;
+}
+
+API_EXPORT(void *) ap_push_array(array_header *arr)
+{
+ if (arr->nelts == arr->nalloc) {
+ int new_size = (arr->nalloc <= 0) ? 1 : arr->nalloc * 2;
+ char *new_data;
+
+ new_data = ap_pcalloc(arr->pool, arr->elt_size * new_size);
+
+ memcpy(new_data, arr->elts, arr->nalloc * arr->elt_size);
+ arr->elts = new_data;
+ arr->nalloc = new_size;
+ }
+
+ ++arr->nelts;
+ return arr->elts + (arr->elt_size * (arr->nelts - 1));
+}
+
+API_EXPORT(void) ap_array_cat(array_header *dst, const array_header *src)
+{
+ int elt_size = dst->elt_size;
+
+ if (dst->nelts + src->nelts > dst->nalloc) {
+ int new_size = (dst->nalloc <= 0) ? 1 : dst->nalloc * 2;
+ char *new_data;
+
+ while (dst->nelts + src->nelts > new_size)
+ new_size *= 2;
+
+ new_data = ap_pcalloc(dst->pool, elt_size * new_size);
+ memcpy(new_data, dst->elts, dst->nalloc * elt_size);
+
+ dst->elts = new_data;
+ dst->nalloc = new_size;
+ }
+
+ memcpy(dst->elts + dst->nelts * elt_size, src->elts, elt_size * src->nelts);
+ dst->nelts += src->nelts;
+}
+
+API_EXPORT(array_header *) ap_copy_array(pool *p, const array_header *arr)
+{
+ array_header *res = ap_make_array(p, arr->nalloc, arr->elt_size);
+
+ memcpy(res->elts, arr->elts, arr->elt_size * arr->nelts);
+ res->nelts = arr->nelts;
+ return res;
+}
+
+/* This cute function copies the array header *only*, but arranges
+ * for the data section to be copied on the first push or arraycat.
+ * It's useful when the elements of the array being copied are
+ * read only, but new stuff *might* get added on the end; we have the
+ * overhead of the full copy only where it is really needed.
+ */
+
+static ap_inline void copy_array_hdr_core(array_header *res,
+ const array_header *arr)
+{
+ res->elts = arr->elts;
+ res->elt_size = arr->elt_size;
+ res->nelts = arr->nelts;
+ res->nalloc = arr->nelts; /* Force overflow on push */
+}
+
+API_EXPORT(array_header *) ap_copy_array_hdr(pool *p, const array_header *arr)
+{
+ array_header *res = (array_header *) ap_palloc(p, sizeof(array_header));
+
+ res->pool = p;
+ copy_array_hdr_core(res, arr);
+ return res;
+}
+
+/* The above is used here to avoid consing multiple new array bodies... */
+
+API_EXPORT(array_header *) ap_append_arrays(pool *p,
+ const array_header *first,
+ const array_header *second)
+{
+ array_header *res = ap_copy_array_hdr(p, first);
+
+ ap_array_cat(res, second);
+ return res;
+}
+
+/* ap_array_pstrcat generates a new string from the pool containing
+ * the concatenated sequence of substrings referenced as elements within
+ * the array. The string will be empty if all substrings are empty or null,
+ * or if there are no elements in the array.
+ * If sep is non-NUL, it will be inserted between elements as a separator.
+ */
+API_EXPORT(char *) ap_array_pstrcat(pool *p, const array_header *arr,
+ const char sep)
+{
+ char *cp, *res, **strpp;
+ int i, len;
+
+ if (arr->nelts <= 0 || arr->elts == NULL) /* Empty table? */
+ return (char *) ap_pcalloc(p, 1);
+
+ /* Pass one --- find length of required string */
+
+ len = 0;
+ for (i = 0, strpp = (char **) arr->elts; ; ++strpp) {
+ if (strpp && *strpp != NULL) {
+ len += strlen(*strpp);
+ }
+ if (++i >= arr->nelts)
+ break;
+ if (sep)
+ ++len;
+ }
+
+ /* Allocate the required string */
+
+ res = (char *) ap_palloc(p, len + 1);
+ cp = res;
+
+ /* Pass two --- copy the argument strings into the result space */
+
+ for (i = 0, strpp = (char **) arr->elts; ; ++strpp) {
+ if (strpp && *strpp != NULL) {
+ len = strlen(*strpp);
+ memcpy(cp, *strpp, len);
+ cp += len;
+ }
+ if (++i >= arr->nelts)
+ break;
+ if (sep)
+ *cp++ = sep;
+ }
+
+ *cp = '\0';
+
+ /* Return the result string */
+
+ return res;
+}
+
+
+/*****************************************************************
+ *
+ * The "table" functions.
+ */
+
+/* XXX: if you tweak this you should look at is_empty_table() and table_elts()
+ * in ap_alloc.h */
+struct table {
+ /* This has to be first to promote backwards compatibility with
+ * older modules which cast a table * to an array_header *...
+ * they should use the table_elts() function for most of the
+ * cases they do this for.
+ */
+ array_header a;
+#ifdef MAKE_TABLE_PROFILE
+ void *creator;
+#endif
+};
+
+#ifdef MAKE_TABLE_PROFILE
+static table_entry *table_push(table *t)
+{
+ if (t->a.nelts == t->a.nalloc) {
+ fprintf(stderr,
+ "table_push: table created by %p hit limit of %u\n",
+ t->creator, t->a.nalloc);
+ }
+ return (table_entry *) ap_push_array(&t->a);
+}
+#else
+#define table_push(t) ((table_entry *) ap_push_array(&(t)->a))
+#endif
+
+
+API_EXPORT(table *) ap_make_table(pool *p, int nelts)
+{
+ table *t = ap_palloc(p, sizeof(table));
+
+ make_array_core(&t->a, p, nelts, sizeof(table_entry));
+#ifdef MAKE_TABLE_PROFILE
+ t->creator = __builtin_return_address(0);
+#endif
+ return t;
+}
+
+API_EXPORT(table *) ap_copy_table(pool *p, const table *t)
+{
+ table *new = ap_palloc(p, sizeof(table));
+
+#ifdef POOL_DEBUG
+ /* we don't copy keys and values, so it's necessary that t->a.pool
+ * have a life span at least as long as p
+ */
+ if (!ap_pool_is_ancestor(t->a.pool, p)) {
+ fprintf(stderr, "copy_table: t's pool is not an ancestor of p\n");
+ abort();
+ }
+#endif
+ make_array_core(&new->a, p, t->a.nalloc, sizeof(table_entry));
+ memcpy(new->a.elts, t->a.elts, t->a.nelts * sizeof(table_entry));
+ new->a.nelts = t->a.nelts;
+ return new;
+}
+
+API_EXPORT(void) ap_clear_table(table *t)
+{
+ t->a.nelts = 0;
+}
+
+API_EXPORT(const char *) ap_table_get(const table *t, const char *key)
+{
+ table_entry *elts = (table_entry *) t->a.elts;
+ int i;
+
+ if (key == NULL)
+ return NULL;
+
+ for (i = 0; i < t->a.nelts; ++i)
+ if (!strcasecmp(elts[i].key, key))
+ return elts[i].val;
+
+ return NULL;
+}
+
+API_EXPORT(void) ap_table_set(table *t, const char *key, const char *val)
+{
+ register int i, j, k;
+ table_entry *elts = (table_entry *) t->a.elts;
+ int done = 0;
+
+ for (i = 0; i < t->a.nelts; ) {
+ if (!strcasecmp(elts[i].key, key)) {
+ if (!done) {
+ elts[i].val = ap_pstrdup(t->a.pool, val);
+ done = 1;
+ ++i;
+ }
+ else { /* delete an extraneous element */
+ for (j = i, k = i + 1; k < t->a.nelts; ++j, ++k) {
+ elts[j].key = elts[k].key;
+ elts[j].val = elts[k].val;
+ }
+ --t->a.nelts;
+ }
+ }
+ else {
+ ++i;
+ }
+ }
+
+ if (!done) {
+ elts = (table_entry *) table_push(t);
+ elts->key = ap_pstrdup(t->a.pool, key);
+ elts->val = ap_pstrdup(t->a.pool, val);
+ }
+}
+
+API_EXPORT(void) ap_table_setn(table *t, const char *key, const char *val)
+{
+ register int i, j, k;
+ table_entry *elts = (table_entry *) t->a.elts;
+ int done = 0;
+
+#ifdef POOL_DEBUG
+ {
+ if (!ap_pool_is_ancestor(ap_find_pool(key), t->a.pool)) {
+ fprintf(stderr, "table_set: key not in ancestor pool of t\n");
+ abort();
+ }
+ if (!ap_pool_is_ancestor(ap_find_pool(val), t->a.pool)) {
+ fprintf(stderr, "table_set: val not in ancestor pool of t\n");
+ abort();
+ }
+ }
+#endif
+
+ for (i = 0; i < t->a.nelts; ) {
+ if (!strcasecmp(elts[i].key, key)) {
+ if (!done) {
+ elts[i].val = (char *)val;
+ done = 1;
+ ++i;
+ }
+ else { /* delete an extraneous element */
+ for (j = i, k = i + 1; k < t->a.nelts; ++j, ++k) {
+ elts[j].key = elts[k].key;
+ elts[j].val = elts[k].val;
+ }
+ --t->a.nelts;
+ }
+ }
+ else {
+ ++i;
+ }
+ }
+
+ if (!done) {
+ elts = (table_entry *) table_push(t);
+ elts->key = (char *)key;
+ elts->val = (char *)val;
+ }
+}
+
+API_EXPORT(void) ap_table_unset(table *t, const char *key)
+{
+ register int i, j, k;
+ table_entry *elts = (table_entry *) t->a.elts;
+
+ for (i = 0; i < t->a.nelts;) {
+ if (!strcasecmp(elts[i].key, key)) {
+
+ /* found an element to skip over
+ * there are any number of ways to remove an element from
+ * a contiguous block of memory. I've chosen one that
+ * doesn't do a memcpy/bcopy/array_delete, *shrug*...
+ */
+ for (j = i, k = i + 1; k < t->a.nelts; ++j, ++k) {
+ elts[j].key = elts[k].key;
+ elts[j].val = elts[k].val;
+ }
+ --t->a.nelts;
+ }
+ else {
+ ++i;
+ }
+ }
+}
+
+API_EXPORT(void) ap_table_merge(table *t, const char *key, const char *val)
+{
+ table_entry *elts = (table_entry *) t->a.elts;
+ int i;
+
+ for (i = 0; i < t->a.nelts; ++i)
+ if (!strcasecmp(elts[i].key, key)) {
+ elts[i].val = ap_pstrcat(t->a.pool, elts[i].val, ", ", val, NULL);
+ return;
+ }
+
+ elts = (table_entry *) table_push(t);
+ elts->key = ap_pstrdup(t->a.pool, key);
+ elts->val = ap_pstrdup(t->a.pool, val);
+}
+
+API_EXPORT(void) ap_table_mergen(table *t, const char *key, const char *val)
+{
+ table_entry *elts = (table_entry *) t->a.elts;
+ int i;
+
+#ifdef POOL_DEBUG
+ {
+ if (!ap_pool_is_ancestor(ap_find_pool(key), t->a.pool)) {
+ fprintf(stderr, "table_set: key not in ancestor pool of t\n");
+ abort();
+ }
+ if (!ap_pool_is_ancestor(ap_find_pool(val), t->a.pool)) {
+ fprintf(stderr, "table_set: key not in ancestor pool of t\n");
+ abort();
+ }
+ }
+#endif
+
+ for (i = 0; i < t->a.nelts; ++i) {
+ if (!strcasecmp(elts[i].key, key)) {
+ elts[i].val = ap_pstrcat(t->a.pool, elts[i].val, ", ", val, NULL);
+ return;
+ }
+ }
+
+ elts = (table_entry *) table_push(t);
+ elts->key = (char *)key;
+ elts->val = (char *)val;
+}
+
+API_EXPORT(void) ap_table_add(table *t, const char *key, const char *val)
+{
+ table_entry *elts = (table_entry *) t->a.elts;
+
+ elts = (table_entry *) table_push(t);
+ elts->key = ap_pstrdup(t->a.pool, key);
+ elts->val = ap_pstrdup(t->a.pool, val);
+}
+
+API_EXPORT(void) ap_table_addn(table *t, const char *key, const char *val)
+{
+ table_entry *elts = (table_entry *) t->a.elts;
+
+#ifdef POOL_DEBUG
+ {
+ if (!ap_pool_is_ancestor(ap_find_pool(key), t->a.pool)) {
+ fprintf(stderr, "table_set: key not in ancestor pool of t\n");
+ abort();
+ }
+ if (!ap_pool_is_ancestor(ap_find_pool(val), t->a.pool)) {
+ fprintf(stderr, "table_set: key not in ancestor pool of t\n");
+ abort();
+ }
+ }
+#endif
+
+ elts = (table_entry *) table_push(t);
+ elts->key = (char *)key;
+ elts->val = (char *)val;
+}
+
+API_EXPORT(table *) ap_overlay_tables(pool *p, const table *overlay, const table *base)
+{
+ table *res;
+
+#ifdef POOL_DEBUG
+ /* we don't copy keys and values, so it's necessary that
+ * overlay->a.pool and base->a.pool have a life span at least
+ * as long as p
+ */
+ if (!ap_pool_is_ancestor(overlay->a.pool, p)) {
+ fprintf(stderr, "overlay_tables: overlay's pool is not an ancestor of p\n");
+ abort();
+ }
+ if (!ap_pool_is_ancestor(base->a.pool, p)) {
+ fprintf(stderr, "overlay_tables: base's pool is not an ancestor of p\n");
+ abort();
+ }
+#endif
+
+ res = ap_palloc(p, sizeof(table));
+ /* behave like append_arrays */
+ res->a.pool = p;
+ copy_array_hdr_core(&res->a, &overlay->a);
+ ap_array_cat(&res->a, &base->a);
+
+ return res;
+}
+
+/* And now for something completely abstract ...
+
+ * For each key value given as a vararg:
+ * run the function pointed to as
+ * int comp(void *r, char *key, char *value);
+ * on each valid key-value pair in the table t that matches the vararg key,
+ * or once for every valid key-value pair if the vararg list is empty,
+ * until the function returns false (0) or we finish the table.
+ *
+ * Note that we restart the traversal for each vararg, which means that
+ * duplicate varargs will result in multiple executions of the function
+ * for each matching key. Note also that if the vararg list is empty,
+ * only one traversal will be made and will cut short if comp returns 0.
+ *
+ * Note that the table_get and table_merge functions assume that each key in
+ * the table is unique (i.e., no multiple entries with the same key). This
+ * function does not make that assumption, since it (unfortunately) isn't
+ * true for some of Apache's tables.
+ *
+ * Note that rec is simply passed-on to the comp function, so that the
+ * caller can pass additional info for the task.
+ */
+API_EXPORT_NONSTD(void) ap_table_do(int (*comp) (void *, const char *, const char *),
+ void *rec, const table *t,...)
+{
+ va_list vp;
+ char *argp;
+ table_entry *elts = (table_entry *) t->a.elts;
+ int rv, i;
+
+ va_start(vp, t);
+
+ argp = va_arg(vp, char *);
+
+ do {
+ for (rv = 1, i = 0; rv && (i < t->a.nelts); ++i) {
+ if (elts[i].key && (!argp || !strcasecmp(elts[i].key, argp))) {
+ rv = (*comp) (rec, elts[i].key, elts[i].val);
+ }
+ }
+ } while (argp && ((argp = va_arg(vp, char *)) != NULL));
+
+ va_end(vp);
+}
+
+/* Curse libc and the fact that it doesn't guarantee a stable sort. We
+ * have to enforce stability ourselves by using the order field. If it
+ * provided a stable sort then we wouldn't even need temporary storage to
+ * do the work below. -djg
+ *
+ * ("stable sort" means that equal keys retain their original relative
+ * ordering in the output.)
+ */
+typedef struct {
+ char *key;
+ char *val;
+ int order;
+} overlap_key;
+
+static int sort_overlap(const void *va, const void *vb)
+{
+ const overlap_key *a = va;
+ const overlap_key *b = vb;
+ int r;
+
+ r = strcasecmp(a->key, b->key);
+ if (r) {
+ return r;
+ }
+ return a->order - b->order;
+}
+
+/* prefer to use the stack for temp storage for overlaps smaller than this */
+#ifndef AP_OVERLAP_TABLES_ON_STACK
+#define AP_OVERLAP_TABLES_ON_STACK (512)
+#endif
+
+API_EXPORT(void) ap_overlap_tables(table *a, const table *b, unsigned flags)
+{
+ overlap_key cat_keys_buf[AP_OVERLAP_TABLES_ON_STACK];
+ overlap_key *cat_keys;
+ int nkeys;
+ table_entry *e;
+ table_entry *last_e;
+ overlap_key *left;
+ overlap_key *right;
+ overlap_key *last;
+
+ nkeys = a->a.nelts + b->a.nelts;
+ if (nkeys < AP_OVERLAP_TABLES_ON_STACK) {
+ cat_keys = cat_keys_buf;
+ }
+ else {
+ /* XXX: could use scratch free space in a or b's pool instead...
+ * which could save an allocation in b's pool.
+ */
+ cat_keys = ap_palloc(b->a.pool, sizeof(overlap_key) * nkeys);
+ }
+
+ nkeys = 0;
+
+ /* Create a list of the entries from a concatenated with the entries
+ * from b.
+ */
+ e = (table_entry *)a->a.elts;
+ last_e = e + a->a.nelts;
+ while (e < last_e) {
+ cat_keys[nkeys].key = e->key;
+ cat_keys[nkeys].val = e->val;
+ cat_keys[nkeys].order = nkeys;
+ ++nkeys;
+ ++e;
+ }
+
+ e = (table_entry *)b->a.elts;
+ last_e = e + b->a.nelts;
+ while (e < last_e) {
+ cat_keys[nkeys].key = e->key;
+ cat_keys[nkeys].val = e->val;
+ cat_keys[nkeys].order = nkeys;
+ ++nkeys;
+ ++e;
+ }
+
+ qsort(cat_keys, nkeys, sizeof(overlap_key), sort_overlap);
+
+ /* Now iterate over the sorted list and rebuild a.
+ * Start by making sure it has enough space.
+ */
+ a->a.nelts = 0;
+ if (a->a.nalloc < nkeys) {
+ a->a.elts = ap_palloc(a->a.pool, a->a.elt_size * nkeys * 2);
+ a->a.nalloc = nkeys * 2;
+ }
+
+ /*
+ * In both the merge and set cases we retain the invariant:
+ *
+ * left->key, (left+1)->key, (left+2)->key, ..., (right-1)->key
+ * are all equal keys. (i.e. strcasecmp returns 0)
+ *
+ * We essentially need to find the maximal
+ * right for each key, then we can do a quick merge or set as
+ * appropriate.
+ */
+
+ if (flags & AP_OVERLAP_TABLES_MERGE) {
+ left = cat_keys;
+ last = left + nkeys;
+ while (left < last) {
+ right = left + 1;
+ if (right == last
+ || strcasecmp(left->key, right->key)) {
+ ap_table_addn(a, left->key, left->val);
+ left = right;
+ }
+ else {
+ char *strp;
+ char *value;
+ size_t len;
+
+ /* Have to merge some headers. Let's re-use the order field,
+ * since it's handy... we'll store the length of val there.
+ */
+ left->order = strlen(left->val);
+ len = left->order;
+ do {
+ right->order = strlen(right->val);
+ len += 2 + right->order;
+ ++right;
+ } while (right < last
+ && !strcasecmp(left->key, right->key));
+ /* right points one past the last header to merge */
+ value = ap_palloc(a->a.pool, len + 1);
+ strp = value;
+ for (;;) {
+ memcpy(strp, left->val, left->order);
+ strp += left->order;
+ ++left;
+ if (left == right) break;
+ *strp++ = ',';
+ *strp++ = ' ';
+ }
+ *strp = 0;
+ ap_table_addn(a, (left-1)->key, value);
+ }
+ }
+ }
+ else {
+ left = cat_keys;
+ last = left + nkeys;
+ while (left < last) {
+ right = left + 1;
+ while (right < last && !strcasecmp(left->key, right->key)) {
+ ++right;
+ }
+ ap_table_addn(a, (right-1)->key, (right-1)->val);
+ left = right;
+ }
+ }
+}
+
+/*****************************************************************
+ *
+ * Managing generic cleanups.
+ */
+
+struct cleanup {
+ void *data;
+ void (*plain_cleanup) (void *);
+ void (*child_cleanup) (void *);
+ struct cleanup *next;
+};
+
+API_EXPORT(void) ap_register_cleanup_ex(pool *p, void *data,
+ void (*plain_cleanup) (void *),
+ void (*child_cleanup) (void *),
+ int (*magic_cleanup) (void *))
+{
+ struct cleanup *c;
+ if (p) {
+ c = (struct cleanup *) ap_palloc(p, sizeof(struct cleanup));
+ c->data = data;
+ c->plain_cleanup = plain_cleanup;
+ c->child_cleanup = child_cleanup;
+ c->next = p->cleanups;
+ p->cleanups = c;
+ }
+ /* attempt to do magic even if not passed a pool. Allows us
+ * to perform the magic, therefore, "whenever" we want/need */
+ if (magic_cleanup) {
+ if (!magic_cleanup(data))
+ ap_log_error(APLOG_MARK, APLOG_WARNING, NULL,
+ "exec() may not be safe");
+ }
+}
+
+API_EXPORT(void) ap_register_cleanup(pool *p, void *data,
+ void (*plain_cleanup) (void *),
+ void (*child_cleanup) (void *))
+{
+ ap_register_cleanup_ex(p, data, plain_cleanup, child_cleanup, NULL);
+}
+
+API_EXPORT(void) ap_kill_cleanup(pool *p, void *data, void (*cleanup) (void *))
+{
+ struct cleanup *c = p->cleanups;
+ struct cleanup **lastp = &p->cleanups;
+
+ while (c) {
+ if (c->data == data && c->plain_cleanup == cleanup) {
+ *lastp = c->next;
+ break;
+ }
+
+ lastp = &c->next;
+ c = c->next;
+ }
+}
+
+API_EXPORT(void) ap_run_cleanup(pool *p, void *data, void (*cleanup) (void *))
+{
+ ap_block_alarms(); /* Run cleanup only once! */
+ (*cleanup) (data);
+ ap_kill_cleanup(p, data, cleanup);
+ ap_unblock_alarms();
+}
+
+static void run_cleanups(struct cleanup *c)
+{
+ while (c) {
+ (*c->plain_cleanup) (c->data);
+ c = c->next;
+ }
+}
+
+static void run_child_cleanups(struct cleanup *c)
+{
+ while (c) {
+ (*c->child_cleanup) (c->data);
+ c = c->next;
+ }
+}
+
+static void cleanup_pool_for_exec(pool *p)
+{
+ run_child_cleanups(p->cleanups);
+ p->cleanups = NULL;
+
+ for (p = p->sub_pools; p; p = p->sub_next)
+ cleanup_pool_for_exec(p);
+}
+
+API_EXPORT(void) ap_cleanup_for_exec(void)
+{
+#if !defined(WIN32) && !defined(OS2) && !defined(NETWARE)
+ /*
+ * Don't need to do anything on NT, NETWARE or OS/2, because I
+ * am actually going to spawn the new process - not
+ * exec it. All handles that are not inheritable, will
+ * be automajically closed. The only problem is with
+ * file handles that are open, but there isn't much
+ * I can do about that (except if the child decides
+ * to go out and close them
+ */
+ ap_block_alarms();
+ cleanup_pool_for_exec(permanent_pool);
+ ap_unblock_alarms();
+#endif /* ndef WIN32 */
+}
+
+API_EXPORT_NONSTD(void) ap_null_cleanup(void *data)
+{
+ /* do nothing cleanup routine */
+}
+
+/*****************************************************************
+ *
+ * Files and file descriptors; these are just an application of the
+ * generic cleanup interface.
+ */
+
+#if defined(WIN32)
+/* Provided by service.c, internal to the core library (not exported) */
+BOOL isWindowsNT(void);
+
+int ap_close_handle_on_exec(HANDLE nth)
+{
+ /* Protect the fd so that it will not be inherited by child processes */
+ if (isWindowsNT()) {
+ DWORD hinfo;
+ if (!GetHandleInformation(nth, &hinfo)) {
+ ap_log_error(APLOG_MARK, APLOG_ERR, NULL, "GetHandleInformation"
+ "(%08x) failed", nth);
+ return 0;
+ }
+ if ((hinfo & HANDLE_FLAG_INHERIT)
+ && !SetHandleInformation(nth, HANDLE_FLAG_INHERIT, 0)) {
+ ap_log_error(APLOG_MARK, APLOG_ERR, NULL, "SetHandleInformation"
+ "(%08x, HANDLE_FLAG_INHERIT, 0) failed", nth);
+ return 0;
+ }
+ return 1;
+ }
+ else /* Win9x */ {
+ /* XXX: This API doesn't work... you can't change the handle by just
+ * 'touching' it... you must duplicat to a second handle and close
+ * the original.
+ */
+ return 0;
+ }
+}
+
+int ap_close_fd_on_exec(int fd)
+{
+ return ap_close_handle_on_exec((HANDLE)_get_osfhandle(fd));
+}
+
+#else
+
+int ap_close_fd_on_exec(int fd)
+{
+#if defined(F_SETFD) && defined(FD_CLOEXEC)
+ /* Protect the fd so that it will not be inherited by child processes */
+ if(fcntl(fd, F_SETFD, FD_CLOEXEC) < 0) {
+ ap_log_error(APLOG_MARK, APLOG_ERR, NULL,
+ "fcntl(%d, F_SETFD, FD_CLOEXEC) failed", fd);
+ return 0;
+ }
+
+ return 1;
+#else
+ return 0;
+#endif
+}
+
+#endif /* ndef(WIN32) */
+
+static void fd_cleanup(void *fdv)
+{
+ close((int) (long) fdv);
+}
+
+static int fd_magic_cleanup(void *fdv)
+{
+ return ap_close_fd_on_exec((int) (long) fdv);
+}
+
+API_EXPORT(void) ap_note_cleanups_for_fd_ex(pool *p, int fd, int domagic)
+{
+#if defined(NETWARE)
+ domagic = 0; /* skip magic for NetWare, at least for now */
+#endif
+ ap_register_cleanup_ex(p, (void *) (long) fd, fd_cleanup, fd_cleanup,
+ domagic ? fd_magic_cleanup : NULL);
+}
+
+API_EXPORT(void) ap_note_cleanups_for_fd(pool *p, int fd)
+{
+ ap_note_cleanups_for_fd_ex(p, fd, 0);
+}
+
+API_EXPORT(void) ap_kill_cleanups_for_fd(pool *p, int fd)
+{
+ ap_kill_cleanup(p, (void *) (long) fd, fd_cleanup);
+}
+
+API_EXPORT(int) ap_popenf_ex(pool *a, const char *name, int flg, int mode,
+ int domagic)
+{
+ int fd;
+ int save_errno;
+
+ ap_block_alarms();
+ fd = open(name, flg, mode);
+ save_errno = errno;
+ if (fd >= 0) {
+ fd = ap_slack(fd, AP_SLACK_HIGH);
+ ap_note_cleanups_for_fd_ex(a, fd, domagic);
+ }
+ ap_unblock_alarms();
+ errno = save_errno;
+ return fd;
+}
+
+API_EXPORT(int) ap_popenf(pool *a, const char *name, int flg, int mode)
+{
+ return ap_popenf_ex(a, name, flg, mode, 0);
+}
+
+API_EXPORT(int) ap_pclosef(pool *a, int fd)
+{
+ int res;
+ int save_errno;
+
+ ap_block_alarms();
+ res = close(fd);
+ save_errno = errno;
+ ap_kill_cleanup(a, (void *) (long) fd, fd_cleanup);
+ ap_unblock_alarms();
+ errno = save_errno;
+ return res;
+}
+
+#ifdef WIN32
+static void h_cleanup(void *nth)
+{
+ CloseHandle((HANDLE) nth);
+}
+
+static int h_magic_cleanup(void *nth)
+{
+ /* Set handle not-inherited
+ */
+ return ap_close_handle_on_exec((HANDLE) nth);
+}
+
+API_EXPORT(void) ap_note_cleanups_for_h_ex(pool *p, HANDLE nth, int domagic)
+{
+ ap_register_cleanup_ex(p, (void *) nth, h_cleanup, h_cleanup,
+ domagic ? h_magic_cleanup : NULL);
+}
+
+API_EXPORT(void) ap_note_cleanups_for_h(pool *p, HANDLE nth)
+{
+ ap_note_cleanups_for_h_ex(p, nth, 0);
+}
+
+API_EXPORT(int) ap_pcloseh(pool *a, HANDLE hDevice)
+{
+ int res=0;
+ int save_errno;
+
+ ap_block_alarms();
+
+ if (!CloseHandle(hDevice)) {
+ res = GetLastError();
+ }
+
+ save_errno = errno;
+ ap_kill_cleanup(a, (void *) hDevice, h_cleanup);
+ ap_unblock_alarms();
+ errno = save_errno;
+ return res;
+}
+#endif
+
+/* Note that we have separate plain_ and child_ cleanups for FILE *s,
+ * since fclose() would flush I/O buffers, which is extremely undesirable;
+ * we just close the descriptor.
+ */
+
+static void file_cleanup(void *fpv)
+{
+ fclose((FILE *) fpv);
+}
+
+static void file_child_cleanup(void *fpv)
+{
+ close(fileno((FILE *) fpv));
+}
+
+static int file_magic_cleanup(void *fpv)
+{
+ return ap_close_fd_on_exec(fileno((FILE *) fpv));
+}
+
+API_EXPORT(void) ap_note_cleanups_for_file_ex(pool *p, FILE *fp, int domagic)
+{
+#if defined(NETWARE)
+ domagic = 0; /* skip magic for NetWare, at least for now */
+#endif
+ ap_register_cleanup_ex(p, (void *) fp, file_cleanup, file_child_cleanup,
+ domagic ? file_magic_cleanup : NULL);
+}
+
+API_EXPORT(void) ap_note_cleanups_for_file(pool *p, FILE *fp)
+{
+ ap_note_cleanups_for_file_ex(p, fp, 0);
+}
+
+API_EXPORT(FILE *) ap_pfopen(pool *a, const char *name, const char *mode)
+{
+ FILE *fd = NULL;
+ int baseFlag, desc;
+ int modeFlags = 0;
+ int saved_errno;
+
+#ifdef WIN32
+ modeFlags = _S_IREAD | _S_IWRITE;
+#else
+ modeFlags = S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP | S_IROTH | S_IWOTH;
+#endif
+
+ ap_block_alarms();
+
+ if (*mode == 'a') {
+ /* Work around faulty implementations of fopen */
+ baseFlag = (*(mode + 1) == '+') ? O_RDWR : O_WRONLY;
+ desc = open(name, baseFlag | O_APPEND | O_CREAT,
+ modeFlags);
+ if (desc >= 0) {
+ desc = ap_slack(desc, AP_SLACK_LOW);
+ fd = ap_fdopen(desc, mode);
+ }
+ }
+ else {
+ fd = fopen(name, mode);
+ }
+ saved_errno = errno;
+ if (fd != NULL)
+ ap_note_cleanups_for_file(a, fd);
+ ap_unblock_alarms();
+ errno = saved_errno;
+ return fd;
+}
+
+API_EXPORT(FILE *) ap_pfdopen(pool *a, int fd, const char *mode)
+{
+ FILE *f;
+ int saved_errno;
+
+ ap_block_alarms();
+ f = ap_fdopen(fd, mode);
+ saved_errno = errno;
+ if (f != NULL)
+ ap_note_cleanups_for_file(a, f);
+ ap_unblock_alarms();
+ errno = saved_errno;
+ return f;
+}
+
+
+API_EXPORT(int) ap_pfclose(pool *a, FILE *fd)
+{
+ int res;
+
+ ap_block_alarms();
+ res = fclose(fd);
+ ap_kill_cleanup(a, (void *) fd, file_cleanup);
+ ap_unblock_alarms();
+ return res;
+}
+
+/*
+ * DIR * with cleanup
+ */
+
+static void dir_cleanup(void *dv)
+{
+ closedir((DIR *) dv);
+}
+
+API_EXPORT(DIR *) ap_popendir(pool *p, const char *name)
+{
+ DIR *d;
+ int save_errno;
+
+ ap_block_alarms();
+ d = opendir(name);
+ if (d == NULL) {
+ save_errno = errno;
+ ap_unblock_alarms();
+ errno = save_errno;
+ return NULL;
+ }
+ ap_register_cleanup(p, (void *) d, dir_cleanup, dir_cleanup);
+ ap_unblock_alarms();
+ return d;
+}
+
+API_EXPORT(void) ap_pclosedir(pool *p, DIR * d)
+{
+ ap_block_alarms();
+ ap_kill_cleanup(p, (void *) d, dir_cleanup);
+ closedir(d);
+ ap_unblock_alarms();
+}
+
+/*****************************************************************
+ *
+ * Files and file descriptors; these are just an application of the
+ * generic cleanup interface.
+ */
+
+static void socket_cleanup(void *fdv)
+{
+ closesocket((int) (long) fdv);
+}
+
+static int socket_magic_cleanup(void *fpv)
+{
+#ifdef WIN32
+ return ap_close_handle_on_exec((HANDLE) fpv);
+#else
+ return ap_close_fd_on_exec((int) (long) fpv);
+#endif
+}
+
+API_EXPORT(void) ap_note_cleanups_for_socket_ex(pool *p, int fd, int domagic)
+{
+#if defined(TPF) || defined(NETWARE)
+ domagic = 0; /* skip magic (fcntl) for TPF sockets, at least for now */
+#endif
+ ap_register_cleanup_ex(p, (void *) (long) fd, socket_cleanup,
+ socket_cleanup,
+ domagic ? socket_magic_cleanup : NULL);
+}
+
+API_EXPORT(void) ap_note_cleanups_for_socket(pool *p, int fd)
+{
+ ap_note_cleanups_for_socket_ex(p, fd, 0);
+}
+
+API_EXPORT(void) ap_kill_cleanups_for_socket(pool *p, int sock)
+{
+ ap_kill_cleanup(p, (void *) (long) sock, socket_cleanup);
+}
+
+API_EXPORT(int) ap_psocket_ex(pool *p, int domain, int type, int protocol,
+ int domagic)
+{
+ int fd;
+
+ ap_block_alarms();
+ fd = socket(domain, type, protocol);
+ if (fd == -1) {
+ int save_errno = errno;
+ ap_unblock_alarms();
+ errno = save_errno;
+ return -1;
+ }
+ ap_note_cleanups_for_socket_ex(p, fd, domagic);
+ ap_unblock_alarms();
+ return fd;
+}
+
+API_EXPORT(int) ap_psocket(pool *p, int domain, int type, int protocol)
+{
+ return ap_psocket_ex(p, domain, type, protocol, 0);
+}
+
+API_EXPORT(int) ap_pclosesocket(pool *a, int sock)
+{
+ int res;
+ int save_errno;
+
+ ap_block_alarms();
+ res = closesocket(sock);
+#if defined(WIN32) || defined(NETWARE)
+ errno = WSAGetLastError();
+#endif /* WIN32 */
+ save_errno = errno;
+ ap_kill_cleanup(a, (void *) (long) sock, socket_cleanup);
+ ap_unblock_alarms();
+ errno = save_errno;
+ return res;
+}
+
+
+/*
+ * Here's a pool-based interface to POSIX regex's regcomp().
+ * Note that we return regex_t instead of being passed one.
+ * The reason is that if you use an already-used regex_t structure,
+ * the memory that you've already allocated gets forgotten, and
+ * regfree() doesn't clear it. So we don't allow it.
+ */
+
+static void regex_cleanup(void *preg)
+{
+ regfree((regex_t *) preg);
+}
+
+API_EXPORT(regex_t *) ap_pregcomp(pool *p, const char *pattern, int cflags)
+{
+ regex_t *preg = ap_palloc(p, sizeof(regex_t));
+
+ if (regcomp(preg, pattern, cflags))
+ return NULL;
+
+ ap_register_cleanup(p, (void *) preg, regex_cleanup, regex_cleanup);
+
+ return preg;
+}
+
+
+API_EXPORT(void) ap_pregfree(pool *p, regex_t * reg)
+{
+ ap_block_alarms();
+ regfree(reg);
+ ap_kill_cleanup(p, (void *) reg, regex_cleanup);
+ ap_unblock_alarms();
+}
+
+/*****************************************************************
+ *
+ * More grotty system stuff... subprocesses. Frump. These don't use
+ * the generic cleanup interface because I don't want multiple
+ * subprocesses to result in multiple three-second pauses; the
+ * subprocesses have to be "freed" all at once. If someone comes
+ * along with another resource they want to allocate which has the
+ * same property, we might want to fold support for that into the
+ * generic interface, but for now, it's a special case
+ */
+
+struct process_chain {
+ pid_t pid;
+ enum kill_conditions kill_how;
+ struct process_chain *next;
+};
+
+API_EXPORT(void) ap_note_subprocess(pool *a, pid_t pid, enum kill_conditions
+how) {
+ struct process_chain *new =
+ (struct process_chain *) ap_palloc(a, sizeof(struct process_chain));
+
+ new->pid = pid;
+ new->kill_how = how;
+ new->next = a->subprocesses;
+ a->subprocesses = new;
+}
+
+#ifdef WIN32
+#define os_pipe(fds) _pipe(fds, 512, O_BINARY | O_NOINHERIT)
+#else
+#define os_pipe(fds) pipe(fds)
+#endif /* WIN32 */
+
+/* for ap_fdopen, to get binary mode */
+#if defined (OS2) || defined (WIN32) || defined (NETWARE)
+#define BINMODE "b"
+#else
+#define BINMODE
+#endif
+
+static pid_t spawn_child_core(pool *p, int (*func) (void *, child_info *),
+ void *data,enum kill_conditions kill_how,
+ int *pipe_in, int *pipe_out, int *pipe_err)
+{
+ pid_t pid;
+ int in_fds[2];
+ int out_fds[2];
+ int err_fds[2];
+ int save_errno;
+
+ if (pipe_in && os_pipe(in_fds) < 0) {
+ return 0;
+ }
+
+ if (pipe_out && os_pipe(out_fds) < 0) {
+ save_errno = errno;
+ if (pipe_in) {
+ close(in_fds[0]);
+ close(in_fds[1]);
+ }
+ errno = save_errno;
+ return 0;
+ }
+
+ if (pipe_err && os_pipe(err_fds) < 0) {
+ save_errno = errno;
+ if (pipe_in) {
+ close(in_fds[0]);
+ close(in_fds[1]);
+ }
+ if (pipe_out) {
+ close(out_fds[0]);
+ close(out_fds[1]);
+ }
+ errno = save_errno;
+ return 0;
+ }
+
+#ifdef WIN32
+
+ {
+ HANDLE thread_handle;
+ int hStdIn, hStdOut, hStdErr;
+ int old_priority;
+ child_info info;
+
+ (void) ap_acquire_mutex(spawn_mutex);
+ thread_handle = GetCurrentThread(); /* doesn't need to be closed */
+ old_priority = GetThreadPriority(thread_handle);
+ SetThreadPriority(thread_handle, THREAD_PRIORITY_HIGHEST);
+ /* Now do the right thing with your pipes */
+ if (pipe_in) {
+ hStdIn = dup(fileno(stdin));
+ if(dup2(in_fds[0], fileno(stdin)))
+ ap_log_error(APLOG_MARK, APLOG_ERR, NULL, "dup2(stdin) failed");
+ close(in_fds[0]);
+ }
+ if (pipe_out) {
+ hStdOut = dup(fileno(stdout));
+ close(fileno(stdout));
+ if(dup2(out_fds[1], fileno(stdout)))
+ ap_log_error(APLOG_MARK, APLOG_ERR, NULL, "dup2(stdout) failed");
+ close(out_fds[1]);
+ }
+ if (pipe_err) {
+ hStdErr = dup(fileno(stderr));
+ if(dup2(err_fds[1], fileno(stderr)))
+ ap_log_error(APLOG_MARK, APLOG_ERR, NULL, "dup2(stderr) failed");
+ close(err_fds[1]);
+ }
+
+ info.hPipeInputRead = GetStdHandle(STD_INPUT_HANDLE);
+ info.hPipeOutputWrite = GetStdHandle(STD_OUTPUT_HANDLE);
+ info.hPipeErrorWrite = GetStdHandle(STD_ERROR_HANDLE);
+
+ pid = (*func) (data, &info);
+ if (pid == -1) pid = 0; /* map Win32 error code onto Unix default */
+
+ if (!pid) {
+ save_errno = errno;
+ close(in_fds[1]);
+ close(out_fds[0]);
+ close(err_fds[0]);
+ }
+
+ /* restore the original stdin, stdout and stderr */
+ if (pipe_in) {
+ dup2(hStdIn, fileno(stdin));
+ close(hStdIn);
+ }
+ if (pipe_out) {
+ dup2(hStdOut, fileno(stdout));
+ close(hStdOut);
+ }
+ if (pipe_err) {
+ dup2(hStdErr, fileno(stderr));
+ close(hStdErr);
+ }
+
+ if (pid) {
+ ap_note_subprocess(p, pid, kill_how);
+ if (pipe_in) {
+ *pipe_in = in_fds[1];
+ }
+ if (pipe_out) {
+ *pipe_out = out_fds[0];
+ }
+ if (pipe_err) {
+ *pipe_err = err_fds[0];
+ }
+ }
+ SetThreadPriority(thread_handle, old_priority);
+ (void) ap_release_mutex(spawn_mutex);
+ /*
+ * go on to the end of the function, where you can
+ * return the pid
+ */
+ if (!pid) {
+ errno = save_errno;
+ }
+ }
+#elif defined(NETWARE)
+ /* NetWare currently has no pipes yet. This will
+ be solved with the new libc for NetWare soon. */
+ pid = 0;
+#elif defined(OS2)
+ {
+ int save_in=-1, save_out=-1, save_err=-1;
+
+ if (pipe_out) {
+ save_out = dup(STDOUT_FILENO);
+ dup2(out_fds[1], STDOUT_FILENO);
+ close(out_fds[1]);
+ DosSetFHState(out_fds[0], OPEN_FLAGS_NOINHERIT);
+ }
+
+ if (pipe_in) {
+ save_in = dup(STDIN_FILENO);
+ dup2(in_fds[0], STDIN_FILENO);
+ close(in_fds[0]);
+ DosSetFHState(in_fds[1], OPEN_FLAGS_NOINHERIT);
+ }
+
+ if (pipe_err) {
+ save_err = dup(STDERR_FILENO);
+ dup2(err_fds[1], STDERR_FILENO);
+ close(err_fds[1]);
+ DosSetFHState(err_fds[0], OPEN_FLAGS_NOINHERIT);
+ }
+
+ pid = func(data, NULL);
+
+ if ( pid )
+ ap_note_subprocess(p, pid, kill_how);
+
+ if (pipe_out) {
+ close(STDOUT_FILENO);
+ dup2(save_out, STDOUT_FILENO);
+ close(save_out);
+ *pipe_out = out_fds[0];
+ }
+
+ if (pipe_in) {
+ close(STDIN_FILENO);
+ dup2(save_in, STDIN_FILENO);
+ close(save_in);
+ *pipe_in = in_fds[1];
+ }
+
+ if (pipe_err) {
+ close(STDERR_FILENO);
+ dup2(save_err, STDERR_FILENO);
+ close(save_err);
+ *pipe_err = err_fds[0];
+ }
+ }
+#elif defined(TPF)
+ return (pid = ap_tpf_spawn_child(p, func, data, kill_how,
+ pipe_in, pipe_out, pipe_err, out_fds, in_fds, err_fds));
+#else
+
+ if ((pid = fork()) < 0) {
+ save_errno = errno;
+ if (pipe_in) {
+ close(in_fds[0]);
+ close(in_fds[1]);
+ }
+ if (pipe_out) {
+ close(out_fds[0]);
+ close(out_fds[1]);
+ }
+ if (pipe_err) {
+ close(err_fds[0]);
+ close(err_fds[1]);
+ }
+ errno = save_errno;
+ return 0;
+ }
+
+ if (!pid) {
+ /* Child process */
+ RAISE_SIGSTOP(SPAWN_CHILD);
+
+ if (pipe_out) {
+ close(out_fds[0]);
+ dup2(out_fds[1], STDOUT_FILENO);
+ close(out_fds[1]);
+ }
+
+ if (pipe_in) {
+ close(in_fds[1]);
+ dup2(in_fds[0], STDIN_FILENO);
+ close(in_fds[0]);
+ }
+
+ if (pipe_err) {
+ close(err_fds[0]);
+ dup2(err_fds[1], STDERR_FILENO);
+ close(err_fds[1]);
+ }
+
+ /* HP-UX SIGCHLD fix goes here, if someone will remind me what it is... */
+ signal(SIGCHLD, SIG_DFL); /* Was that it? */
+
+ func(data, NULL);
+ exit(1); /* Should only get here if the exec in func() failed */
+ }
+
+ /* Parent process */
+
+ ap_note_subprocess(p, pid, kill_how);
+
+ if (pipe_out) {
+ close(out_fds[1]);
+ *pipe_out = out_fds[0];
+ }
+
+ if (pipe_in) {
+ close(in_fds[0]);
+ *pipe_in = in_fds[1];
+ }
+
+ if (pipe_err) {
+ close(err_fds[1]);
+ *pipe_err = err_fds[0];
+ }
+#endif /* WIN32 */
+
+ return pid;
+}
+
+
+API_EXPORT(int) ap_spawn_child(pool *p, int (*func) (void *, child_info *),
+ void *data, enum kill_conditions kill_how,
+ FILE **pipe_in, FILE **pipe_out,
+ FILE **pipe_err)
+{
+ int fd_in, fd_out, fd_err;
+ pid_t pid;
+ int save_errno;
+
+ ap_block_alarms();
+
+ pid = spawn_child_core(p, func, data, kill_how,
+ pipe_in ? &fd_in : NULL,
+ pipe_out ? &fd_out : NULL,
+ pipe_err ? &fd_err : NULL);
+
+ if (pid == 0) {
+ save_errno = errno;
+ ap_unblock_alarms();
+ errno = save_errno;
+ return 0;
+ }
+
+ if (pipe_out) {
+ *pipe_out = ap_fdopen(fd_out, "r" BINMODE);
+ if (*pipe_out)
+ ap_note_cleanups_for_file(p, *pipe_out);
+ else
+ close(fd_out);
+ }
+
+ if (pipe_in) {
+ *pipe_in = ap_fdopen(fd_in, "w" BINMODE);
+ if (*pipe_in)
+ ap_note_cleanups_for_file(p, *pipe_in);
+ else
+ close(fd_in);
+ }
+
+ if (pipe_err) {
+ *pipe_err = ap_fdopen(fd_err, "r" BINMODE);
+ if (*pipe_err)
+ ap_note_cleanups_for_file(p, *pipe_err);
+ else
+ close(fd_err);
+ }
+
+ ap_unblock_alarms();
+ return pid;
+}
+
+API_EXPORT(int) ap_bspawn_child(pool *p, int (*func) (void *, child_info *), void *data,
+ enum kill_conditions kill_how,
+ BUFF **pipe_in, BUFF **pipe_out, BUFF **pipe_err)
+{
+#ifdef WIN32
+ SECURITY_ATTRIBUTES sa = {0};
+ HANDLE hPipeOutputRead = NULL;
+ HANDLE hPipeOutputWrite = NULL;
+ HANDLE hPipeInputRead = NULL;
+ HANDLE hPipeInputWrite = NULL;
+ HANDLE hPipeErrorRead = NULL;
+ HANDLE hPipeErrorWrite = NULL;
+ HANDLE hPipeInputWriteDup = NULL;
+ HANDLE hPipeOutputReadDup = NULL;
+ HANDLE hPipeErrorReadDup = NULL;
+ HANDLE hCurrentProcess;
+ pid_t pid = 0;
+ child_info info;
+
+
+ ap_block_alarms();
+
+ /*
+ * First thing to do is to create the pipes that we will use for stdin, stdout, and
+ * stderr in the child process.
+ */
+ sa.nLength = sizeof(sa);
+ sa.bInheritHandle = TRUE;
+ sa.lpSecurityDescriptor = NULL;
+
+
+ /* Create pipes for standard input/output/error redirection. */
+ if (pipe_in && !CreatePipe(&hPipeInputRead, &hPipeInputWrite, &sa, 0))
+ return 0;
+
+ if (pipe_out && !CreatePipe(&hPipeOutputRead, &hPipeOutputWrite, &sa, 0)) {
+ if(pipe_in) {
+ CloseHandle(hPipeInputRead);
+ CloseHandle(hPipeInputWrite);
+ }
+ return 0;
+ }
+
+ if (pipe_err && !CreatePipe(&hPipeErrorRead, &hPipeErrorWrite, &sa, 0)) {
+ if(pipe_in) {
+ CloseHandle(hPipeInputRead);
+ CloseHandle(hPipeInputWrite);
+ }
+ if(pipe_out) {
+ CloseHandle(hPipeOutputRead);
+ CloseHandle(hPipeOutputWrite);
+ }
+ return 0;
+ }
+ /*
+ * When the pipe handles are created, the security descriptor
+ * indicates that the handle can be inherited. However, we do not
+ * want the server side handles to the pipe to be inherited by the
+ * child CGI process. If the child CGI does inherit the server
+ * side handles, then the child may be left around if the server
+ * closes its handles (e.g. if the http connection is aborted),
+ * because the child will have a valid copy of handles to both
+ * sides of the pipes, and no I/O error will occur. Microsoft
+ * recommends using DuplicateHandle to turn off the inherit bit
+ * under NT and Win95.
+ */
+ hCurrentProcess = GetCurrentProcess();
+ if ((pipe_in && !DuplicateHandle(hCurrentProcess, hPipeInputWrite,
+ hCurrentProcess,
+ &hPipeInputWriteDup, 0, FALSE,
+ DUPLICATE_SAME_ACCESS))
+ || (pipe_out && !DuplicateHandle(hCurrentProcess, hPipeOutputRead,
+ hCurrentProcess, &hPipeOutputReadDup,
+ 0, FALSE, DUPLICATE_SAME_ACCESS))
+ || (pipe_err && !DuplicateHandle(hCurrentProcess, hPipeErrorRead,
+ hCurrentProcess, &hPipeErrorReadDup,
+ 0, FALSE, DUPLICATE_SAME_ACCESS))) {
+ if (pipe_in) {
+ CloseHandle(hPipeInputRead);
+ CloseHandle(hPipeInputWrite);
+ }
+ if (pipe_out) {
+ CloseHandle(hPipeOutputRead);
+ CloseHandle(hPipeOutputWrite);
+ }
+ if (pipe_err) {
+ CloseHandle(hPipeErrorRead);
+ CloseHandle(hPipeErrorWrite);
+ }
+ return 0;
+ }
+ else {
+ if (pipe_in) {
+ CloseHandle(hPipeInputWrite);
+ hPipeInputWrite = hPipeInputWriteDup;
+ }
+ if (pipe_out) {
+ CloseHandle(hPipeOutputRead);
+ hPipeOutputRead = hPipeOutputReadDup;
+ }
+ if (pipe_err) {
+ CloseHandle(hPipeErrorRead);
+ hPipeErrorRead = hPipeErrorReadDup;
+ }
+ }
+
+ /* The script writes stdout to this pipe handle */
+ info.hPipeOutputWrite = hPipeOutputWrite;
+
+ /* The script reads stdin from this pipe handle */
+ info.hPipeInputRead = hPipeInputRead;
+
+ /* The script writes stderr to this pipe handle */
+ info.hPipeErrorWrite = hPipeErrorWrite;
+
+ /*
+ * Try to launch the CGI. Under the covers, this call
+ * will try to pick up the appropriate interpreter if
+ * one is needed.
+ */
+ pid = func(data, &info);
+ if (pid == -1) {
+ /* Things didn't work, so cleanup */
+ pid = 0; /* map Win32 error code onto Unix default */
+ CloseHandle(hPipeOutputRead);
+ CloseHandle(hPipeInputWrite);
+ CloseHandle(hPipeErrorRead);
+ }
+ else {
+ if (pipe_out) {
+ /*
+ * This pipe represents stdout for the script,
+ * so we read from this pipe.
+ */
+ /* Create a read buffer */
+ *pipe_out = ap_bcreate(p, B_RD);
+
+ /* Setup the cleanup routine for the handle */
+ ap_note_cleanups_for_h_ex(p, hPipeOutputRead, 1);
+
+ /* Associate the handle with the new buffer */
+ ap_bpushh(*pipe_out, hPipeOutputRead);
+ }
+
+ if (pipe_in) {
+ /*
+ * This pipe represents stdin for the script, so we
+ * write to this pipe.
+ */
+ /* Create a write buffer */
+ *pipe_in = ap_bcreate(p, B_WR);
+
+ /* Setup the cleanup routine for the handle */
+ ap_note_cleanups_for_h_ex(p, hPipeInputWrite, 1);
+
+ /* Associate the handle with the new buffer */
+ ap_bpushh(*pipe_in, hPipeInputWrite);
+
+ }
+
+ if (pipe_err) {
+ /*
+ * This pipe represents stderr for the script, so
+ * we read from this pipe.
+ */
+ /* Create a read buffer */
+ *pipe_err = ap_bcreate(p, B_RD);
+
+ /* Setup the cleanup routine for the handle */
+ ap_note_cleanups_for_h_ex(p, hPipeErrorRead, 1);
+
+ /* Associate the handle with the new buffer */
+ ap_bpushh(*pipe_err, hPipeErrorRead);
+ }
+ }
+
+
+ /*
+ * Now that handles have been inherited, close them to be safe.
+ * You don't want to read or write to them accidentally, and we
+ * sure don't want to have a handle leak.
+ */
+ CloseHandle(hPipeOutputWrite);
+ CloseHandle(hPipeInputRead);
+ CloseHandle(hPipeErrorWrite);
+
+#else
+ int fd_in, fd_out, fd_err;
+ pid_t pid;
+ int save_errno;
+
+ ap_block_alarms();
+
+ pid = spawn_child_core(p, func, data, kill_how,
+ pipe_in ? &fd_in : NULL,
+ pipe_out ? &fd_out : NULL,
+ pipe_err ? &fd_err : NULL);
+
+ if (pid == 0) {
+ save_errno = errno;
+ ap_unblock_alarms();
+ errno = save_errno;
+ return 0;
+ }
+
+ if (pipe_out) {
+ *pipe_out = ap_bcreate(p, B_RD);
+ ap_note_cleanups_for_fd_ex(p, fd_out, 0);
+ ap_bpushfd(*pipe_out, fd_out, fd_out);
+ }
+
+ if (pipe_in) {
+ *pipe_in = ap_bcreate(p, B_WR);
+ ap_note_cleanups_for_fd_ex(p, fd_in, 0);
+ ap_bpushfd(*pipe_in, fd_in, fd_in);
+ }
+
+ if (pipe_err) {
+ *pipe_err = ap_bcreate(p, B_RD);
+ ap_note_cleanups_for_fd_ex(p, fd_err, 0);
+ ap_bpushfd(*pipe_err, fd_err, fd_err);
+ }
+#endif
+
+ ap_unblock_alarms();
+ return pid;
+}
+
+
+/*
+ * Timing constants for killing subprocesses
+ * There is a total 3-second delay between sending a SIGINT
+ * and sending of the final SIGKILL.
+ * TIMEOUT_INTERVAL should be set to TIMEOUT_USECS / 64
+ * for the exponential timeout algorithm.
+ */
+#define TIMEOUT_USECS 3000000
+#define TIMEOUT_INTERVAL 46875
+
+static void free_proc_chain(struct process_chain *procs)
+{
+ /* Dispose of the subprocesses we've spawned off in the course of
+ * whatever it was we're cleaning up now. This may involve killing
+ * some of them off...
+ */
+ struct process_chain *p;
+ int need_timeout = 0;
+ int status;
+#if !defined(WIN32) && !defined(NETWARE)
+ int timeout_interval;
+ struct timeval tv;
+#endif
+
+ if (procs == NULL)
+ return; /* No work. Whew! */
+
+ /* First, check to see if we need to do the SIGTERM, sleep, SIGKILL
+ * dance with any of the processes we're cleaning up. If we've got
+ * any kill-on-sight subprocesses, ditch them now as well, so they
+ * don't waste any more cycles doing whatever it is that they shouldn't
+ * be doing anymore.
+ */
+#ifdef WIN32
+ /* Pick up all defunct processes */
+ for (p = procs; p; p = p->next) {
+ if (GetExitCodeProcess((HANDLE) p->pid, &status)) {
+ p->kill_how = kill_never;
+ }
+ }
+
+
+ for (p = procs; p; p = p->next) {
+ if (p->kill_how == kill_after_timeout) {
+ need_timeout = 1;
+ }
+ else if (p->kill_how == kill_always) {
+ TerminateProcess((HANDLE) p->pid, 1);
+ }
+ }
+ /* Sleep only if we have to... */
+
+ if (need_timeout)
+ sleep(3);
+
+ /* OK, the scripts we just timed out for have had a chance to clean up
+ * --- now, just get rid of them, and also clean up the system accounting
+ * goop...
+ */
+
+ for (p = procs; p; p = p->next) {
+ if (p->kill_how == kill_after_timeout)
+ TerminateProcess((HANDLE) p->pid, 1);
+ }
+
+ for (p = procs; p; p = p->next) {
+ CloseHandle((HANDLE) p->pid);
+ }
+#elif defined(NETWARE)
+#else
+#ifndef NEED_WAITPID
+ /* Pick up all defunct processes */
+ for (p = procs; p; p = p->next) {
+ if (waitpid(p->pid, (int *) 0, WNOHANG) > 0) {
+ p->kill_how = kill_never;
+ }
+ }
+#endif
+
+ for (p = procs; p; p = p->next) {
+ if ((p->kill_how == kill_after_timeout)
+ || (p->kill_how == kill_only_once)) {
+ /*
+ * This is totally bogus, but seems to be the
+ * only portable (as in reliable) way to accomplish
+ * this. Note that this implies an unavoidable
+ * delay.
+ */
+ ap_os_kill(p->pid, SIGTERM);
+ need_timeout = 1;
+ }
+ else if (p->kill_how == kill_always) {
+ kill(p->pid, SIGKILL);
+ }
+ }
+
+ /* Sleep only if we have to. The sleep algorithm grows
+ * by a factor of two on each iteration. TIMEOUT_INTERVAL
+ * is equal to TIMEOUT_USECS / 64.
+ */
+ if (need_timeout) {
+ timeout_interval = TIMEOUT_INTERVAL;
+ tv.tv_sec = 0;
+ tv.tv_usec = timeout_interval;
+ ap_select(0, NULL, NULL, NULL, &tv);
+
+ do {
+ need_timeout = 0;
+ for (p = procs; p; p = p->next) {
+ if (p->kill_how == kill_after_timeout) {
+ if (waitpid(p->pid, (int *) 0, WNOHANG | WUNTRACED) > 0)
+ p->kill_how = kill_never;
+ else
+ need_timeout = 1;
+ }
+ }
+ if (need_timeout) {
+ if (timeout_interval >= TIMEOUT_USECS) {
+ break;
+ }
+ tv.tv_sec = timeout_interval / 1000000;
+ tv.tv_usec = timeout_interval % 1000000;
+ ap_select(0, NULL, NULL, NULL, &tv);
+ timeout_interval *= 2;
+ }
+ } while (need_timeout);
+ }
+
+ /* OK, the scripts we just timed out for have had a chance to clean up
+ * --- now, just get rid of them, and also clean up the system accounting
+ * goop...
+ */
+
+ for (p = procs; p; p = p->next) {
+ if (p->kill_how == kill_after_timeout)
+ kill(p->pid, SIGKILL);
+
+ if (p->kill_how != kill_never)
+ waitpid(p->pid, &status, 0);
+ }
+#endif /* !WIN32 && !NETWARE*/
+}