Please note that diffs are not public domain; they are subject to the copyright notices on the relevant files. =================================================================== RCS file: /ftp/cvs/cvsroot/src/sys/kern/subr_pool.c,v rcsdiff: /ftp/cvs/cvsroot/src/sys/kern/subr_pool.c,v: warning: Unknown phrases like `commitid ...;' are present. retrieving revision 1.60.2.4 retrieving revision 1.68 diff -u -p -r1.60.2.4 -r1.68 --- src/sys/kern/subr_pool.c 2002/10/10 18:43:14 1.60.2.4 +++ src/sys/kern/subr_pool.c 2002/03/08 21:41:59 1.68 @@ -1,4 +1,4 @@ -/* $NetBSD: subr_pool.c,v 1.60.2.4 2002/10/10 18:43:14 jdolecek Exp $ */ +/* $NetBSD: subr_pool.c,v 1.68 2002/03/08 21:41:59 thorpej Exp $ */ /*- * Copyright (c) 1997, 1999, 2000 The NetBSD Foundation, Inc. @@ -38,7 +38,7 @@ */ #include -__KERNEL_RCSID(0, "$NetBSD: subr_pool.c,v 1.60.2.4 2002/10/10 18:43:14 jdolecek Exp $"); +__KERNEL_RCSID(0, "$NetBSD: subr_pool.c,v 1.68 2002/03/08 21:41:59 thorpej Exp $"); #include "opt_pool.h" #include "opt_poollog.h" @@ -94,7 +94,7 @@ struct pool_item_header { TAILQ_HEAD(,pool_item) ph_itemlist; /* chunk list for this page */ LIST_ENTRY(pool_item_header) ph_hashlist; /* Off-page page headers */ - unsigned int ph_nmissing; /* # of chunks in use */ + int ph_nmissing; /* # of chunks in use */ caddr_t ph_page; /* this page's address */ struct timeval ph_time; /* last referenced */ }; @@ -145,6 +145,16 @@ struct pool_item { /* The cache group pool. */ static struct pool pcgpool; +/* The pool cache group. */ +#define PCG_NOBJECTS 16 +struct pool_cache_group { + TAILQ_ENTRY(pool_cache_group) + pcg_list; /* link in the pool cache's group list */ + u_int pcg_avail; /* # available objects */ + /* pointers to the objects */ + void *pcg_objects[PCG_NOBJECTS]; +}; + static void pool_cache_reclaim(struct pool_cache *); static int pool_catchup(struct pool *); @@ -415,7 +425,7 @@ pool_init(struct pool *pp, size_t size, if (size < sizeof(struct pool_item)) size = sizeof(struct pool_item); - size = roundup(size, align); + size = ALIGN(size); #ifdef DIAGNOSTIC if (size > palloc->pa_pagesz) panic("pool_init: pool item size (%lu) too large", @@ -564,14 +574,15 @@ pool_destroy(struct pool *pp) #ifdef DIAGNOSTIC if (pp->pr_nout != 0) { pr_printlog(pp, NULL, printf); - panic("pool_destroy: pool busy: still out: %u", + panic("pool_destroy: pool busy: still out: %u\n", pp->pr_nout); } #endif /* Remove all pages */ - while ((ph = TAILQ_FIRST(&pp->pr_pagelist)) != NULL) - pr_rmpage(pp, ph, NULL); + if ((pp->pr_roflags & PR_STATIC) == 0) + while ((ph = TAILQ_FIRST(&pp->pr_pagelist)) != NULL) + pr_rmpage(pp, ph, NULL); /* Remove from global pool list */ simple_lock(&pool_head_slock); @@ -634,9 +645,15 @@ pool_get(struct pool *pp, int flags) void *v; #ifdef DIAGNOSTIC + if (__predict_false((pp->pr_roflags & PR_STATIC) && + (flags & PR_MALLOCOK))) { + pr_printlog(pp, NULL, printf); + panic("pool_get: static"); + } + if (__predict_false(curproc == NULL && doing_shutdown == 0 && (flags & PR_WAITOK) != 0)) - panic("pool_get: %s: must have NOWAIT", pp->pr_wchan); + panic("pool_get: must have NOWAIT"); #ifdef LOCKDEBUG if (flags & PR_WAITOK) @@ -715,7 +732,7 @@ pool_get(struct pool *pp, int flags) simple_unlock(&pp->pr_slock); printf("pool_get: %s: curpage NULL, nitems %u\n", pp->pr_wchan, pp->pr_nitems); - panic("pool_get: nitems inconsistent"); + panic("pool_get: nitems inconsistent\n"); } #endif @@ -785,7 +802,7 @@ pool_get(struct pool *pp, int flags) simple_unlock(&pp->pr_slock); printf("pool_get: %s: items on itemlist, nitems %u\n", pp->pr_wchan, pp->pr_nitems); - panic("pool_get: nitems inconsistent"); + panic("pool_get: nitems inconsistent\n"); } #endif @@ -916,7 +933,6 @@ pool_do_put(struct pool *pp, void *v) #endif TAILQ_INSERT_HEAD(&ph->ph_itemlist, pi, pi_list); - KDASSERT(ph->ph_nmissing != 0); ph->ph_nmissing--; pp->pr_nput++; pp->pr_nitems++; @@ -950,8 +966,7 @@ pool_do_put(struct pool *pp, void *v) */ if (ph->ph_nmissing == 0) { pp->pr_nidle++; - if (pp->pr_npages > pp->pr_maxpages || - (pp->pr_alloc->pa_flags & PA_WANT) != 0) { + if (pp->pr_npages > pp->pr_maxpages) { pr_rmpage(pp, ph, NULL); } else { TAILQ_REMOVE(&pp->pr_pagelist, ph, ph_pagelist); @@ -1039,7 +1054,7 @@ pool_prime(struct pool *pp, int n) { struct pool_item_header *ph; caddr_t cp; - int newpages; + int newpages, error = 0; simple_lock(&pp->pr_slock); @@ -1053,6 +1068,7 @@ pool_prime(struct pool *pp, int n) simple_lock(&pp->pr_slock); if (__predict_false(cp == NULL || ph == NULL)) { + error = ENOMEM; if (cp != NULL) pool_allocator_free(pp, cp); break; @@ -1126,8 +1142,6 @@ pool_prime_page(struct pool *pp, caddr_t while (n--) { pi = (struct pool_item *)cp; - KASSERT(((((vaddr_t)pi) + ioff) & (align - 1)) == 0); - /* Insert on page list */ TAILQ_INSERT_TAIL(&ph->ph_itemlist, pi, pi_list); #ifdef DIAGNOSTIC @@ -1152,7 +1166,9 @@ pool_prime_page(struct pool *pp, caddr_t * * Note 1, we never wait for memory here, we let the caller decide what to do. * - * Note 2, we must be called with the pool already locked, and we return + * Note 2, this doesn't work with static pools. + * + * Note 3, we must be called with the pool already locked, and we return * with it locked. */ static int @@ -1162,6 +1178,18 @@ pool_catchup(struct pool *pp) caddr_t cp; int error = 0; + if (pp->pr_roflags & PR_STATIC) { + /* + * We dropped below the low water mark, and this is not a + * good thing. Log a warning. + * + * XXX: rate-limit this? + */ + printf("WARNING: static pool `%s' dropped below low water " + "mark\n", pp->pr_wchan); + return (0); + } + while (POOL_NEEDS_CATCHUP(pp)) { /* * Call the page back-end allocator for more memory. @@ -1190,6 +1218,7 @@ pool_catchup(struct pool *pp) void pool_setlowat(struct pool *pp, int n) { + int error; simple_lock(&pp->pr_slock); @@ -1199,7 +1228,7 @@ pool_setlowat(struct pool *pp, int n) : roundup(n, pp->pr_itemsperpage) / pp->pr_itemsperpage; /* Make sure we're caught up with the newly-set low water mark. */ - if (POOL_NEEDS_CATCHUP(pp) && pool_catchup(pp) != 0) { + if (POOL_NEEDS_CATCHUP(pp) && (error = pool_catchup(pp) != 0)) { /* * XXX: Should we log a warning? Should we set up a timeout * to try again in a second or so? The latter could break @@ -1262,6 +1291,9 @@ pool_reclaim(struct pool *pp) struct pool_pagelist pq; int s; + if (pp->pr_roflags & PR_STATIC) + return (0); + if (pp->pr_drain_hook != NULL) { /* * The drain hook must be called with the pool unlocked. @@ -1945,9 +1977,9 @@ pool_allocator_alloc(struct pool *org, i TAILQ_INSERT_TAIL(&pa->pa_list, pp, pr_alloc_list); if (pp == org) continue; - simple_unlock(&pa->pa_slock); + simple_unlock(&pa->pa_list); freed = pool_reclaim(pp); - simple_lock(&pa->pa_slock); + simple_lock(&pa->pa_list); } while ((pp = TAILQ_FIRST(&pa->pa_list)) != start && freed == 0); @@ -1988,7 +2020,6 @@ pool_allocator_free(struct pool *pp, voi pp->pr_flags &= ~PR_WANTED; wakeup(pp); } - simple_unlock(&pp->pr_slock); } pa->pa_flags &= ~PA_WANT; simple_unlock(&pa->pa_slock);