version 1.111.2.8, 2007/10/27 11:35:38 |
version 1.152, 2008/02/27 19:55:59 |
Line 251 sokvafree(vaddr_t sva, vsize_t len) |
|
Line 251 sokvafree(vaddr_t sva, vsize_t len) |
|
} |
} |
|
|
static void |
static void |
sodoloanfree(struct vm_page **pgs, void *buf, size_t size, bool mapped) |
sodoloanfree(struct vm_page **pgs, void *buf, size_t size) |
{ |
{ |
vaddr_t sva, eva; |
vaddr_t va, sva, eva; |
vsize_t len; |
vsize_t len; |
int npgs; |
paddr_t pa; |
|
int i, npgs; |
KASSERT(pgs != NULL); |
|
|
|
eva = round_page((vaddr_t) buf + size); |
eva = round_page((vaddr_t) buf + size); |
sva = trunc_page((vaddr_t) buf); |
sva = trunc_page((vaddr_t) buf); |
len = eva - sva; |
len = eva - sva; |
npgs = len >> PAGE_SHIFT; |
npgs = len >> PAGE_SHIFT; |
|
|
if (mapped) { |
if (__predict_false(pgs == NULL)) { |
pmap_kremove(sva, len); |
pgs = alloca(npgs * sizeof(*pgs)); |
pmap_update(pmap_kernel()); |
|
|
for (i = 0, va = sva; va < eva; i++, va += PAGE_SIZE) { |
|
if (pmap_extract(pmap_kernel(), va, &pa) == false) |
|
panic("sodoloanfree: va 0x%lx not mapped", va); |
|
pgs[i] = PHYS_TO_VM_PAGE(pa); |
|
} |
} |
} |
|
|
|
pmap_kremove(sva, len); |
|
pmap_update(pmap_kernel()); |
uvm_unloan(pgs, npgs, UVM_LOAN_TOPAGE); |
uvm_unloan(pgs, npgs, UVM_LOAN_TOPAGE); |
sokvafree(sva, len); |
sokvafree(sva, len); |
} |
} |
|
|
static size_t |
static size_t |
sodopendfree() |
sodopendfree(void) |
{ |
{ |
size_t rv; |
size_t rv; |
|
|
|
|
*/ |
*/ |
|
|
static size_t |
static size_t |
sodopendfreel() |
sodopendfreel(void) |
{ |
{ |
struct mbuf *m, *next; |
struct mbuf *m, *next; |
size_t rv = 0; |
size_t rv = 0; |
int s; |
|
|
|
KASSERT(mutex_owned(&so_pendfree_lock)); |
KASSERT(mutex_owned(&so_pendfree_lock)); |
|
|
|
|
|
|
for (; m != NULL; m = next) { |
for (; m != NULL; m = next) { |
next = m->m_next; |
next = m->m_next; |
KASSERT((~m->m_flags & (M_EXT|M_EXT_PAGES)) == 0); |
|
KASSERT(m->m_ext.ext_refcnt == 0); |
|
|
|
rv += m->m_ext.ext_size; |
rv += m->m_ext.ext_size; |
sodoloanfree(m->m_ext.ext_pgs, m->m_ext.ext_buf, |
sodoloanfree((m->m_flags & M_EXT_PAGES) ? |
m->m_ext.ext_size, |
m->m_ext.ext_pgs : NULL, m->m_ext.ext_buf, |
(m->m_ext.ext_flags & M_EXT_LAZY) == 0); |
m->m_ext.ext_size); |
s = splvm(); |
pool_cache_put(mb_cache, m); |
pool_cache_put(&mbpool_cache, m); |
|
splx(s); |
|
} |
} |
|
|
mutex_enter(&so_pendfree_lock); |
mutex_enter(&so_pendfree_lock); |
|
|
soloanfree(struct mbuf *m, void *buf, size_t size, void *arg) |
soloanfree(struct mbuf *m, void *buf, size_t size, void *arg) |
{ |
{ |
|
|
KASSERT(m != NULL); |
if (m == NULL) { |
|
|
|
/* |
|
* called from MEXTREMOVE. |
|
*/ |
|
|
|
sodoloanfree(NULL, buf, size); |
|
return; |
|
} |
|
|
/* |
/* |
* postpone freeing mbuf. |
* postpone freeing mbuf. |
Line 351 sosend_loan(struct socket *so, struct ui |
|
Line 361 sosend_loan(struct socket *so, struct ui |
|
struct iovec *iov = uio->uio_iov; |
struct iovec *iov = uio->uio_iov; |
vaddr_t sva, eva; |
vaddr_t sva, eva; |
vsize_t len; |
vsize_t len; |
vaddr_t lva; |
vaddr_t lva, va; |
int npgs, error; |
int npgs, i, error; |
#if !defined(__HAVE_LAZY_MBUF) |
|
vaddr_t va; |
|
int i; |
|
#endif /* !defined(__HAVE_LAZY_MBUF) */ |
|
|
|
if (VMSPACE_IS_KERNEL_P(uio->uio_vmspace)) |
if (VMSPACE_IS_KERNEL_P(uio->uio_vmspace)) |
return (0); |
return (0); |
Line 385 sosend_loan(struct socket *so, struct ui |
|
Line 391 sosend_loan(struct socket *so, struct ui |
|
return (0); |
return (0); |
} |
} |
|
|
#if !defined(__HAVE_LAZY_MBUF) |
|
for (i = 0, va = lva; i < npgs; i++, va += PAGE_SIZE) |
for (i = 0, va = lva; i < npgs; i++, va += PAGE_SIZE) |
pmap_kenter_pa(va, VM_PAGE_TO_PHYS(m->m_ext.ext_pgs[i]), |
pmap_kenter_pa(va, VM_PAGE_TO_PHYS(m->m_ext.ext_pgs[i]), |
VM_PROT_READ); |
VM_PROT_READ); |
pmap_update(pmap_kernel()); |
pmap_update(pmap_kernel()); |
#endif /* !defined(__HAVE_LAZY_MBUF) */ |
|
|
|
lva += (vaddr_t) iov->iov_base & PAGE_MASK; |
lva += (vaddr_t) iov->iov_base & PAGE_MASK; |
|
|
MEXTADD(m, (void *) lva, space, M_MBUF, soloanfree, so); |
MEXTADD(m, (void *) lva, space, M_MBUF, soloanfree, so); |
m->m_flags |= M_EXT_PAGES | M_EXT_ROMAP; |
m->m_flags |= M_EXT_PAGES | M_EXT_ROMAP; |
|
|
#if defined(__HAVE_LAZY_MBUF) |
|
m->m_flags |= M_EXT_LAZY; |
|
m->m_ext.ext_flags |= M_EXT_LAZY; |
|
#endif /* defined(__HAVE_LAZY_MBUF) */ |
|
|
|
uio->uio_resid -= space; |
uio->uio_resid -= space; |
/* uio_offset not updated, not set/used for write(2) */ |
/* uio_offset not updated, not set/used for write(2) */ |
uio->uio_iov->iov_base = (char *)uio->uio_iov->iov_base + space; |
uio->uio_iov->iov_base = (char *)uio->uio_iov->iov_base + space; |
Line 429 sokva_reclaim_callback(struct callback_e |
|
Line 428 sokva_reclaim_callback(struct callback_e |
|
} |
} |
|
|
struct mbuf * |
struct mbuf * |
getsombuf(struct socket *so) |
getsombuf(struct socket *so, int type) |
{ |
{ |
struct mbuf *m; |
struct mbuf *m; |
|
|
m = m_get(M_WAIT, MT_SONAME); |
m = m_get(M_WAIT, type); |
MCLAIM(m, so->so_mowner); |
MCLAIM(m, so->so_mowner); |
return m; |
return m; |
} |
} |
Line 443 m_intopt(struct socket *so, int val) |
|
Line 442 m_intopt(struct socket *so, int val) |
|
{ |
{ |
struct mbuf *m; |
struct mbuf *m; |
|
|
m = getsombuf(so); |
m = getsombuf(so, MT_SOOPTS); |
m->m_len = sizeof(int); |
m->m_len = sizeof(int); |
*mtod(m, int *) = val; |
*mtod(m, int *) = val; |
return m; |
return m; |
|
|
soinit(void) |
soinit(void) |
{ |
{ |
|
|
mutex_init(&so_pendfree_lock, MUTEX_DRIVER, IPL_VM); |
mutex_init(&so_pendfree_lock, MUTEX_DEFAULT, IPL_VM); |
cv_init(&socurkva_cv, "sokva"); |
cv_init(&socurkva_cv, "sokva"); |
|
|
/* Set the initial adjusted socket buffer size. */ |
/* Set the initial adjusted socket buffer size. */ |
Line 581 sobind(struct socket *so, struct mbuf *n |
|
Line 580 sobind(struct socket *so, struct mbuf *n |
|
} |
} |
|
|
int |
int |
solisten(struct socket *so, int backlog) |
solisten(struct socket *so, int backlog, struct lwp *l) |
{ |
{ |
int s, error; |
int s, error; |
|
|
s = splsoftnet(); |
s = splsoftnet(); |
error = (*so->so_proto->pr_usrreq)(so, PRU_LISTEN, NULL, |
error = (*so->so_proto->pr_usrreq)(so, PRU_LISTEN, NULL, |
NULL, NULL, NULL); |
NULL, NULL, l); |
if (error != 0) { |
if (error != 0) { |
splx(s); |
splx(s); |
return error; |
return error; |
Line 661 soclose(struct socket *so) |
|
Line 660 soclose(struct socket *so) |
|
goto drop; |
goto drop; |
} |
} |
if (so->so_options & SO_LINGER) { |
if (so->so_options & SO_LINGER) { |
if ((so->so_state & SS_ISDISCONNECTING) && |
if ((so->so_state & SS_ISDISCONNECTING) && so->so_nbio) |
(so->so_state & SS_NBIO)) |
|
goto drop; |
goto drop; |
while (so->so_state & SS_ISCONNECTED) { |
while (so->so_state & SS_ISCONNECTED) { |
error = tsleep((void *)&so->so_timeo, |
error = tsleep((void *)&so->so_timeo, |
Line 871 sosend(struct socket *so, struct mbuf *a |
|
Line 869 sosend(struct socket *so, struct mbuf *a |
|
snderr(EMSGSIZE); |
snderr(EMSGSIZE); |
if (space < resid + clen && |
if (space < resid + clen && |
(atomic || space < so->so_snd.sb_lowat || space < clen)) { |
(atomic || space < so->so_snd.sb_lowat || space < clen)) { |
if (so->so_state & SS_NBIO) |
if (so->so_nbio) |
snderr(EWOULDBLOCK); |
snderr(EWOULDBLOCK); |
sbunlock(&so->so_snd); |
sbunlock(&so->so_snd); |
error = sbwait(&so->so_snd); |
error = sbwait(&so->so_snd); |
Line 1011 soreceive(struct socket *so, struct mbuf |
|
Line 1009 soreceive(struct socket *so, struct mbuf |
|
{ |
{ |
struct lwp *l = curlwp; |
struct lwp *l = curlwp; |
struct mbuf *m, **mp; |
struct mbuf *m, **mp; |
int flags, len, error, s, offset, moff, type, orig_resid; |
int atomic, flags, len, error, s, offset, moff, type, orig_resid; |
const struct protosw *pr; |
const struct protosw *pr; |
struct mbuf *nextrecord; |
struct mbuf *nextrecord; |
int mbuf_removed = 0; |
int mbuf_removed = 0; |
|
const struct domain *dom; |
|
|
pr = so->so_proto; |
pr = so->so_proto; |
|
atomic = pr->pr_flags & PR_ATOMIC; |
|
dom = pr->pr_domain; |
mp = mp0; |
mp = mp0; |
type = 0; |
type = 0; |
orig_resid = uio->uio_resid; |
orig_resid = uio->uio_resid; |
Line 1077 soreceive(struct socket *so, struct mbuf |
|
Line 1078 soreceive(struct socket *so, struct mbuf |
|
(so->so_rcv.sb_cc < so->so_rcv.sb_lowat || |
(so->so_rcv.sb_cc < so->so_rcv.sb_lowat || |
((flags & MSG_WAITALL) && |
((flags & MSG_WAITALL) && |
uio->uio_resid <= so->so_rcv.sb_hiwat)) && |
uio->uio_resid <= so->so_rcv.sb_hiwat)) && |
m->m_nextpkt == NULL && |
m->m_nextpkt == NULL && !atomic)) { |
(pr->pr_flags & PR_ATOMIC) == 0)) { |
|
#ifdef DIAGNOSTIC |
#ifdef DIAGNOSTIC |
if (m == NULL && so->so_rcv.sb_cc) |
if (m == NULL && so->so_rcv.sb_cc) |
panic("receive 1"); |
panic("receive 1"); |
Line 1109 soreceive(struct socket *so, struct mbuf |
|
Line 1109 soreceive(struct socket *so, struct mbuf |
|
} |
} |
if (uio->uio_resid == 0) |
if (uio->uio_resid == 0) |
goto release; |
goto release; |
if ((so->so_state & SS_NBIO) || (flags & MSG_DONTWAIT)) { |
if (so->so_nbio || (flags & MSG_DONTWAIT)) { |
error = EWOULDBLOCK; |
error = EWOULDBLOCK; |
goto release; |
goto release; |
} |
} |
Line 1167 soreceive(struct socket *so, struct mbuf |
|
Line 1167 soreceive(struct socket *so, struct mbuf |
|
sbfree(&so->so_rcv, m); |
sbfree(&so->so_rcv, m); |
mbuf_removed = 1; |
mbuf_removed = 1; |
if (controlp != NULL) { |
if (controlp != NULL) { |
struct domain *dom = pr->pr_domain; |
|
if (dom->dom_externalize && l && |
if (dom->dom_externalize && l && |
mtod(m, struct cmsghdr *)->cmsg_type == |
mtod(m, struct cmsghdr *)->cmsg_type == |
SCM_RIGHTS) |
SCM_RIGHTS) |
Line 1181 soreceive(struct socket *so, struct mbuf |
|
Line 1180 soreceive(struct socket *so, struct mbuf |
|
* Dispose of any SCM_RIGHTS message that went |
* Dispose of any SCM_RIGHTS message that went |
* through the read path rather than recv. |
* through the read path rather than recv. |
*/ |
*/ |
if (pr->pr_domain->dom_dispose && |
if (dom->dom_dispose && |
mtod(m, struct cmsghdr *)->cmsg_type == SCM_RIGHTS) |
mtod(m, struct cmsghdr *)->cmsg_type == SCM_RIGHTS) |
(*pr->pr_domain->dom_dispose)(m); |
(*dom->dom_dispose)(m); |
MFREE(m, so->so_rcv.sb_mb); |
MFREE(m, so->so_rcv.sb_mb); |
m = so->so_rcv.sb_mb; |
m = so->so_rcv.sb_mb; |
} |
} |
Line 1270 soreceive(struct socket *so, struct mbuf |
|
Line 1269 soreceive(struct socket *so, struct mbuf |
|
* This avoids a later panic("receive 1a") |
* This avoids a later panic("receive 1a") |
* when compiled with DIAGNOSTIC. |
* when compiled with DIAGNOSTIC. |
*/ |
*/ |
if (m && mbuf_removed |
if (m && mbuf_removed && atomic) |
&& (pr->pr_flags & PR_ATOMIC)) |
|
(void) sbdroprecord(&so->so_rcv); |
(void) sbdroprecord(&so->so_rcv); |
|
|
goto release; |
goto release; |
Line 1374 soreceive(struct socket *so, struct mbuf |
|
Line 1372 soreceive(struct socket *so, struct mbuf |
|
} |
} |
} |
} |
|
|
if (m && pr->pr_flags & PR_ATOMIC) { |
if (m && atomic) { |
flags |= MSG_TRUNC; |
flags |= MSG_TRUNC; |
if ((flags & MSG_PEEK) == 0) |
if ((flags & MSG_PEEK) == 0) |
(void) sbdroprecord(&so->so_rcv); |
(void) sbdroprecord(&so->so_rcv); |
Line 1796 soo_kqfilter(struct file *fp, struct kno |
|
Line 1794 soo_kqfilter(struct file *fp, struct kno |
|
sb = &so->so_snd; |
sb = &so->so_snd; |
break; |
break; |
default: |
default: |
return (1); |
return (EINVAL); |
} |
} |
SLIST_INSERT_HEAD(&sb->sb_sel.sel_klist, kn, kn_selnext); |
SLIST_INSERT_HEAD(&sb->sb_sel.sel_klist, kn, kn_selnext); |
sb->sb_flags |= SB_KNOTE; |
sb->sb_flags |= SB_KNOTE; |