version 1.111.2.13, 2008/02/27 09:24:06 |
version 1.146, 2007/11/24 07:46:11 |
Line 253 sokvafree(vaddr_t sva, vsize_t len) |
|
Line 253 sokvafree(vaddr_t sva, vsize_t len) |
|
static void |
static void |
sodoloanfree(struct vm_page **pgs, void *buf, size_t size) |
sodoloanfree(struct vm_page **pgs, void *buf, size_t size) |
{ |
{ |
vaddr_t sva, eva; |
vaddr_t va, sva, eva; |
vsize_t len; |
vsize_t len; |
int npgs; |
paddr_t pa; |
|
int i, npgs; |
KASSERT(pgs != NULL); |
|
|
|
eva = round_page((vaddr_t) buf + size); |
eva = round_page((vaddr_t) buf + size); |
sva = trunc_page((vaddr_t) buf); |
sva = trunc_page((vaddr_t) buf); |
len = eva - sva; |
len = eva - sva; |
npgs = len >> PAGE_SHIFT; |
npgs = len >> PAGE_SHIFT; |
|
|
|
if (__predict_false(pgs == NULL)) { |
|
pgs = alloca(npgs * sizeof(*pgs)); |
|
|
|
for (i = 0, va = sva; va < eva; i++, va += PAGE_SIZE) { |
|
if (pmap_extract(pmap_kernel(), va, &pa) == false) |
|
panic("sodoloanfree: va 0x%lx not mapped", va); |
|
pgs[i] = PHYS_TO_VM_PAGE(pa); |
|
} |
|
} |
|
|
pmap_kremove(sva, len); |
pmap_kremove(sva, len); |
pmap_update(pmap_kernel()); |
pmap_update(pmap_kernel()); |
uvm_unloan(pgs, npgs, UVM_LOAN_TOPAGE); |
uvm_unloan(pgs, npgs, UVM_LOAN_TOPAGE); |
|
|
|
|
for (; m != NULL; m = next) { |
for (; m != NULL; m = next) { |
next = m->m_next; |
next = m->m_next; |
KASSERT((~m->m_flags & (M_EXT|M_EXT_PAGES)) == 0); |
|
KASSERT(m->m_ext.ext_refcnt == 0); |
|
|
|
rv += m->m_ext.ext_size; |
rv += m->m_ext.ext_size; |
sodoloanfree(m->m_ext.ext_pgs, m->m_ext.ext_buf, |
sodoloanfree((m->m_flags & M_EXT_PAGES) ? |
|
m->m_ext.ext_pgs : NULL, m->m_ext.ext_buf, |
m->m_ext.ext_size); |
m->m_ext.ext_size); |
pool_cache_put(mb_cache, m); |
pool_cache_put(mb_cache, m); |
} |
} |
|
|
soloanfree(struct mbuf *m, void *buf, size_t size, void *arg) |
soloanfree(struct mbuf *m, void *buf, size_t size, void *arg) |
{ |
{ |
|
|
KASSERT(m != NULL); |
if (m == NULL) { |
|
|
|
/* |
|
* called from MEXTREMOVE. |
|
*/ |
|
|
|
sodoloanfree(NULL, buf, size); |
|
return; |
|
} |
|
|
/* |
/* |
* postpone freeing mbuf. |
* postpone freeing mbuf. |
Line 345 sosend_loan(struct socket *so, struct ui |
|
Line 361 sosend_loan(struct socket *so, struct ui |
|
struct iovec *iov = uio->uio_iov; |
struct iovec *iov = uio->uio_iov; |
vaddr_t sva, eva; |
vaddr_t sva, eva; |
vsize_t len; |
vsize_t len; |
vaddr_t lva; |
vaddr_t lva, va; |
int npgs, error; |
int npgs, i, error; |
vaddr_t va; |
|
int i; |
|
|
|
if (VMSPACE_IS_KERNEL_P(uio->uio_vmspace)) |
if (VMSPACE_IS_KERNEL_P(uio->uio_vmspace)) |
return (0); |
return (0); |
Line 414 sokva_reclaim_callback(struct callback_e |
|
Line 428 sokva_reclaim_callback(struct callback_e |
|
} |
} |
|
|
struct mbuf * |
struct mbuf * |
getsombuf(struct socket *so, int type) |
getsombuf(struct socket *so) |
{ |
{ |
struct mbuf *m; |
struct mbuf *m; |
|
|
m = m_get(M_WAIT, type); |
m = m_get(M_WAIT, MT_SONAME); |
MCLAIM(m, so->so_mowner); |
MCLAIM(m, so->so_mowner); |
return m; |
return m; |
} |
} |
Line 428 m_intopt(struct socket *so, int val) |
|
Line 442 m_intopt(struct socket *so, int val) |
|
{ |
{ |
struct mbuf *m; |
struct mbuf *m; |
|
|
m = getsombuf(so, MT_SOOPTS); |
m = getsombuf(so); |
m->m_len = sizeof(int); |
m->m_len = sizeof(int); |
*mtod(m, int *) = val; |
*mtod(m, int *) = val; |
return m; |
return m; |
|
|
soinit(void) |
soinit(void) |
{ |
{ |
|
|
mutex_init(&so_pendfree_lock, MUTEX_DEFAULT, IPL_VM); |
mutex_init(&so_pendfree_lock, MUTEX_DRIVER, IPL_VM); |
cv_init(&socurkva_cv, "sokva"); |
cv_init(&socurkva_cv, "sokva"); |
|
|
/* Set the initial adjusted socket buffer size. */ |
/* Set the initial adjusted socket buffer size. */ |
Line 566 sobind(struct socket *so, struct mbuf *n |
|
Line 580 sobind(struct socket *so, struct mbuf *n |
|
} |
} |
|
|
int |
int |
solisten(struct socket *so, int backlog, struct lwp *l) |
solisten(struct socket *so, int backlog) |
{ |
{ |
int s, error; |
int s, error; |
|
|
s = splsoftnet(); |
s = splsoftnet(); |
error = (*so->so_proto->pr_usrreq)(so, PRU_LISTEN, NULL, |
error = (*so->so_proto->pr_usrreq)(so, PRU_LISTEN, NULL, |
NULL, NULL, l); |
NULL, NULL, NULL); |
if (error != 0) { |
if (error != 0) { |
splx(s); |
splx(s); |
return error; |
return error; |
Line 646 soclose(struct socket *so) |
|
Line 660 soclose(struct socket *so) |
|
goto drop; |
goto drop; |
} |
} |
if (so->so_options & SO_LINGER) { |
if (so->so_options & SO_LINGER) { |
if ((so->so_state & SS_ISDISCONNECTING) && so->so_nbio) |
if ((so->so_state & SS_ISDISCONNECTING) && |
|
(so->so_state & SS_NBIO)) |
goto drop; |
goto drop; |
while (so->so_state & SS_ISCONNECTED) { |
while (so->so_state & SS_ISCONNECTED) { |
error = tsleep((void *)&so->so_timeo, |
error = tsleep((void *)&so->so_timeo, |
Line 855 sosend(struct socket *so, struct mbuf *a |
|
Line 870 sosend(struct socket *so, struct mbuf *a |
|
snderr(EMSGSIZE); |
snderr(EMSGSIZE); |
if (space < resid + clen && |
if (space < resid + clen && |
(atomic || space < so->so_snd.sb_lowat || space < clen)) { |
(atomic || space < so->so_snd.sb_lowat || space < clen)) { |
if (so->so_nbio) |
if (so->so_state & SS_NBIO) |
snderr(EWOULDBLOCK); |
snderr(EWOULDBLOCK); |
sbunlock(&so->so_snd); |
sbunlock(&so->so_snd); |
error = sbwait(&so->so_snd); |
error = sbwait(&so->so_snd); |
Line 1095 soreceive(struct socket *so, struct mbuf |
|
Line 1110 soreceive(struct socket *so, struct mbuf |
|
} |
} |
if (uio->uio_resid == 0) |
if (uio->uio_resid == 0) |
goto release; |
goto release; |
if (so->so_nbio || (flags & MSG_DONTWAIT)) { |
if ((so->so_state & SS_NBIO) || (flags & MSG_DONTWAIT)) { |
error = EWOULDBLOCK; |
error = EWOULDBLOCK; |
goto release; |
goto release; |
} |
} |
Line 1780 soo_kqfilter(struct file *fp, struct kno |
|
Line 1795 soo_kqfilter(struct file *fp, struct kno |
|
sb = &so->so_snd; |
sb = &so->so_snd; |
break; |
break; |
default: |
default: |
return (EINVAL); |
return (1); |
} |
} |
SLIST_INSERT_HEAD(&sb->sb_sel.sel_klist, kn, kn_selnext); |
SLIST_INSERT_HEAD(&sb->sb_sel.sel_klist, kn, kn_selnext); |
sb->sb_flags |= SB_KNOTE; |
sb->sb_flags |= SB_KNOTE; |