version 1.159.2.1, 2008/05/18 12:35:11 |
version 1.235.2.6, 2016/10/05 20:56:03 |
|
|
/* $NetBSD$ */ |
/* $NetBSD$ */ |
|
|
/*- |
/*- |
* Copyright (c) 2002, 2007, 2008 The NetBSD Foundation, Inc. |
* Copyright (c) 2002, 2007, 2008, 2009 The NetBSD Foundation, Inc. |
* All rights reserved. |
* All rights reserved. |
* |
* |
* This code is derived from software contributed to The NetBSD Foundation |
* This code is derived from software contributed to The NetBSD Foundation |
* by Jason R. Thorpe of Wasabi Systems, Inc. |
* by Jason R. Thorpe of Wasabi Systems, Inc, and by Andrew Doran. |
* |
* |
* Redistribution and use in source and binary forms, with or without |
* Redistribution and use in source and binary forms, with or without |
* modification, are permitted provided that the following conditions |
* modification, are permitted provided that the following conditions |
|
|
* @(#)uipc_socket.c 8.6 (Berkeley) 5/2/95 |
* @(#)uipc_socket.c 8.6 (Berkeley) 5/2/95 |
*/ |
*/ |
|
|
|
/* |
|
* Socket operation routines. |
|
* |
|
* These routines are called by the routines in sys_socket.c or from a |
|
* system process, and implement the semantics of socket operations by |
|
* switching out to the protocol specific routines. |
|
*/ |
|
|
#include <sys/cdefs.h> |
#include <sys/cdefs.h> |
__KERNEL_RCSID(0, "$NetBSD$"); |
__KERNEL_RCSID(0, "$NetBSD$"); |
|
|
|
#ifdef _KERNEL_OPT |
|
#include "opt_compat_netbsd.h" |
#include "opt_sock_counters.h" |
#include "opt_sock_counters.h" |
#include "opt_sosend_loan.h" |
#include "opt_sosend_loan.h" |
#include "opt_mbuftrace.h" |
#include "opt_mbuftrace.h" |
#include "opt_somaxkva.h" |
#include "opt_somaxkva.h" |
|
#include "opt_multiprocessor.h" /* XXX */ |
|
#include "opt_sctp.h" |
|
#endif |
|
|
#include <sys/param.h> |
#include <sys/param.h> |
#include <sys/systm.h> |
#include <sys/systm.h> |
#include <sys/proc.h> |
#include <sys/proc.h> |
#include <sys/file.h> |
#include <sys/file.h> |
#include <sys/filedesc.h> |
#include <sys/filedesc.h> |
#include <sys/malloc.h> |
#include <sys/kmem.h> |
#include <sys/mbuf.h> |
#include <sys/mbuf.h> |
#include <sys/domain.h> |
#include <sys/domain.h> |
#include <sys/kernel.h> |
#include <sys/kernel.h> |
Line 84 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 97 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#include <sys/socketvar.h> |
#include <sys/socketvar.h> |
#include <sys/signalvar.h> |
#include <sys/signalvar.h> |
#include <sys/resourcevar.h> |
#include <sys/resourcevar.h> |
|
#include <sys/uidinfo.h> |
#include <sys/event.h> |
#include <sys/event.h> |
#include <sys/poll.h> |
#include <sys/poll.h> |
#include <sys/kauth.h> |
#include <sys/kauth.h> |
#include <sys/mutex.h> |
#include <sys/mutex.h> |
#include <sys/condvar.h> |
#include <sys/condvar.h> |
|
#include <sys/kthread.h> |
|
|
#include <uvm/uvm.h> |
#ifdef COMPAT_50 |
|
#include <compat/sys/time.h> |
|
#include <compat/sys/socket.h> |
|
#endif |
|
|
|
#include <uvm/uvm_extern.h> |
|
#include <uvm/uvm_loan.h> |
|
#include <uvm/uvm_page.h> |
|
|
MALLOC_DEFINE(M_SOOPTS, "soopts", "socket options"); |
|
MALLOC_DEFINE(M_SONAME, "soname", "socket name"); |
MALLOC_DEFINE(M_SONAME, "soname", "socket name"); |
|
|
extern const struct fileops socketops; |
extern const struct fileops socketops; |
Line 125 EVCNT_ATTACH_STATIC(sosend_kvalimit); |
|
Line 146 EVCNT_ATTACH_STATIC(sosend_kvalimit); |
|
|
|
#endif /* SOSEND_COUNTERS */ |
#endif /* SOSEND_COUNTERS */ |
|
|
static struct callback_entry sokva_reclaimerentry; |
#if defined(SOSEND_NO_LOAN) || defined(MULTIPROCESSOR) |
|
|
#ifdef SOSEND_NO_LOAN |
|
int sock_loan_thresh = -1; |
int sock_loan_thresh = -1; |
#else |
#else |
int sock_loan_thresh = 4096; |
int sock_loan_thresh = 4096; |
#endif |
#endif |
|
|
static kmutex_t so_pendfree_lock; |
static kmutex_t so_pendfree_lock; |
static struct mbuf *so_pendfree; |
static struct mbuf *so_pendfree = NULL; |
|
|
#ifndef SOMAXKVA |
#ifndef SOMAXKVA |
#define SOMAXKVA (16 * 1024 * 1024) |
#define SOMAXKVA (16 * 1024 * 1024) |
Line 143 int somaxkva = SOMAXKVA; |
|
Line 162 int somaxkva = SOMAXKVA; |
|
static int socurkva; |
static int socurkva; |
static kcondvar_t socurkva_cv; |
static kcondvar_t socurkva_cv; |
|
|
|
static kauth_listener_t socket_listener; |
|
|
#define SOCK_LOAN_CHUNK 65536 |
#define SOCK_LOAN_CHUNK 65536 |
|
|
static size_t sodopendfree(void); |
static void sopendfree_thread(void *); |
static size_t sodopendfreel(void); |
static kcondvar_t pendfree_thread_cv; |
|
static lwp_t *sopendfree_lwp; |
|
|
|
static void sysctl_kern_socket_setup(void); |
|
static struct sysctllog *socket_sysctllog; |
|
|
static vsize_t |
static vsize_t |
sokvareserve(struct socket *so, vsize_t len) |
sokvareserve(struct socket *so, vsize_t len) |
Line 155 sokvareserve(struct socket *so, vsize_t |
|
Line 180 sokvareserve(struct socket *so, vsize_t |
|
|
|
mutex_enter(&so_pendfree_lock); |
mutex_enter(&so_pendfree_lock); |
while (socurkva + len > somaxkva) { |
while (socurkva + len > somaxkva) { |
size_t freed; |
|
|
|
/* |
|
* try to do pendfree. |
|
*/ |
|
|
|
freed = sodopendfreel(); |
|
|
|
/* |
|
* if some kva was freed, try again. |
|
*/ |
|
|
|
if (freed) |
|
continue; |
|
|
|
SOSEND_COUNTER_INCR(&sosend_kvalimit); |
SOSEND_COUNTER_INCR(&sosend_kvalimit); |
error = cv_wait_sig(&socurkva_cv, &so_pendfree_lock); |
error = cv_wait_sig(&socurkva_cv, &so_pendfree_lock); |
if (error) { |
if (error) { |
Line 197 sokvaunreserve(vsize_t len) |
|
Line 207 sokvaunreserve(vsize_t len) |
|
*/ |
*/ |
|
|
vaddr_t |
vaddr_t |
sokvaalloc(vsize_t len, struct socket *so) |
sokvaalloc(vaddr_t sva, vsize_t len, struct socket *so) |
{ |
{ |
vaddr_t lva; |
vaddr_t lva; |
|
|
Line 212 sokvaalloc(vsize_t len, struct socket *s |
|
Line 222 sokvaalloc(vsize_t len, struct socket *s |
|
* allocate kva. |
* allocate kva. |
*/ |
*/ |
|
|
lva = uvm_km_alloc(kernel_map, len, 0, UVM_KMF_VAONLY | UVM_KMF_WAITVA); |
lva = uvm_km_alloc(kernel_map, len, atop(sva) & uvmexp.colormask, |
|
UVM_KMF_COLORMATCH | UVM_KMF_VAONLY | UVM_KMF_WAITVA); |
if (lva == 0) { |
if (lva == 0) { |
sokvaunreserve(len); |
sokvaunreserve(len); |
return (0); |
return (0); |
Line 262 sodoloanfree(struct vm_page **pgs, void |
|
Line 273 sodoloanfree(struct vm_page **pgs, void |
|
sokvafree(sva, len); |
sokvafree(sva, len); |
} |
} |
|
|
static size_t |
|
sodopendfree(void) |
|
{ |
|
size_t rv; |
|
|
|
if (__predict_true(so_pendfree == NULL)) |
|
return 0; |
|
|
|
mutex_enter(&so_pendfree_lock); |
|
rv = sodopendfreel(); |
|
mutex_exit(&so_pendfree_lock); |
|
|
|
return rv; |
|
} |
|
|
|
/* |
/* |
* sodopendfreel: free mbufs on "pendfree" list. |
* sopendfree_thread: free mbufs on "pendfree" list. |
* unlock and relock so_pendfree_lock when freeing mbufs. |
* unlock and relock so_pendfree_lock when freeing mbufs. |
* |
|
* => called with so_pendfree_lock held. |
|
*/ |
*/ |
|
|
static size_t |
static void |
sodopendfreel(void) |
sopendfree_thread(void *v) |
{ |
{ |
struct mbuf *m, *next; |
struct mbuf *m, *next; |
size_t rv = 0; |
size_t rv; |
|
|
KASSERT(mutex_owned(&so_pendfree_lock)); |
|
|
|
while (so_pendfree != NULL) { |
mutex_enter(&so_pendfree_lock); |
m = so_pendfree; |
|
so_pendfree = NULL; |
|
mutex_exit(&so_pendfree_lock); |
|
|
|
for (; m != NULL; m = next) { |
for (;;) { |
next = m->m_next; |
rv = 0; |
KASSERT((~m->m_flags & (M_EXT|M_EXT_PAGES)) == 0); |
while (so_pendfree != NULL) { |
KASSERT(m->m_ext.ext_refcnt == 0); |
m = so_pendfree; |
|
so_pendfree = NULL; |
|
mutex_exit(&so_pendfree_lock); |
|
|
|
for (; m != NULL; m = next) { |
|
next = m->m_next; |
|
KASSERT((~m->m_flags & (M_EXT|M_EXT_PAGES)) == 0); |
|
KASSERT(m->m_ext.ext_refcnt == 0); |
|
|
|
rv += m->m_ext.ext_size; |
|
sodoloanfree(m->m_ext.ext_pgs, m->m_ext.ext_buf, |
|
m->m_ext.ext_size); |
|
pool_cache_put(mb_cache, m); |
|
} |
|
|
rv += m->m_ext.ext_size; |
mutex_enter(&so_pendfree_lock); |
sodoloanfree(m->m_ext.ext_pgs, m->m_ext.ext_buf, |
|
m->m_ext.ext_size); |
|
pool_cache_put(mb_cache, m); |
|
} |
} |
|
if (rv) |
mutex_enter(&so_pendfree_lock); |
cv_broadcast(&socurkva_cv); |
|
cv_wait(&pendfree_thread_cv, &so_pendfree_lock); |
} |
} |
|
panic("sopendfree_thread"); |
return (rv); |
/* NOTREACHED */ |
} |
} |
|
|
void |
void |
Line 330 soloanfree(struct mbuf *m, void *buf, si |
|
Line 330 soloanfree(struct mbuf *m, void *buf, si |
|
mutex_enter(&so_pendfree_lock); |
mutex_enter(&so_pendfree_lock); |
m->m_next = so_pendfree; |
m->m_next = so_pendfree; |
so_pendfree = m; |
so_pendfree = m; |
cv_broadcast(&socurkva_cv); |
cv_signal(&pendfree_thread_cv); |
mutex_exit(&so_pendfree_lock); |
mutex_exit(&so_pendfree_lock); |
} |
} |
|
|
Line 360 sosend_loan(struct socket *so, struct ui |
|
Line 360 sosend_loan(struct socket *so, struct ui |
|
|
|
KASSERT(npgs <= M_EXT_MAXPAGES); |
KASSERT(npgs <= M_EXT_MAXPAGES); |
|
|
lva = sokvaalloc(len, so); |
lva = sokvaalloc(sva, len, so); |
if (lva == 0) |
if (lva == 0) |
return 0; |
return 0; |
|
|
Line 373 sosend_loan(struct socket *so, struct ui |
|
Line 373 sosend_loan(struct socket *so, struct ui |
|
|
|
for (i = 0, va = lva; i < npgs; i++, va += PAGE_SIZE) |
for (i = 0, va = lva; i < npgs; i++, va += PAGE_SIZE) |
pmap_kenter_pa(va, VM_PAGE_TO_PHYS(m->m_ext.ext_pgs[i]), |
pmap_kenter_pa(va, VM_PAGE_TO_PHYS(m->m_ext.ext_pgs[i]), |
VM_PROT_READ); |
VM_PROT_READ, 0); |
pmap_update(pmap_kernel()); |
pmap_update(pmap_kernel()); |
|
|
lva += (vaddr_t) iov->iov_base & PAGE_MASK; |
lva += (vaddr_t) iov->iov_base & PAGE_MASK; |
Line 393 sosend_loan(struct socket *so, struct ui |
|
Line 393 sosend_loan(struct socket *so, struct ui |
|
return (space); |
return (space); |
} |
} |
|
|
static int |
|
sokva_reclaim_callback(struct callback_entry *ce, void *obj, void *arg) |
|
{ |
|
|
|
KASSERT(ce == &sokva_reclaimerentry); |
|
KASSERT(obj == NULL); |
|
|
|
sodopendfree(); |
|
if (!vm_map_starved_p(kernel_map)) { |
|
return CALLBACK_CHAIN_ABORT; |
|
} |
|
return CALLBACK_CHAIN_CONTINUE; |
|
} |
|
|
|
struct mbuf * |
struct mbuf * |
getsombuf(struct socket *so, int type) |
getsombuf(struct socket *so, int type) |
{ |
{ |
Line 417 getsombuf(struct socket *so, int type) |
|
Line 403 getsombuf(struct socket *so, int type) |
|
return m; |
return m; |
} |
} |
|
|
struct mbuf * |
static int |
m_intopt(struct socket *so, int val) |
socket_listener_cb(kauth_cred_t cred, kauth_action_t action, void *cookie, |
|
void *arg0, void *arg1, void *arg2, void *arg3) |
{ |
{ |
struct mbuf *m; |
int result; |
|
enum kauth_network_req req; |
|
|
m = getsombuf(so, MT_SOOPTS); |
result = KAUTH_RESULT_DEFER; |
m->m_len = sizeof(int); |
req = (enum kauth_network_req)arg0; |
*mtod(m, int *) = val; |
|
return m; |
if ((action != KAUTH_NETWORK_SOCKET) && |
|
(action != KAUTH_NETWORK_BIND)) |
|
return result; |
|
|
|
switch (req) { |
|
case KAUTH_REQ_NETWORK_BIND_PORT: |
|
result = KAUTH_RESULT_ALLOW; |
|
break; |
|
|
|
case KAUTH_REQ_NETWORK_SOCKET_DROP: { |
|
/* Normal users can only drop their own connections. */ |
|
struct socket *so = (struct socket *)arg1; |
|
|
|
if (so->so_cred && proc_uidmatch(cred, so->so_cred) == 0) |
|
result = KAUTH_RESULT_ALLOW; |
|
|
|
break; |
|
} |
|
|
|
case KAUTH_REQ_NETWORK_SOCKET_OPEN: |
|
/* We allow "raw" routing/bluetooth sockets to anyone. */ |
|
if ((u_long)arg1 == PF_ROUTE || (u_long)arg1 == PF_OROUTE |
|
|| (u_long)arg1 == PF_BLUETOOTH) { |
|
result = KAUTH_RESULT_ALLOW; |
|
} else { |
|
/* Privileged, let secmodel handle this. */ |
|
if ((u_long)arg2 == SOCK_RAW) |
|
break; |
|
} |
|
|
|
result = KAUTH_RESULT_ALLOW; |
|
|
|
break; |
|
|
|
case KAUTH_REQ_NETWORK_SOCKET_CANSEE: |
|
result = KAUTH_RESULT_ALLOW; |
|
|
|
break; |
|
|
|
default: |
|
break; |
|
} |
|
|
|
return result; |
} |
} |
|
|
void |
void |
soinit(void) |
soinit(void) |
{ |
{ |
|
|
|
sysctl_kern_socket_setup(); |
|
|
mutex_init(&so_pendfree_lock, MUTEX_DEFAULT, IPL_VM); |
mutex_init(&so_pendfree_lock, MUTEX_DEFAULT, IPL_VM); |
softnet_lock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_NONE); |
softnet_lock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_NONE); |
cv_init(&socurkva_cv, "sokva"); |
cv_init(&socurkva_cv, "sokva"); |
|
cv_init(&pendfree_thread_cv, "sopendfr"); |
|
soinit2(); |
|
|
/* Set the initial adjusted socket buffer size. */ |
/* Set the initial adjusted socket buffer size. */ |
if (sb_max_set(sb_max)) |
if (sb_max_set(sb_max)) |
panic("bad initial sb_max value: %lu", sb_max); |
panic("bad initial sb_max value: %lu", sb_max); |
|
|
callback_register(&vm_map_to_kernel(kernel_map)->vmk_reclaim_callback, |
socket_listener = kauth_listen_scope(KAUTH_SCOPE_NETWORK, |
&sokva_reclaimerentry, NULL, sokva_reclaim_callback); |
socket_listener_cb, NULL); |
|
} |
|
|
|
void |
|
soinit1(void) |
|
{ |
|
int error = kthread_create(PRI_NONE, KTHREAD_MPSAFE, NULL, |
|
sopendfree_thread, NULL, &sopendfree_lwp, "sopendfree"); |
|
if (error) |
|
panic("soinit1 %d", error); |
} |
} |
|
|
/* |
/* |
* Socket operation routines. |
* socreate: create a new socket of the specified type and the protocol. |
* These routines are called by the routines in |
* |
* sys_socket.c or from a system process, and |
* => Caller may specify another socket for lock sharing (must not be held). |
* implement the semantics of socket operations by |
* => Returns the new socket without lock held. |
* switching out to the protocol specific routines. |
|
*/ |
*/ |
/*ARGSUSED*/ |
|
int |
int |
socreate(int dom, struct socket **aso, int type, int proto, struct lwp *l, |
socreate(int dom, struct socket **aso, int type, int proto, struct lwp *l, |
struct socket *lockso) |
struct socket *lockso) |
Line 481 socreate(int dom, struct socket **aso, i |
|
Line 523 socreate(int dom, struct socket **aso, i |
|
return EPROTOTYPE; |
return EPROTOTYPE; |
return EPROTONOSUPPORT; |
return EPROTONOSUPPORT; |
} |
} |
if (prp->pr_usrreq == NULL) |
if (prp->pr_usrreqs == NULL) |
return EPROTONOSUPPORT; |
return EPROTONOSUPPORT; |
if (prp->pr_type != type) |
if (prp->pr_type != type) |
return EPROTOTYPE; |
return EPROTOTYPE; |
Line 498 socreate(int dom, struct socket **aso, i |
|
Line 540 socreate(int dom, struct socket **aso, i |
|
#endif |
#endif |
uid = kauth_cred_geteuid(l->l_cred); |
uid = kauth_cred_geteuid(l->l_cred); |
so->so_uidinfo = uid_find(uid); |
so->so_uidinfo = uid_find(uid); |
if (lockso != NULL) { |
so->so_cpid = l->l_proc->p_pid; |
/* Caller wants us to share a lock. */ |
|
|
/* |
|
* Lock assigned and taken during PCB attach, unless we share |
|
* the lock with another socket, e.g. socketpair(2) case. |
|
*/ |
|
if (lockso) { |
lock = lockso->so_lock; |
lock = lockso->so_lock; |
so->so_lock = lock; |
so->so_lock = lock; |
mutex_obj_hold(lock); |
mutex_obj_hold(lock); |
mutex_enter(lock); |
mutex_enter(lock); |
} else { |
|
/* Lock assigned and taken during PRU_ATTACH. */ |
|
} |
} |
error = (*prp->pr_usrreq)(so, PRU_ATTACH, NULL, |
|
(struct mbuf *)(long)proto, NULL, l); |
/* Attach the PCB (returns with the socket lock held). */ |
|
error = (*prp->pr_usrreqs->pr_attach)(so, proto); |
KASSERT(solocked(so)); |
KASSERT(solocked(so)); |
if (error != 0) { |
|
|
if (error) { |
|
KASSERT(so->so_pcb == NULL); |
so->so_state |= SS_NOFDREF; |
so->so_state |= SS_NOFDREF; |
sofree(so); |
sofree(so); |
return error; |
return error; |
} |
} |
|
so->so_cred = kauth_cred_dup(l->l_cred); |
sounlock(so); |
sounlock(so); |
|
|
*aso = so; |
*aso = so; |
return 0; |
return 0; |
} |
} |
|
|
/* On success, write file descriptor to fdout and return zero. On |
/* |
* failure, return non-zero; *fdout will be undefined. |
* fsocreate: create a socket and a file descriptor associated with it. |
|
* |
|
* => On success, write file descriptor to fdout and return zero. |
|
* => On failure, return non-zero; *fdout will be undefined. |
*/ |
*/ |
int |
int |
fsocreate(int domain, struct socket **sop, int type, int protocol, |
fsocreate(int domain, struct socket **sop, int type, int proto, int *fdout) |
struct lwp *l, int *fdout) |
|
{ |
{ |
struct socket *so; |
lwp_t *l = curlwp; |
struct file *fp; |
int error, fd, flags; |
int fd, error; |
struct socket *so; |
|
struct file *fp; |
|
|
if ((error = fd_allocfile(&fp, &fd)) != 0) |
if ((error = fd_allocfile(&fp, &fd)) != 0) { |
return (error); |
return error; |
fp->f_flag = FREAD|FWRITE; |
} |
|
flags = type & SOCK_FLAGS_MASK; |
|
fd_set_exclose(l, fd, (flags & SOCK_CLOEXEC) != 0); |
|
fp->f_flag = FREAD|FWRITE|((flags & SOCK_NONBLOCK) ? FNONBLOCK : 0)| |
|
((flags & SOCK_NOSIGPIPE) ? FNOSIGPIPE : 0); |
fp->f_type = DTYPE_SOCKET; |
fp->f_type = DTYPE_SOCKET; |
fp->f_ops = &socketops; |
fp->f_ops = &socketops; |
error = socreate(domain, &so, type, protocol, l, NULL); |
|
if (error != 0) { |
type &= ~SOCK_FLAGS_MASK; |
|
error = socreate(domain, &so, type, proto, l, NULL); |
|
if (error) { |
fd_abort(curproc, fp, fd); |
fd_abort(curproc, fp, fd); |
} else { |
return error; |
if (sop != NULL) |
} |
*sop = so; |
if (flags & SOCK_NONBLOCK) { |
fp->f_data = so; |
so->so_state |= SS_NBIO; |
fd_affix(curproc, fp, fd); |
|
*fdout = fd; |
|
} |
} |
|
fp->f_socket = so; |
|
fd_affix(curproc, fp, fd); |
|
|
|
if (sop != NULL) { |
|
*sop = so; |
|
} |
|
*fdout = fd; |
return error; |
return error; |
} |
} |
|
|
int |
int |
sobind(struct socket *so, struct mbuf *nam, struct lwp *l) |
sofamily(const struct socket *so) |
|
{ |
|
const struct protosw *pr; |
|
const struct domain *dom; |
|
|
|
if ((pr = so->so_proto) == NULL) |
|
return AF_UNSPEC; |
|
if ((dom = pr->pr_domain) == NULL) |
|
return AF_UNSPEC; |
|
return dom->dom_family; |
|
} |
|
|
|
int |
|
sobind(struct socket *so, struct sockaddr *nam, struct lwp *l) |
{ |
{ |
int error; |
int error; |
|
|
solock(so); |
solock(so); |
error = (*so->so_proto->pr_usrreq)(so, PRU_BIND, NULL, nam, NULL, l); |
if (nam->sa_family != so->so_proto->pr_domain->dom_family) { |
|
sounlock(so); |
|
return EAFNOSUPPORT; |
|
} |
|
error = (*so->so_proto->pr_usrreqs->pr_bind)(so, nam, l); |
sounlock(so); |
sounlock(so); |
return error; |
return error; |
} |
} |
|
|
solisten(struct socket *so, int backlog, struct lwp *l) |
solisten(struct socket *so, int backlog, struct lwp *l) |
{ |
{ |
int error; |
int error; |
|
short oldopt, oldqlimit; |
|
|
solock(so); |
solock(so); |
if ((so->so_state & (SS_ISCONNECTED | SS_ISCONNECTING | |
if ((so->so_state & (SS_ISCONNECTED | SS_ISCONNECTING | |
SS_ISDISCONNECTING)) != 0) { |
SS_ISDISCONNECTING)) != 0) { |
sounlock(so); |
|
return (EOPNOTSUPP); |
|
} |
|
error = (*so->so_proto->pr_usrreq)(so, PRU_LISTEN, NULL, |
|
NULL, NULL, l); |
|
if (error != 0) { |
|
sounlock(so); |
sounlock(so); |
return error; |
return EINVAL; |
} |
} |
|
oldopt = so->so_options; |
|
oldqlimit = so->so_qlimit; |
if (TAILQ_EMPTY(&so->so_q)) |
if (TAILQ_EMPTY(&so->so_q)) |
so->so_options |= SO_ACCEPTCONN; |
so->so_options |= SO_ACCEPTCONN; |
if (backlog < 0) |
if (backlog < 0) |
backlog = 0; |
backlog = 0; |
so->so_qlimit = min(backlog, somaxconn); |
so->so_qlimit = min(backlog, somaxconn); |
|
|
|
error = (*so->so_proto->pr_usrreqs->pr_listen)(so, l); |
|
if (error != 0) { |
|
so->so_options = oldopt; |
|
so->so_qlimit = oldqlimit; |
|
sounlock(so); |
|
return error; |
|
} |
sounlock(so); |
sounlock(so); |
return 0; |
return 0; |
} |
} |
Line 620 sofree(struct socket *so) |
|
Line 706 sofree(struct socket *so) |
|
KASSERT(!cv_has_waiters(&so->so_snd.sb_cv)); |
KASSERT(!cv_has_waiters(&so->so_snd.sb_cv)); |
sorflush(so); |
sorflush(so); |
refs = so->so_aborting; /* XXX */ |
refs = so->so_aborting; /* XXX */ |
|
/* Remove acccept filter if one is present. */ |
|
if (so->so_accf != NULL) |
|
(void)accept_filt_clear(so); |
sounlock(so); |
sounlock(so); |
if (refs == 0) /* XXX */ |
if (refs == 0) /* XXX */ |
soput(so); |
soput(so); |
} |
} |
|
|
/* |
/* |
* Close a socket on last file table reference removal. |
* soclose: close a socket on last file table reference removal. |
* Initiate disconnect if connected. |
* Initiate disconnect if connected. Free socket when disconnect complete. |
* Free socket when disconnect complete. |
|
*/ |
*/ |
int |
int |
soclose(struct socket *so) |
soclose(struct socket *so) |
{ |
{ |
struct socket *so2; |
struct socket *so2; |
int error; |
int error = 0; |
int error2; |
|
|
|
error = 0; |
|
solock(so); |
solock(so); |
if (so->so_options & SO_ACCEPTCONN) { |
if (so->so_options & SO_ACCEPTCONN) { |
do { |
for (;;) { |
if ((so2 = TAILQ_FIRST(&so->so_q0)) != 0) { |
if ((so2 = TAILQ_FIRST(&so->so_q0)) != 0) { |
KASSERT(solocked2(so, so2)); |
KASSERT(solocked2(so, so2)); |
(void) soqremque(so2, 0); |
(void) soqremque(so2, 0); |
Line 657 soclose(struct socket *so) |
|
Line 743 soclose(struct socket *so) |
|
solock(so); |
solock(so); |
continue; |
continue; |
} |
} |
} while (0); |
break; |
|
} |
} |
} |
if (so->so_pcb == 0) |
if (so->so_pcb == NULL) |
goto discard; |
goto discard; |
if (so->so_state & SS_ISCONNECTED) { |
if (so->so_state & SS_ISCONNECTED) { |
if ((so->so_state & SS_ISDISCONNECTING) == 0) { |
if ((so->so_state & SS_ISDISCONNECTING) == 0) { |
Line 668 soclose(struct socket *so) |
|
Line 755 soclose(struct socket *so) |
|
goto drop; |
goto drop; |
} |
} |
if (so->so_options & SO_LINGER) { |
if (so->so_options & SO_LINGER) { |
if ((so->so_state & SS_ISDISCONNECTING) && so->so_nbio) |
if ((so->so_state & (SS_ISDISCONNECTING|SS_NBIO)) == |
|
(SS_ISDISCONNECTING|SS_NBIO)) |
goto drop; |
goto drop; |
while (so->so_state & SS_ISCONNECTED) { |
while (so->so_state & SS_ISCONNECTED) { |
error = sowait(so, so->so_linger * hz); |
error = sowait(so, true, so->so_linger * hz); |
if (error) |
if (error) |
break; |
break; |
} |
} |
Line 679 soclose(struct socket *so) |
|
Line 767 soclose(struct socket *so) |
|
} |
} |
drop: |
drop: |
if (so->so_pcb) { |
if (so->so_pcb) { |
error2 = (*so->so_proto->pr_usrreq)(so, PRU_DETACH, |
KASSERT(solocked(so)); |
NULL, NULL, NULL, NULL); |
(*so->so_proto->pr_usrreqs->pr_detach)(so); |
if (error == 0) |
|
error = error2; |
|
} |
} |
discard: |
discard: |
if (so->so_state & SS_NOFDREF) |
KASSERT((so->so_state & SS_NOFDREF) == 0); |
panic("soclose: NOFDREF"); |
kauth_cred_free(so->so_cred); |
so->so_state |= SS_NOFDREF; |
so->so_state |= SS_NOFDREF; |
sofree(so); |
sofree(so); |
return (error); |
return error; |
} |
} |
|
|
/* |
/* |
Line 700 soabort(struct socket *so) |
|
Line 786 soabort(struct socket *so) |
|
{ |
{ |
u_int refs; |
u_int refs; |
int error; |
int error; |
|
|
KASSERT(solocked(so)); |
KASSERT(solocked(so)); |
KASSERT(so->so_head == NULL); |
KASSERT(so->so_head == NULL); |
|
|
so->so_aborting++; /* XXX */ |
so->so_aborting++; /* XXX */ |
error = (*so->so_proto->pr_usrreq)(so, PRU_ABORT, NULL, |
error = (*so->so_proto->pr_usrreqs->pr_abort)(so); |
NULL, NULL, NULL); |
|
refs = --so->so_aborting; /* XXX */ |
refs = --so->so_aborting; /* XXX */ |
if (error || (refs == 0)) { |
if (error || (refs == 0)) { |
sofree(so); |
sofree(so); |
Line 717 soabort(struct socket *so) |
|
Line 802 soabort(struct socket *so) |
|
} |
} |
|
|
int |
int |
soaccept(struct socket *so, struct mbuf *nam) |
soaccept(struct socket *so, struct sockaddr *nam) |
{ |
{ |
int error; |
int error; |
|
|
KASSERT(solocked(so)); |
KASSERT(solocked(so)); |
|
KASSERT((so->so_state & SS_NOFDREF) != 0); |
|
|
error = 0; |
|
if ((so->so_state & SS_NOFDREF) == 0) |
|
panic("soaccept: !NOFDREF"); |
|
so->so_state &= ~SS_NOFDREF; |
so->so_state &= ~SS_NOFDREF; |
if ((so->so_state & SS_ISDISCONNECTED) == 0 || |
if ((so->so_state & SS_ISDISCONNECTED) == 0 || |
(so->so_proto->pr_flags & PR_ABRTACPTDIS) == 0) |
(so->so_proto->pr_flags & PR_ABRTACPTDIS) == 0) |
error = (*so->so_proto->pr_usrreq)(so, PRU_ACCEPT, |
error = (*so->so_proto->pr_usrreqs->pr_accept)(so, nam); |
NULL, nam, NULL, NULL); |
|
else |
else |
error = ECONNABORTED; |
error = ECONNABORTED; |
|
|
return (error); |
return error; |
} |
} |
|
|
int |
int |
soconnect(struct socket *so, struct mbuf *nam, struct lwp *l) |
soconnect(struct socket *so, struct sockaddr *nam, struct lwp *l) |
{ |
{ |
int error; |
int error; |
|
|
KASSERT(solocked(so)); |
KASSERT(solocked(so)); |
|
|
if (so->so_options & SO_ACCEPTCONN) |
if (so->so_options & SO_ACCEPTCONN) |
return (EOPNOTSUPP); |
return EOPNOTSUPP; |
/* |
/* |
* If protocol is connection-based, can only connect once. |
* If protocol is connection-based, can only connect once. |
* Otherwise, if connected, try to disconnect first. |
* Otherwise, if connected, try to disconnect first. |
Line 754 soconnect(struct socket *so, struct mbuf |
|
Line 836 soconnect(struct socket *so, struct mbuf |
|
*/ |
*/ |
if (so->so_state & (SS_ISCONNECTED|SS_ISCONNECTING) && |
if (so->so_state & (SS_ISCONNECTED|SS_ISCONNECTING) && |
((so->so_proto->pr_flags & PR_CONNREQUIRED) || |
((so->so_proto->pr_flags & PR_CONNREQUIRED) || |
(error = sodisconnect(so)))) |
(error = sodisconnect(so)))) { |
error = EISCONN; |
error = EISCONN; |
else |
} else { |
error = (*so->so_proto->pr_usrreq)(so, PRU_CONNECT, |
if (nam->sa_family != so->so_proto->pr_domain->dom_family) { |
NULL, nam, NULL, l); |
return EAFNOSUPPORT; |
return (error); |
} |
|
error = (*so->so_proto->pr_usrreqs->pr_connect)(so, nam, l); |
|
} |
|
|
|
return error; |
} |
} |
|
|
int |
int |
soconnect2(struct socket *so1, struct socket *so2) |
soconnect2(struct socket *so1, struct socket *so2) |
{ |
{ |
int error; |
|
|
|
KASSERT(solocked2(so1, so2)); |
KASSERT(solocked2(so1, so2)); |
|
|
error = (*so1->so_proto->pr_usrreq)(so1, PRU_CONNECT2, |
return (*so1->so_proto->pr_usrreqs->pr_connect2)(so1, so2); |
NULL, (struct mbuf *)so2, NULL, NULL); |
|
return (error); |
|
} |
} |
|
|
int |
int |
Line 786 sodisconnect(struct socket *so) |
|
Line 868 sodisconnect(struct socket *so) |
|
} else if (so->so_state & SS_ISDISCONNECTING) { |
} else if (so->so_state & SS_ISDISCONNECTING) { |
error = EALREADY; |
error = EALREADY; |
} else { |
} else { |
error = (*so->so_proto->pr_usrreq)(so, PRU_DISCONNECT, |
error = (*so->so_proto->pr_usrreqs->pr_disconnect)(so); |
NULL, NULL, NULL, NULL); |
|
} |
} |
sodopendfree(); |
|
return (error); |
return (error); |
} |
} |
|
|
Line 812 sodisconnect(struct socket *so) |
|
Line 892 sodisconnect(struct socket *so) |
|
* Data and control buffers are freed on return. |
* Data and control buffers are freed on return. |
*/ |
*/ |
int |
int |
sosend(struct socket *so, struct mbuf *addr, struct uio *uio, struct mbuf *top, |
sosend(struct socket *so, struct sockaddr *addr, struct uio *uio, |
struct mbuf *control, int flags, struct lwp *l) |
struct mbuf *top, struct mbuf *control, int flags, struct lwp *l) |
{ |
{ |
struct mbuf **mp, *m; |
struct mbuf **mp, *m; |
struct proc *p; |
|
long space, len, resid, clen, mlen; |
long space, len, resid, clen, mlen; |
int error, s, dontroute, atomic; |
int error, s, dontroute, atomic; |
|
short wakeup_state = 0; |
|
|
p = l->l_proc; |
|
sodopendfree(); |
|
clen = 0; |
clen = 0; |
|
|
/* |
/* |
Line 850 sosend(struct socket *so, struct mbuf *a |
|
Line 928 sosend(struct socket *so, struct mbuf *a |
|
dontroute = |
dontroute = |
(flags & MSG_DONTROUTE) && (so->so_options & SO_DONTROUTE) == 0 && |
(flags & MSG_DONTROUTE) && (so->so_options & SO_DONTROUTE) == 0 && |
(so->so_proto->pr_flags & PR_ATOMIC); |
(so->so_proto->pr_flags & PR_ATOMIC); |
if (l) |
l->l_ru.ru_msgsnd++; |
l->l_ru.ru_msgsnd++; |
|
if (control) |
if (control) |
clen = control->m_len; |
clen = control->m_len; |
restart: |
restart: |
Line 869 sosend(struct socket *so, struct mbuf *a |
|
Line 946 sosend(struct socket *so, struct mbuf *a |
|
} |
} |
if ((so->so_state & SS_ISCONNECTED) == 0) { |
if ((so->so_state & SS_ISCONNECTED) == 0) { |
if (so->so_proto->pr_flags & PR_CONNREQUIRED) { |
if (so->so_proto->pr_flags & PR_CONNREQUIRED) { |
if ((so->so_state & SS_ISCONFIRMING) == 0 && |
if (resid || clen == 0) { |
!(resid == 0 && clen != 0)) { |
|
error = ENOTCONN; |
error = ENOTCONN; |
goto release; |
goto release; |
} |
} |
} else if (addr == 0) { |
} else if (addr == NULL) { |
error = EDESTADDRREQ; |
error = EDESTADDRREQ; |
goto release; |
goto release; |
} |
} |
Line 889 sosend(struct socket *so, struct mbuf *a |
|
Line 965 sosend(struct socket *so, struct mbuf *a |
|
} |
} |
if (space < resid + clen && |
if (space < resid + clen && |
(atomic || space < so->so_snd.sb_lowat || space < clen)) { |
(atomic || space < so->so_snd.sb_lowat || space < clen)) { |
if (so->so_nbio) { |
if ((so->so_state & SS_NBIO) || (flags & MSG_NBIO)) { |
error = EWOULDBLOCK; |
error = EWOULDBLOCK; |
goto release; |
goto release; |
} |
} |
sbunlock(&so->so_snd); |
sbunlock(&so->so_snd); |
|
if (wakeup_state & SS_RESTARTSYS) { |
|
error = ERESTART; |
|
goto out; |
|
} |
error = sbwait(&so->so_snd); |
error = sbwait(&so->so_snd); |
if (error) |
if (error) |
goto out; |
goto out; |
|
wakeup_state = so->so_state; |
goto restart; |
goto restart; |
} |
} |
|
wakeup_state = 0; |
mp = ⊤ |
mp = ⊤ |
space -= clen; |
space -= clen; |
do { |
do { |
Line 916 sosend(struct socket *so, struct mbuf *a |
|
Line 998 sosend(struct socket *so, struct mbuf *a |
|
m = m_gethdr(M_WAIT, MT_DATA); |
m = m_gethdr(M_WAIT, MT_DATA); |
mlen = MHLEN; |
mlen = MHLEN; |
m->m_pkthdr.len = 0; |
m->m_pkthdr.len = 0; |
m->m_pkthdr.rcvif = NULL; |
m_reset_rcvif(m); |
} else { |
} else { |
m = m_get(M_WAIT, MT_DATA); |
m = m_get(M_WAIT, MT_DATA); |
mlen = MLEN; |
mlen = MLEN; |
Line 933 sosend(struct socket *so, struct mbuf *a |
|
Line 1015 sosend(struct socket *so, struct mbuf *a |
|
} |
} |
if (resid >= MINCLSIZE && space >= MCLBYTES) { |
if (resid >= MINCLSIZE && space >= MCLBYTES) { |
SOSEND_COUNTER_INCR(&sosend_copy_big); |
SOSEND_COUNTER_INCR(&sosend_copy_big); |
m_clget(m, M_WAIT); |
m_clget(m, M_DONTWAIT); |
if ((m->m_flags & M_EXT) == 0) |
if ((m->m_flags & M_EXT) == 0) |
goto nopages; |
goto nopages; |
mlen = MCLBYTES; |
mlen = MCLBYTES; |
Line 982 sosend(struct socket *so, struct mbuf *a |
|
Line 1064 sosend(struct socket *so, struct mbuf *a |
|
so->so_options |= SO_DONTROUTE; |
so->so_options |= SO_DONTROUTE; |
if (resid > 0) |
if (resid > 0) |
so->so_state |= SS_MORETOCOME; |
so->so_state |= SS_MORETOCOME; |
error = (*so->so_proto->pr_usrreq)(so, |
if (flags & MSG_OOB) { |
(flags & MSG_OOB) ? PRU_SENDOOB : PRU_SEND, |
error = (*so->so_proto->pr_usrreqs->pr_sendoob)(so, |
top, addr, control, curlwp); |
top, control); |
|
} else { |
|
error = (*so->so_proto->pr_usrreqs->pr_send)(so, |
|
top, addr, control, l); |
|
} |
if (dontroute) |
if (dontroute) |
so->so_options &= ~SO_DONTROUTE; |
so->so_options &= ~SO_DONTROUTE; |
if (resid > 0) |
if (resid > 0) |
Line 1069 soreceive(struct socket *so, struct mbuf |
|
Line 1155 soreceive(struct socket *so, struct mbuf |
|
{ |
{ |
struct lwp *l = curlwp; |
struct lwp *l = curlwp; |
struct mbuf *m, **mp, *mt; |
struct mbuf *m, **mp, *mt; |
int atomic, flags, len, error, s, offset, moff, type, orig_resid; |
size_t len, offset, moff, orig_resid; |
|
int atomic, flags, error, s, type; |
const struct protosw *pr; |
const struct protosw *pr; |
struct mbuf *nextrecord; |
struct mbuf *nextrecord; |
int mbuf_removed = 0; |
int mbuf_removed = 0; |
const struct domain *dom; |
const struct domain *dom; |
|
short wakeup_state = 0; |
|
|
pr = so->so_proto; |
pr = so->so_proto; |
atomic = pr->pr_flags & PR_ATOMIC; |
atomic = pr->pr_flags & PR_ATOMIC; |
Line 1091 soreceive(struct socket *so, struct mbuf |
|
Line 1179 soreceive(struct socket *so, struct mbuf |
|
else |
else |
flags = 0; |
flags = 0; |
|
|
if ((flags & MSG_DONTWAIT) == 0) |
|
sodopendfree(); |
|
|
|
if (flags & MSG_OOB) { |
if (flags & MSG_OOB) { |
m = m_get(M_WAIT, MT_DATA); |
m = m_get(M_WAIT, MT_DATA); |
solock(so); |
solock(so); |
error = (*pr->pr_usrreq)(so, PRU_RCVOOB, m, |
error = (*pr->pr_usrreqs->pr_recvoob)(so, m, flags & MSG_PEEK); |
(struct mbuf *)(long)(flags & MSG_PEEK), NULL, l); |
|
sounlock(so); |
sounlock(so); |
if (error) |
if (error) |
goto bad; |
goto bad; |
do { |
do { |
error = uiomove(mtod(m, void *), |
error = uiomove(mtod(m, void *), |
(int) min(uio->uio_resid, m->m_len), uio); |
MIN(uio->uio_resid, m->m_len), uio); |
m = m_free(m); |
m = m_free(m); |
} while (uio->uio_resid > 0 && error == 0 && m); |
} while (uio->uio_resid > 0 && error == 0 && m); |
bad: |
bad: |
Line 1122 soreceive(struct socket *so, struct mbuf |
|
Line 1206 soreceive(struct socket *so, struct mbuf |
|
*/ |
*/ |
s = splsoftnet(); |
s = splsoftnet(); |
solock(so); |
solock(so); |
if (so->so_state & SS_ISCONFIRMING && uio->uio_resid) |
|
(*pr->pr_usrreq)(so, PRU_RCVD, NULL, NULL, NULL, l); |
|
|
|
restart: |
restart: |
if ((error = sblock(&so->so_rcv, SBLOCKWAIT(flags))) != 0) { |
if ((error = sblock(&so->so_rcv, SBLOCKWAIT(flags))) != 0) { |
sounlock(so); |
sounlock(so); |
Line 1181 soreceive(struct socket *so, struct mbuf |
|
Line 1262 soreceive(struct socket *so, struct mbuf |
|
} |
} |
if (uio->uio_resid == 0) |
if (uio->uio_resid == 0) |
goto release; |
goto release; |
if (so->so_nbio || (flags & MSG_DONTWAIT)) { |
if ((so->so_state & SS_NBIO) || |
|
(flags & (MSG_DONTWAIT|MSG_NBIO))) { |
error = EWOULDBLOCK; |
error = EWOULDBLOCK; |
goto release; |
goto release; |
} |
} |
SBLASTRECORDCHK(&so->so_rcv, "soreceive sbwait 1"); |
SBLASTRECORDCHK(&so->so_rcv, "soreceive sbwait 1"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive sbwait 1"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive sbwait 1"); |
sbunlock(&so->so_rcv); |
sbunlock(&so->so_rcv); |
error = sbwait(&so->so_rcv); |
if (wakeup_state & SS_RESTARTSYS) |
|
error = ERESTART; |
|
else |
|
error = sbwait(&so->so_rcv); |
if (error != 0) { |
if (error != 0) { |
sounlock(so); |
sounlock(so); |
splx(s); |
splx(s); |
return error; |
return error; |
} |
} |
|
wakeup_state = so->so_state; |
goto restart; |
goto restart; |
} |
} |
dontblock: |
dontblock: |
Line 1238 soreceive(struct socket *so, struct mbuf |
|
Line 1324 soreceive(struct socket *so, struct mbuf |
|
m->m_next = NULL; |
m->m_next = NULL; |
m = so->so_rcv.sb_mb; |
m = so->so_rcv.sb_mb; |
} else { |
} else { |
MFREE(m, so->so_rcv.sb_mb); |
m = so->so_rcv.sb_mb = m_free(m); |
m = so->so_rcv.sb_mb; |
|
} |
} |
sbsync(&so->so_rcv, nextrecord); |
sbsync(&so->so_rcv, nextrecord); |
} |
} |
} |
} |
|
if (pr->pr_flags & PR_ADDR_OPT) { |
|
/* |
|
* For SCTP we may be getting a |
|
* whole message OR a partial delivery. |
|
*/ |
|
if (m->m_type == MT_SONAME) { |
|
orig_resid = 0; |
|
if (flags & MSG_PEEK) { |
|
if (paddr) |
|
*paddr = m_copy(m, 0, m->m_len); |
|
m = m->m_next; |
|
} else { |
|
sbfree(&so->so_rcv, m); |
|
if (paddr) { |
|
*paddr = m; |
|
so->so_rcv.sb_mb = m->m_next; |
|
m->m_next = 0; |
|
m = so->so_rcv.sb_mb; |
|
} else { |
|
m = so->so_rcv.sb_mb = m_free(m); |
|
} |
|
} |
|
} |
|
} |
|
|
/* |
/* |
* Process one or more MT_CONTROL mbufs present before any data mbufs |
* Process one or more MT_CONTROL mbufs present before any data mbufs |
Line 1282 soreceive(struct socket *so, struct mbuf |
|
Line 1391 soreceive(struct socket *so, struct mbuf |
|
type == SCM_RIGHTS) { |
type == SCM_RIGHTS) { |
sounlock(so); |
sounlock(so); |
splx(s); |
splx(s); |
error = (*dom->dom_externalize)(cm, l); |
error = (*dom->dom_externalize)(cm, l, |
|
(flags & MSG_CMSG_CLOEXEC) ? |
|
O_CLOEXEC : 0); |
s = splsoftnet(); |
s = splsoftnet(); |
solock(so); |
solock(so); |
} |
} |
Line 1332 soreceive(struct socket *so, struct mbuf |
|
Line 1443 soreceive(struct socket *so, struct mbuf |
|
panic("receive 3"); |
panic("receive 3"); |
#endif |
#endif |
so->so_state &= ~SS_RCVATMARK; |
so->so_state &= ~SS_RCVATMARK; |
|
wakeup_state = 0; |
len = uio->uio_resid; |
len = uio->uio_resid; |
if (so->so_oobmark && len > so->so_oobmark - offset) |
if (so->so_oobmark && len > so->so_oobmark - offset) |
len = so->so_oobmark - offset; |
len = so->so_oobmark - offset; |
Line 1350 soreceive(struct socket *so, struct mbuf |
|
Line 1462 soreceive(struct socket *so, struct mbuf |
|
SBLASTMBUFCHK(&so->so_rcv, "soreceive uiomove"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive uiomove"); |
sounlock(so); |
sounlock(so); |
splx(s); |
splx(s); |
error = uiomove(mtod(m, char *) + moff, (int)len, uio); |
error = uiomove(mtod(m, char *) + moff, len, uio); |
s = splsoftnet(); |
s = splsoftnet(); |
solock(so); |
solock(so); |
if (error != 0) { |
if (error != 0) { |
Line 1375 soreceive(struct socket *so, struct mbuf |
|
Line 1487 soreceive(struct socket *so, struct mbuf |
|
if (len == m->m_len - moff) { |
if (len == m->m_len - moff) { |
if (m->m_flags & M_EOR) |
if (m->m_flags & M_EOR) |
flags |= MSG_EOR; |
flags |= MSG_EOR; |
|
#ifdef SCTP |
|
if (m->m_flags & M_NOTIFICATION) |
|
flags |= MSG_NOTIFICATION; |
|
#endif /* SCTP */ |
if (flags & MSG_PEEK) { |
if (flags & MSG_PEEK) { |
m = m->m_next; |
m = m->m_next; |
moff = 0; |
moff = 0; |
Line 1387 soreceive(struct socket *so, struct mbuf |
|
Line 1503 soreceive(struct socket *so, struct mbuf |
|
so->so_rcv.sb_mb = m = m->m_next; |
so->so_rcv.sb_mb = m = m->m_next; |
*mp = NULL; |
*mp = NULL; |
} else { |
} else { |
MFREE(m, so->so_rcv.sb_mb); |
m = so->so_rcv.sb_mb = m_free(m); |
m = so->so_rcv.sb_mb; |
|
} |
} |
/* |
/* |
* If m != NULL, we also know that |
* If m != NULL, we also know that |
Line 1460 soreceive(struct socket *so, struct mbuf |
|
Line 1575 soreceive(struct socket *so, struct mbuf |
|
* get it filled again. |
* get it filled again. |
*/ |
*/ |
if ((pr->pr_flags & PR_WANTRCVD) && so->so_pcb) |
if ((pr->pr_flags & PR_WANTRCVD) && so->so_pcb) |
(*pr->pr_usrreq)(so, PRU_RCVD, |
(*pr->pr_usrreqs->pr_rcvd)(so, flags, l); |
NULL, (struct mbuf *)(long)flags, NULL, l); |
|
SBLASTRECORDCHK(&so->so_rcv, "soreceive sbwait 2"); |
SBLASTRECORDCHK(&so->so_rcv, "soreceive sbwait 2"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive sbwait 2"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive sbwait 2"); |
error = sbwait(&so->so_rcv); |
if (wakeup_state & SS_RESTARTSYS) |
|
error = ERESTART; |
|
else |
|
error = sbwait(&so->so_rcv); |
if (error != 0) { |
if (error != 0) { |
sbunlock(&so->so_rcv); |
sbunlock(&so->so_rcv); |
sounlock(so); |
sounlock(so); |
Line 1473 soreceive(struct socket *so, struct mbuf |
|
Line 1590 soreceive(struct socket *so, struct mbuf |
|
} |
} |
if ((m = so->so_rcv.sb_mb) != NULL) |
if ((m = so->so_rcv.sb_mb) != NULL) |
nextrecord = m->m_nextpkt; |
nextrecord = m->m_nextpkt; |
|
wakeup_state = so->so_state; |
} |
} |
} |
} |
|
|
Line 1498 soreceive(struct socket *so, struct mbuf |
|
Line 1616 soreceive(struct socket *so, struct mbuf |
|
SBLASTRECORDCHK(&so->so_rcv, "soreceive 4"); |
SBLASTRECORDCHK(&so->so_rcv, "soreceive 4"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive 4"); |
SBLASTMBUFCHK(&so->so_rcv, "soreceive 4"); |
if (pr->pr_flags & PR_WANTRCVD && so->so_pcb) |
if (pr->pr_flags & PR_WANTRCVD && so->so_pcb) |
(*pr->pr_usrreq)(so, PRU_RCVD, NULL, |
(*pr->pr_usrreqs->pr_rcvd)(so, flags, l); |
(struct mbuf *)(long)flags, NULL, l); |
|
} |
} |
if (orig_resid == uio->uio_resid && orig_resid && |
if (orig_resid == uio->uio_resid && orig_resid && |
(flags & MSG_EOR) == 0 && (so->so_state & SS_CANTRCVMORE) == 0) { |
(flags & MSG_EOR) == 0 && (so->so_state & SS_CANTRCVMORE) == 0) { |
Line 1533 soshutdown(struct socket *so, int how) |
|
Line 1650 soshutdown(struct socket *so, int how) |
|
error = 0; |
error = 0; |
} |
} |
if (how == SHUT_WR || how == SHUT_RDWR) |
if (how == SHUT_WR || how == SHUT_RDWR) |
error = (*pr->pr_usrreq)(so, PRU_SHUTDOWN, NULL, |
error = (*pr->pr_usrreqs->pr_shutdown)(so); |
NULL, NULL, NULL); |
|
|
|
return error; |
return error; |
} |
} |
|
|
void |
void |
|
sorestart(struct socket *so) |
|
{ |
|
/* |
|
* An application has called close() on an fd on which another |
|
* of its threads has called a socket system call. |
|
* Mark this and wake everyone up, and code that would block again |
|
* instead returns ERESTART. |
|
* On system call re-entry the fd is validated and EBADF returned. |
|
* Any other fd will block again on the 2nd syscall. |
|
*/ |
|
solock(so); |
|
so->so_state |= SS_RESTARTSYS; |
|
cv_broadcast(&so->so_cv); |
|
cv_broadcast(&so->so_snd.sb_cv); |
|
cv_broadcast(&so->so_rcv.sb_cv); |
|
sounlock(so); |
|
} |
|
|
|
void |
sorflush(struct socket *so) |
sorflush(struct socket *so) |
{ |
{ |
struct sockbuf *sb, asb; |
struct sockbuf *sb, asb; |
Line 1568 sorflush(struct socket *so) |
|
Line 1703 sorflush(struct socket *so) |
|
sbrelease(&asb, so); |
sbrelease(&asb, so); |
} |
} |
|
|
|
/* |
|
* internal set SOL_SOCKET options |
|
*/ |
static int |
static int |
sosetopt1(struct socket *so, int level, int optname, struct mbuf *m) |
sosetopt1(struct socket *so, const struct sockopt *sopt) |
{ |
{ |
int optval, val; |
int error = EINVAL, opt; |
struct linger *l; |
int optval = 0; /* XXX: gcc */ |
struct sockbuf *sb; |
struct linger l; |
struct timeval *tv; |
struct timeval tv; |
|
|
switch (optname) { |
switch ((opt = sopt->sopt_name)) { |
|
|
case SO_LINGER: |
case SO_ACCEPTFILTER: |
if (m == NULL || m->m_len != sizeof(struct linger)) |
error = accept_filt_setopt(so, sopt); |
return EINVAL; |
KASSERT(solocked(so)); |
l = mtod(m, struct linger *); |
|
if (l->l_linger < 0 || l->l_linger > USHRT_MAX || |
|
l->l_linger > (INT_MAX / hz)) |
|
return EDOM; |
|
so->so_linger = l->l_linger; |
|
if (l->l_onoff) |
|
so->so_options |= SO_LINGER; |
|
else |
|
so->so_options &= ~SO_LINGER; |
|
break; |
break; |
|
|
|
case SO_LINGER: |
|
error = sockopt_get(sopt, &l, sizeof(l)); |
|
solock(so); |
|
if (error) |
|
break; |
|
if (l.l_linger < 0 || l.l_linger > USHRT_MAX || |
|
l.l_linger > (INT_MAX / hz)) { |
|
error = EDOM; |
|
break; |
|
} |
|
so->so_linger = l.l_linger; |
|
if (l.l_onoff) |
|
so->so_options |= SO_LINGER; |
|
else |
|
so->so_options &= ~SO_LINGER; |
|
break; |
|
|
case SO_DEBUG: |
case SO_DEBUG: |
case SO_KEEPALIVE: |
case SO_KEEPALIVE: |
case SO_DONTROUTE: |
case SO_DONTROUTE: |
Line 1601 sosetopt1(struct socket *so, int level, |
|
Line 1747 sosetopt1(struct socket *so, int level, |
|
case SO_REUSEPORT: |
case SO_REUSEPORT: |
case SO_OOBINLINE: |
case SO_OOBINLINE: |
case SO_TIMESTAMP: |
case SO_TIMESTAMP: |
if (m == NULL || m->m_len < sizeof(int)) |
case SO_NOSIGPIPE: |
return EINVAL; |
#ifdef SO_OTIMESTAMP |
if (*mtod(m, int *)) |
case SO_OTIMESTAMP: |
so->so_options |= optname; |
#endif |
|
error = sockopt_getint(sopt, &optval); |
|
solock(so); |
|
if (error) |
|
break; |
|
if (optval) |
|
so->so_options |= opt; |
else |
else |
so->so_options &= ~optname; |
so->so_options &= ~opt; |
break; |
break; |
|
|
case SO_SNDBUF: |
case SO_SNDBUF: |
case SO_RCVBUF: |
case SO_RCVBUF: |
case SO_SNDLOWAT: |
case SO_SNDLOWAT: |
case SO_RCVLOWAT: |
case SO_RCVLOWAT: |
if (m == NULL || m->m_len < sizeof(int)) |
error = sockopt_getint(sopt, &optval); |
return EINVAL; |
solock(so); |
|
if (error) |
|
break; |
|
|
/* |
/* |
* Values < 1 make no sense for any of these |
* Values < 1 make no sense for any of these |
* options, so disallow them. |
* options, so disallow them. |
*/ |
*/ |
optval = *mtod(m, int *); |
if (optval < 1) { |
if (optval < 1) |
error = EINVAL; |
return EINVAL; |
break; |
|
} |
switch (optname) { |
|
|
|
|
switch (opt) { |
case SO_SNDBUF: |
case SO_SNDBUF: |
|
if (sbreserve(&so->so_snd, (u_long)optval, so) == 0) { |
|
error = ENOBUFS; |
|
break; |
|
} |
|
so->so_snd.sb_flags &= ~SB_AUTOSIZE; |
|
break; |
|
|
case SO_RCVBUF: |
case SO_RCVBUF: |
sb = (optname == SO_SNDBUF) ? |
if (sbreserve(&so->so_rcv, (u_long)optval, so) == 0) { |
&so->so_snd : &so->so_rcv; |
error = ENOBUFS; |
if (sbreserve(sb, (u_long)optval, so) == 0) |
break; |
return ENOBUFS; |
} |
sb->sb_flags &= ~SB_AUTOSIZE; |
so->so_rcv.sb_flags &= ~SB_AUTOSIZE; |
break; |
break; |
|
|
/* |
/* |
Line 1640 sosetopt1(struct socket *so, int level, |
|
Line 1801 sosetopt1(struct socket *so, int level, |
|
* the high-water. |
* the high-water. |
*/ |
*/ |
case SO_SNDLOWAT: |
case SO_SNDLOWAT: |
so->so_snd.sb_lowat = |
if (optval > so->so_snd.sb_hiwat) |
(optval > so->so_snd.sb_hiwat) ? |
optval = so->so_snd.sb_hiwat; |
so->so_snd.sb_hiwat : optval; |
|
|
so->so_snd.sb_lowat = optval; |
break; |
break; |
|
|
case SO_RCVLOWAT: |
case SO_RCVLOWAT: |
so->so_rcv.sb_lowat = |
if (optval > so->so_rcv.sb_hiwat) |
(optval > so->so_rcv.sb_hiwat) ? |
optval = so->so_rcv.sb_hiwat; |
so->so_rcv.sb_hiwat : optval; |
|
|
so->so_rcv.sb_lowat = optval; |
break; |
break; |
} |
} |
break; |
break; |
|
|
|
#ifdef COMPAT_50 |
|
case SO_OSNDTIMEO: |
|
case SO_ORCVTIMEO: { |
|
struct timeval50 otv; |
|
error = sockopt_get(sopt, &otv, sizeof(otv)); |
|
if (error) { |
|
solock(so); |
|
break; |
|
} |
|
timeval50_to_timeval(&otv, &tv); |
|
opt = opt == SO_OSNDTIMEO ? SO_SNDTIMEO : SO_RCVTIMEO; |
|
error = 0; |
|
/*FALLTHROUGH*/ |
|
} |
|
#endif /* COMPAT_50 */ |
|
|
case SO_SNDTIMEO: |
case SO_SNDTIMEO: |
case SO_RCVTIMEO: |
case SO_RCVTIMEO: |
if (m == NULL || m->m_len < sizeof(*tv)) |
if (error) |
return EINVAL; |
error = sockopt_get(sopt, &tv, sizeof(tv)); |
tv = mtod(m, struct timeval *); |
solock(so); |
if (tv->tv_sec > (INT_MAX - tv->tv_usec / tick) / hz) |
if (error) |
return EDOM; |
break; |
val = tv->tv_sec * hz + tv->tv_usec / tick; |
|
if (val == 0 && tv->tv_usec != 0) |
if (tv.tv_sec > (INT_MAX - tv.tv_usec / tick) / hz) { |
val = 1; |
error = EDOM; |
|
break; |
|
} |
|
|
switch (optname) { |
optval = tv.tv_sec * hz + tv.tv_usec / tick; |
|
if (optval == 0 && tv.tv_usec != 0) |
|
optval = 1; |
|
|
|
switch (opt) { |
case SO_SNDTIMEO: |
case SO_SNDTIMEO: |
so->so_snd.sb_timeo = val; |
so->so_snd.sb_timeo = optval; |
break; |
break; |
case SO_RCVTIMEO: |
case SO_RCVTIMEO: |
so->so_rcv.sb_timeo = val; |
so->so_rcv.sb_timeo = optval; |
break; |
break; |
} |
} |
break; |
break; |
|
|
default: |
default: |
return ENOPROTOOPT; |
solock(so); |
|
error = ENOPROTOOPT; |
|
break; |
} |
} |
return 0; |
KASSERT(solocked(so)); |
|
return error; |
} |
} |
|
|
int |
int |
sosetopt(struct socket *so, int level, int optname, struct mbuf *m) |
sosetopt(struct socket *so, struct sockopt *sopt) |
{ |
{ |
int error, prerr; |
int error, prerr; |
|
|
solock(so); |
if (sopt->sopt_level == SOL_SOCKET) { |
if (level == SOL_SOCKET) |
error = sosetopt1(so, sopt); |
error = sosetopt1(so, level, optname, m); |
KASSERT(solocked(so)); |
else |
} else { |
error = ENOPROTOOPT; |
error = ENOPROTOOPT; |
|
solock(so); |
|
} |
|
|
if ((error == 0 || error == ENOPROTOOPT) && |
if ((error == 0 || error == ENOPROTOOPT) && |
so->so_proto != NULL && so->so_proto->pr_ctloutput != NULL) { |
so->so_proto != NULL && so->so_proto->pr_ctloutput != NULL) { |
/* give the protocol stack a shot */ |
/* give the protocol stack a shot */ |
prerr = (*so->so_proto->pr_ctloutput)(PRCO_SETOPT, so, level, |
prerr = (*so->so_proto->pr_ctloutput)(PRCO_SETOPT, so, sopt); |
optname, &m); |
|
if (prerr == 0) |
if (prerr == 0) |
error = 0; |
error = 0; |
else if (prerr != ENOPROTOOPT) |
else if (prerr != ENOPROTOOPT) |
error = prerr; |
error = prerr; |
} else if (m != NULL) |
} |
(void)m_free(m); |
|
sounlock(so); |
sounlock(so); |
return error; |
return error; |
} |
} |
|
|
|
/* |
|
* so_setsockopt() is a wrapper providing a sockopt structure for sosetopt() |
|
*/ |
|
int |
|
so_setsockopt(struct lwp *l, struct socket *so, int level, int name, |
|
const void *val, size_t valsize) |
|
{ |
|
struct sockopt sopt; |
|
int error; |
|
|
|
KASSERT(valsize == 0 || val != NULL); |
|
|
|
sockopt_init(&sopt, level, name, valsize); |
|
sockopt_set(&sopt, val, valsize); |
|
|
|
error = sosetopt(so, &sopt); |
|
|
|
sockopt_destroy(&sopt); |
|
|
|
return error; |
|
} |
|
|
|
/* |
|
* internal get SOL_SOCKET options |
|
*/ |
|
static int |
|
sogetopt1(struct socket *so, struct sockopt *sopt) |
|
{ |
|
int error, optval, opt; |
|
struct linger l; |
|
struct timeval tv; |
|
|
|
switch ((opt = sopt->sopt_name)) { |
|
|
|
case SO_ACCEPTFILTER: |
|
error = accept_filt_getopt(so, sopt); |
|
break; |
|
|
|
case SO_LINGER: |
|
l.l_onoff = (so->so_options & SO_LINGER) ? 1 : 0; |
|
l.l_linger = so->so_linger; |
|
|
|
error = sockopt_set(sopt, &l, sizeof(l)); |
|
break; |
|
|
|
case SO_USELOOPBACK: |
|
case SO_DONTROUTE: |
|
case SO_DEBUG: |
|
case SO_KEEPALIVE: |
|
case SO_REUSEADDR: |
|
case SO_REUSEPORT: |
|
case SO_BROADCAST: |
|
case SO_OOBINLINE: |
|
case SO_TIMESTAMP: |
|
case SO_NOSIGPIPE: |
|
#ifdef SO_OTIMESTAMP |
|
case SO_OTIMESTAMP: |
|
#endif |
|
case SO_ACCEPTCONN: |
|
error = sockopt_setint(sopt, (so->so_options & opt) ? 1 : 0); |
|
break; |
|
|
|
case SO_TYPE: |
|
error = sockopt_setint(sopt, so->so_type); |
|
break; |
|
|
|
case SO_ERROR: |
|
error = sockopt_setint(sopt, so->so_error); |
|
so->so_error = 0; |
|
break; |
|
|
|
case SO_SNDBUF: |
|
error = sockopt_setint(sopt, so->so_snd.sb_hiwat); |
|
break; |
|
|
|
case SO_RCVBUF: |
|
error = sockopt_setint(sopt, so->so_rcv.sb_hiwat); |
|
break; |
|
|
|
case SO_SNDLOWAT: |
|
error = sockopt_setint(sopt, so->so_snd.sb_lowat); |
|
break; |
|
|
|
case SO_RCVLOWAT: |
|
error = sockopt_setint(sopt, so->so_rcv.sb_lowat); |
|
break; |
|
|
|
#ifdef COMPAT_50 |
|
case SO_OSNDTIMEO: |
|
case SO_ORCVTIMEO: { |
|
struct timeval50 otv; |
|
|
|
optval = (opt == SO_OSNDTIMEO ? |
|
so->so_snd.sb_timeo : so->so_rcv.sb_timeo); |
|
|
|
otv.tv_sec = optval / hz; |
|
otv.tv_usec = (optval % hz) * tick; |
|
|
|
error = sockopt_set(sopt, &otv, sizeof(otv)); |
|
break; |
|
} |
|
#endif /* COMPAT_50 */ |
|
|
|
case SO_SNDTIMEO: |
|
case SO_RCVTIMEO: |
|
optval = (opt == SO_SNDTIMEO ? |
|
so->so_snd.sb_timeo : so->so_rcv.sb_timeo); |
|
|
|
tv.tv_sec = optval / hz; |
|
tv.tv_usec = (optval % hz) * tick; |
|
|
|
error = sockopt_set(sopt, &tv, sizeof(tv)); |
|
break; |
|
|
|
case SO_OVERFLOWED: |
|
error = sockopt_setint(sopt, so->so_rcv.sb_overflowed); |
|
break; |
|
|
|
default: |
|
error = ENOPROTOOPT; |
|
break; |
|
} |
|
|
|
return (error); |
|
} |
|
|
int |
int |
sogetopt(struct socket *so, int level, int optname, struct mbuf **mp) |
sogetopt(struct socket *so, struct sockopt *sopt) |
{ |
{ |
struct mbuf *m; |
|
int error; |
int error; |
|
|
solock(so); |
solock(so); |
if (level != SOL_SOCKET) { |
if (sopt->sopt_level != SOL_SOCKET) { |
if (so->so_proto && so->so_proto->pr_ctloutput) { |
if (so->so_proto && so->so_proto->pr_ctloutput) { |
error = ((*so->so_proto->pr_ctloutput) |
error = ((*so->so_proto->pr_ctloutput) |
(PRCO_GETOPT, so, level, optname, mp)); |
(PRCO_GETOPT, so, sopt)); |
} else |
} else |
error = (ENOPROTOOPT); |
error = (ENOPROTOOPT); |
} else { |
} else { |
m = m_get(M_WAIT, MT_SOOPTS); |
error = sogetopt1(so, sopt); |
m->m_len = sizeof(int); |
} |
|
sounlock(so); |
|
return (error); |
|
} |
|
|
switch (optname) { |
/* |
|
* alloc sockopt data buffer buffer |
|
* - will be released at destroy |
|
*/ |
|
static int |
|
sockopt_alloc(struct sockopt *sopt, size_t len, km_flag_t kmflag) |
|
{ |
|
|
case SO_LINGER: |
KASSERT(sopt->sopt_size == 0); |
m->m_len = sizeof(struct linger); |
|
mtod(m, struct linger *)->l_onoff = |
|
(so->so_options & SO_LINGER) ? 1 : 0; |
|
mtod(m, struct linger *)->l_linger = so->so_linger; |
|
break; |
|
|
|
case SO_USELOOPBACK: |
if (len > sizeof(sopt->sopt_buf)) { |
case SO_DONTROUTE: |
sopt->sopt_data = kmem_zalloc(len, kmflag); |
case SO_DEBUG: |
if (sopt->sopt_data == NULL) |
case SO_KEEPALIVE: |
return ENOMEM; |
case SO_REUSEADDR: |
} else |
case SO_REUSEPORT: |
sopt->sopt_data = sopt->sopt_buf; |
case SO_BROADCAST: |
|
case SO_OOBINLINE: |
|
case SO_TIMESTAMP: |
|
*mtod(m, int *) = (so->so_options & optname) ? 1 : 0; |
|
break; |
|
|
|
case SO_TYPE: |
sopt->sopt_size = len; |
*mtod(m, int *) = so->so_type; |
return 0; |
break; |
} |
|
|
case SO_ERROR: |
/* |
*mtod(m, int *) = so->so_error; |
* initialise sockopt storage |
so->so_error = 0; |
* - MAY sleep during allocation |
break; |
*/ |
|
void |
|
sockopt_init(struct sockopt *sopt, int level, int name, size_t size) |
|
{ |
|
|
case SO_SNDBUF: |
memset(sopt, 0, sizeof(*sopt)); |
*mtod(m, int *) = so->so_snd.sb_hiwat; |
|
break; |
|
|
|
case SO_RCVBUF: |
sopt->sopt_level = level; |
*mtod(m, int *) = so->so_rcv.sb_hiwat; |
sopt->sopt_name = name; |
break; |
(void)sockopt_alloc(sopt, size, KM_SLEEP); |
|
} |
|
|
case SO_SNDLOWAT: |
/* |
*mtod(m, int *) = so->so_snd.sb_lowat; |
* destroy sockopt storage |
break; |
* - will release any held memory references |
|
*/ |
|
void |
|
sockopt_destroy(struct sockopt *sopt) |
|
{ |
|
|
case SO_RCVLOWAT: |
if (sopt->sopt_data != sopt->sopt_buf) |
*mtod(m, int *) = so->so_rcv.sb_lowat; |
kmem_free(sopt->sopt_data, sopt->sopt_size); |
break; |
|
|
|
case SO_SNDTIMEO: |
memset(sopt, 0, sizeof(*sopt)); |
case SO_RCVTIMEO: |
} |
{ |
|
int val = (optname == SO_SNDTIMEO ? |
|
so->so_snd.sb_timeo : so->so_rcv.sb_timeo); |
|
|
|
m->m_len = sizeof(struct timeval); |
|
mtod(m, struct timeval *)->tv_sec = val / hz; |
|
mtod(m, struct timeval *)->tv_usec = |
|
(val % hz) * tick; |
|
break; |
|
} |
|
|
|
case SO_OVERFLOWED: |
/* |
*mtod(m, int *) = so->so_rcv.sb_overflowed; |
* set sockopt value |
break; |
* - value is copied into sockopt |
|
* - memory is allocated when necessary, will not sleep |
|
*/ |
|
int |
|
sockopt_set(struct sockopt *sopt, const void *buf, size_t len) |
|
{ |
|
int error; |
|
|
default: |
if (sopt->sopt_size == 0) { |
sounlock(so); |
error = sockopt_alloc(sopt, len, KM_NOSLEEP); |
(void)m_free(m); |
if (error) |
return (ENOPROTOOPT); |
return error; |
|
} |
|
|
|
KASSERT(sopt->sopt_size == len); |
|
memcpy(sopt->sopt_data, buf, len); |
|
return 0; |
|
} |
|
|
|
/* |
|
* common case of set sockopt integer value |
|
*/ |
|
int |
|
sockopt_setint(struct sockopt *sopt, int val) |
|
{ |
|
|
|
return sockopt_set(sopt, &val, sizeof(int)); |
|
} |
|
|
|
/* |
|
* get sockopt value |
|
* - correct size must be given |
|
*/ |
|
int |
|
sockopt_get(const struct sockopt *sopt, void *buf, size_t len) |
|
{ |
|
|
|
if (sopt->sopt_size != len) |
|
return EINVAL; |
|
|
|
memcpy(buf, sopt->sopt_data, len); |
|
return 0; |
|
} |
|
|
|
/* |
|
* common case of get sockopt integer value |
|
*/ |
|
int |
|
sockopt_getint(const struct sockopt *sopt, int *valp) |
|
{ |
|
|
|
return sockopt_get(sopt, valp, sizeof(int)); |
|
} |
|
|
|
/* |
|
* set sockopt value from mbuf |
|
* - ONLY for legacy code |
|
* - mbuf is released by sockopt |
|
* - will not sleep |
|
*/ |
|
int |
|
sockopt_setmbuf(struct sockopt *sopt, struct mbuf *m) |
|
{ |
|
size_t len; |
|
int error; |
|
|
|
len = m_length(m); |
|
|
|
if (sopt->sopt_size == 0) { |
|
error = sockopt_alloc(sopt, len, KM_NOSLEEP); |
|
if (error) |
|
return error; |
|
} |
|
|
|
KASSERT(sopt->sopt_size == len); |
|
m_copydata(m, 0, len, sopt->sopt_data); |
|
m_freem(m); |
|
|
|
return 0; |
|
} |
|
|
|
/* |
|
* get sockopt value into mbuf |
|
* - ONLY for legacy code |
|
* - mbuf to be released by the caller |
|
* - will not sleep |
|
*/ |
|
struct mbuf * |
|
sockopt_getmbuf(const struct sockopt *sopt) |
|
{ |
|
struct mbuf *m; |
|
|
|
if (sopt->sopt_size > MCLBYTES) |
|
return NULL; |
|
|
|
m = m_get(M_DONTWAIT, MT_SOOPTS); |
|
if (m == NULL) |
|
return NULL; |
|
|
|
if (sopt->sopt_size > MLEN) { |
|
MCLGET(m, M_DONTWAIT); |
|
if ((m->m_flags & M_EXT) == 0) { |
|
m_free(m); |
|
return NULL; |
} |
} |
*mp = m; |
|
error = 0; |
|
} |
} |
|
|
sounlock(so); |
memcpy(mtod(m, void *), sopt->sopt_data, sopt->sopt_size); |
return (error); |
m->m_len = sopt->sopt_size; |
|
|
|
return m; |
} |
} |
|
|
void |
void |
Line 1804 sohasoutofband(struct socket *so) |
|
Line 2209 sohasoutofband(struct socket *so) |
|
{ |
{ |
|
|
fownsignal(so->so_pgid, SIGURG, POLL_PRI, POLLPRI|POLLRDBAND, so); |
fownsignal(so->so_pgid, SIGURG, POLL_PRI, POLLPRI|POLLRDBAND, so); |
selnotify(&so->so_rcv.sb_sel, POLLPRI | POLLRDBAND, 0); |
selnotify(&so->so_rcv.sb_sel, POLLPRI | POLLRDBAND, NOTE_SUBMIT); |
} |
} |
|
|
static void |
static void |
Line 1812 filt_sordetach(struct knote *kn) |
|
Line 2217 filt_sordetach(struct knote *kn) |
|
{ |
{ |
struct socket *so; |
struct socket *so; |
|
|
so = ((file_t *)kn->kn_obj)->f_data; |
so = ((file_t *)kn->kn_obj)->f_socket; |
solock(so); |
solock(so); |
SLIST_REMOVE(&so->so_rcv.sb_sel.sel_klist, kn, knote, kn_selnext); |
SLIST_REMOVE(&so->so_rcv.sb_sel.sel_klist, kn, knote, kn_selnext); |
if (SLIST_EMPTY(&so->so_rcv.sb_sel.sel_klist)) |
if (SLIST_EMPTY(&so->so_rcv.sb_sel.sel_klist)) |
Line 1827 filt_soread(struct knote *kn, long hint) |
|
Line 2232 filt_soread(struct knote *kn, long hint) |
|
struct socket *so; |
struct socket *so; |
int rv; |
int rv; |
|
|
so = ((file_t *)kn->kn_obj)->f_data; |
so = ((file_t *)kn->kn_obj)->f_socket; |
if (hint != NOTE_SUBMIT) |
if (hint != NOTE_SUBMIT) |
solock(so); |
solock(so); |
kn->kn_data = so->so_rcv.sb_cc; |
kn->kn_data = so->so_rcv.sb_cc; |
Line 1839 filt_soread(struct knote *kn, long hint) |
|
Line 2244 filt_soread(struct knote *kn, long hint) |
|
rv = 1; |
rv = 1; |
else if (kn->kn_sfflags & NOTE_LOWAT) |
else if (kn->kn_sfflags & NOTE_LOWAT) |
rv = (kn->kn_data >= kn->kn_sdata); |
rv = (kn->kn_data >= kn->kn_sdata); |
else |
else |
rv = (kn->kn_data >= so->so_rcv.sb_lowat); |
rv = (kn->kn_data >= so->so_rcv.sb_lowat); |
if (hint != NOTE_SUBMIT) |
if (hint != NOTE_SUBMIT) |
sounlock(so); |
sounlock(so); |
Line 1851 filt_sowdetach(struct knote *kn) |
|
Line 2256 filt_sowdetach(struct knote *kn) |
|
{ |
{ |
struct socket *so; |
struct socket *so; |
|
|
so = ((file_t *)kn->kn_obj)->f_data; |
so = ((file_t *)kn->kn_obj)->f_socket; |
solock(so); |
solock(so); |
SLIST_REMOVE(&so->so_snd.sb_sel.sel_klist, kn, knote, kn_selnext); |
SLIST_REMOVE(&so->so_snd.sb_sel.sel_klist, kn, knote, kn_selnext); |
if (SLIST_EMPTY(&so->so_snd.sb_sel.sel_klist)) |
if (SLIST_EMPTY(&so->so_snd.sb_sel.sel_klist)) |
Line 1866 filt_sowrite(struct knote *kn, long hint |
|
Line 2271 filt_sowrite(struct knote *kn, long hint |
|
struct socket *so; |
struct socket *so; |
int rv; |
int rv; |
|
|
so = ((file_t *)kn->kn_obj)->f_data; |
so = ((file_t *)kn->kn_obj)->f_socket; |
if (hint != NOTE_SUBMIT) |
if (hint != NOTE_SUBMIT) |
solock(so); |
solock(so); |
kn->kn_data = sbspace(&so->so_snd); |
kn->kn_data = sbspace(&so->so_snd); |
Line 1895 filt_solisten(struct knote *kn, long hin |
|
Line 2300 filt_solisten(struct knote *kn, long hin |
|
struct socket *so; |
struct socket *so; |
int rv; |
int rv; |
|
|
so = ((file_t *)kn->kn_obj)->f_data; |
so = ((file_t *)kn->kn_obj)->f_socket; |
|
|
/* |
/* |
* Set kn_data to number of incoming connections, not |
* Set kn_data to number of incoming connections, not |
Line 1923 soo_kqfilter(struct file *fp, struct kno |
|
Line 2328 soo_kqfilter(struct file *fp, struct kno |
|
struct socket *so; |
struct socket *so; |
struct sockbuf *sb; |
struct sockbuf *sb; |
|
|
so = ((file_t *)kn->kn_obj)->f_data; |
so = ((file_t *)kn->kn_obj)->f_socket; |
solock(so); |
solock(so); |
switch (kn->kn_filter) { |
switch (kn->kn_filter) { |
case EVFILT_READ: |
case EVFILT_READ: |
Line 2005 sopoll(struct socket *so, int events) |
|
Line 2410 sopoll(struct socket *so, int events) |
|
#include <sys/sysctl.h> |
#include <sys/sysctl.h> |
|
|
static int sysctl_kern_somaxkva(SYSCTLFN_PROTO); |
static int sysctl_kern_somaxkva(SYSCTLFN_PROTO); |
|
static int sysctl_kern_sbmax(SYSCTLFN_PROTO); |
|
|
/* |
/* |
* sysctl helper routine for kern.somaxkva. ensures that the given |
* sysctl helper routine for kern.somaxkva. ensures that the given |
Line 2035 sysctl_kern_somaxkva(SYSCTLFN_ARGS) |
|
Line 2441 sysctl_kern_somaxkva(SYSCTLFN_ARGS) |
|
return (error); |
return (error); |
} |
} |
|
|
SYSCTL_SETUP(sysctl_kern_somaxkva_setup, "sysctl kern.somaxkva setup") |
/* |
|
* sysctl helper routine for kern.sbmax. Basically just ensures that |
|
* any new value is not too small. |
|
*/ |
|
static int |
|
sysctl_kern_sbmax(SYSCTLFN_ARGS) |
|
{ |
|
int error, new_sbmax; |
|
struct sysctlnode node; |
|
|
|
new_sbmax = sb_max; |
|
node = *rnode; |
|
node.sysctl_data = &new_sbmax; |
|
error = sysctl_lookup(SYSCTLFN_CALL(&node)); |
|
if (error || newp == NULL) |
|
return (error); |
|
|
|
KERNEL_LOCK(1, NULL); |
|
error = sb_max_set(new_sbmax); |
|
KERNEL_UNLOCK_ONE(NULL); |
|
|
|
return (error); |
|
} |
|
|
|
static void |
|
sysctl_kern_socket_setup(void) |
{ |
{ |
|
|
sysctl_createv(clog, 0, NULL, NULL, |
KASSERT(socket_sysctllog == NULL); |
CTLFLAG_PERMANENT, |
|
CTLTYPE_NODE, "kern", NULL, |
|
NULL, 0, NULL, 0, |
|
CTL_KERN, CTL_EOL); |
|
|
|
sysctl_createv(clog, 0, NULL, NULL, |
sysctl_createv(&socket_sysctllog, 0, NULL, NULL, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLTYPE_INT, "somaxkva", |
CTLTYPE_INT, "somaxkva", |
SYSCTL_DESCR("Maximum amount of kernel memory to be " |
SYSCTL_DESCR("Maximum amount of kernel memory to be " |
"used for socket buffers"), |
"used for socket buffers"), |
sysctl_kern_somaxkva, 0, NULL, 0, |
sysctl_kern_somaxkva, 0, NULL, 0, |
CTL_KERN, KERN_SOMAXKVA, CTL_EOL); |
CTL_KERN, KERN_SOMAXKVA, CTL_EOL); |
|
|
|
sysctl_createv(&socket_sysctllog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "sbmax", |
|
SYSCTL_DESCR("Maximum socket buffer size"), |
|
sysctl_kern_sbmax, 0, NULL, 0, |
|
CTL_KERN, KERN_SBMAX, CTL_EOL); |
} |
} |