version 1.253.2.1, 2007/10/06 15:29:02 |
version 1.335, 2016/07/06 08:42:34 |
|
|
* 2. Redistributions in binary form must reproduce the above copyright |
* 2. Redistributions in binary form must reproduce the above copyright |
* notice, this list of conditions and the following disclaimer in the |
* notice, this list of conditions and the following disclaimer in the |
* documentation and/or other materials provided with the distribution. |
* documentation and/or other materials provided with the distribution. |
* 3. All advertising materials mentioning features or use of this software |
|
* must display the following acknowledgement: |
|
* This product includes software developed by the NetBSD |
|
* Foundation, Inc. and its contributors. |
|
* 4. Neither the name of The NetBSD Foundation nor the names of its |
|
* contributors may be used to endorse or promote products derived |
|
* from this software without specific prior written permission. |
|
* |
* |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
|
|
#include <sys/cdefs.h> |
#include <sys/cdefs.h> |
__KERNEL_RCSID(0, "$NetBSD$"); |
__KERNEL_RCSID(0, "$NetBSD$"); |
|
|
|
#ifdef _KERNEL_OPT |
#include "opt_inet.h" |
#include "opt_inet.h" |
|
#include "opt_compat_netbsd.h" |
#include "opt_gateway.h" |
#include "opt_gateway.h" |
#include "opt_pfil_hooks.h" |
|
#include "opt_ipsec.h" |
#include "opt_ipsec.h" |
#include "opt_mrouting.h" |
#include "opt_mrouting.h" |
#include "opt_mbuftrace.h" |
#include "opt_mbuftrace.h" |
#include "opt_inet_csum.h" |
#include "opt_inet_csum.h" |
|
#endif |
|
|
|
#include "arp.h" |
|
|
#include <sys/param.h> |
#include <sys/param.h> |
#include <sys/systm.h> |
#include <sys/systm.h> |
#include <sys/malloc.h> |
#include <sys/cpu.h> |
#include <sys/mbuf.h> |
#include <sys/mbuf.h> |
#include <sys/domain.h> |
#include <sys/domain.h> |
#include <sys/protosw.h> |
#include <sys/protosw.h> |
Line 126 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 123 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#include <net/if.h> |
#include <net/if.h> |
#include <net/if_dl.h> |
#include <net/if_dl.h> |
#include <net/route.h> |
#include <net/route.h> |
|
#include <net/pktqueue.h> |
#include <net/pfil.h> |
#include <net/pfil.h> |
|
|
#include <netinet/in.h> |
#include <netinet/in.h> |
Line 135 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 133 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#include <netinet/in_proto.h> |
#include <netinet/in_proto.h> |
#include <netinet/in_var.h> |
#include <netinet/in_var.h> |
#include <netinet/ip_var.h> |
#include <netinet/ip_var.h> |
|
#include <netinet/ip_private.h> |
#include <netinet/ip_icmp.h> |
#include <netinet/ip_icmp.h> |
/* just for gif_ttl */ |
/* just for gif_ttl */ |
#include <netinet/in_gif.h> |
#include <netinet/in_gif.h> |
Line 145 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 144 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#ifdef MROUTING |
#ifdef MROUTING |
#include <netinet/ip_mroute.h> |
#include <netinet/ip_mroute.h> |
#endif |
#endif |
|
#include <netinet/portalgo.h> |
|
|
#ifdef IPSEC |
#ifdef IPSEC |
#include <netinet6/ipsec.h> |
|
#include <netkey/key.h> |
|
#endif |
|
#ifdef FAST_IPSEC |
|
#include <netipsec/ipsec.h> |
#include <netipsec/ipsec.h> |
#include <netipsec/key.h> |
#endif |
#endif /* FAST_IPSEC*/ |
|
|
|
#ifndef IPFORWARDING |
#ifndef IPFORWARDING |
#ifdef GATEWAY |
#ifdef GATEWAY |
Line 178 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 173 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#define IPMTUDISCTIMEOUT (10 * 60) /* as per RFC 1191 */ |
#define IPMTUDISCTIMEOUT (10 * 60) /* as per RFC 1191 */ |
#endif |
#endif |
|
|
|
#ifdef COMPAT_50 |
|
#include <compat/sys/time.h> |
|
#include <compat/sys/socket.h> |
|
#endif |
|
|
/* |
/* |
* Note: DIRECTED_BROADCAST is handled this way so that previous |
* Note: DIRECTED_BROADCAST is handled this way so that previous |
* configuration using this option will Just Work. |
* configuration using this option will Just Work. |
Line 216 int ip_do_randomid = 0; |
|
Line 216 int ip_do_randomid = 0; |
|
* to the loopback interface instead of the interface where the |
* to the loopback interface instead of the interface where the |
* packets for those addresses are received. |
* packets for those addresses are received. |
*/ |
*/ |
int ip_checkinterface = 0; |
static int ip_checkinterface __read_mostly = 0; |
|
|
|
|
struct rttimer_queue *ip_mtudisc_timeout_q = NULL; |
struct rttimer_queue *ip_mtudisc_timeout_q = NULL; |
|
|
int ipqmaxlen = IFQ_MAXLEN; |
pktqueue_t * ip_pktq __read_mostly; |
u_long in_ifaddrhash; /* size of hash table - 1 */ |
pfil_head_t * inet_pfil_hook __read_mostly; |
int in_ifaddrentries; /* total number of addrs */ |
ipid_state_t * ip_ids __read_mostly; |
struct in_ifaddrhead in_ifaddrhead; |
percpu_t * ipstat_percpu __read_mostly; |
struct in_ifaddrhashhead *in_ifaddrhashtbl; |
|
u_long in_multihash; /* size of hash table - 1 */ |
|
int in_multientries; /* total number of addrs */ |
|
struct in_multihashhead *in_multihashtbl; |
|
struct ifqueue ipintrq; |
|
struct ipstat ipstat; |
|
uint16_t ip_id; |
|
|
|
#ifdef PFIL_HOOKS |
|
struct pfil_head inet_pfil_hook; |
|
#endif |
|
|
|
/* |
|
* Cached copy of nmbclusters. If nbclusters is different, |
|
* recalculate IP parameters derived from nmbclusters. |
|
*/ |
|
static int ip_nmbclusters; /* copy of nmbclusters */ |
|
static void ip_nmbclusters_changed(void); /* recalc limits */ |
|
|
|
#define CHECK_NMBCLUSTER_PARAMS() \ |
|
do { \ |
|
if (__predict_false(ip_nmbclusters != nmbclusters)) \ |
|
ip_nmbclusters_changed(); \ |
|
} while (/*CONSTCOND*/0) |
|
|
|
/* IP datagram reassembly queues (hashed) */ |
|
#define IPREASS_NHASH_LOG2 6 |
|
#define IPREASS_NHASH (1 << IPREASS_NHASH_LOG2) |
|
#define IPREASS_HMASK (IPREASS_NHASH - 1) |
|
#define IPREASS_HASH(x,y) \ |
|
(((((x) & 0xF) | ((((x) >> 8) & 0xF) << 4)) ^ (y)) & IPREASS_HMASK) |
|
struct ipqhead ipq[IPREASS_NHASH]; |
|
int ipq_locked; |
|
static int ip_nfragpackets; /* packets in reass queue */ |
|
static int ip_nfrags; /* total fragments in reass queues */ |
|
|
|
int ip_maxfragpackets = 200; /* limit on packets. XXX sysctl */ |
|
int ip_maxfrags; /* limit on fragments. XXX sysctl */ |
|
|
|
|
static struct route ipforward_rt __cacheline_aligned; |
|
|
/* |
uint16_t ip_id; |
* Additive-Increase/Multiplicative-Decrease (AIMD) strategy for |
|
* IP reassembly queue buffer managment. |
|
* |
|
* We keep a count of total IP fragments (NB: not fragmented packets!) |
|
* awaiting reassembly (ip_nfrags) and a limit (ip_maxfrags) on fragments. |
|
* If ip_nfrags exceeds ip_maxfrags the limit, we drop half the |
|
* total fragments in reassembly queues.This AIMD policy avoids |
|
* repeatedly deleting single packets under heavy fragmentation load |
|
* (e.g., from lossy NFS peers). |
|
*/ |
|
static u_int ip_reass_ttl_decr(u_int ticks); |
|
static void ip_reass_drophalf(void); |
|
|
|
|
|
static inline int ipq_lock_try(void); |
|
static inline void ipq_unlock(void); |
|
|
|
static inline int |
|
ipq_lock_try(void) |
|
{ |
|
int s; |
|
|
|
/* |
|
* Use splvm() -- we're blocking things that would cause |
|
* mbuf allocation. |
|
*/ |
|
s = splvm(); |
|
if (ipq_locked) { |
|
splx(s); |
|
return (0); |
|
} |
|
ipq_locked = 1; |
|
splx(s); |
|
return (1); |
|
} |
|
|
|
static inline void |
|
ipq_unlock(void) |
|
{ |
|
int s; |
|
|
|
s = splvm(); |
|
ipq_locked = 0; |
|
splx(s); |
|
} |
|
|
|
#ifdef DIAGNOSTIC |
|
#define IPQ_LOCK() \ |
|
do { \ |
|
if (ipq_lock_try() == 0) { \ |
|
printf("%s:%d: ipq already locked\n", __FILE__, __LINE__); \ |
|
panic("ipq_lock"); \ |
|
} \ |
|
} while (/*CONSTCOND*/ 0) |
|
#define IPQ_LOCK_CHECK() \ |
|
do { \ |
|
if (ipq_locked == 0) { \ |
|
printf("%s:%d: ipq lock not held\n", __FILE__, __LINE__); \ |
|
panic("ipq lock check"); \ |
|
} \ |
|
} while (/*CONSTCOND*/ 0) |
|
#else |
|
#define IPQ_LOCK() (void) ipq_lock_try() |
|
#define IPQ_LOCK_CHECK() /* nothing */ |
|
#endif |
|
|
|
#define IPQ_UNLOCK() ipq_unlock() |
|
|
|
POOL_INIT(inmulti_pool, sizeof(struct in_multi), 0, 0, 0, "inmltpl", NULL, |
|
IPL_SOFTNET); |
|
POOL_INIT(ipqent_pool, sizeof(struct ipqent), 0, 0, 0, "ipqepl", NULL, |
|
IPL_VM); |
|
|
|
#ifdef INET_CSUM_COUNTERS |
#ifdef INET_CSUM_COUNTERS |
#include <sys/device.h> |
#include <sys/device.h> |
Line 368 EVCNT_ATTACH_STATIC(ip_swcsum); |
|
Line 258 EVCNT_ATTACH_STATIC(ip_swcsum); |
|
* maintenance when the remote end is on a network that is not known |
* maintenance when the remote end is on a network that is not known |
* to us. |
* to us. |
*/ |
*/ |
int ip_nhops = 0; |
|
|
static int ip_nhops = 0; |
|
|
static struct ip_srcrt { |
static struct ip_srcrt { |
struct in_addr dst; /* final destination */ |
struct in_addr dst; /* final destination */ |
char nop; /* one NOP to align */ |
char nop; /* one NOP to align */ |
Line 376 static struct ip_srcrt { |
|
Line 268 static struct ip_srcrt { |
|
struct in_addr route[MAX_IPOPTLEN/sizeof(struct in_addr)]; |
struct in_addr route[MAX_IPOPTLEN/sizeof(struct in_addr)]; |
} ip_srcrt; |
} ip_srcrt; |
|
|
|
static int ip_drainwanted; |
|
|
|
struct sockaddr_in ipaddr = { |
|
.sin_len = sizeof(ipaddr), |
|
.sin_family = AF_INET, |
|
}; |
|
|
static void save_rte(u_char *, struct in_addr); |
static void save_rte(u_char *, struct in_addr); |
|
|
#ifdef MBUFTRACE |
#ifdef MBUFTRACE |
Line 383 struct mowner ip_rx_mowner = MOWNER_INIT |
|
Line 282 struct mowner ip_rx_mowner = MOWNER_INIT |
|
struct mowner ip_tx_mowner = MOWNER_INIT("internet", "tx"); |
struct mowner ip_tx_mowner = MOWNER_INIT("internet", "tx"); |
#endif |
#endif |
|
|
/* |
static void ipintr(void *); |
* Compute IP limits derived from the value of nmbclusters. |
static void ip_input(struct mbuf *); |
*/ |
static void ip_forward(struct mbuf *, int, struct ifnet *); |
static void |
static bool ip_dooptions(struct mbuf *); |
ip_nmbclusters_changed(void) |
static struct in_ifaddr *ip_rtaddr(struct in_addr); |
{ |
static void sysctl_net_inet_ip_setup(struct sysctllog **); |
ip_maxfrags = nmbclusters / 4; |
|
ip_nmbclusters = nmbclusters; |
static struct in_ifaddr *ip_match_our_address(struct ifnet *, struct ip *, |
} |
int *); |
|
static struct in_ifaddr *ip_match_our_address_broadcast(struct ifnet *, |
|
struct ip *); |
|
|
|
/* XXX: Not yet enabled. */ |
|
#define SOFTNET_LOCK() KASSERT(mutex_owned(softnet_lock)) |
|
#define SOFTNET_UNLOCK() KASSERT(mutex_owned(softnet_lock)) |
|
|
/* |
/* |
* IP initialization: fill in IP protocol switch table. |
* IP initialization: fill in IP protocol switch table. |
|
|
ip_init(void) |
ip_init(void) |
{ |
{ |
const struct protosw *pr; |
const struct protosw *pr; |
int i; |
|
|
in_init(); |
|
sysctl_net_inet_ip_setup(NULL); |
|
|
pr = pffindproto(PF_INET, IPPROTO_RAW, SOCK_RAW); |
pr = pffindproto(PF_INET, IPPROTO_RAW, SOCK_RAW); |
if (pr == 0) |
KASSERT(pr != NULL); |
panic("ip_init"); |
|
for (i = 0; i < IPPROTO_MAX; i++) |
ip_pktq = pktq_create(IFQ_MAXLEN, ipintr, NULL); |
|
KASSERT(ip_pktq != NULL); |
|
|
|
for (u_int i = 0; i < IPPROTO_MAX; i++) { |
ip_protox[i] = pr - inetsw; |
ip_protox[i] = pr - inetsw; |
|
} |
for (pr = inetdomain.dom_protosw; |
for (pr = inetdomain.dom_protosw; |
pr < inetdomain.dom_protoswNPROTOSW; pr++) |
pr < inetdomain.dom_protoswNPROTOSW; pr++) |
if (pr->pr_domain->dom_family == PF_INET && |
if (pr->pr_domain->dom_family == PF_INET && |
pr->pr_protocol && pr->pr_protocol != IPPROTO_RAW) |
pr->pr_protocol && pr->pr_protocol != IPPROTO_RAW) |
ip_protox[pr->pr_protocol] = pr - inetsw; |
ip_protox[pr->pr_protocol] = pr - inetsw; |
|
|
for (i = 0; i < IPREASS_NHASH; i++) |
ip_reass_init(); |
LIST_INIT(&ipq[i]); |
|
|
|
ip_id = time_second & 0xfffff; |
|
|
|
ipintrq.ifq_maxlen = ipqmaxlen; |
ip_ids = ip_id_init(); |
ip_nmbclusters_changed(); |
ip_id = time_uptime & 0xfffff; |
|
|
TAILQ_INIT(&in_ifaddrhead); |
|
in_ifaddrhashtbl = hashinit(IN_IFADDR_HASH_SIZE, HASH_LIST, M_IFADDR, |
|
M_WAITOK, &in_ifaddrhash); |
|
in_multihashtbl = hashinit(IN_IFADDR_HASH_SIZE, HASH_LIST, M_IPMADDR, |
|
M_WAITOK, &in_multihash); |
|
ip_mtudisc_timeout_q = rt_timer_queue_create(ip_mtudisc_timeout); |
ip_mtudisc_timeout_q = rt_timer_queue_create(ip_mtudisc_timeout); |
#ifdef GATEWAY |
#ifdef GATEWAY |
ipflow_init(ip_hashsize); |
ipflow_init(); |
#endif |
#endif |
|
|
#ifdef PFIL_HOOKS |
|
/* Register our Packet Filter hook. */ |
/* Register our Packet Filter hook. */ |
inet_pfil_hook.ph_type = PFIL_TYPE_AF; |
inet_pfil_hook = pfil_head_create(PFIL_TYPE_AF, (void *)AF_INET); |
inet_pfil_hook.ph_af = AF_INET; |
KASSERT(inet_pfil_hook != NULL); |
i = pfil_head_register(&inet_pfil_hook); |
|
if (i != 0) |
|
printf("ip_init: WARNING: unable to register pfil hook, " |
|
"error %d\n", i); |
|
#endif /* PFIL_HOOKS */ |
|
|
|
#ifdef MBUFTRACE |
#ifdef MBUFTRACE |
MOWNER_ATTACH(&ip_tx_mowner); |
MOWNER_ATTACH(&ip_tx_mowner); |
MOWNER_ATTACH(&ip_rx_mowner); |
MOWNER_ATTACH(&ip_rx_mowner); |
#endif /* MBUFTRACE */ |
#endif /* MBUFTRACE */ |
|
|
|
ipstat_percpu = percpu_alloc(sizeof(uint64_t) * IP_NSTATS); |
} |
} |
|
|
struct sockaddr_in ipaddr = { |
static struct in_ifaddr * |
.sin_len = sizeof(ipaddr), |
ip_match_our_address(struct ifnet *ifp, struct ip *ip, int *downmatch) |
.sin_family = AF_INET, |
{ |
}; |
struct in_ifaddr *ia = NULL; |
struct route ipforward_rt; |
int checkif; |
|
|
|
/* |
|
* Enable a consistency check between the destination address |
|
* and the arrival interface for a unicast packet (the RFC 1122 |
|
* strong ES model) if IP forwarding is disabled and the packet |
|
* is not locally generated. |
|
* |
|
* XXX - Checking also should be disabled if the destination |
|
* address is ipnat'ed to a different interface. |
|
* |
|
* XXX - Checking is incompatible with IP aliases added |
|
* to the loopback interface instead of the interface where |
|
* the packets are received. |
|
* |
|
* XXX - We need to add a per ifaddr flag for this so that |
|
* we get finer grain control. |
|
*/ |
|
checkif = ip_checkinterface && (ipforwarding == 0) && |
|
(ifp->if_flags & IFF_LOOPBACK) == 0; |
|
|
|
IN_ADDRHASH_READER_FOREACH(ia, ip->ip_dst.s_addr) { |
|
if (in_hosteq(ia->ia_addr.sin_addr, ip->ip_dst)) { |
|
if (ia->ia4_flags & IN_IFF_NOTREADY) |
|
continue; |
|
if (checkif && ia->ia_ifp != ifp) |
|
continue; |
|
if ((ia->ia_ifp->if_flags & IFF_UP) != 0) |
|
break; |
|
else |
|
downmatch++; |
|
} |
|
} |
|
|
|
return ia; |
|
} |
|
|
|
static struct in_ifaddr * |
|
ip_match_our_address_broadcast(struct ifnet *ifp, struct ip *ip) |
|
{ |
|
struct in_ifaddr *ia = NULL; |
|
struct ifaddr *ifa; |
|
|
|
IFADDR_FOREACH(ifa, ifp) { |
|
if (ifa->ifa_addr->sa_family != AF_INET) |
|
continue; |
|
ia = ifatoia(ifa); |
|
if (ia->ia4_flags & IN_IFF_NOTREADY) |
|
continue; |
|
if (in_hosteq(ip->ip_dst, ia->ia_broadaddr.sin_addr) || |
|
in_hosteq(ip->ip_dst, ia->ia_netbroadcast) || |
|
/* |
|
* Look for all-0's host part (old broadcast addr), |
|
* either for subnet or net. |
|
*/ |
|
ip->ip_dst.s_addr == ia->ia_subnet || |
|
ip->ip_dst.s_addr == ia->ia_net) |
|
goto matched; |
|
/* |
|
* An interface with IP address zero accepts |
|
* all packets that arrive on that interface. |
|
*/ |
|
if (in_nullhost(ia->ia_addr.sin_addr)) |
|
goto matched; |
|
} |
|
ia = NULL; |
|
|
|
matched: |
|
return ia; |
|
} |
|
|
/* |
/* |
* IP software interrupt routine |
* IP software interrupt routine. |
*/ |
*/ |
void |
static void |
ipintr(void) |
ipintr(void *arg __unused) |
{ |
{ |
int s; |
|
struct mbuf *m; |
struct mbuf *m; |
|
|
while (!IF_IS_EMPTY(&ipintrq)) { |
KASSERT(cpu_softintr_p()); |
s = splnet(); |
|
IF_DEQUEUE(&ipintrq, m); |
mutex_enter(softnet_lock); |
splx(s); |
while ((m = pktq_dequeue(ip_pktq)) != NULL) { |
if (m == 0) |
|
return; |
|
MCLAIM(m, &ip_rx_mowner); |
|
ip_input(m); |
ip_input(m); |
} |
} |
|
mutex_exit(softnet_lock); |
} |
} |
|
|
/* |
/* |
* Ip input routine. Checksum and byte swap header. If fragmented |
* IP input routine. Checksum and byte swap header. If fragmented |
* try to reassemble. Process options. Pass to next level. |
* try to reassemble. Process options. Pass to next level. |
*/ |
*/ |
void |
static void |
ip_input(struct mbuf *m) |
ip_input(struct mbuf *m) |
{ |
{ |
struct ip *ip = NULL; |
struct ip *ip = NULL; |
struct ipq *fp; |
|
struct in_ifaddr *ia; |
struct in_ifaddr *ia; |
struct ifaddr *ifa; |
int hlen = 0, len; |
struct ipqent *ipqe; |
|
int hlen = 0, mff, len; |
|
int downmatch; |
int downmatch; |
int checkif; |
|
int srcrt = 0; |
int srcrt = 0; |
int s; |
ifnet_t *ifp; |
u_int hash; |
struct psref psref; |
#ifdef FAST_IPSEC |
|
struct m_tag *mtag; |
KASSERTMSG(cpu_softintr_p(), "ip_input: not in the software " |
struct tdb_ident *tdbi; |
"interrupt handler; synchronization assumptions violated"); |
struct secpolicy *sp; |
|
int error; |
|
#endif /* FAST_IPSEC */ |
|
|
|
MCLAIM(m, &ip_rx_mowner); |
MCLAIM(m, &ip_rx_mowner); |
#ifdef DIAGNOSTIC |
KASSERT((m->m_flags & M_PKTHDR) != 0); |
if ((m->m_flags & M_PKTHDR) == 0) |
|
panic("ipintr no HDR"); |
ifp = m_get_rcvif_psref(m, &psref); |
#endif |
if (__predict_false(ifp == NULL)) |
|
goto out; |
|
|
/* |
/* |
* If no IP addresses have been set yet but the interfaces |
* If no IP addresses have been set yet but the interfaces |
* are receiving, can't do anything with incoming packets yet. |
* are receiving, can't do anything with incoming packets yet. |
|
* Note: we pre-check without locks held. |
*/ |
*/ |
if (TAILQ_FIRST(&in_ifaddrhead) == 0) |
if (IN_ADDRLIST_READER_EMPTY()) |
goto bad; |
goto out; |
ipstat.ips_total++; |
IP_STATINC(IP_STAT_TOTAL); |
|
|
/* |
/* |
* If the IP header is not aligned, slurp it up into a new |
* If the IP header is not aligned, slurp it up into a new |
* mbuf with space for link headers, in the event we forward |
* mbuf with space for link headers, in the event we forward |
Line 522 ip_input(struct mbuf *m) |
|
Line 483 ip_input(struct mbuf *m) |
|
if ((m = m_copyup(m, sizeof(struct ip), |
if ((m = m_copyup(m, sizeof(struct ip), |
(max_linkhdr + 3) & ~3)) == NULL) { |
(max_linkhdr + 3) & ~3)) == NULL) { |
/* XXXJRT new stat, please */ |
/* XXXJRT new stat, please */ |
ipstat.ips_toosmall++; |
IP_STATINC(IP_STAT_TOOSMALL); |
return; |
goto out; |
} |
} |
} else if (__predict_false(m->m_len < sizeof (struct ip))) { |
} else if (__predict_false(m->m_len < sizeof (struct ip))) { |
if ((m = m_pullup(m, sizeof (struct ip))) == NULL) { |
if ((m = m_pullup(m, sizeof (struct ip))) == NULL) { |
ipstat.ips_toosmall++; |
IP_STATINC(IP_STAT_TOOSMALL); |
return; |
goto out; |
} |
} |
} |
} |
ip = mtod(m, struct ip *); |
ip = mtod(m, struct ip *); |
if (ip->ip_v != IPVERSION) { |
if (ip->ip_v != IPVERSION) { |
ipstat.ips_badvers++; |
IP_STATINC(IP_STAT_BADVERS); |
goto bad; |
goto out; |
} |
} |
hlen = ip->ip_hl << 2; |
hlen = ip->ip_hl << 2; |
if (hlen < sizeof(struct ip)) { /* minimum header length */ |
if (hlen < sizeof(struct ip)) { /* minimum header length */ |
ipstat.ips_badhlen++; |
IP_STATINC(IP_STAT_BADHLEN); |
goto bad; |
goto out; |
} |
} |
if (hlen > m->m_len) { |
if (hlen > m->m_len) { |
if ((m = m_pullup(m, hlen)) == 0) { |
if ((m = m_pullup(m, hlen)) == NULL) { |
ipstat.ips_badhlen++; |
IP_STATINC(IP_STAT_BADHLEN); |
return; |
goto out; |
} |
} |
ip = mtod(m, struct ip *); |
ip = mtod(m, struct ip *); |
} |
} |
Line 554 ip_input(struct mbuf *m) |
|
Line 515 ip_input(struct mbuf *m) |
|
* not allowed. |
* not allowed. |
*/ |
*/ |
if (IN_MULTICAST(ip->ip_src.s_addr)) { |
if (IN_MULTICAST(ip->ip_src.s_addr)) { |
ipstat.ips_badaddr++; |
IP_STATINC(IP_STAT_BADADDR); |
goto bad; |
goto out; |
} |
} |
|
|
/* 127/8 must not appear on wire - RFC1122 */ |
/* 127/8 must not appear on wire - RFC1122 */ |
if ((ntohl(ip->ip_dst.s_addr) >> IN_CLASSA_NSHIFT) == IN_LOOPBACKNET || |
if ((ntohl(ip->ip_dst.s_addr) >> IN_CLASSA_NSHIFT) == IN_LOOPBACKNET || |
(ntohl(ip->ip_src.s_addr) >> IN_CLASSA_NSHIFT) == IN_LOOPBACKNET) { |
(ntohl(ip->ip_src.s_addr) >> IN_CLASSA_NSHIFT) == IN_LOOPBACKNET) { |
if ((m->m_pkthdr.rcvif->if_flags & IFF_LOOPBACK) == 0) { |
if ((ifp->if_flags & IFF_LOOPBACK) == 0) { |
ipstat.ips_badaddr++; |
IP_STATINC(IP_STAT_BADADDR); |
goto bad; |
goto out; |
} |
} |
} |
} |
|
|
switch (m->m_pkthdr.csum_flags & |
switch (m->m_pkthdr.csum_flags & |
((m->m_pkthdr.rcvif->if_csum_flags_rx & M_CSUM_IPv4) | |
((ifp->if_csum_flags_rx & M_CSUM_IPv4) | |
M_CSUM_IPv4_BAD)) { |
M_CSUM_IPv4_BAD)) { |
case M_CSUM_IPv4|M_CSUM_IPv4_BAD: |
case M_CSUM_IPv4|M_CSUM_IPv4_BAD: |
INET_CSUM_COUNTER_INCR(&ip_hwcsum_bad); |
INET_CSUM_COUNTER_INCR(&ip_hwcsum_bad); |
goto badcsum; |
IP_STATINC(IP_STAT_BADSUM); |
|
goto out; |
|
|
case M_CSUM_IPv4: |
case M_CSUM_IPv4: |
/* Checksum was okay. */ |
/* Checksum was okay. */ |
Line 584 ip_input(struct mbuf *m) |
|
Line 546 ip_input(struct mbuf *m) |
|
* Must compute it ourselves. Maybe skip checksum on |
* Must compute it ourselves. Maybe skip checksum on |
* loopback interfaces. |
* loopback interfaces. |
*/ |
*/ |
if (__predict_true(!(m->m_pkthdr.rcvif->if_flags & |
if (__predict_true(!(ifp->if_flags & IFF_LOOPBACK) || |
IFF_LOOPBACK) || ip_do_loopback_cksum)) { |
ip_do_loopback_cksum)) { |
INET_CSUM_COUNTER_INCR(&ip_swcsum); |
INET_CSUM_COUNTER_INCR(&ip_swcsum); |
if (in_cksum(m, hlen) != 0) |
if (in_cksum(m, hlen) != 0) { |
goto badcsum; |
IP_STATINC(IP_STAT_BADSUM); |
|
goto out; |
|
} |
} |
} |
break; |
break; |
} |
} |
Line 600 ip_input(struct mbuf *m) |
|
Line 564 ip_input(struct mbuf *m) |
|
* Check for additional length bogosity |
* Check for additional length bogosity |
*/ |
*/ |
if (len < hlen) { |
if (len < hlen) { |
ipstat.ips_badlen++; |
IP_STATINC(IP_STAT_BADLEN); |
goto bad; |
goto out; |
} |
} |
|
|
/* |
/* |
Line 611 ip_input(struct mbuf *m) |
|
Line 575 ip_input(struct mbuf *m) |
|
* Drop packet if shorter than we expect. |
* Drop packet if shorter than we expect. |
*/ |
*/ |
if (m->m_pkthdr.len < len) { |
if (m->m_pkthdr.len < len) { |
ipstat.ips_tooshort++; |
IP_STATINC(IP_STAT_TOOSHORT); |
goto bad; |
goto out; |
} |
} |
if (m->m_pkthdr.len > len) { |
if (m->m_pkthdr.len > len) { |
if (m->m_len == m->m_pkthdr.len) { |
if (m->m_len == m->m_pkthdr.len) { |
Line 622 ip_input(struct mbuf *m) |
|
Line 586 ip_input(struct mbuf *m) |
|
m_adj(m, len - m->m_pkthdr.len); |
m_adj(m, len - m->m_pkthdr.len); |
} |
} |
|
|
#if defined(IPSEC) |
|
/* ipflow (IP fast forwarding) is not compatible with IPsec. */ |
|
m->m_flags &= ~M_CANFASTFWD; |
|
#else |
|
/* |
/* |
* Assume that we can create a fast-forward IP flow entry |
* Assume that we can create a fast-forward IP flow entry |
* based on this packet. |
* based on this packet. |
*/ |
*/ |
m->m_flags |= M_CANFASTFWD; |
m->m_flags |= M_CANFASTFWD; |
#endif |
|
|
|
#ifdef PFIL_HOOKS |
|
/* |
/* |
* Run through list of hooks for input packets. If there are any |
* Run through list of hooks for input packets. If there are any |
* filters which require that additional packets in the flow are |
* filters which require that additional packets in the flow are |
Line 641 ip_input(struct mbuf *m) |
|
Line 599 ip_input(struct mbuf *m) |
|
* Note that filters must _never_ set this flag, as another filter |
* Note that filters must _never_ set this flag, as another filter |
* in the list may have previously cleared it. |
* in the list may have previously cleared it. |
*/ |
*/ |
/* |
#if defined(IPSEC) |
* let ipfilter look at packet on the wire, |
if (!ipsec_used || !ipsec_indone(m)) |
* not the decapsulated packet. |
|
*/ |
|
#ifdef IPSEC |
|
if (!ipsec_getnhist(m)) |
|
#elif defined(FAST_IPSEC) |
|
if (!ipsec_indone(m)) |
|
#else |
#else |
if (1) |
if (1) |
#endif |
#endif |
{ |
{ |
struct in_addr odst; |
struct in_addr odst = ip->ip_dst; |
|
bool freed; |
|
|
odst = ip->ip_dst; |
SOFTNET_LOCK(); |
if (pfil_run_hooks(&inet_pfil_hook, &m, m->m_pkthdr.rcvif, |
freed = pfil_run_hooks(inet_pfil_hook, &m, ifp, PFIL_IN) != 0; |
PFIL_IN) != 0) |
SOFTNET_UNLOCK(); |
return; |
if (freed || m == NULL) { |
if (m == NULL) |
goto out; |
return; |
} |
ip = mtod(m, struct ip *); |
ip = mtod(m, struct ip *); |
hlen = ip->ip_hl << 2; |
hlen = ip->ip_hl << 2; |
|
|
/* |
/* |
* XXX The setting of "srcrt" here is to prevent ip_forward() |
* XXX The setting of "srcrt" here is to prevent ip_forward() |
* from generating ICMP redirects for packets that have |
* from generating ICMP redirects for packets that have |
Line 679 ip_input(struct mbuf *m) |
|
Line 633 ip_input(struct mbuf *m) |
|
*/ |
*/ |
srcrt = (odst.s_addr != ip->ip_dst.s_addr); |
srcrt = (odst.s_addr != ip->ip_dst.s_addr); |
} |
} |
#endif /* PFIL_HOOKS */ |
|
|
|
#ifdef ALTQ |
#ifdef ALTQ |
/* XXX Temporary until ALTQ is changed to use a pfil hook */ |
/* XXX Temporary until ALTQ is changed to use a pfil hook */ |
if (altq_input != NULL && (*altq_input)(m, AF_INET) == 0) { |
if (altq_input) { |
/* packet dropped by traffic conditioner */ |
SOFTNET_LOCK(); |
return; |
if ((*altq_input)(m, AF_INET) == 0) { |
|
/* Packet dropped by traffic conditioner. */ |
|
SOFTNET_UNLOCK(); |
|
goto out; |
|
} |
|
SOFTNET_UNLOCK(); |
} |
} |
#endif |
#endif |
|
|
Line 697 ip_input(struct mbuf *m) |
|
Line 655 ip_input(struct mbuf *m) |
|
*/ |
*/ |
ip_nhops = 0; /* for source routed packets */ |
ip_nhops = 0; /* for source routed packets */ |
if (hlen > sizeof (struct ip) && ip_dooptions(m)) |
if (hlen > sizeof (struct ip) && ip_dooptions(m)) |
return; |
goto out; |
|
|
/* |
|
* Enable a consistency check between the destination address |
|
* and the arrival interface for a unicast packet (the RFC 1122 |
|
* strong ES model) if IP forwarding is disabled and the packet |
|
* is not locally generated. |
|
* |
|
* XXX - Checking also should be disabled if the destination |
|
* address is ipnat'ed to a different interface. |
|
* |
|
* XXX - Checking is incompatible with IP aliases added |
|
* to the loopback interface instead of the interface where |
|
* the packets are received. |
|
* |
|
* XXX - We need to add a per ifaddr flag for this so that |
|
* we get finer grain control. |
|
*/ |
|
checkif = ip_checkinterface && (ipforwarding == 0) && |
|
(m->m_pkthdr.rcvif != NULL) && |
|
((m->m_pkthdr.rcvif->if_flags & IFF_LOOPBACK) == 0); |
|
|
|
/* |
/* |
* Check our list of addresses, to see if the packet is for us. |
* Check our list of addresses, to see if the packet is for us. |
* |
* |
* Traditional 4.4BSD did not consult IFF_UP at all. |
* Traditional 4.4BSD did not consult IFF_UP at all. |
* The behavior here is to treat addresses on !IFF_UP interface |
* The behavior here is to treat addresses on !IFF_UP interface |
* as not mine. |
* or IN_IFF_NOTREADY addresses as not mine. |
*/ |
*/ |
downmatch = 0; |
downmatch = 0; |
LIST_FOREACH(ia, &IN_IFADDR_HASH(ip->ip_dst.s_addr), ia_hash) { |
ia = ip_match_our_address(ifp, ip, &downmatch); |
if (in_hosteq(ia->ia_addr.sin_addr, ip->ip_dst)) { |
|
if (checkif && ia->ia_ifp != m->m_pkthdr.rcvif) |
|
continue; |
|
if ((ia->ia_ifp->if_flags & IFF_UP) != 0) |
|
break; |
|
else |
|
downmatch++; |
|
} |
|
} |
|
if (ia != NULL) |
if (ia != NULL) |
goto ours; |
goto ours; |
if (m->m_pkthdr.rcvif && m->m_pkthdr.rcvif->if_flags & IFF_BROADCAST) { |
|
IFADDR_FOREACH(ifa, m->m_pkthdr.rcvif) { |
if (ifp->if_flags & IFF_BROADCAST) { |
if (ifa->ifa_addr->sa_family != AF_INET) |
ia = ip_match_our_address_broadcast(ifp, ip); |
continue; |
if (ia != NULL) |
ia = ifatoia(ifa); |
goto ours; |
if (in_hosteq(ip->ip_dst, ia->ia_broadaddr.sin_addr) || |
|
in_hosteq(ip->ip_dst, ia->ia_netbroadcast) || |
|
/* |
|
* Look for all-0's host part (old broadcast addr), |
|
* either for subnet or net. |
|
*/ |
|
ip->ip_dst.s_addr == ia->ia_subnet || |
|
ip->ip_dst.s_addr == ia->ia_net) |
|
goto ours; |
|
/* |
|
* An interface with IP address zero accepts |
|
* all packets that arrive on that interface. |
|
*/ |
|
if (in_nullhost(ia->ia_addr.sin_addr)) |
|
goto ours; |
|
} |
|
} |
} |
|
|
if (IN_MULTICAST(ip->ip_dst.s_addr)) { |
if (IN_MULTICAST(ip->ip_dst.s_addr)) { |
struct in_multi *inm; |
|
#ifdef MROUTING |
#ifdef MROUTING |
extern struct socket *ip_mrouter; |
extern struct socket *ip_mrouter; |
|
|
Line 779 ip_input(struct mbuf *m) |
|
Line 692 ip_input(struct mbuf *m) |
|
* as expected when ip_mforward() is called from |
* as expected when ip_mforward() is called from |
* ip_output().) |
* ip_output().) |
*/ |
*/ |
if (ip_mforward(m, m->m_pkthdr.rcvif) != 0) { |
SOFTNET_LOCK(); |
ipstat.ips_cantforward++; |
if (ip_mforward(m, ifp) != 0) { |
m_freem(m); |
SOFTNET_UNLOCK(); |
return; |
IP_STATINC(IP_STAT_CANTFORWARD); |
|
goto out; |
} |
} |
|
SOFTNET_UNLOCK(); |
|
|
/* |
/* |
* The process-level routing demon needs to receive |
* The process-level routing demon needs to receive |
* all multicast IGMP packets, whether or not this |
* all multicast IGMP packets, whether or not this |
* host belongs to their destination groups. |
* host belongs to their destination groups. |
*/ |
*/ |
if (ip->ip_p == IPPROTO_IGMP) |
if (ip->ip_p == IPPROTO_IGMP) { |
goto ours; |
goto ours; |
ipstat.ips_forward++; |
} |
|
IP_STATINC(IP_STAT_CANTFORWARD); |
} |
} |
#endif |
#endif |
/* |
/* |
* See if we belong to the destination multicast group on the |
* See if we belong to the destination multicast group on the |
* arrival interface. |
* arrival interface. |
*/ |
*/ |
IN_LOOKUP_MULTI(ip->ip_dst, m->m_pkthdr.rcvif, inm); |
if (!in_multi_group(ip->ip_dst, ifp, 0)) { |
if (inm == NULL) { |
IP_STATINC(IP_STAT_CANTFORWARD); |
ipstat.ips_cantforward++; |
goto out; |
m_freem(m); |
|
return; |
|
} |
} |
goto ours; |
goto ours; |
} |
} |
Line 815 ip_input(struct mbuf *m) |
|
Line 729 ip_input(struct mbuf *m) |
|
* Not for us; forward if possible and desirable. |
* Not for us; forward if possible and desirable. |
*/ |
*/ |
if (ipforwarding == 0) { |
if (ipforwarding == 0) { |
ipstat.ips_cantforward++; |
m_put_rcvif_psref(ifp, &psref); |
|
IP_STATINC(IP_STAT_CANTFORWARD); |
m_freem(m); |
m_freem(m); |
} else { |
} else { |
/* |
/* |
Line 825 ip_input(struct mbuf *m) |
|
Line 740 ip_input(struct mbuf *m) |
|
* forwarding loop till TTL goes to 0. |
* forwarding loop till TTL goes to 0. |
*/ |
*/ |
if (downmatch) { |
if (downmatch) { |
|
m_put_rcvif_psref(ifp, &psref); |
icmp_error(m, ICMP_UNREACH, ICMP_UNREACH_HOST, 0, 0); |
icmp_error(m, ICMP_UNREACH, ICMP_UNREACH_HOST, 0, 0); |
ipstat.ips_cantforward++; |
IP_STATINC(IP_STAT_CANTFORWARD); |
return; |
return; |
} |
} |
#ifdef IPSEC |
#ifdef IPSEC |
if (ipsec4_in_reject(m, NULL)) { |
/* Perform IPsec, if any. */ |
ipsecstat.in_polvio++; |
if (ipsec_used) { |
goto bad; |
SOFTNET_LOCK(); |
} |
if (ipsec4_input(m, IP_FORWARDING | |
#endif |
(ip_directedbcast ? IP_ALLOWBROADCAST : 0)) != 0) { |
#ifdef FAST_IPSEC |
SOFTNET_UNLOCK(); |
mtag = m_tag_find(m, PACKET_TAG_IPSEC_IN_DONE, NULL); |
goto out; |
s = splsoftnet(); |
|
if (mtag != NULL) { |
|
tdbi = (struct tdb_ident *)(mtag + 1); |
|
sp = ipsec_getpolicy(tdbi, IPSEC_DIR_INBOUND); |
|
} else { |
|
sp = ipsec_getpolicybyaddr(m, IPSEC_DIR_INBOUND, |
|
IP_FORWARDING, &error); |
|
} |
|
if (sp == NULL) { /* NB: can happen if error */ |
|
splx(s); |
|
/*XXX error stat???*/ |
|
DPRINTF(("ip_input: no SP for forwarding\n")); /*XXX*/ |
|
goto bad; |
|
} |
|
|
|
/* |
|
* Check security policy against packet attributes. |
|
*/ |
|
error = ipsec_in_reject(sp, m); |
|
KEY_FREESP(&sp); |
|
splx(s); |
|
if (error) { |
|
ipstat.ips_cantforward++; |
|
goto bad; |
|
} |
|
|
|
/* |
|
* Peek at the outbound SP for this packet to determine if |
|
* it's a Fast Forward candidate. |
|
*/ |
|
mtag = m_tag_find(m, PACKET_TAG_IPSEC_PENDING_TDB, NULL); |
|
if (mtag != NULL) |
|
m->m_flags &= ~M_CANFASTFWD; |
|
else { |
|
s = splsoftnet(); |
|
sp = ipsec4_checkpolicy(m, IPSEC_DIR_OUTBOUND, |
|
(IP_FORWARDING | |
|
(ip_directedbcast ? IP_ALLOWBROADCAST : 0)), |
|
&error, NULL); |
|
if (sp != NULL) { |
|
m->m_flags &= ~M_CANFASTFWD; |
|
KEY_FREESP(&sp); |
|
} |
} |
splx(s); |
SOFTNET_UNLOCK(); |
} |
} |
#endif /* FAST_IPSEC */ |
#endif |
|
ip_forward(m, srcrt, ifp); |
ip_forward(m, srcrt); |
m_put_rcvif_psref(ifp, &psref); |
} |
} |
return; |
return; |
|
|
ours: |
ours: |
|
m_put_rcvif_psref(ifp, &psref); |
|
ifp = NULL; |
|
|
/* |
/* |
* If offset or IP_MF are set, must reassemble. |
* If offset or IP_MF are set, must reassemble. |
* Otherwise, nothing need be done. |
|
* (We could look in the reassembly queue to see |
|
* if the packet was previously fragmented, |
|
* but it's not worth the time; just let them time out.) |
|
*/ |
*/ |
if (ip->ip_off & ~htons(IP_DF|IP_RF)) { |
if (ip->ip_off & ~htons(IP_DF|IP_RF)) { |
|
|
/* |
|
* Look for queue of fragments |
|
* of this datagram. |
|
*/ |
|
IPQ_LOCK(); |
|
hash = IPREASS_HASH(ip->ip_src.s_addr, ip->ip_id); |
|
LIST_FOREACH(fp, &ipq[hash], ipq_q) { |
|
if (ip->ip_id == fp->ipq_id && |
|
in_hosteq(ip->ip_src, fp->ipq_src) && |
|
in_hosteq(ip->ip_dst, fp->ipq_dst) && |
|
ip->ip_p == fp->ipq_p) |
|
goto found; |
|
|
|
} |
|
fp = 0; |
|
found: |
|
|
|
/* |
/* |
* Adjust ip_len to not reflect header, |
* Pass to IP reassembly mechanism. |
* set ipqe_mff if more fragments are expected, |
|
* convert offset of this to bytes. |
|
*/ |
*/ |
ip->ip_len = htons(ntohs(ip->ip_len) - hlen); |
if (ip_reass_packet(&m, ip) != 0) { |
mff = (ip->ip_off & htons(IP_MF)) != 0; |
/* Failed; invalid fragment(s) or packet. */ |
if (mff) { |
goto out; |
/* |
} |
* Make sure that fragments have a data length |
if (m == NULL) { |
* that's a non-zero multiple of 8 bytes. |
/* More fragments should come; silently return. */ |
*/ |
goto out; |
if (ntohs(ip->ip_len) == 0 || |
|
(ntohs(ip->ip_len) & 0x7) != 0) { |
|
ipstat.ips_badfrags++; |
|
IPQ_UNLOCK(); |
|
goto bad; |
|
} |
|
} |
} |
ip->ip_off = htons((ntohs(ip->ip_off) & IP_OFFMASK) << 3); |
|
|
|
/* |
/* |
* If datagram marked as having more fragments |
* Reassembly is done, we have the final packet. |
* or if this is not the first fragment, |
* Updated cached data in local variable(s). |
* attempt reassembly; if it succeeds, proceed. |
|
*/ |
*/ |
if (mff || ip->ip_off != htons(0)) { |
ip = mtod(m, struct ip *); |
ipstat.ips_fragments++; |
hlen = ip->ip_hl << 2; |
s = splvm(); |
|
ipqe = pool_get(&ipqent_pool, PR_NOWAIT); |
|
splx(s); |
|
if (ipqe == NULL) { |
|
ipstat.ips_rcvmemdrop++; |
|
IPQ_UNLOCK(); |
|
goto bad; |
|
} |
|
ipqe->ipqe_mff = mff; |
|
ipqe->ipqe_m = m; |
|
ipqe->ipqe_ip = ip; |
|
m = ip_reass(ipqe, fp, &ipq[hash]); |
|
if (m == 0) { |
|
IPQ_UNLOCK(); |
|
return; |
|
} |
|
ipstat.ips_reassembled++; |
|
ip = mtod(m, struct ip *); |
|
hlen = ip->ip_hl << 2; |
|
ip->ip_len = htons(ntohs(ip->ip_len) + hlen); |
|
} else |
|
if (fp) |
|
ip_freef(fp); |
|
IPQ_UNLOCK(); |
|
} |
} |
|
|
#if defined(IPSEC) |
#ifdef IPSEC |
/* |
|
* enforce IPsec policy checking if we are seeing last header. |
|
* note that we do not visit this with protocols with pcb layer |
|
* code - like udp/tcp/raw ip. |
|
*/ |
|
if ((inetsw[ip_protox[ip->ip_p]].pr_flags & PR_LASTHDR) != 0 && |
|
ipsec4_in_reject(m, NULL)) { |
|
ipsecstat.in_polvio++; |
|
goto bad; |
|
} |
|
#endif |
|
#ifdef FAST_IPSEC |
|
/* |
/* |
* enforce IPsec policy checking if we are seeing last header. |
* Enforce IPsec policy checking if we are seeing last header. |
* note that we do not visit this with protocols with pcb layer |
* Note that we do not visit this with protocols with PCB layer |
* code - like udp/tcp/raw ip. |
* code - like UDP/TCP/raw IP. |
*/ |
*/ |
if ((inetsw[ip_protox[ip->ip_p]].pr_flags & PR_LASTHDR) != 0) { |
if (ipsec_used && |
/* |
(inetsw[ip_protox[ip->ip_p]].pr_flags & PR_LASTHDR) != 0) { |
* Check if the packet has already had IPsec processing |
SOFTNET_LOCK(); |
* done. If so, then just pass it along. This tag gets |
if (ipsec4_input(m, 0) != 0) { |
* set during AH, ESP, etc. input handling, before the |
SOFTNET_UNLOCK(); |
* packet is returned to the ip input queue for delivery. |
goto out; |
*/ |
|
mtag = m_tag_find(m, PACKET_TAG_IPSEC_IN_DONE, NULL); |
|
s = splsoftnet(); |
|
if (mtag != NULL) { |
|
tdbi = (struct tdb_ident *)(mtag + 1); |
|
sp = ipsec_getpolicy(tdbi, IPSEC_DIR_INBOUND); |
|
} else { |
|
sp = ipsec_getpolicybyaddr(m, IPSEC_DIR_INBOUND, |
|
IP_FORWARDING, &error); |
|
} |
|
if (sp != NULL) { |
|
/* |
|
* Check security policy against packet attributes. |
|
*/ |
|
error = ipsec_in_reject(sp, m); |
|
KEY_FREESP(&sp); |
|
} else { |
|
/* XXX error stat??? */ |
|
error = EINVAL; |
|
DPRINTF(("ip_input: no SP, packet discarded\n"));/*XXX*/ |
|
} |
} |
splx(s); |
SOFTNET_UNLOCK(); |
if (error) |
|
goto bad; |
|
} |
} |
#endif /* FAST_IPSEC */ |
#endif |
|
|
/* |
/* |
* Switch out to protocol's input routine. |
* Switch out to protocol's input routine. |
Line 1027 DPRINTF(("ip_input: no SP, packet discar |
|
Line 813 DPRINTF(("ip_input: no SP, packet discar |
|
if (ia && ip) |
if (ia && ip) |
ia->ia_ifa.ifa_data.ifad_inbytes += ntohs(ip->ip_len); |
ia->ia_ifa.ifa_data.ifad_inbytes += ntohs(ip->ip_len); |
#endif |
#endif |
ipstat.ips_delivered++; |
IP_STATINC(IP_STAT_DELIVERED); |
{ |
|
int off = hlen, nh = ip->ip_p; |
|
|
|
|
const int off = hlen, nh = ip->ip_p; |
|
|
|
SOFTNET_LOCK(); |
(*inetsw[ip_protox[nh]].pr_input)(m, off, nh); |
(*inetsw[ip_protox[nh]].pr_input)(m, off, nh); |
|
SOFTNET_UNLOCK(); |
return; |
return; |
} |
|
bad: |
|
m_freem(m); |
|
return; |
|
|
|
badcsum: |
|
ipstat.ips_badsum++; |
|
m_freem(m); |
|
} |
|
|
|
/* |
|
* Take incoming datagram fragment and try to |
|
* reassemble it into whole datagram. If a chain for |
|
* reassembly of this datagram already exists, then it |
|
* is given as fp; otherwise have to make a chain. |
|
*/ |
|
struct mbuf * |
|
ip_reass(struct ipqent *ipqe, struct ipq *fp, struct ipqhead *ipqhead) |
|
{ |
|
struct mbuf *m = ipqe->ipqe_m; |
|
struct ipqent *nq, *p, *q; |
|
struct ip *ip; |
|
struct mbuf *t; |
|
int hlen = ipqe->ipqe_ip->ip_hl << 2; |
|
int i, next, s; |
|
|
|
IPQ_LOCK_CHECK(); |
out: |
|
m_put_rcvif_psref(ifp, &psref); |
/* |
if (m != NULL) |
* Presence of header sizes in mbufs |
m_freem(m); |
* would confuse code below. |
|
*/ |
|
m->m_data += hlen; |
|
m->m_len -= hlen; |
|
|
|
#ifdef notyet |
|
/* make sure fragment limit is up-to-date */ |
|
CHECK_NMBCLUSTER_PARAMS(); |
|
|
|
/* If we have too many fragments, drop the older half. */ |
|
if (ip_nfrags >= ip_maxfrags) |
|
ip_reass_drophalf(void); |
|
#endif |
|
|
|
/* |
|
* We are about to add a fragment; increment frag count. |
|
*/ |
|
ip_nfrags++; |
|
|
|
/* |
|
* If first fragment to arrive, create a reassembly queue. |
|
*/ |
|
if (fp == 0) { |
|
/* |
|
* Enforce upper bound on number of fragmented packets |
|
* for which we attempt reassembly; |
|
* If maxfrag is 0, never accept fragments. |
|
* If maxfrag is -1, accept all fragments without limitation. |
|
*/ |
|
if (ip_maxfragpackets < 0) |
|
; |
|
else if (ip_nfragpackets >= ip_maxfragpackets) |
|
goto dropfrag; |
|
ip_nfragpackets++; |
|
MALLOC(fp, struct ipq *, sizeof (struct ipq), |
|
M_FTABLE, M_NOWAIT); |
|
if (fp == NULL) |
|
goto dropfrag; |
|
LIST_INSERT_HEAD(ipqhead, fp, ipq_q); |
|
fp->ipq_nfrags = 1; |
|
fp->ipq_ttl = IPFRAGTTL; |
|
fp->ipq_p = ipqe->ipqe_ip->ip_p; |
|
fp->ipq_id = ipqe->ipqe_ip->ip_id; |
|
TAILQ_INIT(&fp->ipq_fragq); |
|
fp->ipq_src = ipqe->ipqe_ip->ip_src; |
|
fp->ipq_dst = ipqe->ipqe_ip->ip_dst; |
|
p = NULL; |
|
goto insert; |
|
} else { |
|
fp->ipq_nfrags++; |
|
} |
|
|
|
/* |
|
* Find a segment which begins after this one does. |
|
*/ |
|
for (p = NULL, q = TAILQ_FIRST(&fp->ipq_fragq); q != NULL; |
|
p = q, q = TAILQ_NEXT(q, ipqe_q)) |
|
if (ntohs(q->ipqe_ip->ip_off) > ntohs(ipqe->ipqe_ip->ip_off)) |
|
break; |
|
|
|
/* |
|
* If there is a preceding segment, it may provide some of |
|
* our data already. If so, drop the data from the incoming |
|
* segment. If it provides all of our data, drop us. |
|
*/ |
|
if (p != NULL) { |
|
i = ntohs(p->ipqe_ip->ip_off) + ntohs(p->ipqe_ip->ip_len) - |
|
ntohs(ipqe->ipqe_ip->ip_off); |
|
if (i > 0) { |
|
if (i >= ntohs(ipqe->ipqe_ip->ip_len)) |
|
goto dropfrag; |
|
m_adj(ipqe->ipqe_m, i); |
|
ipqe->ipqe_ip->ip_off = |
|
htons(ntohs(ipqe->ipqe_ip->ip_off) + i); |
|
ipqe->ipqe_ip->ip_len = |
|
htons(ntohs(ipqe->ipqe_ip->ip_len) - i); |
|
} |
|
} |
|
|
|
/* |
|
* While we overlap succeeding segments trim them or, |
|
* if they are completely covered, dequeue them. |
|
*/ |
|
for (; q != NULL && |
|
ntohs(ipqe->ipqe_ip->ip_off) + ntohs(ipqe->ipqe_ip->ip_len) > |
|
ntohs(q->ipqe_ip->ip_off); q = nq) { |
|
i = (ntohs(ipqe->ipqe_ip->ip_off) + |
|
ntohs(ipqe->ipqe_ip->ip_len)) - ntohs(q->ipqe_ip->ip_off); |
|
if (i < ntohs(q->ipqe_ip->ip_len)) { |
|
q->ipqe_ip->ip_len = |
|
htons(ntohs(q->ipqe_ip->ip_len) - i); |
|
q->ipqe_ip->ip_off = |
|
htons(ntohs(q->ipqe_ip->ip_off) + i); |
|
m_adj(q->ipqe_m, i); |
|
break; |
|
} |
|
nq = TAILQ_NEXT(q, ipqe_q); |
|
m_freem(q->ipqe_m); |
|
TAILQ_REMOVE(&fp->ipq_fragq, q, ipqe_q); |
|
s = splvm(); |
|
pool_put(&ipqent_pool, q); |
|
splx(s); |
|
fp->ipq_nfrags--; |
|
ip_nfrags--; |
|
} |
|
|
|
insert: |
|
/* |
|
* Stick new segment in its place; |
|
* check for complete reassembly. |
|
*/ |
|
if (p == NULL) { |
|
TAILQ_INSERT_HEAD(&fp->ipq_fragq, ipqe, ipqe_q); |
|
} else { |
|
TAILQ_INSERT_AFTER(&fp->ipq_fragq, p, ipqe, ipqe_q); |
|
} |
|
next = 0; |
|
for (p = NULL, q = TAILQ_FIRST(&fp->ipq_fragq); q != NULL; |
|
p = q, q = TAILQ_NEXT(q, ipqe_q)) { |
|
if (ntohs(q->ipqe_ip->ip_off) != next) |
|
return (0); |
|
next += ntohs(q->ipqe_ip->ip_len); |
|
} |
|
if (p->ipqe_mff) |
|
return (0); |
|
|
|
/* |
|
* Reassembly is complete. Check for a bogus message size and |
|
* concatenate fragments. |
|
*/ |
|
q = TAILQ_FIRST(&fp->ipq_fragq); |
|
ip = q->ipqe_ip; |
|
if ((next + (ip->ip_hl << 2)) > IP_MAXPACKET) { |
|
ipstat.ips_toolong++; |
|
ip_freef(fp); |
|
return (0); |
|
} |
|
m = q->ipqe_m; |
|
t = m->m_next; |
|
m->m_next = 0; |
|
m_cat(m, t); |
|
nq = TAILQ_NEXT(q, ipqe_q); |
|
s = splvm(); |
|
pool_put(&ipqent_pool, q); |
|
splx(s); |
|
for (q = nq; q != NULL; q = nq) { |
|
t = q->ipqe_m; |
|
nq = TAILQ_NEXT(q, ipqe_q); |
|
s = splvm(); |
|
pool_put(&ipqent_pool, q); |
|
splx(s); |
|
m_cat(m, t); |
|
} |
|
ip_nfrags -= fp->ipq_nfrags; |
|
|
|
/* |
|
* Create header for new ip packet by |
|
* modifying header of first packet; |
|
* dequeue and discard fragment reassembly header. |
|
* Make header visible. |
|
*/ |
|
ip->ip_len = htons(next); |
|
ip->ip_src = fp->ipq_src; |
|
ip->ip_dst = fp->ipq_dst; |
|
LIST_REMOVE(fp, ipq_q); |
|
FREE(fp, M_FTABLE); |
|
ip_nfragpackets--; |
|
m->m_len += (ip->ip_hl << 2); |
|
m->m_data -= (ip->ip_hl << 2); |
|
/* some debugging cruft by sklower, below, will go away soon */ |
|
if (m->m_flags & M_PKTHDR) { /* XXX this should be done elsewhere */ |
|
int plen = 0; |
|
for (t = m; t; t = t->m_next) |
|
plen += t->m_len; |
|
m->m_pkthdr.len = plen; |
|
m->m_pkthdr.csum_flags = 0; |
|
} |
|
return (m); |
|
|
|
dropfrag: |
|
if (fp != 0) |
|
fp->ipq_nfrags--; |
|
ip_nfrags--; |
|
ipstat.ips_fragdropped++; |
|
m_freem(m); |
|
s = splvm(); |
|
pool_put(&ipqent_pool, ipqe); |
|
splx(s); |
|
return (0); |
|
} |
} |
|
|
/* |
/* |
* Free a fragment reassembly header and all |
* IP timer processing. |
* associated datagrams. |
|
*/ |
*/ |
void |
void |
ip_freef(struct ipq *fp) |
ip_slowtimo(void) |
{ |
|
struct ipqent *q, *p; |
|
u_int nfrags = 0; |
|
int s; |
|
|
|
IPQ_LOCK_CHECK(); |
|
|
|
for (q = TAILQ_FIRST(&fp->ipq_fragq); q != NULL; q = p) { |
|
p = TAILQ_NEXT(q, ipqe_q); |
|
m_freem(q->ipqe_m); |
|
nfrags++; |
|
TAILQ_REMOVE(&fp->ipq_fragq, q, ipqe_q); |
|
s = splvm(); |
|
pool_put(&ipqent_pool, q); |
|
splx(s); |
|
} |
|
|
|
if (nfrags != fp->ipq_nfrags) |
|
printf("ip_freef: nfrags %d != %d\n", fp->ipq_nfrags, nfrags); |
|
ip_nfrags -= nfrags; |
|
LIST_REMOVE(fp, ipq_q); |
|
FREE(fp, M_FTABLE); |
|
ip_nfragpackets--; |
|
} |
|
|
|
/* |
|
* IP reassembly TTL machinery for multiplicative drop. |
|
*/ |
|
static u_int fragttl_histo[(IPFRAGTTL+1)]; |
|
|
|
|
|
/* |
|
* Decrement TTL of all reasembly queue entries by `ticks'. |
|
* Count number of distinct fragments (as opposed to partial, fragmented |
|
* datagrams) in the reassembly queue. While we traverse the entire |
|
* reassembly queue, compute and return the median TTL over all fragments. |
|
*/ |
|
static u_int |
|
ip_reass_ttl_decr(u_int ticks) |
|
{ |
|
u_int nfrags, median, dropfraction, keepfraction; |
|
struct ipq *fp, *nfp; |
|
int i; |
|
|
|
nfrags = 0; |
|
memset(fragttl_histo, 0, sizeof fragttl_histo); |
|
|
|
for (i = 0; i < IPREASS_NHASH; i++) { |
|
for (fp = LIST_FIRST(&ipq[i]); fp != NULL; fp = nfp) { |
|
fp->ipq_ttl = ((fp->ipq_ttl <= ticks) ? |
|
0 : fp->ipq_ttl - ticks); |
|
nfp = LIST_NEXT(fp, ipq_q); |
|
if (fp->ipq_ttl == 0) { |
|
ipstat.ips_fragtimeout++; |
|
ip_freef(fp); |
|
} else { |
|
nfrags += fp->ipq_nfrags; |
|
fragttl_histo[fp->ipq_ttl] += fp->ipq_nfrags; |
|
} |
|
} |
|
} |
|
|
|
KASSERT(ip_nfrags == nfrags); |
|
|
|
/* Find median (or other drop fraction) in histogram. */ |
|
dropfraction = (ip_nfrags / 2); |
|
keepfraction = ip_nfrags - dropfraction; |
|
for (i = IPFRAGTTL, median = 0; i >= 0; i--) { |
|
median += fragttl_histo[i]; |
|
if (median >= keepfraction) |
|
break; |
|
} |
|
|
|
/* Return TTL of median (or other fraction). */ |
|
return (u_int)i; |
|
} |
|
|
|
void |
|
ip_reass_drophalf(void) |
|
{ |
{ |
|
|
u_int median_ticks; |
mutex_enter(softnet_lock); |
/* |
KERNEL_LOCK(1, NULL); |
* Compute median TTL of all fragments, and count frags |
|
* with that TTL or lower (roughly half of all fragments). |
|
*/ |
|
median_ticks = ip_reass_ttl_decr(0); |
|
|
|
/* Drop half. */ |
|
median_ticks = ip_reass_ttl_decr(median_ticks); |
|
|
|
} |
ip_reass_slowtimo(); |
|
|
/* |
KERNEL_UNLOCK_ONE(NULL); |
* IP timer processing; |
mutex_exit(softnet_lock); |
* if a timer expires on a reassembly |
|
* queue, discard it. |
|
*/ |
|
void |
|
ip_slowtimo(void) |
|
{ |
|
static u_int dropscanidx = 0; |
|
u_int i; |
|
u_int median_ttl; |
|
int s = splsoftnet(); |
|
|
|
IPQ_LOCK(); |
|
|
|
/* Age TTL of all fragments by 1 tick .*/ |
|
median_ttl = ip_reass_ttl_decr(1); |
|
|
|
/* make sure fragment limit is up-to-date */ |
|
CHECK_NMBCLUSTER_PARAMS(); |
|
|
|
/* If we have too many fragments, drop the older half. */ |
|
if (ip_nfrags > ip_maxfrags) |
|
ip_reass_ttl_decr(median_ttl); |
|
|
|
/* |
|
* If we are over the maximum number of fragmented packets |
|
* (due to the limit being lowered), drain off |
|
* enough to get down to the new limit. Start draining |
|
* from the reassembly hashqueue most recently drained. |
|
*/ |
|
if (ip_maxfragpackets < 0) |
|
; |
|
else { |
|
int wrapped = 0; |
|
|
|
i = dropscanidx; |
|
while (ip_nfragpackets > ip_maxfragpackets && wrapped == 0) { |
|
while (LIST_FIRST(&ipq[i]) != NULL) |
|
ip_freef(LIST_FIRST(&ipq[i])); |
|
if (++i >= IPREASS_NHASH) { |
|
i = 0; |
|
} |
|
/* |
|
* Dont scan forever even if fragment counters are |
|
* wrong: stop after scanning entire reassembly queue. |
|
*/ |
|
if (i == dropscanidx) |
|
wrapped = 1; |
|
} |
|
dropscanidx = i; |
|
} |
|
IPQ_UNLOCK(); |
|
splx(s); |
|
} |
} |
|
|
/* |
/* |
* Drain off all datagram fragments. |
* IP drain processing. |
*/ |
*/ |
void |
void |
ip_drain(void) |
ip_drain(void) |
{ |
{ |
|
|
/* |
KERNEL_LOCK(1, NULL); |
* We may be called from a device's interrupt context. If |
ip_reass_drain(); |
* the ipq is already busy, just bail out now. |
KERNEL_UNLOCK_ONE(NULL); |
*/ |
|
if (ipq_lock_try() == 0) |
|
return; |
|
|
|
/* |
|
* Drop half the total fragments now. If more mbufs are needed, |
|
* we will be called again soon. |
|
*/ |
|
ip_reass_drophalf(); |
|
|
|
IPQ_UNLOCK(); |
|
} |
} |
|
|
/* |
/* |
* Do option processing on a datagram, |
* ip_dooptions: perform option processing on a datagram, possibly discarding |
* possibly discarding it if bad options are encountered, |
* it if bad options are encountered, or forwarding it if source-routed. |
* or forwarding it if source-routed. |
* |
* Returns 1 if packet has been forwarded/freed, |
* => Returns true if packet has been forwarded/freed. |
* 0 if the packet should be processed further. |
* => Returns false if the packet should be processed further. |
*/ |
*/ |
int |
static bool |
ip_dooptions(struct mbuf *m) |
ip_dooptions(struct mbuf *m) |
{ |
{ |
struct ip *ip = mtod(m, struct ip *); |
struct ip *ip = mtod(m, struct ip *); |
Line 1524 ip_dooptions(struct mbuf *m) |
|
Line 948 ip_dooptions(struct mbuf *m) |
|
/* |
/* |
* locate outgoing interface |
* locate outgoing interface |
*/ |
*/ |
bcopy((void *)(cp + off), (void *)&ipaddr.sin_addr, |
memcpy((void *)&ipaddr.sin_addr, (void *)(cp + off), |
sizeof(ipaddr.sin_addr)); |
sizeof(ipaddr.sin_addr)); |
if (opt == IPOPT_SSRR) |
if (opt == IPOPT_SSRR) |
ia = ifatoia(ifa_ifwithladdr(sintosa(&ipaddr))); |
ia = ifatoia(ifa_ifwithladdr(sintosa(&ipaddr))); |
Line 1560 ip_dooptions(struct mbuf *m) |
|
Line 984 ip_dooptions(struct mbuf *m) |
|
off--; /* 0 origin */ |
off--; /* 0 origin */ |
if ((off + sizeof(struct in_addr)) > optlen) |
if ((off + sizeof(struct in_addr)) > optlen) |
break; |
break; |
bcopy((void *)(&ip->ip_dst), (void *)&ipaddr.sin_addr, |
memcpy((void *)&ipaddr.sin_addr, (void *)(&ip->ip_dst), |
sizeof(ipaddr.sin_addr)); |
sizeof(ipaddr.sin_addr)); |
/* |
/* |
* locate outgoing interface; if we're the destination, |
* locate outgoing interface; if we're the destination, |
Line 1603 ip_dooptions(struct mbuf *m) |
|
Line 1027 ip_dooptions(struct mbuf *m) |
|
case IPOPT_TS_TSONLY: |
case IPOPT_TS_TSONLY: |
break; |
break; |
|
|
case IPOPT_TS_TSANDADDR: |
case IPOPT_TS_TSANDADDR: { |
|
struct ifnet *rcvif; |
|
int s; |
|
|
if (ipt->ipt_ptr - 1 + sizeof(n_time) + |
if (ipt->ipt_ptr - 1 + sizeof(n_time) + |
sizeof(struct in_addr) > ipt->ipt_len) { |
sizeof(struct in_addr) > ipt->ipt_len) { |
code = (u_char *)&ipt->ipt_ptr - |
code = (u_char *)&ipt->ipt_ptr - |
Line 1611 ip_dooptions(struct mbuf *m) |
|
Line 1038 ip_dooptions(struct mbuf *m) |
|
goto bad; |
goto bad; |
} |
} |
ipaddr.sin_addr = dst; |
ipaddr.sin_addr = dst; |
|
rcvif = m_get_rcvif(m, &s); |
ia = ifatoia(ifaof_ifpforaddr(sintosa(&ipaddr), |
ia = ifatoia(ifaof_ifpforaddr(sintosa(&ipaddr), |
m->m_pkthdr.rcvif)); |
rcvif)); |
|
m_put_rcvif(rcvif, &s); |
if (ia == 0) |
if (ia == 0) |
continue; |
continue; |
bcopy(&ia->ia_addr.sin_addr, |
bcopy(&ia->ia_addr.sin_addr, |
cp0, sizeof(struct in_addr)); |
cp0, sizeof(struct in_addr)); |
ipt->ipt_ptr += sizeof(struct in_addr); |
ipt->ipt_ptr += sizeof(struct in_addr); |
break; |
break; |
|
} |
|
|
case IPOPT_TS_PRESPEC: |
case IPOPT_TS_PRESPEC: |
if (ipt->ipt_ptr - 1 + sizeof(n_time) + |
if (ipt->ipt_ptr - 1 + sizeof(n_time) + |
Line 1627 ip_dooptions(struct mbuf *m) |
|
Line 1057 ip_dooptions(struct mbuf *m) |
|
(u_char *)ip; |
(u_char *)ip; |
goto bad; |
goto bad; |
} |
} |
bcopy(cp0, &ipaddr.sin_addr, |
memcpy(&ipaddr.sin_addr, cp0, |
sizeof(struct in_addr)); |
sizeof(struct in_addr)); |
if (ifatoia(ifa_ifwithaddr(sintosa(&ipaddr))) |
if (ifatoia(ifa_ifwithaddr(sintosa(&ipaddr))) |
== NULL) |
== NULL) |
Line 1649 ip_dooptions(struct mbuf *m) |
|
Line 1079 ip_dooptions(struct mbuf *m) |
|
} |
} |
} |
} |
if (forward) { |
if (forward) { |
|
struct ifnet *rcvif; |
|
struct psref psref; |
|
|
if (ip_forwsrcrt == 0) { |
if (ip_forwsrcrt == 0) { |
type = ICMP_UNREACH; |
type = ICMP_UNREACH; |
code = ICMP_UNREACH_SRCFAIL; |
code = ICMP_UNREACH_SRCFAIL; |
goto bad; |
goto bad; |
} |
} |
ip_forward(m, 1); |
|
return (1); |
rcvif = m_get_rcvif_psref(m, &psref); |
|
if (__predict_false(rcvif == NULL)) { |
|
type = ICMP_UNREACH; |
|
code = ICMP_UNREACH_HOST; |
|
goto bad; |
|
} |
|
ip_forward(m, 1, rcvif); |
|
m_put_rcvif_psref(rcvif, &psref); |
|
return true; |
} |
} |
return (0); |
return false; |
bad: |
bad: |
icmp_error(m, type, code, 0, 0); |
icmp_error(m, type, code, 0, 0); |
ipstat.ips_badoptions++; |
IP_STATINC(IP_STAT_BADOPTIONS); |
return (1); |
return true; |
} |
} |
|
|
/* |
/* |
* Given address of next destination (final or next hop), |
* ip_rtaddr: given address of next destination (final or next hop), |
* return internet address info of interface to be used to get there. |
* return internet address info of interface to be used to get there. |
*/ |
*/ |
struct in_ifaddr * |
static struct in_ifaddr * |
ip_rtaddr(struct in_addr dst) |
ip_rtaddr(struct in_addr dst) |
{ |
{ |
struct rtentry *rt; |
struct rtentry *rt; |
Line 1679 ip_rtaddr(struct in_addr dst) |
|
Line 1120 ip_rtaddr(struct in_addr dst) |
|
|
|
sockaddr_in_init(&u.dst4, &dst, 0); |
sockaddr_in_init(&u.dst4, &dst, 0); |
|
|
if ((rt = rtcache_lookup(&ipforward_rt, &u.dst)) == NULL) |
SOFTNET_LOCK(); |
|
rt = rtcache_lookup(&ipforward_rt, &u.dst); |
|
SOFTNET_UNLOCK(); |
|
if (rt == NULL) |
return NULL; |
return NULL; |
|
|
return ifatoia(rt->rt_ifa); |
return ifatoia(rt->rt_ifa); |
} |
} |
|
|
/* |
/* |
* Save incoming source route for use in replies, |
* save_rte: save incoming source route for use in replies, to be picked |
* to be picked up later by ip_srcroute if the receiver is interested. |
* up later by ip_srcroute if the receiver is interested. |
*/ |
*/ |
void |
static void |
save_rte(u_char *option, struct in_addr dst) |
save_rte(u_char *option, struct in_addr dst) |
{ |
{ |
unsigned olen; |
unsigned olen; |
|
|
olen = option[IPOPT_OLEN]; |
olen = option[IPOPT_OLEN]; |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) |
|
printf("save_rte: olen %d\n", olen); |
|
#endif /* 0 */ |
|
if (olen > sizeof(ip_srcrt) - (1 + sizeof(dst))) |
if (olen > sizeof(ip_srcrt) - (1 + sizeof(dst))) |
return; |
return; |
bcopy((void *)option, (void *)ip_srcrt.srcopt, olen); |
memcpy((void *)ip_srcrt.srcopt, (void *)option, olen); |
ip_nhops = (olen - IPOPT_OFFSET - 1) / sizeof(struct in_addr); |
ip_nhops = (olen - IPOPT_OFFSET - 1) / sizeof(struct in_addr); |
ip_srcrt.dst = dst; |
ip_srcrt.dst = dst; |
} |
} |
Line 1729 ip_srcroute(void) |
|
Line 1169 ip_srcroute(void) |
|
/* length is (nhops+1)*sizeof(addr) + sizeof(nop + srcrt header) */ |
/* length is (nhops+1)*sizeof(addr) + sizeof(nop + srcrt header) */ |
m->m_len = ip_nhops * sizeof(struct in_addr) + sizeof(struct in_addr) + |
m->m_len = ip_nhops * sizeof(struct in_addr) + sizeof(struct in_addr) + |
OPTSIZ; |
OPTSIZ; |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) |
|
printf("ip_srcroute: nhops %d mlen %d", ip_nhops, m->m_len); |
|
#endif |
|
|
|
/* |
/* |
* First save first hop for return route |
* First save first hop for return route |
*/ |
*/ |
p = &ip_srcrt.route[ip_nhops - 1]; |
p = &ip_srcrt.route[ip_nhops - 1]; |
*(mtod(m, struct in_addr *)) = *p--; |
*(mtod(m, struct in_addr *)) = *p--; |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) |
|
printf(" hops %x", ntohl(mtod(m, struct in_addr *)->s_addr)); |
|
#endif |
|
|
|
/* |
/* |
* Copy option fields and padding (nop) to mbuf. |
* Copy option fields and padding (nop) to mbuf. |
Line 1759 ip_srcroute(void) |
|
Line 1191 ip_srcroute(void) |
|
* reversing the path (pointers are now aligned). |
* reversing the path (pointers are now aligned). |
*/ |
*/ |
while (p >= ip_srcrt.route) { |
while (p >= ip_srcrt.route) { |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) |
|
printf(" %x", ntohl(q->s_addr)); |
|
#endif |
|
*q++ = *p--; |
*q++ = *p--; |
} |
} |
/* |
/* |
* Last hop goes to final destination. |
* Last hop goes to final destination. |
*/ |
*/ |
*q = ip_srcrt.dst; |
*q = ip_srcrt.dst; |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) |
|
printf(" %x\n", ntohl(q->s_addr)); |
|
#endif |
|
return (m); |
return (m); |
} |
} |
|
|
const int inetctlerrmap[PRC_NCMDS] = { |
const int inetctlerrmap[PRC_NCMDS] = { |
0, 0, 0, 0, |
[PRC_MSGSIZE] = EMSGSIZE, |
0, EMSGSIZE, EHOSTDOWN, EHOSTUNREACH, |
[PRC_HOSTDEAD] = EHOSTDOWN, |
EHOSTUNREACH, EHOSTUNREACH, ECONNREFUSED, ECONNREFUSED, |
[PRC_HOSTUNREACH] = EHOSTUNREACH, |
EMSGSIZE, EHOSTUNREACH, 0, 0, |
[PRC_UNREACH_NET] = EHOSTUNREACH, |
0, 0, 0, 0, |
[PRC_UNREACH_HOST] = EHOSTUNREACH, |
ENOPROTOOPT |
[PRC_UNREACH_PROTOCOL] = ECONNREFUSED, |
|
[PRC_UNREACH_PORT] = ECONNREFUSED, |
|
[PRC_UNREACH_SRCFAIL] = EHOSTUNREACH, |
|
[PRC_PARAMPROB] = ENOPROTOOPT, |
}; |
}; |
|
|
|
void |
|
ip_fasttimo(void) |
|
{ |
|
if (ip_drainwanted) { |
|
ip_drain(); |
|
ip_drainwanted = 0; |
|
} |
|
} |
|
|
|
void |
|
ip_drainstub(void) |
|
{ |
|
ip_drainwanted = 1; |
|
} |
|
|
/* |
/* |
* Forward a packet. If some error occurs return the sender |
* Forward a packet. If some error occurs return the sender |
* an icmp packet. Note we can't always generate a meaningful |
* an icmp packet. Note we can't always generate a meaningful |
Line 1799 const int inetctlerrmap[PRC_NCMDS] = { |
|
Line 1241 const int inetctlerrmap[PRC_NCMDS] = { |
|
* The srcrt parameter indicates whether the packet is being forwarded |
* The srcrt parameter indicates whether the packet is being forwarded |
* via a source route. |
* via a source route. |
*/ |
*/ |
void |
static void |
ip_forward(struct mbuf *m, int srcrt) |
ip_forward(struct mbuf *m, int srcrt, struct ifnet *rcvif) |
{ |
{ |
struct ip *ip = mtod(m, struct ip *); |
struct ip *ip = mtod(m, struct ip *); |
struct rtentry *rt; |
struct rtentry *rt; |
Line 1811 ip_forward(struct mbuf *m, int srcrt) |
|
Line 1253 ip_forward(struct mbuf *m, int srcrt) |
|
struct sockaddr dst; |
struct sockaddr dst; |
struct sockaddr_in dst4; |
struct sockaddr_in dst4; |
} u; |
} u; |
|
uint64_t *ips; |
|
|
|
KASSERTMSG(cpu_softintr_p(), "ip_forward: not in the software " |
|
"interrupt handler; synchronization assumptions violated"); |
|
|
/* |
/* |
* We are now in the output path. |
* We are now in the output path. |
Line 1823 ip_forward(struct mbuf *m, int srcrt) |
|
Line 1269 ip_forward(struct mbuf *m, int srcrt) |
|
m->m_pkthdr.csum_flags = 0; |
m->m_pkthdr.csum_flags = 0; |
|
|
dest = 0; |
dest = 0; |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) { |
|
printf("forward: src %s ", inet_ntoa(ip->ip_src)); |
|
printf("dst %s ttl %x\n", inet_ntoa(ip->ip_dst), ip->ip_ttl); |
|
} |
|
#endif |
|
if (m->m_flags & (M_BCAST|M_MCAST) || in_canforward(ip->ip_dst) == 0) { |
if (m->m_flags & (M_BCAST|M_MCAST) || in_canforward(ip->ip_dst) == 0) { |
ipstat.ips_cantforward++; |
IP_STATINC(IP_STAT_CANTFORWARD); |
m_freem(m); |
m_freem(m); |
return; |
return; |
} |
} |
|
|
|
SOFTNET_LOCK(); |
|
|
if (ip->ip_ttl <= IPTTLDEC) { |
if (ip->ip_ttl <= IPTTLDEC) { |
icmp_error(m, ICMP_TIMXCEED, ICMP_TIMXCEED_INTRANS, dest, 0); |
icmp_error(m, ICMP_TIMXCEED, ICMP_TIMXCEED_INTRANS, dest, 0); |
|
SOFTNET_UNLOCK(); |
return; |
return; |
} |
} |
|
|
sockaddr_in_init(&u.dst4, &ip->ip_dst, 0); |
sockaddr_in_init(&u.dst4, &ip->ip_dst, 0); |
|
|
if ((rt = rtcache_lookup(&ipforward_rt, &u.dst)) == NULL) { |
if ((rt = rtcache_lookup(&ipforward_rt, &u.dst)) == NULL) { |
icmp_error(m, ICMP_UNREACH, ICMP_UNREACH_NET, dest, 0); |
icmp_error(m, ICMP_UNREACH, ICMP_UNREACH_NET, dest, 0); |
|
SOFTNET_UNLOCK(); |
return; |
return; |
} |
} |
|
|
Line 1864 ip_forward(struct mbuf *m, int srcrt) |
|
Line 1310 ip_forward(struct mbuf *m, int srcrt) |
|
* Also, don't send redirect if forwarding using a default route |
* Also, don't send redirect if forwarding using a default route |
* or a route modified by a redirect. |
* or a route modified by a redirect. |
*/ |
*/ |
if (rt->rt_ifp == m->m_pkthdr.rcvif && |
if (rt->rt_ifp == rcvif && |
(rt->rt_flags & (RTF_DYNAMIC|RTF_MODIFIED)) == 0 && |
(rt->rt_flags & (RTF_DYNAMIC|RTF_MODIFIED)) == 0 && |
!in_nullhost(satocsin(rt_getkey(rt))->sin_addr) && |
!in_nullhost(satocsin(rt_getkey(rt))->sin_addr) && |
ipsendredirects && !srcrt) { |
ipsendredirects && !srcrt) { |
Line 1881 ip_forward(struct mbuf *m, int srcrt) |
|
Line 1327 ip_forward(struct mbuf *m, int srcrt) |
|
*/ |
*/ |
type = ICMP_REDIRECT; |
type = ICMP_REDIRECT; |
code = ICMP_REDIRECT_HOST; |
code = ICMP_REDIRECT_HOST; |
#ifdef DIAGNOSTIC |
|
if (ipprintfs) |
|
printf("redirect (%d) to %x\n", code, |
|
(u_int32_t)dest); |
|
#endif |
|
} |
} |
} |
} |
|
|
error = ip_output(m, NULL, &ipforward_rt, |
error = ip_output(m, NULL, &ipforward_rt, |
(IP_FORWARDING | (ip_directedbcast ? IP_ALLOWBROADCAST : 0)), |
(IP_FORWARDING | (ip_directedbcast ? IP_ALLOWBROADCAST : 0)), |
(struct ip_moptions *)NULL, (struct socket *)NULL); |
NULL, NULL); |
|
|
if (error) |
if (error) { |
ipstat.ips_cantforward++; |
IP_STATINC(IP_STAT_CANTFORWARD); |
else { |
goto error; |
ipstat.ips_forward++; |
} |
if (type) |
|
ipstat.ips_redirectsent++; |
ips = IP_STAT_GETREF(); |
else { |
ips[IP_STAT_FORWARD]++; |
if (mcopy) { |
|
|
if (type) { |
|
ips[IP_STAT_REDIRECTSENT]++; |
|
IP_STAT_PUTREF(); |
|
goto redirect; |
|
} |
|
|
|
IP_STAT_PUTREF(); |
|
if (mcopy) { |
#ifdef GATEWAY |
#ifdef GATEWAY |
if (mcopy->m_flags & M_CANFASTFWD) |
if (mcopy->m_flags & M_CANFASTFWD) |
ipflow_create(&ipforward_rt, mcopy); |
ipflow_create(&ipforward_rt, mcopy); |
#endif |
#endif |
m_freem(mcopy); |
m_freem(mcopy); |
} |
|
return; |
|
} |
|
} |
} |
if (mcopy == NULL) |
|
|
SOFTNET_UNLOCK(); |
|
return; |
|
|
|
redirect: |
|
error: |
|
if (mcopy == NULL) { |
|
SOFTNET_UNLOCK(); |
return; |
return; |
|
} |
|
|
switch (error) { |
switch (error) { |
|
|
Line 1931 ip_forward(struct mbuf *m, int srcrt) |
|
Line 1385 ip_forward(struct mbuf *m, int srcrt) |
|
case EMSGSIZE: |
case EMSGSIZE: |
type = ICMP_UNREACH; |
type = ICMP_UNREACH; |
code = ICMP_UNREACH_NEEDFRAG; |
code = ICMP_UNREACH_NEEDFRAG; |
#if !defined(IPSEC) && !defined(FAST_IPSEC) |
|
if (ipforward_rt.ro_rt != NULL) |
|
destmtu = ipforward_rt.ro_rt->rt_ifp->if_mtu; |
|
#else |
|
/* |
|
* If the packet is routed over IPsec tunnel, tell the |
|
* originator the tunnel MTU. |
|
* tunnel MTU = if MTU - sizeof(IP) - ESP/AH hdrsiz |
|
* XXX quickhack!!! |
|
*/ |
|
if (ipforward_rt.ro_rt != NULL) { |
|
struct secpolicy *sp; |
|
int ipsecerror; |
|
size_t ipsechdr; |
|
struct route *ro; |
|
|
|
sp = ipsec4_getpolicybyaddr(mcopy, |
|
IPSEC_DIR_OUTBOUND, IP_FORWARDING, |
|
&ipsecerror); |
|
|
|
if (sp == NULL) |
|
destmtu = ipforward_rt.ro_rt->rt_ifp->if_mtu; |
|
else { |
|
/* count IPsec header size */ |
|
ipsechdr = ipsec4_hdrsiz(mcopy, |
|
IPSEC_DIR_OUTBOUND, NULL); |
|
|
|
/* |
if ((rt = rtcache_validate(&ipforward_rt)) != NULL) |
* find the correct route for outer IPv4 |
destmtu = rt->rt_ifp->if_mtu; |
* header, compute tunnel MTU. |
#ifdef IPSEC |
*/ |
if (ipsec_used) |
|
(void)ipsec4_forward(mcopy, &destmtu); |
if (sp->req != NULL |
|
&& sp->req->sav != NULL |
|
&& sp->req->sav->sah != NULL) { |
|
ro = &sp->req->sav->sah->sa_route; |
|
if (ro->ro_rt && ro->ro_rt->rt_ifp) { |
|
destmtu = |
|
ro->ro_rt->rt_rmx.rmx_mtu ? |
|
ro->ro_rt->rt_rmx.rmx_mtu : |
|
ro->ro_rt->rt_ifp->if_mtu; |
|
destmtu -= ipsechdr; |
|
} |
|
} |
|
|
|
#ifdef IPSEC |
|
key_freesp(sp); |
|
#else |
|
KEY_FREESP(&sp); |
|
#endif |
#endif |
} |
IP_STATINC(IP_STAT_CANTFRAG); |
} |
|
#endif /*IPSEC*/ |
|
ipstat.ips_cantfrag++; |
|
break; |
break; |
|
|
case ENOBUFS: |
case ENOBUFS: |
#if 1 |
|
/* |
/* |
* a router should not generate ICMP_SOURCEQUENCH as |
* Do not generate ICMP_SOURCEQUENCH as required in RFC 1812, |
* required in RFC1812 Requirements for IP Version 4 Routers. |
* Requirements for IP Version 4 Routers. Source quench can |
* source quench could be a big problem under DoS attacks, |
* big problem under DoS attacks or if the underlying |
* or if the underlying interface is rate-limited. |
* interface is rate-limited. |
*/ |
*/ |
if (mcopy) |
if (mcopy) |
m_freem(mcopy); |
m_freem(mcopy); |
|
SOFTNET_UNLOCK(); |
return; |
return; |
#else |
|
type = ICMP_SOURCEQUENCH; |
|
code = 0; |
|
break; |
|
#endif |
|
} |
} |
icmp_error(mcopy, type, code, dest, destmtu); |
icmp_error(mcopy, type, code, dest, destmtu); |
|
SOFTNET_UNLOCK(); |
} |
} |
|
|
void |
void |
ip_savecontrol(struct inpcb *inp, struct mbuf **mp, struct ip *ip, |
ip_savecontrol(struct inpcb *inp, struct mbuf **mp, struct ip *ip, |
struct mbuf *m) |
struct mbuf *m) |
{ |
{ |
|
struct socket *so = inp->inp_socket; |
if (inp->inp_socket->so_options & SO_TIMESTAMP) { |
ifnet_t *ifp; |
|
int inpflags = inp->inp_flags; |
|
struct psref psref; |
|
|
|
ifp = m_get_rcvif_psref(m, &psref); |
|
if (__predict_false(ifp == NULL)) |
|
return; /* XXX should report error? */ |
|
|
|
if (so->so_options & SO_TIMESTAMP |
|
#ifdef SO_OTIMESTAMP |
|
|| so->so_options & SO_OTIMESTAMP |
|
#endif |
|
) { |
struct timeval tv; |
struct timeval tv; |
|
|
microtime(&tv); |
microtime(&tv); |
|
#ifdef SO_OTIMESTAMP |
|
if (so->so_options & SO_OTIMESTAMP) { |
|
struct timeval50 tv50; |
|
timeval_to_timeval50(&tv, &tv50); |
|
*mp = sbcreatecontrol((void *) &tv50, sizeof(tv50), |
|
SCM_OTIMESTAMP, SOL_SOCKET); |
|
} else |
|
#endif |
*mp = sbcreatecontrol((void *) &tv, sizeof(tv), |
*mp = sbcreatecontrol((void *) &tv, sizeof(tv), |
SCM_TIMESTAMP, SOL_SOCKET); |
SCM_TIMESTAMP, SOL_SOCKET); |
if (*mp) |
if (*mp) |
mp = &(*mp)->m_next; |
mp = &(*mp)->m_next; |
} |
} |
if (inp->inp_flags & INP_RECVDSTADDR) { |
if (inpflags & INP_RECVDSTADDR) { |
*mp = sbcreatecontrol((void *) &ip->ip_dst, |
*mp = sbcreatecontrol((void *) &ip->ip_dst, |
sizeof(struct in_addr), IP_RECVDSTADDR, IPPROTO_IP); |
sizeof(struct in_addr), IP_RECVDSTADDR, IPPROTO_IP); |
if (*mp) |
if (*mp) |
mp = &(*mp)->m_next; |
mp = &(*mp)->m_next; |
} |
} |
#ifdef notyet |
if (inpflags & INP_RECVPKTINFO) { |
/* |
struct in_pktinfo ipi; |
* XXX |
ipi.ipi_addr = ip->ip_src; |
* Moving these out of udp_input() made them even more broken |
ipi.ipi_ifindex = ifp->if_index; |
* than they already were. |
*mp = sbcreatecontrol((void *) &ipi, |
* - fenner@parc.xerox.com |
sizeof(ipi), IP_RECVPKTINFO, IPPROTO_IP); |
*/ |
|
/* options were tossed already */ |
|
if (inp->inp_flags & INP_RECVOPTS) { |
|
*mp = sbcreatecontrol((void *) opts_deleted_above, |
|
sizeof(struct in_addr), IP_RECVOPTS, IPPROTO_IP); |
|
if (*mp) |
if (*mp) |
mp = &(*mp)->m_next; |
mp = &(*mp)->m_next; |
} |
} |
/* ip_srcroute doesn't do what we want here, need to fix */ |
if (inpflags & INP_PKTINFO) { |
if (inp->inp_flags & INP_RECVRETOPTS) { |
struct in_pktinfo ipi; |
*mp = sbcreatecontrol((void *) ip_srcroute(), |
ipi.ipi_addr = ip->ip_dst; |
sizeof(struct in_addr), IP_RECVRETOPTS, IPPROTO_IP); |
ipi.ipi_ifindex = ifp->if_index; |
|
*mp = sbcreatecontrol((void *) &ipi, |
|
sizeof(ipi), IP_PKTINFO, IPPROTO_IP); |
if (*mp) |
if (*mp) |
mp = &(*mp)->m_next; |
mp = &(*mp)->m_next; |
} |
} |
#endif |
if (inpflags & INP_RECVIF) { |
if (inp->inp_flags & INP_RECVIF) { |
|
struct sockaddr_dl sdl; |
struct sockaddr_dl sdl; |
|
|
sockaddr_dl_init(&sdl, sizeof(sdl), |
sockaddr_dl_init(&sdl, sizeof(sdl), ifp ? |
(m->m_pkthdr.rcvif != NULL) |
ifp->if_index : 0, 0, NULL, 0, NULL, 0); |
? m->m_pkthdr.rcvif->if_index |
|
: 0, |
|
0, NULL, 0, NULL, 0); |
|
*mp = sbcreatecontrol(&sdl, sdl.sdl_len, IP_RECVIF, IPPROTO_IP); |
*mp = sbcreatecontrol(&sdl, sdl.sdl_len, IP_RECVIF, IPPROTO_IP); |
if (*mp) |
if (*mp) |
mp = &(*mp)->m_next; |
mp = &(*mp)->m_next; |
} |
} |
|
if (inpflags & INP_RECVTTL) { |
|
*mp = sbcreatecontrol((void *) &ip->ip_ttl, |
|
sizeof(uint8_t), IP_RECVTTL, IPPROTO_IP); |
|
if (*mp) |
|
mp = &(*mp)->m_next; |
|
} |
|
m_put_rcvif_psref(ifp, &psref); |
} |
} |
|
|
/* |
/* |
Line 2079 sysctl_net_inet_ip_forwsrcrt(SYSCTLFN_AR |
|
Line 1503 sysctl_net_inet_ip_forwsrcrt(SYSCTLFN_AR |
|
if (error || newp == NULL) |
if (error || newp == NULL) |
return (error); |
return (error); |
|
|
if (kauth_authorize_network(l->l_cred, KAUTH_NETWORK_FORWSRCRT, |
error = kauth_authorize_network(l->l_cred, KAUTH_NETWORK_FORWSRCRT, |
0, NULL, NULL, NULL)) |
0, NULL, NULL, NULL); |
return (EPERM); |
if (error) |
|
return (error); |
|
|
ip_forwsrcrt = tmp; |
ip_forwsrcrt = tmp; |
|
|
Line 2107 sysctl_net_inet_ip_pmtudto(SYSCTLFN_ARGS |
|
Line 1532 sysctl_net_inet_ip_pmtudto(SYSCTLFN_ARGS |
|
if (tmp < 0) |
if (tmp < 0) |
return (EINVAL); |
return (EINVAL); |
|
|
|
mutex_enter(softnet_lock); |
|
|
ip_mtudisc_timeout = tmp; |
ip_mtudisc_timeout = tmp; |
rt_timer_queue_change(ip_mtudisc_timeout_q, ip_mtudisc_timeout); |
rt_timer_queue_change(ip_mtudisc_timeout_q, ip_mtudisc_timeout); |
|
|
return (0); |
mutex_exit(softnet_lock); |
} |
|
|
|
#ifdef GATEWAY |
|
/* |
|
* sysctl helper routine for net.inet.ip.maxflows. |
|
*/ |
|
static int |
|
sysctl_net_inet_ip_maxflows(SYSCTLFN_ARGS) |
|
{ |
|
int s; |
|
|
|
s = sysctl_lookup(SYSCTLFN_CALL(rnode)); |
|
if (s || newp == NULL) |
|
return (s); |
|
|
|
s = splsoftnet(); |
|
ipflow_reap(0); |
|
splx(s); |
|
|
|
return (0); |
return (0); |
} |
} |
|
|
static int |
static int |
sysctl_net_inet_ip_hashsize(SYSCTLFN_ARGS) |
sysctl_net_inet_ip_stats(SYSCTLFN_ARGS) |
{ |
{ |
int error, tmp; |
|
struct sysctlnode node; |
|
|
|
node = *rnode; |
|
tmp = ip_hashsize; |
|
node.sysctl_data = &tmp; |
|
error = sysctl_lookup(SYSCTLFN_CALL(&node)); |
|
if (error || newp == NULL) |
|
return (error); |
|
|
|
if ((tmp & (tmp - 1)) == 0 && tmp != 0) { |
|
/* |
|
* Can only fail due to malloc() |
|
*/ |
|
if (ipflow_invalidate_all(tmp)) |
|
return ENOMEM; |
|
} else { |
|
/* |
|
* EINVAL if not a power of 2 |
|
*/ |
|
return EINVAL; |
|
} |
|
|
|
return (0); |
return (NETSTAT_SYSCTL(ipstat_percpu, IP_NSTATS)); |
} |
} |
#endif /* GATEWAY */ |
|
|
|
|
|
SYSCTL_SETUP(sysctl_net_inet_ip_setup, "sysctl net.inet.ip subtree setup") |
static void |
|
sysctl_net_inet_ip_setup(struct sysctllog **clog) |
{ |
{ |
extern int subnetsarelocal, hostzeroisbroadcast; |
|
|
|
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT, |
|
CTLTYPE_NODE, "net", NULL, |
|
NULL, 0, NULL, 0, |
|
CTL_NET, CTL_EOL); |
|
sysctl_createv(clog, 0, NULL, NULL, |
sysctl_createv(clog, 0, NULL, NULL, |
CTLFLAG_PERMANENT, |
CTLFLAG_PERMANENT, |
CTLTYPE_NODE, "inet", |
CTLTYPE_NODE, "inet", |
Line 2238 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
Line 1617 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
NULL, 0, &ip_allowsrcrt, 0, |
NULL, 0, &ip_allowsrcrt, 0, |
CTL_NET, PF_INET, IPPROTO_IP, |
CTL_NET, PF_INET, IPPROTO_IP, |
IPCTL_ALLOWSRCRT, CTL_EOL); |
IPCTL_ALLOWSRCRT, CTL_EOL); |
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "subnetsarelocal", |
|
SYSCTL_DESCR("Whether logical subnets are considered " |
|
"local"), |
|
NULL, 0, &subnetsarelocal, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
IPCTL_SUBNETSARELOCAL, CTL_EOL); |
|
sysctl_createv(clog, 0, NULL, NULL, |
sysctl_createv(clog, 0, NULL, NULL, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLTYPE_INT, "mtudisc", |
CTLTYPE_INT, "mtudisc", |
Line 2271 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
Line 1643 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLTYPE_INT, "mtudisctimeout", |
CTLTYPE_INT, "mtudisctimeout", |
SYSCTL_DESCR("Lifetime of a Path MTU Discovered route"), |
SYSCTL_DESCR("Lifetime of a Path MTU Discovered route"), |
sysctl_net_inet_ip_pmtudto, 0, &ip_mtudisc_timeout, 0, |
sysctl_net_inet_ip_pmtudto, 0, (void *)&ip_mtudisc_timeout, 0, |
CTL_NET, PF_INET, IPPROTO_IP, |
CTL_NET, PF_INET, IPPROTO_IP, |
IPCTL_MTUDISCTIMEOUT, CTL_EOL); |
IPCTL_MTUDISCTIMEOUT, CTL_EOL); |
#ifdef GATEWAY |
|
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "maxflows", |
|
SYSCTL_DESCR("Number of flows for fast forwarding"), |
|
sysctl_net_inet_ip_maxflows, 0, &ip_maxflows, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
IPCTL_MAXFLOWS, CTL_EOL); |
|
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "hashsize", |
|
SYSCTL_DESCR("Size of hash table for fast forwarding (IPv4)"), |
|
sysctl_net_inet_ip_hashsize, 0, &ip_hashsize, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
CTL_CREATE, CTL_EOL); |
|
#endif /* GATEWAY */ |
|
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "hostzerobroadcast", |
|
SYSCTL_DESCR("All zeroes address is broadcast address"), |
|
NULL, 0, &hostzeroisbroadcast, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
IPCTL_HOSTZEROBROADCAST, CTL_EOL); |
|
#if NGIF > 0 |
|
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "gifttl", |
|
SYSCTL_DESCR("Default TTL for a gif tunnel datagram"), |
|
NULL, 0, &ip_gif_ttl, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
IPCTL_GIF_TTL, CTL_EOL); |
|
#endif /* NGIF */ |
|
#ifndef IPNOPRIVPORTS |
#ifndef IPNOPRIVPORTS |
sysctl_createv(clog, 0, NULL, NULL, |
sysctl_createv(clog, 0, NULL, NULL, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
Line 2324 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
Line 1664 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
CTL_NET, PF_INET, IPPROTO_IP, |
CTL_NET, PF_INET, IPPROTO_IP, |
IPCTL_LOWPORTMAX, CTL_EOL); |
IPCTL_LOWPORTMAX, CTL_EOL); |
#endif /* IPNOPRIVPORTS */ |
#endif /* IPNOPRIVPORTS */ |
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "maxfragpackets", |
|
SYSCTL_DESCR("Maximum number of fragments to retain for " |
|
"possible reassembly"), |
|
NULL, 0, &ip_maxfragpackets, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
IPCTL_MAXFRAGPACKETS, CTL_EOL); |
|
#if NGRE > 0 |
#if NGRE > 0 |
sysctl_createv(clog, 0, NULL, NULL, |
sysctl_createv(clog, 0, NULL, NULL, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
Line 2367 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
Line 1699 SYSCTL_SETUP(sysctl_net_inet_ip_setup, " |
|
CTLFLAG_PERMANENT, |
CTLFLAG_PERMANENT, |
CTLTYPE_STRUCT, "stats", |
CTLTYPE_STRUCT, "stats", |
SYSCTL_DESCR("IP statistics"), |
SYSCTL_DESCR("IP statistics"), |
NULL, 0, &ipstat, sizeof(ipstat), |
sysctl_net_inet_ip_stats, 0, NULL, 0, |
CTL_NET, PF_INET, IPPROTO_IP, IPCTL_STATS, |
CTL_NET, PF_INET, IPPROTO_IP, IPCTL_STATS, |
CTL_EOL); |
CTL_EOL); |
|
#if NARP |
|
sysctl_createv(clog, 0, NULL, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_INT, "dad_count", |
|
SYSCTL_DESCR("Number of Duplicate Address Detection " |
|
"probes to send"), |
|
NULL, 0, &ip_dad_count, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, |
|
IPCTL_DAD_COUNT, CTL_EOL); |
|
#endif |
|
|
|
/* anonportalgo RFC6056 subtree */ |
|
const struct sysctlnode *portalgo_node; |
|
sysctl_createv(clog, 0, NULL, &portalgo_node, |
|
CTLFLAG_PERMANENT, |
|
CTLTYPE_NODE, "anonportalgo", |
|
SYSCTL_DESCR("Anonymous Port Algorithm Selection (RFC 6056)"), |
|
NULL, 0, NULL, 0, |
|
CTL_NET, PF_INET, IPPROTO_IP, CTL_CREATE, CTL_EOL); |
|
sysctl_createv(clog, 0, &portalgo_node, NULL, |
|
CTLFLAG_PERMANENT, |
|
CTLTYPE_STRING, "available", |
|
SYSCTL_DESCR("available algorithms"), |
|
sysctl_portalgo_available, 0, NULL, PORTALGO_MAXLEN, |
|
CTL_CREATE, CTL_EOL); |
|
sysctl_createv(clog, 0, &portalgo_node, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_STRING, "selected", |
|
SYSCTL_DESCR("selected algorithm"), |
|
sysctl_portalgo_selected4, 0, NULL, PORTALGO_MAXLEN, |
|
CTL_CREATE, CTL_EOL); |
|
sysctl_createv(clog, 0, &portalgo_node, NULL, |
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE, |
|
CTLTYPE_STRUCT, "reserve", |
|
SYSCTL_DESCR("bitmap of reserved ports"), |
|
sysctl_portalgo_reserve4, 0, NULL, 0, |
|
CTL_CREATE, CTL_EOL); |
|
} |
|
|
|
void |
|
ip_statinc(u_int stat) |
|
{ |
|
|
|
KASSERT(stat < IP_NSTATS); |
|
IP_STATINC(stat); |
} |
} |