/* $NetBSD: ip_fil_netbsd.c,v 1.31.4.1 2019/10/03 17:18:32 martin Exp $ */ /* * Copyright (C) 2012 by Darren Reed. * * See the IPFILTER.LICENCE file for details on licencing. */ #if !defined(lint) #if defined(__NetBSD__) #include __KERNEL_RCSID(0, "$NetBSD: ip_fil_netbsd.c,v 1.31.4.1 2019/10/03 17:18:32 martin Exp $"); #else static const char sccsid[] = "@(#)ip_fil.c 2.41 6/5/96 (C) 1993-2000 Darren Reed"; static const char rcsid[] = "@(#)Id: ip_fil_netbsd.c,v 1.1.1.2 2012/07/22 13:45:17 darrenr Exp"; #endif #endif #if defined(KERNEL) || defined(_KERNEL) # undef KERNEL # undef _KERNEL # define KERNEL 1 # define _KERNEL 1 #endif #include #if (NetBSD >= 199905) && !defined(IPFILTER_LKM) # if (__NetBSD_Version__ >= 799003000) # ifdef _KERNEL_OPT # include "opt_ipsec.h" # endif # else # include "opt_ipsec.h" # endif #endif #include #include #include #include #include #include #include #if (NetBSD > 199609) # include #else # include #endif #if (__NetBSD_Version__ >= 599005900) # include #endif #include #include #include #include #if (__NetBSD_Version__ >= 399002000) # include #endif #if (__NetBSD_Version__ >= 799003000) #include #include #endif #if defined(__NetBSD__) #include #endif #include #include #include #include #include #include #include #include #if __NetBSD_Version__ >= 105190000 /* 1.5T */ # include # include #endif #include #include #include "netinet/ip_compat.h" #ifdef USE_INET6 # include # if (__NetBSD_Version__ >= 106000000) # include # endif # if __NetBSD_Version__ >= 499001100 # include # include # endif #endif #include "netinet/ip_fil.h" #include "netinet/ip_nat.h" #include "netinet/ip_frag.h" #include "netinet/ip_state.h" #include "netinet/ip_proxy.h" #include "netinet/ip_auth.h" #include "netinet/ip_sync.h" #include "netinet/ip_lookup.h" #include "netinet/ip_dstlist.h" #ifdef IPFILTER_SCAN #include "netinet/ip_scan.h" #endif #include #include #include #ifdef INET extern int ip_optcopy (struct ip *, struct ip *); #endif #ifdef IPFILTER_M_IPFILTER MALLOC_DEFINE(M_IPFILTER, "IP Filter", "IP Filter packet filter data structures"); #endif #if __NetBSD_Version__ >= 105009999 # define csuminfo csum_flags #endif #if __NetBSD_Version__ < 200000000 extern struct protosw inetsw[]; #endif #if (__NetBSD_Version__ >= 599002000) static kauth_listener_t ipf_listener; #endif #if (__NetBSD_Version__ < 399001400) extern int ip6_getpmtu (struct route_in6 *, struct route_in6 *, struct ifnet *, struct in6_addr *, u_long *, int *); #endif #if (NetBSD >= 199511) static int ipfopen(dev_t dev, int flags, int devtype, PROC_T *p); static int ipfclose(dev_t dev, int flags, int devtype, PROC_T *p); #else # if (__NetBSD_Version__ >= 399001400) static int ipfopen(dev_t dev, int flags, struct lwp *); static int ipfclose(dev_t dev, int flags, struct lwp *); # else static int ipfopen(dev_t dev, int flags); static int ipfclose(dev_t dev, int flags); # endif /* __NetBSD_Version__ >= 399001400 */ #endif static int ipfread(dev_t, struct uio *, int ioflag); static int ipfwrite(dev_t, struct uio *, int ioflag); static int ipfpoll(dev_t, int events, PROC_T *); static void ipf_timer_func(void *ptr); const struct cdevsw ipl_cdevsw = { .d_open = ipfopen, .d_close = ipfclose, .d_read = ipfread, .d_write = ipfwrite, .d_ioctl = ipfioctl, .d_stop = nostop, .d_tty = notty, .d_poll = ipfpoll, .d_mmap = nommap, #if (__NetBSD_Version__ >= 200000000) .d_kqfilter = nokqfilter, #endif .d_discard = nodiscard, #ifdef D_OTHER .d_flag = D_OTHER #else .d_flag = 0 #endif }; #if (__NetBSD_Version__ >= 799003000) kmutex_t ipf_ref_mutex; int ipf_active; #endif ipf_main_softc_t ipfmain; static u_short ipid = 0; static int (*ipf_savep)(void *, ip_t *, int, void *, int, struct mbuf **); static int ipf_send_ip(fr_info_t *, mb_t *); #ifdef USE_INET6 static int ipf_fastroute6(struct mbuf *, struct mbuf **, fr_info_t *, frdest_t *); #endif #if defined(NETBSD_PF) # include /* * We provide the ipf_checkp name just to minimize changes later. */ int (*ipf_checkp)(void *, ip_t *ip, int hlen, void *ifp, int out, mb_t **mp); #endif /* NETBSD_PF */ #if defined(__NetBSD_Version__) && (__NetBSD_Version__ >= 105110000) # include static int ipf_check_wrapper(void *, struct mbuf **, struct ifnet *, int ); static int ipf_check_wrapper(void *arg, struct mbuf **mp, struct ifnet *ifp, int dir) { struct ip *ip; int rv, hlen; #if __NetBSD_Version__ >= 200080000 /* * ensure that mbufs are writable beforehand * as it's assumed by ipf code. * XXX inefficient */ int error = m_makewritable(mp, 0, M_COPYALL, M_DONTWAIT); if (error) { m_freem(*mp); *mp = NULL; return error; } #endif ip = mtod(*mp, struct ip *); hlen = ip->ip_hl << 2; #ifdef INET #if defined(M_CSUM_TCPv4) /* * If the packet is out-bound, we can't delay checksums * here. For in-bound, the checksum has already been * validated. */ if (dir == PFIL_OUT) { if ((*mp)->m_pkthdr.csum_flags & (M_CSUM_TCPv4|M_CSUM_UDPv4)) { in_undefer_cksum_tcpudp(*mp); (*mp)->m_pkthdr.csum_flags &= ~(M_CSUM_TCPv4|M_CSUM_UDPv4); } } #endif /* M_CSUM_TCPv4 */ #endif /* INET */ /* * Note, we don't need to update the checksum, because * it has already been verified. */ rv = ipf_check(&ipfmain, ip, hlen, ifp, (dir == PFIL_OUT), mp); return (rv); } # ifdef USE_INET6 # include static int ipf_check_wrapper6(void *, struct mbuf **, struct ifnet *, int ); static int ipf_check_wrapper6(void *arg, struct mbuf **mp, struct ifnet *ifp, int dir) { #if defined(INET6) # if defined(M_CSUM_TCPv6) && (__NetBSD_Version__ > 200000000) /* * If the packet is out-bound, we can't delay checksums * here. For in-bound, the checksum has already been * validated. */ if (dir == PFIL_OUT) { if ((*mp)->m_pkthdr.csum_flags & (M_CSUM_TCPv6|M_CSUM_UDPv6)) { # if (__NetBSD_Version__ > 399000600) in6_undefer_cksum_tcpudp(*mp); # endif (*mp)->m_pkthdr.csum_flags &= ~(M_CSUM_TCPv6| M_CSUM_UDPv6); } } # endif #endif /* INET6 */ return (ipf_check(&ipfmain, mtod(*mp, struct ip *), sizeof(struct ip6_hdr), ifp, (dir == PFIL_OUT), mp)); } # endif # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) # if (__NetBSD_Version__ >= 799000400) static void ipf_pfilsync(void *, unsigned long, void *); static void ipf_pfilsync(void *hdr, unsigned long cmd, void *arg2) { /* * The interface pointer is useless for create (we have nothing to * compare it to) and at detach, the interface name is still in the * list of active NICs (albeit, down, but that's not any real * indicator) and doing ifunit() on the name will still return the * pointer, so it's not much use then, either. */ ipf_sync(&ipfmain, NULL); } # else static int ipf_pfilsync(void *, struct mbuf **, struct ifnet *, int); static int ipf_pfilsync(void *hdr, struct mbuf **mp, struct ifnet *ifp, int dir) { ipf_sync(&ipfmain, NULL); return 0; } # endif # endif #endif /* __NetBSD_Version__ >= 105110000 */ #if defined(IPFILTER_LKM) int ipf_identify(s) char *s; { if (strcmp(s, "ipl") == 0) return 1; return 0; } #endif /* IPFILTER_LKM */ #if (__NetBSD_Version__ >= 599002000) static int ipf_listener_cb(kauth_cred_t cred, kauth_action_t action, void *cookie, void *arg0, void *arg1, void *arg2, void *arg3) { int result; enum kauth_network_req req; result = KAUTH_RESULT_DEFER; req = (enum kauth_network_req)arg0; if (action != KAUTH_NETWORK_FIREWALL) return result; /* These must have came from device context. */ if ((req == KAUTH_REQ_NETWORK_FIREWALL_FW) || (req == KAUTH_REQ_NETWORK_FIREWALL_NAT)) result = KAUTH_RESULT_ALLOW; return result; } #endif /* * Try to detect the case when compiling for NetBSD with pseudo-device */ void ipfilterattach(int count) { #if (__NetBSD_Version__ >= 799003000) return; #else #if (__NetBSD_Version__ >= 599002000) ipf_listener = kauth_listen_scope(KAUTH_SCOPE_NETWORK, ipf_listener_cb, NULL); #endif if (ipf_load_all() == 0) (void) ipf_create_all(&ipfmain); #endif } int ipfattach(ipf_main_softc_t *softc) { SPL_INT(s); #if (__NetBSD_Version__ >= 499005500) int i; #endif #if defined(NETBSD_PF) && (__NetBSD_Version__ >= 104200000) int error = 0; # if defined(__NetBSD_Version__) && (__NetBSD_Version__ >= 105110000) pfil_head_t *ph_inet; # ifdef USE_INET6 pfil_head_t *ph_inet6; # endif # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) pfil_head_t *ph_ifsync; # endif # endif #endif SPL_NET(s); if ((softc->ipf_running > 0) || (ipf_checkp == ipf_check)) { printf("IP Filter: already initialized\n"); SPL_X(s); IPFERROR(130017); return EBUSY; } if (ipf_init_all(softc) < 0) { SPL_X(s); IPFERROR(130015); return EIO; } #ifdef NETBSD_PF # if (__NetBSD_Version__ >= 104200000) # if __NetBSD_Version__ >= 105110000 ph_inet = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET); # ifdef USE_INET6 ph_inet6 = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET6); # endif # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) ph_ifsync = pfil_head_get(PFIL_TYPE_IFNET, 0); # endif if (ph_inet == NULL # ifdef USE_INET6 && ph_inet6 == NULL # endif # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) && ph_ifsync == NULL # endif ) { SPL_X(s); IPFERROR(130016); return ENODEV; } if (ph_inet != NULL) error = pfil_add_hook((void *)ipf_check_wrapper, NULL, PFIL_IN|PFIL_OUT, ph_inet); else error = 0; # else error = pfil_add_hook((void *)ipf_check, PFIL_IN|PFIL_OUT, &inetsw[ip_protox[IPPROTO_IP]].pr_pfh); # endif if (error) { IPFERROR(130013); goto pfil_error; } # else pfil_add_hook((void *)ipf_check, PFIL_IN|PFIL_OUT); # endif # ifdef USE_INET6 # if __NetBSD_Version__ >= 105110000 if (ph_inet6 != NULL) error = pfil_add_hook((void *)ipf_check_wrapper6, NULL, PFIL_IN|PFIL_OUT, ph_inet6); else error = 0; if (error) { pfil_remove_hook((void *)ipf_check_wrapper6, NULL, PFIL_IN|PFIL_OUT, ph_inet6); ipfmain.ipf_interror = 130014; goto pfil_error; } # else error = pfil_add_hook((void *)ipf_check, PFIL_IN|PFIL_OUT, &inetsw[ip_protox[IPPROTO_IPV6]].pr_pfh); if (error) { pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT, &inetsw[ip_protox[IPPROTO_IP]].pr_pfh); IPFERROR(130014); goto pfil_error; } # endif # endif # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) if (ph_ifsync != NULL) #if (__NetBSD_Version__ >= 799000400) (void) pfil_add_ihook((void *)ipf_pfilsync, NULL, PFIL_IFNET, ph_ifsync); #else (void) pfil_add_hook((void *)ipf_pfilsync, NULL, PFIL_IFNET, ph_ifsync); #endif # endif #endif #if (__NetBSD_Version__ >= 499005500) for (i = 0; i < IPL_LOGSIZE; i++) selinit(&ipfmain.ipf_selwait[i]); #else bzero((char *)ipfmain.ipf_selwait, sizeof(ipfmain.ipf_selwait)); #endif ipf_savep = ipf_checkp; ipf_checkp = ipf_check; #ifdef INET if (softc->ipf_control_forwarding & 1) ipforwarding = 1; #endif ipid = 0; SPL_X(s); #if (__NetBSD_Version__ >= 104010000) # if (__NetBSD_Version__ >= 499002000) callout_init(&softc->ipf_slow_ch, 0); # else callout_init(&softc->ipf_slow_ch); # endif callout_reset(&softc->ipf_slow_ch, (hz / IPF_HZ_DIVIDE) * IPF_HZ_MULT, ipf_timer_func, softc); #else timeout(ipf_timer_func, softc, (hz / IPF_HZ_DIVIDE) * IPF_HZ_MULT); #endif return 0; #if __NetBSD_Version__ >= 105110000 pfil_error: SPL_X(s); ipf_fini_all(softc); return error; #endif } static void ipf_timer_func(void *ptr) { ipf_main_softc_t *softc = ptr; SPL_INT(s); SPL_NET(s); READ_ENTER(&softc->ipf_global); if (softc->ipf_running > 0) ipf_slowtimer(softc); if (softc->ipf_running == -1 || softc->ipf_running == 1) { #if NETBSD_GE_REV(104240000) callout_reset(&softc->ipf_slow_ch, hz / 2, ipf_timer_func, softc); #else timeout(ipf_timer_func, softc, (hz / IPF_HZ_DIVIDE) * IPF_HZ_MULT); #endif } RWLOCK_EXIT(&softc->ipf_global); SPL_X(s); } /* * Disable the filter by removing the hooks from the IP input/output * stream. */ int ipfdetach(ipf_main_softc_t *softc) { SPL_INT(s); #if (__NetBSD_Version__ >= 499005500) int i; #endif #if defined(NETBSD_PF) && (__NetBSD_Version__ >= 104200000) int error = 0; # if __NetBSD_Version__ >= 105150000 pfil_head_t *ph_inet = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET); # ifdef USE_INET6 pfil_head_t *ph_inet6 = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET6); # endif # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) struct pfil_head *ph_ifsync = pfil_head_get(PFIL_TYPE_IFNET, 0); # endif # endif #endif SPL_NET(s); #if (__NetBSD_Version__ >= 104010000) if (softc->ipf_running > 0) callout_stop(&softc->ipf_slow_ch); #else untimeout(ipf_slowtimer, NULL); #endif /* NetBSD */ ipf_checkp = ipf_savep; (void) ipf_flush(softc, IPL_LOGIPF, FR_INQUE|FR_OUTQUE|FR_INACTIVE); (void) ipf_flush(softc, IPL_LOGIPF, FR_INQUE|FR_OUTQUE); #ifdef INET if (softc->ipf_control_forwarding & 2) ipforwarding = 0; #endif #ifdef NETBSD_PF # if (__NetBSD_Version__ >= 104200000) # if __NetBSD_Version__ >= 105110000 # if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET) # if __NetBSD_Version__ >= 799000400 (void) pfil_remove_ihook((void *)ipf_pfilsync, NULL, PFIL_IFNET, ph_ifsync); # else (void) pfil_remove_hook((void *)ipf_pfilsync, NULL, PFIL_IFNET, ph_ifsync); # endif # endif if (ph_inet != NULL) error = pfil_remove_hook((void *)ipf_check_wrapper, NULL, PFIL_IN|PFIL_OUT, ph_inet); else error = 0; # else error = pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT, &inetsw[ip_protox[IPPROTO_IP]].pr_pfh); # endif if (error) { SPL_X(s); IPFERROR(130011); return error; } # else pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT); # endif # ifdef USE_INET6 # if __NetBSD_Version__ >= 105110000 if (ph_inet6 != NULL) error = pfil_remove_hook((void *)ipf_check_wrapper6, NULL, PFIL_IN|PFIL_OUT, ph_inet6); else error = 0; # else error = pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT, &inetsw[ip_protox[IPPROTO_IPV6]].pr_pfh); # endif if (error) { SPL_X(s); IPFERROR(130012); return error; } # endif #endif SPL_X(s); #if (__NetBSD_Version__ >= 499005500) for (i = 0; i < IPL_LOGSIZE; i++) seldestroy(&ipfmain.ipf_selwait[i]); #endif ipf_fini_all(softc); return 0; } /* * Filter ioctl interface. */ int ipfioctl(dev_t dev, u_long cmd, #if (__NetBSD_Version__ >= 499001000) void *data, #else caddr_t data, #endif int mode #if (NetBSD >= 199511) # if (__NetBSD_Version__ >= 399001400) , struct lwp *p # if (__NetBSD_Version__ >= 399002000) # define UID(l) kauth_cred_getuid((l)->l_cred) # else # define UID(l) ((l)->l_proc->p_cred->p_ruid) # endif # else , struct proc *p # define UID(p) ((p)->p_cred->p_ruid) # endif #endif ) { int error = 0, unit = 0; SPL_INT(s); #if (__NetBSD_Version__ >= 399002000) if ((mode & FWRITE) && kauth_authorize_network(p->l_cred, KAUTH_NETWORK_FIREWALL, KAUTH_REQ_NETWORK_FIREWALL_FW, NULL, NULL, NULL)) { ipfmain.ipf_interror = 130005; return EPERM; } #else if ((securelevel >= 2) && (mode & FWRITE)) { ipfmain.ipf_interror = 130001; return EPERM; } #endif unit = GET_MINOR(dev); if ((IPL_LOGMAX < unit) || (unit < 0)) { ipfmain.ipf_interror = 130002; return ENXIO; } if (ipfmain.ipf_running <= 0) { if (unit != IPL_LOGIPF && cmd != SIOCIPFINTERROR) { ipfmain.ipf_interror = 130003; return EIO; } if (cmd != SIOCIPFGETNEXT && cmd != SIOCIPFGET && cmd != SIOCIPFSET && cmd != SIOCFRENB && cmd != SIOCGETFS && cmd != SIOCGETFF && cmd != SIOCIPFINTERROR) { ipfmain.ipf_interror = 130004; return EIO; } } SPL_NET(s); error = ipf_ioctlswitch(&ipfmain, unit, data, cmd, mode, UID(p), p); if (error != -1) { SPL_X(s); return error; } SPL_X(s); return error; } /* * ipf_send_reset - this could conceivably be a call to tcp_respond(), but that * requires a large amount of setting up and isn't any more efficient. */ int ipf_send_reset(fr_info_t *fin) { struct tcphdr *tcp, *tcp2; int tlen = 0, hlen; struct mbuf *m; #ifdef USE_INET6 ip6_t *ip6; #endif ip_t *ip; tcp = fin->fin_dp; if (tcp->th_flags & TH_RST) return -1; /* feedback loop */ if (ipf_checkl4sum(fin) == -1) return -1; tlen = fin->fin_dlen - (TCP_OFF(tcp) << 2) + ((tcp->th_flags & TH_SYN) ? 1 : 0) + ((tcp->th_flags & TH_FIN) ? 1 : 0); #ifdef USE_INET6 hlen = (fin->fin_v == 6) ? sizeof(ip6_t) : sizeof(ip_t); #else hlen = sizeof(ip_t); #endif #ifdef MGETHDR MGETHDR(m, M_DONTWAIT, MT_HEADER); #else MGET(m, M_DONTWAIT, MT_HEADER); #endif if (m == NULL) return -1; if (sizeof(*tcp2) + hlen > MHLEN) { MCLGET(m, M_DONTWAIT); if (m == NULL) return -1; if ((m->m_flags & M_EXT) == 0) { FREE_MB_T(m); return -1; } } m->m_len = sizeof(*tcp2) + hlen; m->m_data += max_linkhdr; m->m_pkthdr.len = m->m_len; m_reset_rcvif(m); ip = mtod(m, struct ip *); bzero((char *)ip, hlen); #ifdef USE_INET6 ip6 = (ip6_t *)ip; #endif bzero((char *)ip, sizeof(*tcp2) + hlen); tcp2 = (struct tcphdr *)((char *)ip + hlen); tcp2->th_sport = tcp->th_dport; tcp2->th_dport = tcp->th_sport; if (tcp->th_flags & TH_ACK) { tcp2->th_seq = tcp->th_ack; tcp2->th_flags = TH_RST; tcp2->th_ack = 0; } else { tcp2->th_seq = 0; tcp2->th_ack = ntohl(tcp->th_seq); tcp2->th_ack += tlen; tcp2->th_ack = htonl(tcp2->th_ack); tcp2->th_flags = TH_RST|TH_ACK; } tcp2->th_x2 = 0; TCP_OFF_A(tcp2, sizeof(*tcp2) >> 2); tcp2->th_win = tcp->th_win; tcp2->th_sum = 0; tcp2->th_urp = 0; #ifdef USE_INET6 if (fin->fin_v == 6) { ip6->ip6_flow = ((ip6_t *)fin->fin_ip)->ip6_flow; ip6->ip6_plen = htons(sizeof(struct tcphdr)); ip6->ip6_nxt = IPPROTO_TCP; ip6->ip6_hlim = 0; ip6->ip6_src = fin->fin_dst6.in6; ip6->ip6_dst = fin->fin_src6.in6; tcp2->th_sum = in6_cksum(m, IPPROTO_TCP, sizeof(*ip6), sizeof(*tcp2)); return ipf_send_ip(fin, m); } #endif #ifdef INET ip->ip_p = IPPROTO_TCP; ip->ip_len = htons(sizeof(struct tcphdr)); ip->ip_src.s_addr = fin->fin_daddr; ip->ip_dst.s_addr = fin->fin_saddr; tcp2->th_sum = in_cksum(m, hlen + sizeof(*tcp2)); ip->ip_len = hlen + sizeof(*tcp2); return ipf_send_ip(fin, m); #else return 0; #endif } /* * Expects ip_len to be in host byte order when called. */ static int ipf_send_ip(fr_info_t *fin, mb_t *m) { fr_info_t fnew; #ifdef INET ip_t *oip; #endif ip_t *ip; int hlen; ip = mtod(m, ip_t *); bzero((char *)&fnew, sizeof(fnew)); fnew.fin_main_soft = fin->fin_main_soft; IP_V_A(ip, fin->fin_v); switch (fin->fin_v) { #ifdef INET case 4 : oip = fin->fin_ip; hlen = sizeof(*oip); fnew.fin_v = 4; fnew.fin_p = ip->ip_p; fnew.fin_plen = ntohs(ip->ip_len); HTONS(ip->ip_len); IP_HL_A(ip, sizeof(*oip) >> 2); ip->ip_tos = oip->ip_tos; ip->ip_id = ipf_nextipid(fin); ip->ip_off = htons(ip_mtudisc ? IP_DF : 0); ip->ip_ttl = ip_defttl; ip->ip_sum = 0; break; #endif #ifdef USE_INET6 case 6 : { ip6_t *ip6 = (ip6_t *)ip; ip6->ip6_vfc = 0x60; ip6->ip6_hlim = IPDEFTTL; hlen = sizeof(*ip6); fnew.fin_p = ip6->ip6_nxt; fnew.fin_v = 6; fnew.fin_plen = ntohs(ip6->ip6_plen) + hlen; break; } #endif default : return EINVAL; } #ifdef KAME_IPSEC m_reset_rcvif(m); #endif fnew.fin_ifp = fin->fin_ifp; fnew.fin_flx = FI_NOCKSUM; fnew.fin_m = m; fnew.fin_ip = ip; fnew.fin_mp = &m; fnew.fin_hlen = hlen; fnew.fin_dp = (char *)ip + hlen; (void) ipf_makefrip(hlen, ip, &fnew); return ipf_fastroute(m, &m, &fnew, NULL); } int ipf_send_icmp_err(int type, fr_info_t *fin, int dst) { int err, hlen, xtra, iclen, ohlen, avail; struct in_addr dst4; struct icmp *icmp; struct mbuf *m; i6addr_t dst6; void *ifp; #ifdef USE_INET6 int code; ip6_t *ip6; #endif ip_t *ip, *ip2; if ((type < 0) || (type > ICMP_MAXTYPE)) return -1; #ifdef USE_INET6 code = fin->fin_icode; if ((code < 0) || (code >= sizeof(icmptoicmp6unreach)/sizeof(int))) return -1; #endif if (ipf_checkl4sum(fin) == -1) return -1; #ifdef MGETHDR MGETHDR(m, M_DONTWAIT, MT_HEADER); #else MGET(m, M_DONTWAIT, MT_HEADER); #endif if (m == NULL) return -1; avail = MHLEN; xtra = 0; hlen = 0; ohlen = 0; dst4.s_addr = 0; ifp = fin->fin_ifp; if (fin->fin_v == 4) { if ((fin->fin_p == IPPROTO_ICMP) && !(fin->fin_flx & FI_SHORT)) switch (ntohs(fin->fin_data[0]) >> 8) { case ICMP_ECHO : case ICMP_TSTAMP : case ICMP_IREQ : case ICMP_MASKREQ : break; default : FREE_MB_T(m); return 0; } if (dst == 0) { if (ipf_ifpaddr(&ipfmain, 4, FRI_NORMAL, ifp, &dst6, NULL) == -1) { FREE_MB_T(m); return -1; } dst4 = dst6.in4; } else dst4.s_addr = fin->fin_daddr; hlen = sizeof(ip_t); ohlen = fin->fin_hlen; iclen = hlen + offsetof(struct icmp, icmp_ip) + ohlen; if (fin->fin_hlen < fin->fin_plen) xtra = MIN(fin->fin_dlen, 8); else xtra = 0; } #ifdef USE_INET6 else if (fin->fin_v == 6) { hlen = sizeof(ip6_t); ohlen = sizeof(ip6_t); iclen = hlen + offsetof(struct icmp, icmp_ip) + ohlen; type = icmptoicmp6types[type]; if (type == ICMP6_DST_UNREACH) code = icmptoicmp6unreach[code]; if (iclen + max_linkhdr + fin->fin_plen > avail) { MCLGET(m, M_DONTWAIT); if (m == NULL) return -1; if ((m->m_flags & M_EXT) == 0) { FREE_MB_T(m); return -1; } avail = MCLBYTES; } xtra = MIN(fin->fin_plen, avail - iclen - max_linkhdr); xtra = MIN(xtra, IPV6_MMTU - iclen); if (dst == 0 && !IN6_IS_ADDR_LINKLOCAL(&fin->fin_dst6.in6)) { if (ipf_ifpaddr(&ipfmain, 6, FRI_NORMAL, ifp, &dst6, NULL) == -1) { FREE_MB_T(m); return -1; } } else dst6 = fin->fin_dst6; } #endif else { FREE_MB_T(m); return -1; } avail -= (max_linkhdr + iclen); if (avail < 0) { FREE_MB_T(m); return -1; } if (xtra > avail) xtra = avail; iclen += xtra; m->m_data += max_linkhdr; m_reset_rcvif(m); m->m_pkthdr.len = iclen; m->m_len = iclen; ip = mtod(m, ip_t *); icmp = (struct icmp *)((char *)ip + hlen); ip2 = (ip_t *)&icmp->icmp_ip; icmp->icmp_type = type; icmp->icmp_code = fin->fin_icode; icmp->icmp_cksum = 0; #ifdef icmp_nextmtu if (type == ICMP_UNREACH && fin->fin_icode == ICMP_UNREACH_NEEDFRAG) { if (fin->fin_mtu != 0) { icmp->icmp_nextmtu = htons(fin->fin_mtu); } else if (ifp != NULL) { icmp->icmp_nextmtu = htons(GETIFMTU_4(ifp)); } else { /* make up a number... */ icmp->icmp_nextmtu = htons(fin->fin_plen - 20); } } #endif bcopy((char *)fin->fin_ip, (char *)ip2, ohlen); #if defined(M_CSUM_IPv4) /* * Clear any in-bound checksum flags for this packet. */ m->m_pkthdr.csuminfo = 0; #endif /* __NetBSD__ && M_CSUM_IPv4 */ #ifdef USE_INET6 ip6 = (ip6_t *)ip; if (fin->fin_v == 6) { ip6->ip6_flow = ((ip6_t *)fin->fin_ip)->ip6_flow; ip6->ip6_plen = htons(iclen - hlen); ip6->ip6_nxt = IPPROTO_ICMPV6; ip6->ip6_hlim = 0; ip6->ip6_src = dst6.in6; ip6->ip6_dst = fin->fin_src6.in6; if (xtra > 0) bcopy((char *)fin->fin_ip + ohlen, (char *)&icmp->icmp_ip + ohlen, xtra); icmp->icmp_cksum = in6_cksum(m, IPPROTO_ICMPV6, sizeof(*ip6), iclen - hlen); } else #endif { ip->ip_p = IPPROTO_ICMP; ip->ip_src.s_addr = dst4.s_addr; ip->ip_dst.s_addr = fin->fin_saddr; if (xtra > 0) bcopy((char *)fin->fin_ip + ohlen, (char *)&icmp->icmp_ip + ohlen, xtra); icmp->icmp_cksum = ipf_cksum((u_short *)icmp, sizeof(*icmp) + 8); ip->ip_len = iclen; ip->ip_p = IPPROTO_ICMP; } err = ipf_send_ip(fin, m); return err; } /* * m0 - pointer to mbuf where the IP packet starts * mpp - pointer to the mbuf pointer that is the start of the mbuf chain */ int ipf_fastroute(mb_t *m0, mb_t **mpp, fr_info_t *fin, frdest_t *fdp) { register struct ip *ip, *mhip; register struct mbuf *m = *mpp; register struct route *ro; int len, off, error = 0, hlen, code; struct ifnet *ifp, *sifp; ipf_main_softc_t *softc; #if __NetBSD_Version__ >= 499001100 union { struct sockaddr dst; struct sockaddr_in dst4; } u; #else struct sockaddr_in *dst4; #endif struct sockaddr *dst; u_short ip_off, ip_len; struct route iproute; struct rtentry *rt; frdest_t node; frentry_t *fr; if (fin->fin_v == 6) { #ifdef USE_INET6 error = ipf_fastroute6(m0, mpp, fin, fdp); #else error = EPROTONOSUPPORT; #endif if ((error != 0) && (*mpp != NULL)) FREE_MB_T(*mpp); return error; } #ifndef INET FREE_MB_T(*mpp); return EPROTONOSUPPORT; #else hlen = fin->fin_hlen; ip = mtod(m0, struct ip *); softc = fin->fin_main_soft; rt = NULL; ifp = NULL; # if defined(M_CSUM_IPv4) /* * Clear any in-bound checksum flags for this packet. */ m0->m_pkthdr.csuminfo = 0; # endif /* __NetBSD__ && M_CSUM_IPv4 */ /* * Route packet. */ ro = &iproute; memset(ro, 0, sizeof(*ro)); fr = fin->fin_fr; if ((fr != NULL) && !(fr->fr_flags & FR_KEEPSTATE) && (fdp != NULL) && (fdp->fd_type == FRD_DSTLIST)) { if (ipf_dstlist_select_node(fin, fdp->fd_ptr, NULL, &node) == 0) fdp = &node; } if (fdp != NULL) ifp = fdp->fd_ptr; else ifp = fin->fin_ifp; if ((ifp == NULL) && ((fr == NULL) || !(fr->fr_flags & FR_FASTROUTE))) { error = -2; goto bad; } # if __NetBSD_Version__ >= 499001100 if ((fdp != NULL) && (fdp->fd_ip.s_addr != 0)) sockaddr_in_init(&u.dst4, &fdp->fd_ip, 0); else sockaddr_in_init(&u.dst4, &ip->ip_dst, 0); dst = &u.dst; rtcache_setdst(ro, dst); rt = rtcache_init(ro); # else dst4 = (struct sockaddr_in *)&ro->ro_dst; dst = (struct sockaddr *)dst4; dst4->sin_family = AF_INET; dst4->sin_addr = ip->ip_dst; if ((fdp != NULL) && (fdp->fd_ip.s_addr != 0)) dst4->sin_addr = fdp->fd_ip; dst4->sin_len = sizeof(*dst); rtalloc(ro); rt = ro->ro_rt; # endif if ((ifp == NULL) && (rt != NULL)) ifp = rt->rt_ifp; if ((rt == NULL) || (ifp == NULL)) { #ifdef INET if (in_localaddr(ip->ip_dst)) error = EHOSTUNREACH; else #endif error = ENETUNREACH; goto bad; } if (rt->rt_flags & RTF_GATEWAY) dst = rt->rt_gateway; rt->rt_use++; /* * For input packets which are being "fastrouted", they won't * go back through output filtering and miss their chance to get * NAT'd and counted. Duplicated packets aren't considered to be * part of the normal packet stream, so do not NAT them or pass * them through stateful checking, etc. */ if ((fdp != &fr->fr_dif) && (fin->fin_out == 0)) { sifp = fin->fin_ifp; fin->fin_ifp = ifp; fin->fin_out = 1; (void) ipf_acctpkt(fin, NULL); fin->fin_fr = NULL; if (!fr || !(fr->fr_flags & FR_RETMASK)) { u_32_t pass; (void) ipf_state_check(fin, &pass); } switch (ipf_nat_checkout(fin, NULL)) { case 0 : break; case 1 : ip->ip_sum = 0; break; case -1 : error = -1; goto bad; break; } fin->fin_ifp = sifp; fin->fin_out = 0; } else ip->ip_sum = 0; /* * If small enough for interface, can just send directly. */ m_set_rcvif(m, ifp); ip_len = ntohs(ip->ip_len); if (ip_len <= ifp->if_mtu) { # if defined(M_CSUM_IPv4) # if (__NetBSD_Version__ >= 105009999) if (ifp->if_csum_flags_tx & M_CSUM_IPv4) m->m_pkthdr.csuminfo |= M_CSUM_IPv4; # else if (ifp->if_capabilities & IFCAP_CSUM_IPv4) m->m_pkthdr.csuminfo |= M_CSUM_IPv4; # endif /* (__NetBSD_Version__ >= 105009999) */ else if (ip->ip_sum == 0) ip->ip_sum = in_cksum(m, hlen); # else if (!ip->ip_sum) ip->ip_sum = in_cksum(m, hlen); # endif /* M_CSUM_IPv4 */ error = if_output_lock(ifp, ifp, m, dst, rt); goto done; } /* * Too large for interface; fragment if possible. * Must be able to put at least 8 bytes per fragment. */ ip_off = ntohs(ip->ip_off); if (ip_off & IP_DF) { error = EMSGSIZE; goto bad; } len = (ifp->if_mtu - hlen) &~ 7; if (len < 8) { error = EMSGSIZE; goto bad; } { int mhlen, firstlen = len; struct mbuf **mnext = &m->m_act; /* * Loop through length of segment after first fragment, * make new header and copy data of each part and link onto chain. */ m0 = m; mhlen = sizeof (struct ip); for (off = hlen + len; off < ip_len; off += len) { # ifdef MGETHDR MGETHDR(m, M_DONTWAIT, MT_HEADER); # else MGET(m, M_DONTWAIT, MT_HEADER); # endif if (m == 0) { m = m0; error = ENOBUFS; goto bad; } m->m_data += max_linkhdr; mhip = mtod(m, struct ip *); bcopy((char *)ip, (char *)mhip, sizeof(*ip)); #ifdef INET if (hlen > sizeof (struct ip)) { mhlen = ip_optcopy(ip, mhip) + sizeof (struct ip); IP_HL_A(mhip, mhlen >> 2); } #endif m->m_len = mhlen; mhip->ip_off = ((off - hlen) >> 3) + ip_off; if (off + len >= ip_len) len = ip_len - off; else mhip->ip_off |= IP_MF; mhip->ip_len = htons((u_short)(len + mhlen)); m->m_next = m_copym(m0, off, len, M_DONTWAIT); if (m->m_next == 0) { error = ENOBUFS; /* ??? */ goto sendorfree; } m->m_pkthdr.len = mhlen + len; m_reset_rcvif(m); mhip->ip_off = htons((u_short)mhip->ip_off); mhip->ip_sum = 0; #ifdef INET mhip->ip_sum = in_cksum(m, mhlen); #endif *mnext = m; mnext = &m->m_act; } /* * Update first fragment by trimming what's been copied out * and updating header, then send each fragment (in order). */ m_adj(m0, hlen + firstlen - ip_len); ip->ip_len = htons((u_short)(hlen + firstlen)); ip->ip_off = htons((u_short)IP_MF); ip->ip_sum = 0; #ifdef INET ip->ip_sum = in_cksum(m0, hlen); #endif sendorfree: for (m = m0; m; m = m0) { m0 = m->m_act; m->m_act = 0; if (error == 0) { KERNEL_LOCK(1, NULL); error = (*ifp->if_output)(ifp, m, dst, rt); KERNEL_UNLOCK_ONE(NULL); } else { FREE_MB_T(m); } } } done: if (!error) softc->ipf_frouteok[0]++; else softc->ipf_frouteok[1]++; # if __NetBSD_Version__ >= 499001100 rtcache_unref(rt, ro); rtcache_free(ro); # else if (rt) { RTFREE(rt); } # endif return error; bad: if (error == EMSGSIZE) { sifp = fin->fin_ifp; code = fin->fin_icode; fin->fin_icode = ICMP_UNREACH_NEEDFRAG; fin->fin_ifp = ifp; (void) ipf_send_icmp_err(ICMP_UNREACH, fin, 1); fin->fin_ifp = sifp; fin->fin_icode = code; } FREE_MB_T(m); goto done; #endif /* INET */ } #if defined(USE_INET6) /* * This is the IPv6 specific fastroute code. It doesn't clean up the mbuf's * or ensure that it is an IPv6 packet that is being forwarded, those are * expected to be done by the called (ipf_fastroute). */ static int ipf_fastroute6(struct mbuf *m0, struct mbuf **mpp, fr_info_t *fin, frdest_t *fdp) { # if __NetBSD_Version__ >= 499001100 struct route ip6route; const struct sockaddr *dst; union { struct sockaddr dst; struct sockaddr_in6 dst6; } u; struct route *ro; # else struct route_in6 ip6route; struct sockaddr_in6 *dst6; struct route_in6 *ro; # endif struct rtentry *rt; struct ifnet *ifp; u_long mtu; int error; error = 0; ro = &ip6route; if (fdp != NULL) ifp = fdp->fd_ptr; else ifp = fin->fin_ifp; memset(ro, 0, sizeof(*ro)); # if __NetBSD_Version__ >= 499001100 if (fdp != NULL && IP6_NOTZERO(&fdp->fd_ip6)) sockaddr_in6_init(&u.dst6, &fdp->fd_ip6.in6, 0, 0, 0); else sockaddr_in6_init(&u.dst6, &fin->fin_fi.fi_dst.in6, 0, 0, 0); if ((error = in6_setscope(&u.dst6.sin6_addr, ifp, &u.dst6.sin6_scope_id)) != 0) return error; if ((error = sa6_embedscope(&u.dst6, 0)) != 0) return error; dst = &u.dst; rtcache_setdst(ro, dst); rt = rtcache_init(ro); if ((ifp == NULL) && (rt != NULL)) ifp = rt->rt_ifp; # else dst6 = (struct sockaddr_in6 *)&ro->ro_dst; dst6->sin6_family = AF_INET6; dst6->sin6_len = sizeof(struct sockaddr_in6); dst6->sin6_addr = fin->fin_fi.fi_dst.in6; /* KAME */ if (IN6_IS_ADDR_LINKLOCAL(&dst6->sin6_addr)) dst6->sin6_addr.s6_addr16[1] = htons(ifp->if_index); if (fdp != NULL) { if (IP6_NOTZERO(&fdp->fd_ip6)) dst6->sin6_addr = fdp->fd_ip6.in6; } rtalloc((struct route *)ro); if ((ifp == NULL) && (ro->ro_rt != NULL)) ifp = ro->ro_rt->rt_ifp; rt = ro->ro_rt; # endif if ((rt == NULL) || (ifp == NULL)) { error = EHOSTUNREACH; goto bad; } { # if (__NetBSD_Version__ >= 106010000) && !defined(IN6_LINKMTU) struct in6_ifextra *ife; # endif if (rt->rt_flags & RTF_GATEWAY) # if __NetBSD_Version__ >= 499001100 dst = rt->rt_gateway; # else dst6 = (struct sockaddr_in6 *)rt->rt_gateway; # endif rt->rt_use++; /* Determine path MTU. */ # if (__NetBSD_Version__ <= 106009999) mtu = nd_ifinfo[ifp->if_index].linkmtu; # else # ifdef IN6_LINKMTU mtu = IN6_LINKMTU(ifp); # else ife = (struct in6_ifextra *)(ifp)->if_afdata[AF_INET6]; mtu = ife->nd_ifinfo[ifp->if_index].linkmtu; # endif # endif if ((error == 0) && (m0->m_pkthdr.len <= mtu)) { # if __NetBSD_Version__ >= 499001100 error = ip6_if_output(ifp, ifp, m0, satocsin6(dst), rt); # else error = nd6_output(ifp, ifp, m0, dst6, rt); # endif if (error) *mpp = NULL; /* m0 has been freed */ } else { error = EMSGSIZE; } } bad: # if __NetBSD_Version__ >= 499001100 rtcache_unref(rt, ro); rtcache_free(ro); # else if (ro->ro_rt != NULL) { RTFREE(((struct route *)ro)->ro_rt); } # endif return error; } #endif /* INET6 */ int ipf_verifysrc(fr_info_t *fin) { #if __NetBSD_Version__ >= 499001100 union { struct sockaddr dst; struct sockaddr_in dst4; } u; struct rtentry *rt; #else struct sockaddr_in *dst; #endif struct route iproute; int rc; #if __NetBSD_Version__ >= 499001100 sockaddr_in_init(&u.dst4, &fin->fin_src, 0); rtcache_setdst(&iproute, &u.dst); rt = rtcache_init(&iproute); if (rt == NULL) rc = 0; else rc = (fin->fin_ifp == rt->rt_ifp); rtcache_unref(rt, &iproute); rtcache_free(&iproute); #else dst = (struct sockaddr_in *)&iproute.ro_dst; dst->sin_len = sizeof(*dst); dst->sin_family = AF_INET; dst->sin_addr = fin->fin_src; rtalloc(&iproute); if (iproute.ro_rt == NULL) return 0; rc = (fin->fin_ifp == iproute.ro_rt->rt_ifp); RTFREE(iproute.ro_rt); #endif return rc; } /* * return the first IP Address associated with an interface */ int ipf_ifpaddr(ipf_main_softc_t *softc, int v, int atype, void *ifptr, i6addr_t *inp, i6addr_t *inpmask) { #ifdef USE_INET6 struct in6_addr *inp6 = NULL; #endif struct sockaddr *sock, *mask; struct sockaddr_in *sin; struct ifaddr *ifa; struct ifnet *ifp; if ((ifptr == NULL) || (ifptr == (void *)-1)) return -1; ifp = ifptr; mask = NULL; if (v == 4) inp->in4.s_addr = 0; #ifdef USE_INET6 else if (v == 6) bzero((char *)inp, sizeof(*inp)); #endif ifa = IFADDR_READER_FIRST(ifp); sock = ifa ? ifa->ifa_addr : NULL; while (sock != NULL && ifa != NULL) { sin = (struct sockaddr_in *)sock; if ((v == 4) && (sin->sin_family == AF_INET)) break; #ifdef USE_INET6 if ((v == 6) && (sin->sin_family == AF_INET6)) { inp6 = &((struct sockaddr_in6 *)sin)->sin6_addr; if (!IN6_IS_ADDR_LINKLOCAL(inp6) && !IN6_IS_ADDR_LOOPBACK(inp6)) break; } #endif ifa = IFADDR_READER_NEXT(ifa); if (ifa != NULL) sock = ifa->ifa_addr; } if (ifa == NULL || sock == NULL) return -1; mask = ifa->ifa_netmask; if (atype == FRI_BROADCAST) sock = ifa->ifa_broadaddr; else if (atype == FRI_PEERADDR) sock = ifa->ifa_dstaddr; #ifdef USE_INET6 if (v == 6) return ipf_ifpfillv6addr(atype, (struct sockaddr_in6 *)sock, (struct sockaddr_in6 *)mask, inp, inpmask); #endif return ipf_ifpfillv4addr(atype, (struct sockaddr_in *)sock, (struct sockaddr_in *)mask, &inp->in4, &inpmask->in4); } u_32_t ipf_newisn(fr_info_t *fin) { #if __NetBSD_Version__ >= 105190000 /* 1.5T */ size_t asz; if (fin->fin_v == 4) asz = sizeof(struct in_addr); else if (fin->fin_v == 6) asz = sizeof(fin->fin_src); else /* XXX: no way to return error */ return 0; #ifdef INET return tcp_new_iss1((void *)&fin->fin_src, (void *)&fin->fin_dst, fin->fin_sport, fin->fin_dport, asz, 0); #else return ENOSYS; #endif #else static int iss_seq_off = 0; u_char hash[16]; u_32_t newiss; MD5_CTX ctx; /* * Compute the base value of the ISS. It is a hash * of (saddr, sport, daddr, dport, secret). */ MD5Init(&ctx); MD5Update(&ctx, (u_char *) &fin->fin_fi.fi_src, sizeof(fin->fin_fi.fi_src)); MD5Update(&ctx, (u_char *) &fin->fin_fi.fi_dst, sizeof(fin->fin_fi.fi_dst)); MD5Update(&ctx, (u_char *) &fin->fin_dat, sizeof(fin->fin_dat)); MD5Update(&ctx, ipf_iss_secret, sizeof(ipf_iss_secret)); MD5Final(hash, &ctx); memcpy(&newiss, hash, sizeof(newiss)); /* * Now increment our "timer", and add it in to * the computed value. * * XXX Use `addin'? * XXX TCP_ISSINCR too large to use? */ iss_seq_off += 0x00010000; newiss += iss_seq_off; return newiss; #endif } /* ------------------------------------------------------------------------ */ /* Function: ipf_nextipid */ /* Returns: int - 0 == success, -1 == error (packet should be droppped) */ /* Parameters: fin(I) - pointer to packet information */ /* */ /* Returns the next IPv4 ID to use for this packet. */ /* ------------------------------------------------------------------------ */ u_short ipf_nextipid(fr_info_t *fin) { #ifdef USE_MUTEXES ipf_main_softc_t *softc = fin->fin_main_soft; #endif u_short id; MUTEX_ENTER(&softc->ipf_rw); id = ipid++; MUTEX_EXIT(&softc->ipf_rw); return id; } EXTERN_INLINE int ipf_checkv4sum(fr_info_t *fin) { #ifdef M_CSUM_TCP_UDP_BAD int manual, pflag, cflags, active; mb_t *m; if ((fin->fin_flx & FI_NOCKSUM) != 0) return 0; if ((fin->fin_flx & FI_SHORT) != 0) return 1; if (fin->fin_cksum != FI_CK_NEEDED) return (fin->fin_cksum > FI_CK_NEEDED) ? 0 : -1; manual = 0; m = fin->fin_m; if (m == NULL) { manual = 1; goto skipauto; } switch (fin->fin_p) { case IPPROTO_UDP : pflag = M_CSUM_UDPv4; break; case IPPROTO_TCP : pflag = M_CSUM_TCPv4; break; default : pflag = 0; manual = 1; break; } active = ((struct ifnet *)fin->fin_ifp)->if_csum_flags_rx & pflag; active |= M_CSUM_TCP_UDP_BAD | M_CSUM_DATA; cflags = m->m_pkthdr.csum_flags & active; if (pflag != 0) { if (cflags == (pflag | M_CSUM_TCP_UDP_BAD)) { fin->fin_flx |= FI_BAD; fin->fin_cksum = FI_CK_BAD; } else if (cflags == (pflag | M_CSUM_DATA)) { if ((m->m_pkthdr.csum_data ^ 0xffff) != 0) { fin->fin_flx |= FI_BAD; fin->fin_cksum = FI_CK_BAD; } else { fin->fin_cksum = FI_CK_SUMOK; } } else if (cflags == pflag) { fin->fin_cksum = FI_CK_SUMOK; } else { manual = 1; } } skipauto: if (manual != 0) { if (ipf_checkl4sum(fin) == -1) { fin->fin_flx |= FI_BAD; return -1; } } #else if (ipf_checkl4sum(fin) == -1) { fin->fin_flx |= FI_BAD; return -1; } #endif return 0; } #ifdef USE_INET6 EXTERN_INLINE int ipf_checkv6sum(fr_info_t *fin) { # ifdef M_CSUM_TCP_UDP_BAD int manual, pflag, cflags, active; mb_t *m; if ((fin->fin_flx & FI_NOCKSUM) != 0) return 0; if ((fin->fin_flx & FI_SHORT) != 0) return 1; if (fin->fin_cksum != FI_CK_SUMOK) return (fin->fin_cksum > FI_CK_NEEDED) ? 0 : -1; manual = 0; m = fin->fin_m; switch (fin->fin_p) { case IPPROTO_UDP : pflag = M_CSUM_UDPv6; break; case IPPROTO_TCP : pflag = M_CSUM_TCPv6; break; default : pflag = 0; manual = 1; break; } active = ((struct ifnet *)fin->fin_ifp)->if_csum_flags_rx & pflag; active |= M_CSUM_TCP_UDP_BAD | M_CSUM_DATA; cflags = m->m_pkthdr.csum_flags & active; if (pflag != 0) { if (cflags == (pflag | M_CSUM_TCP_UDP_BAD)) { fin->fin_flx |= FI_BAD; } else if (cflags == (pflag | M_CSUM_DATA)) { if ((m->m_pkthdr.csum_data ^ 0xffff) != 0) fin->fin_flx |= FI_BAD; } else if (cflags == pflag) { ; } else { manual = 1; } } if (manual != 0) { if (ipf_checkl4sum(fin) == -1) { fin->fin_flx |= FI_BAD; return -1; } } # else if (ipf_checkl4sum(fin) == -1) { fin->fin_flx |= FI_BAD; return -1; } # endif return 0; } #endif /* USE_INET6 */ size_t mbufchainlen(struct mbuf *m0) { size_t len; if ((m0->m_flags & M_PKTHDR) != 0) { len = m0->m_pkthdr.len; } else { struct mbuf *m; for (m = m0, len = 0; m != NULL; m = m->m_next) len += m->m_len; } return len; } /* ------------------------------------------------------------------------ */ /* Function: ipf_pullup */ /* Returns: NULL == pullup failed, else pointer to protocol header */ /* Parameters: xmin(I)- pointer to buffer where data packet starts */ /* fin(I) - pointer to packet information */ /* len(I) - number of bytes to pullup */ /* */ /* Attempt to move at least len bytes (from the start of the buffer) into a */ /* single buffer for ease of access. Operating system native functions are */ /* used to manage buffers - if necessary. If the entire packet ends up in */ /* a single buffer, set the FI_COALESCE flag even though ipf_coalesce() has */ /* not been called. Both fin_ip and fin_dp are updated before exiting _IF_ */ /* and ONLY if the pullup succeeds. */ /* */ /* We assume that 'xmin' is a pointer to a buffer that is part of the chain */ /* of buffers that starts at *fin->fin_mp. */ /* ------------------------------------------------------------------------ */ void * ipf_pullup(mb_t *xmin, fr_info_t *fin, int len) { int dpoff, ipoff; mb_t *m = xmin; char *ip; if (m == NULL) return NULL; ip = (char *)fin->fin_ip; if ((fin->fin_flx & FI_COALESCE) != 0) return ip; ipoff = fin->fin_ipoff; if (fin->fin_dp != NULL) dpoff = (char *)fin->fin_dp - (char *)ip; else dpoff = 0; if (M_LEN(m) < len) { mb_t *n = *fin->fin_mp; /* * Assume that M_PKTHDR is set and just work with what is left * rather than check.. * Should not make any real difference, anyway. */ if (m != n) { /* * Record the mbuf that points to the mbuf that we're * about to go to work on so that we can update the * m_next appropriately later. */ for (; n->m_next != m; n = n->m_next) ; } else { n = NULL; } #ifdef MHLEN if (len > MHLEN) #else if (len > MLEN) #endif { #ifdef HAVE_M_PULLDOWN if (m_pulldown(m, 0, len, NULL) == NULL) m = NULL; #else FREE_MB_T(*fin->fin_mp); m = NULL; n = NULL; #endif } else { m = m_pullup(m, len); } if (n != NULL) n->m_next = m; if (m == NULL) { /* * When n is non-NULL, it indicates that m pointed to * a sub-chain (tail) of the mbuf and that the head * of this chain has not yet been free'd. */ if (n != NULL) { FREE_MB_T(*fin->fin_mp); } *fin->fin_mp = NULL; fin->fin_m = NULL; return NULL; } if (n == NULL) *fin->fin_mp = m; while (M_LEN(m) == 0) { m = m->m_next; } fin->fin_m = m; ip = MTOD(m, char *) + ipoff; fin->fin_ip = (ip_t *)ip; if (fin->fin_dp != NULL) fin->fin_dp = (char *)fin->fin_ip + dpoff; if (fin->fin_fraghdr != NULL) fin->fin_fraghdr = (char *)ip + ((char *)fin->fin_fraghdr - (char *)fin->fin_ip); } if (len == fin->fin_plen) fin->fin_flx |= FI_COALESCE; return ip; } int ipf_inject(fr_info_t *fin, mb_t *m) { int error; if (fin->fin_out == 0) { if (__predict_false(!pktq_enqueue(ip_pktq, m, 0))) { FREE_MB_T(m); error = ENOBUFS; } else { error = 0; } } else { error = ip_output(m, NULL, NULL, IP_FORWARDING, NULL, NULL); } return error; } u_32_t ipf_random(void) { int number; #ifdef _CPRNG_H number = cprng_fast32(); #else number = arc4random(); #endif return number; } /* * routines below for saving IP headers to buffer */ static int ipfopen(dev_t dev, int flags #if (NetBSD >= 199511) , int devtype, PROC_T *p #endif ) { u_int unit = GET_MINOR(dev); int error; if (IPL_LOGMAX < unit) { error = ENXIO; } else { switch (unit) { case IPL_LOGIPF : case IPL_LOGNAT : case IPL_LOGSTATE : case IPL_LOGAUTH : case IPL_LOGLOOKUP : case IPL_LOGSYNC : #ifdef IPFILTER_SCAN case IPL_LOGSCAN : #endif error = 0; break; default : error = ENXIO; break; } } #if (__NetBSD_Version__ >= 799003000) if (error == 0) { mutex_enter(&ipf_ref_mutex); ipf_active = 1; mutex_exit(&ipf_ref_mutex); } #endif return error; } static int ipfclose(dev_t dev, int flags #if (NetBSD >= 199511) , int devtype, PROC_T *p #endif ) { u_int unit = GET_MINOR(dev); if (IPL_LOGMAX < unit) return ENXIO; else { #if (__NetBSD_Version__ >= 799003000) mutex_enter(&ipf_ref_mutex); ipf_active = 0; mutex_exit(&ipf_ref_mutex); #endif return 0; } } /* * ipfread/ipflog * both of these must operate with at least splnet() lest they be * called during packet processing and cause an inconsistancy to appear in * the filter lists. */ static int ipfread(dev_t dev, struct uio *uio, int ioflag) { if (ipfmain.ipf_running < 1) { ipfmain.ipf_interror = 130006; return EIO; } if (GET_MINOR(dev) == IPL_LOGSYNC) return ipf_sync_read(&ipfmain, uio); #ifdef IPFILTER_LOG return ipf_log_read(&ipfmain, GET_MINOR(dev), uio); #else ipfmain.ipf_interror = 130007; return ENXIO; #endif } /* * ipfwrite * both of these must operate with at least splnet() lest they be * called during packet processing and cause an inconsistancy to appear in * the filter lists. */ static int ipfwrite(dev_t dev, struct uio *uio, int ioflag) { if (ipfmain.ipf_running < 1) { ipfmain.ipf_interror = 130008; return EIO; } if (GET_MINOR(dev) == IPL_LOGSYNC) return ipf_sync_write(&ipfmain, uio); ipfmain.ipf_interror = 130009; return ENXIO; } static int ipfpoll(dev_t dev, int events, PROC_T *p) { u_int unit = GET_MINOR(dev); int revents = 0; if (IPL_LOGMAX < unit) { ipfmain.ipf_interror = 130010; return ENXIO; } switch (unit) { case IPL_LOGIPF : case IPL_LOGNAT : case IPL_LOGSTATE : #ifdef IPFILTER_LOG if ((events & (POLLIN | POLLRDNORM)) && ipf_log_canread(&ipfmain, unit)) revents |= events & (POLLIN | POLLRDNORM); #endif break; case IPL_LOGAUTH : if ((events & (POLLIN | POLLRDNORM)) && ipf_auth_waiting(&ipfmain)) revents |= events & (POLLIN | POLLRDNORM); break; case IPL_LOGSYNC : if ((events & (POLLIN | POLLRDNORM)) && ipf_sync_canread(&ipfmain)) revents |= events & (POLLIN | POLLRDNORM); if ((events & (POLLOUT | POLLWRNORM)) && ipf_sync_canwrite(&ipfmain)) revents |= events & (POLLOUT | POLLWRNORM); break; case IPL_LOGSCAN : case IPL_LOGLOOKUP : default : break; } if ((revents == 0) && (((events & (POLLIN|POLLRDNORM)) != 0))) selrecord(p, &ipfmain.ipf_selwait[unit]); return revents; } u_int ipf_pcksum(fr_info_t *fin, int hlen, u_int sum) { struct mbuf *m; u_int sum2; int off; m = fin->fin_m; off = (char *)fin->fin_dp - (char *)fin->fin_ip; m->m_data += hlen; m->m_len -= hlen; sum2 = in_cksum(fin->fin_m, fin->fin_plen - off); m->m_len += hlen; m->m_data -= hlen; /* * Both sum and sum2 are partial sums, so combine them together. */ sum += ~sum2 & 0xffff; while (sum > 0xffff) sum = (sum & 0xffff) + (sum >> 16); sum2 = ~sum & 0xffff; return sum2; } #if (__NetBSD_Version__ >= 799003000) /* NetBSD module interface */ MODULE(MODULE_CLASS_DRIVER, ipl, "bpf_filter"); static int ipl_init(void *); static int ipl_fini(void *); static int ipl_modcmd(modcmd_t, void *); #ifdef _MODULE static devmajor_t ipl_cmaj = -1, ipl_bmaj = -1; #endif static int ipl_modcmd(modcmd_t cmd, void *opaque) { switch (cmd) { case MODULE_CMD_INIT: return ipl_init(opaque); case MODULE_CMD_FINI: return ipl_fini(opaque); default: return ENOTTY; } } static int ipl_init(void *opaque) { int error; ipf_listener = kauth_listen_scope(KAUTH_SCOPE_NETWORK, ipf_listener_cb, NULL); if ((error = ipf_load_all()) != 0) return error; if (ipf_create_all(&ipfmain) == NULL) { ipf_unload_all(); return ENODEV; } /* Initialize our mutex and reference count */ mutex_init(&ipf_ref_mutex, MUTEX_DEFAULT, IPL_NONE); ipf_active = 0; #ifdef _MODULE /* * Insert ourself into the cdevsw list. */ error = devsw_attach("ipl", NULL, &ipl_bmaj, &ipl_cdevsw, &ipl_cmaj); if (error) ipl_fini(opaque); #endif return error; } static int ipl_fini(void *opaque) { #ifdef _MODULE (void)devsw_detach(NULL, &ipl_cdevsw); #endif /* * Grab the mutex, verify that there are no references * and that there are no running filters. If either * of these exists, reinsert our cdevsw entry and return * an error. */ mutex_enter(&ipf_ref_mutex); if (ipf_active != 0 || ipfmain.ipf_running > 0) { #ifdef _MODULE (void)devsw_attach("ipl", NULL, &ipl_bmaj, &ipl_cdevsw, &ipl_cmaj); #endif mutex_exit(&ipf_ref_mutex); return EBUSY; } /* Clean up the rest of our state before being unloaded */ mutex_exit(&ipf_ref_mutex); mutex_destroy(&ipf_ref_mutex); ipf_destroy_all(&ipfmain); ipf_unload_all(); kauth_unlisten_scope(ipf_listener); return 0; } #endif /* (__NetBSD_Version__ >= 799003000) */