File: | src/usr.sbin/ospf6d/kroute.c |
Warning: | line 1098, column 8 Although the value stored to 'iface' is used in the enclosing expression, the value is never actually read from 'iface' |
Press '?' to see keyboard shortcuts
Keyboard shortcuts:
1 | /* $OpenBSD: kroute.c,v 1.68 2023/06/21 09:47:03 sthen Exp $ */ |
2 | |
3 | /* |
4 | * Copyright (c) 2004 Esben Norby <norby@openbsd.org> |
5 | * Copyright (c) 2003, 2004 Henning Brauer <henning@openbsd.org> |
6 | * |
7 | * Permission to use, copy, modify, and distribute this software for any |
8 | * purpose with or without fee is hereby granted, provided that the above |
9 | * copyright notice and this permission notice appear in all copies. |
10 | * |
11 | * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES |
12 | * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF |
13 | * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR |
14 | * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES |
15 | * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN |
16 | * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF |
17 | * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
18 | */ |
19 | |
20 | #include <sys/types.h> |
21 | #include <sys/socket.h> |
22 | #include <sys/sysctl.h> |
23 | #include <sys/tree.h> |
24 | #include <sys/uio.h> |
25 | #include <netinet/in.h> |
26 | #include <arpa/inet.h> |
27 | #include <net/if.h> |
28 | #include <net/if_dl.h> |
29 | #include <net/if_types.h> |
30 | #include <net/route.h> |
31 | #include <err.h> |
32 | #include <errno(*__errno()).h> |
33 | #include <fcntl.h> |
34 | #include <stdio.h> |
35 | #include <stdlib.h> |
36 | #include <string.h> |
37 | #include <unistd.h> |
38 | #include <limits.h> |
39 | |
40 | #include "ospf6d.h" |
41 | #include "ospfe.h" |
42 | #include "log.h" |
43 | |
44 | struct { |
45 | u_int32_t rtseq; |
46 | pid_t pid; |
47 | int fib_sync; |
48 | int fib_serial; |
49 | u_int8_t fib_prio; |
50 | int fd; |
51 | struct event ev; |
52 | struct event reload; |
53 | u_int rdomain; |
54 | #define KR_RELOAD_IDLE0 0 |
55 | #define KR_RELOAD_FETCH1 1 |
56 | #define KR_RELOAD_HOLD2 2 |
57 | int reload_state; |
58 | } kr_state; |
59 | |
60 | struct kroute_node { |
61 | RB_ENTRY(kroute_node)struct { struct kroute_node *rbe_left; struct kroute_node *rbe_right ; struct kroute_node *rbe_parent; int rbe_color; } entry; |
62 | struct kroute_node *next; |
63 | struct kroute r; |
64 | int serial; |
65 | }; |
66 | |
67 | void kr_redist_remove(struct kroute_node *, struct kroute_node *); |
68 | int kr_redist_eval(struct kroute *, struct kroute *); |
69 | void kr_redistribute(struct kroute_node *); |
70 | int kroute_compare(struct kroute_node *, struct kroute_node *); |
71 | int kr_change_fib(struct kroute_node *, struct kroute *, int, int); |
72 | int kr_delete_fib(struct kroute_node *); |
73 | |
74 | struct kroute_node *kroute_find(const struct in6_addr *, u_int8_t, |
75 | u_int8_t); |
76 | struct kroute_node *kroute_matchgw(struct kroute_node *, |
77 | struct in6_addr *, unsigned int); |
78 | int kroute_insert(struct kroute_node *); |
79 | int kroute_remove(struct kroute_node *); |
80 | void kroute_clear(void); |
81 | |
82 | struct iface *kif_update(u_short, int, struct if_data *, |
83 | struct sockaddr_dl *); |
84 | int kif_validate(u_short); |
85 | |
86 | struct kroute_node *kroute_match(struct in6_addr *); |
87 | |
88 | int protect_lo(void); |
89 | void get_rtaddrs(int, struct sockaddr *, struct sockaddr **); |
90 | void if_change(u_short, int, struct if_data *, struct sockaddr_dl *); |
91 | void if_newaddr(u_short, struct sockaddr_in6 *, |
92 | struct sockaddr_in6 *, struct sockaddr_in6 *); |
93 | void if_deladdr(u_short, struct sockaddr_in6 *, |
94 | struct sockaddr_in6 *, struct sockaddr_in6 *); |
95 | void if_announce(void *); |
96 | |
97 | int send_rtmsg(int, int, struct kroute *); |
98 | int dispatch_rtmsg(void); |
99 | int fetchtable(void); |
100 | int refetchtable(void); |
101 | int rtmsg_process(char *, size_t); |
102 | void kr_fib_reload_timer(int, short, void *); |
103 | void kr_fib_reload_arm_timer(int); |
104 | |
105 | RB_HEAD(kroute_tree, kroute_node)struct kroute_tree { struct kroute_node *rbh_root; } krt; |
106 | RB_PROTOTYPE(kroute_tree, kroute_node, entry, kroute_compare)void kroute_tree_RB_INSERT_COLOR(struct kroute_tree *, struct kroute_node *); void kroute_tree_RB_REMOVE_COLOR(struct kroute_tree *, struct kroute_node *, struct kroute_node *); struct kroute_node *kroute_tree_RB_REMOVE(struct kroute_tree *, struct kroute_node *); struct kroute_node *kroute_tree_RB_INSERT(struct kroute_tree *, struct kroute_node *); struct kroute_node *kroute_tree_RB_FIND (struct kroute_tree *, struct kroute_node *); struct kroute_node *kroute_tree_RB_NFIND(struct kroute_tree *, struct kroute_node *); struct kroute_node *kroute_tree_RB_NEXT(struct kroute_node *); struct kroute_node *kroute_tree_RB_PREV(struct kroute_node *); struct kroute_node *kroute_tree_RB_MINMAX(struct kroute_tree *, int); |
107 | RB_GENERATE(kroute_tree, kroute_node, entry, kroute_compare)void kroute_tree_RB_INSERT_COLOR(struct kroute_tree *head, struct kroute_node *elm) { struct kroute_node *parent, *gparent, *tmp ; while ((parent = (elm)->entry.rbe_parent) && (parent )->entry.rbe_color == 1) { gparent = (parent)->entry.rbe_parent ; if (parent == (gparent)->entry.rbe_left) { tmp = (gparent )->entry.rbe_right; if (tmp && (tmp)->entry.rbe_color == 1) { (tmp)->entry.rbe_color = 0; do { (parent)->entry .rbe_color = 0; (gparent)->entry.rbe_color = 1; } while (0 ); elm = gparent; continue; } if ((parent)->entry.rbe_right == elm) { do { (tmp) = (parent)->entry.rbe_right; if (((parent )->entry.rbe_right = (tmp)->entry.rbe_left)) { ((tmp)-> entry.rbe_left)->entry.rbe_parent = (parent); } do {} while (0); if (((tmp)->entry.rbe_parent = (parent)->entry.rbe_parent )) { if ((parent) == ((parent)->entry.rbe_parent)->entry .rbe_left) ((parent)->entry.rbe_parent)->entry.rbe_left = (tmp); else ((parent)->entry.rbe_parent)->entry.rbe_right = (tmp); } else (head)->rbh_root = (tmp); (tmp)->entry .rbe_left = (parent); (parent)->entry.rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent)) do {} while ( 0); } while (0); tmp = parent; parent = elm; elm = tmp; } do { (parent)->entry.rbe_color = 0; (gparent)->entry.rbe_color = 1; } while (0); do { (tmp) = (gparent)->entry.rbe_left; if (((gparent)->entry.rbe_left = (tmp)->entry.rbe_right )) { ((tmp)->entry.rbe_right)->entry.rbe_parent = (gparent ); } do {} while (0); if (((tmp)->entry.rbe_parent = (gparent )->entry.rbe_parent)) { if ((gparent) == ((gparent)->entry .rbe_parent)->entry.rbe_left) ((gparent)->entry.rbe_parent )->entry.rbe_left = (tmp); else ((gparent)->entry.rbe_parent )->entry.rbe_right = (tmp); } else (head)->rbh_root = ( tmp); (tmp)->entry.rbe_right = (gparent); (gparent)->entry .rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent )) do {} while (0); } while (0); } else { tmp = (gparent)-> entry.rbe_left; if (tmp && (tmp)->entry.rbe_color == 1) { (tmp)->entry.rbe_color = 0; do { (parent)->entry. rbe_color = 0; (gparent)->entry.rbe_color = 1; } while (0) ; elm = gparent; continue; } if ((parent)->entry.rbe_left == elm) { do { (tmp) = (parent)->entry.rbe_left; if (((parent )->entry.rbe_left = (tmp)->entry.rbe_right)) { ((tmp)-> entry.rbe_right)->entry.rbe_parent = (parent); } do {} while (0); if (((tmp)->entry.rbe_parent = (parent)->entry.rbe_parent )) { if ((parent) == ((parent)->entry.rbe_parent)->entry .rbe_left) ((parent)->entry.rbe_parent)->entry.rbe_left = (tmp); else ((parent)->entry.rbe_parent)->entry.rbe_right = (tmp); } else (head)->rbh_root = (tmp); (tmp)->entry .rbe_right = (parent); (parent)->entry.rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent)) do {} while (0); } while (0); tmp = parent; parent = elm; elm = tmp; } do { (parent)->entry.rbe_color = 0; (gparent)->entry.rbe_color = 1; } while (0); do { (tmp) = (gparent)->entry.rbe_right ; if (((gparent)->entry.rbe_right = (tmp)->entry.rbe_left )) { ((tmp)->entry.rbe_left)->entry.rbe_parent = (gparent ); } do {} while (0); if (((tmp)->entry.rbe_parent = (gparent )->entry.rbe_parent)) { if ((gparent) == ((gparent)->entry .rbe_parent)->entry.rbe_left) ((gparent)->entry.rbe_parent )->entry.rbe_left = (tmp); else ((gparent)->entry.rbe_parent )->entry.rbe_right = (tmp); } else (head)->rbh_root = ( tmp); (tmp)->entry.rbe_left = (gparent); (gparent)->entry .rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent )) do {} while (0); } while (0); } } (head->rbh_root)-> entry.rbe_color = 0; } void kroute_tree_RB_REMOVE_COLOR(struct kroute_tree *head, struct kroute_node *parent, struct kroute_node *elm) { struct kroute_node *tmp; while ((elm == ((void *)0) || (elm)->entry.rbe_color == 0) && elm != (head)-> rbh_root) { if ((parent)->entry.rbe_left == elm) { tmp = ( parent)->entry.rbe_right; if ((tmp)->entry.rbe_color == 1) { do { (tmp)->entry.rbe_color = 0; (parent)->entry. rbe_color = 1; } while (0); do { (tmp) = (parent)->entry.rbe_right ; if (((parent)->entry.rbe_right = (tmp)->entry.rbe_left )) { ((tmp)->entry.rbe_left)->entry.rbe_parent = (parent ); } do {} while (0); if (((tmp)->entry.rbe_parent = (parent )->entry.rbe_parent)) { if ((parent) == ((parent)->entry .rbe_parent)->entry.rbe_left) ((parent)->entry.rbe_parent )->entry.rbe_left = (tmp); else ((parent)->entry.rbe_parent )->entry.rbe_right = (tmp); } else (head)->rbh_root = ( tmp); (tmp)->entry.rbe_left = (parent); (parent)->entry .rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent )) do {} while (0); } while (0); tmp = (parent)->entry.rbe_right ; } if (((tmp)->entry.rbe_left == ((void *)0) || ((tmp)-> entry.rbe_left)->entry.rbe_color == 0) && ((tmp)-> entry.rbe_right == ((void *)0) || ((tmp)->entry.rbe_right) ->entry.rbe_color == 0)) { (tmp)->entry.rbe_color = 1; elm = parent; parent = (elm)->entry.rbe_parent; } else { if ( (tmp)->entry.rbe_right == ((void *)0) || ((tmp)->entry. rbe_right)->entry.rbe_color == 0) { struct kroute_node *oleft ; if ((oleft = (tmp)->entry.rbe_left)) (oleft)->entry.rbe_color = 0; (tmp)->entry.rbe_color = 1; do { (oleft) = (tmp)-> entry.rbe_left; if (((tmp)->entry.rbe_left = (oleft)->entry .rbe_right)) { ((oleft)->entry.rbe_right)->entry.rbe_parent = (tmp); } do {} while (0); if (((oleft)->entry.rbe_parent = (tmp)->entry.rbe_parent)) { if ((tmp) == ((tmp)->entry .rbe_parent)->entry.rbe_left) ((tmp)->entry.rbe_parent) ->entry.rbe_left = (oleft); else ((tmp)->entry.rbe_parent )->entry.rbe_right = (oleft); } else (head)->rbh_root = (oleft); (oleft)->entry.rbe_right = (tmp); (tmp)->entry .rbe_parent = (oleft); do {} while (0); if (((oleft)->entry .rbe_parent)) do {} while (0); } while (0); tmp = (parent)-> entry.rbe_right; } (tmp)->entry.rbe_color = (parent)->entry .rbe_color; (parent)->entry.rbe_color = 0; if ((tmp)->entry .rbe_right) ((tmp)->entry.rbe_right)->entry.rbe_color = 0; do { (tmp) = (parent)->entry.rbe_right; if (((parent)-> entry.rbe_right = (tmp)->entry.rbe_left)) { ((tmp)->entry .rbe_left)->entry.rbe_parent = (parent); } do {} while (0) ; if (((tmp)->entry.rbe_parent = (parent)->entry.rbe_parent )) { if ((parent) == ((parent)->entry.rbe_parent)->entry .rbe_left) ((parent)->entry.rbe_parent)->entry.rbe_left = (tmp); else ((parent)->entry.rbe_parent)->entry.rbe_right = (tmp); } else (head)->rbh_root = (tmp); (tmp)->entry .rbe_left = (parent); (parent)->entry.rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent)) do {} while ( 0); } while (0); elm = (head)->rbh_root; break; } } else { tmp = (parent)->entry.rbe_left; if ((tmp)->entry.rbe_color == 1) { do { (tmp)->entry.rbe_color = 0; (parent)->entry .rbe_color = 1; } while (0); do { (tmp) = (parent)->entry. rbe_left; if (((parent)->entry.rbe_left = (tmp)->entry. rbe_right)) { ((tmp)->entry.rbe_right)->entry.rbe_parent = (parent); } do {} while (0); if (((tmp)->entry.rbe_parent = (parent)->entry.rbe_parent)) { if ((parent) == ((parent )->entry.rbe_parent)->entry.rbe_left) ((parent)->entry .rbe_parent)->entry.rbe_left = (tmp); else ((parent)->entry .rbe_parent)->entry.rbe_right = (tmp); } else (head)->rbh_root = (tmp); (tmp)->entry.rbe_right = (parent); (parent)-> entry.rbe_parent = (tmp); do {} while (0); if (((tmp)->entry .rbe_parent)) do {} while (0); } while (0); tmp = (parent)-> entry.rbe_left; } if (((tmp)->entry.rbe_left == ((void *)0 ) || ((tmp)->entry.rbe_left)->entry.rbe_color == 0) && ((tmp)->entry.rbe_right == ((void *)0) || ((tmp)->entry .rbe_right)->entry.rbe_color == 0)) { (tmp)->entry.rbe_color = 1; elm = parent; parent = (elm)->entry.rbe_parent; } else { if ((tmp)->entry.rbe_left == ((void *)0) || ((tmp)-> entry.rbe_left)->entry.rbe_color == 0) { struct kroute_node *oright; if ((oright = (tmp)->entry.rbe_right)) (oright)-> entry.rbe_color = 0; (tmp)->entry.rbe_color = 1; do { (oright ) = (tmp)->entry.rbe_right; if (((tmp)->entry.rbe_right = (oright)->entry.rbe_left)) { ((oright)->entry.rbe_left )->entry.rbe_parent = (tmp); } do {} while (0); if (((oright )->entry.rbe_parent = (tmp)->entry.rbe_parent)) { if (( tmp) == ((tmp)->entry.rbe_parent)->entry.rbe_left) ((tmp )->entry.rbe_parent)->entry.rbe_left = (oright); else ( (tmp)->entry.rbe_parent)->entry.rbe_right = (oright); } else (head)->rbh_root = (oright); (oright)->entry.rbe_left = (tmp); (tmp)->entry.rbe_parent = (oright); do {} while ( 0); if (((oright)->entry.rbe_parent)) do {} while (0); } while (0); tmp = (parent)->entry.rbe_left; } (tmp)->entry.rbe_color = (parent)->entry.rbe_color; (parent)->entry.rbe_color = 0; if ((tmp)->entry.rbe_left) ((tmp)->entry.rbe_left )->entry.rbe_color = 0; do { (tmp) = (parent)->entry.rbe_left ; if (((parent)->entry.rbe_left = (tmp)->entry.rbe_right )) { ((tmp)->entry.rbe_right)->entry.rbe_parent = (parent ); } do {} while (0); if (((tmp)->entry.rbe_parent = (parent )->entry.rbe_parent)) { if ((parent) == ((parent)->entry .rbe_parent)->entry.rbe_left) ((parent)->entry.rbe_parent )->entry.rbe_left = (tmp); else ((parent)->entry.rbe_parent )->entry.rbe_right = (tmp); } else (head)->rbh_root = ( tmp); (tmp)->entry.rbe_right = (parent); (parent)->entry .rbe_parent = (tmp); do {} while (0); if (((tmp)->entry.rbe_parent )) do {} while (0); } while (0); elm = (head)->rbh_root; break ; } } } if (elm) (elm)->entry.rbe_color = 0; } struct kroute_node * kroute_tree_RB_REMOVE(struct kroute_tree *head, struct kroute_node *elm) { struct kroute_node *child, *parent, *old = elm; int color ; if ((elm)->entry.rbe_left == ((void *)0)) child = (elm)-> entry.rbe_right; else if ((elm)->entry.rbe_right == ((void *)0)) child = (elm)->entry.rbe_left; else { struct kroute_node *left; elm = (elm)->entry.rbe_right; while ((left = (elm) ->entry.rbe_left)) elm = left; child = (elm)->entry.rbe_right ; parent = (elm)->entry.rbe_parent; color = (elm)->entry .rbe_color; if (child) (child)->entry.rbe_parent = parent; if (parent) { if ((parent)->entry.rbe_left == elm) (parent )->entry.rbe_left = child; else (parent)->entry.rbe_right = child; do {} while (0); } else (head)->rbh_root = child ; if ((elm)->entry.rbe_parent == old) parent = elm; (elm)-> entry = (old)->entry; if ((old)->entry.rbe_parent) { if (((old)->entry.rbe_parent)->entry.rbe_left == old) ((old )->entry.rbe_parent)->entry.rbe_left = elm; else ((old) ->entry.rbe_parent)->entry.rbe_right = elm; do {} while (0); } else (head)->rbh_root = elm; ((old)->entry.rbe_left )->entry.rbe_parent = elm; if ((old)->entry.rbe_right) ( (old)->entry.rbe_right)->entry.rbe_parent = elm; if (parent ) { left = parent; do { do {} while (0); } while ((left = (left )->entry.rbe_parent)); } goto color; } parent = (elm)-> entry.rbe_parent; color = (elm)->entry.rbe_color; if (child ) (child)->entry.rbe_parent = parent; if (parent) { if ((parent )->entry.rbe_left == elm) (parent)->entry.rbe_left = child ; else (parent)->entry.rbe_right = child; do {} while (0); } else (head)->rbh_root = child; color: if (color == 0) kroute_tree_RB_REMOVE_COLOR (head, parent, child); return (old); } struct kroute_node * kroute_tree_RB_INSERT (struct kroute_tree *head, struct kroute_node *elm) { struct kroute_node *tmp; struct kroute_node *parent = ((void *)0); int comp = 0 ; tmp = (head)->rbh_root; while (tmp) { parent = tmp; comp = (kroute_compare)(elm, parent); if (comp < 0) tmp = (tmp )->entry.rbe_left; else if (comp > 0) tmp = (tmp)->entry .rbe_right; else return (tmp); } do { (elm)->entry.rbe_parent = parent; (elm)->entry.rbe_left = (elm)->entry.rbe_right = ((void *)0); (elm)->entry.rbe_color = 1; } while (0); if (parent != ((void *)0)) { if (comp < 0) (parent)->entry .rbe_left = elm; else (parent)->entry.rbe_right = elm; do { } while (0); } else (head)->rbh_root = elm; kroute_tree_RB_INSERT_COLOR (head, elm); return (((void *)0)); } struct kroute_node * kroute_tree_RB_FIND (struct kroute_tree *head, struct kroute_node *elm) { struct kroute_node *tmp = (head)->rbh_root; int comp; while (tmp) { comp = kroute_compare (elm, tmp); if (comp < 0) tmp = (tmp)->entry.rbe_left; else if (comp > 0) tmp = (tmp)->entry.rbe_right; else return (tmp); } return (((void *)0)); } struct kroute_node * kroute_tree_RB_NFIND (struct kroute_tree *head, struct kroute_node *elm) { struct kroute_node *tmp = (head)->rbh_root; struct kroute_node *res = ((void *)0); int comp; while (tmp) { comp = kroute_compare(elm, tmp ); if (comp < 0) { res = tmp; tmp = (tmp)->entry.rbe_left ; } else if (comp > 0) tmp = (tmp)->entry.rbe_right; else return (tmp); } return (res); } struct kroute_node * kroute_tree_RB_NEXT (struct kroute_node *elm) { if ((elm)->entry.rbe_right) { elm = (elm)->entry.rbe_right; while ((elm)->entry.rbe_left ) elm = (elm)->entry.rbe_left; } else { if ((elm)->entry .rbe_parent && (elm == ((elm)->entry.rbe_parent)-> entry.rbe_left)) elm = (elm)->entry.rbe_parent; else { while ((elm)->entry.rbe_parent && (elm == ((elm)->entry .rbe_parent)->entry.rbe_right)) elm = (elm)->entry.rbe_parent ; elm = (elm)->entry.rbe_parent; } } return (elm); } struct kroute_node * kroute_tree_RB_PREV(struct kroute_node *elm) { if ((elm)->entry.rbe_left) { elm = (elm)->entry.rbe_left ; while ((elm)->entry.rbe_right) elm = (elm)->entry.rbe_right ; } else { if ((elm)->entry.rbe_parent && (elm == ( (elm)->entry.rbe_parent)->entry.rbe_right)) elm = (elm) ->entry.rbe_parent; else { while ((elm)->entry.rbe_parent && (elm == ((elm)->entry.rbe_parent)->entry.rbe_left )) elm = (elm)->entry.rbe_parent; elm = (elm)->entry.rbe_parent ; } } return (elm); } struct kroute_node * kroute_tree_RB_MINMAX (struct kroute_tree *head, int val) { struct kroute_node *tmp = (head)->rbh_root; struct kroute_node *parent = ((void * )0); while (tmp) { parent = tmp; if (val < 0) tmp = (tmp)-> entry.rbe_left; else tmp = (tmp)->entry.rbe_right; } return (parent); } |
108 | |
109 | int |
110 | kr_init(int fs, u_int rdomain, int redis_label_or_prefix, u_int8_t fib_prio) |
111 | { |
112 | int opt = 0, rcvbuf, default_rcvbuf; |
113 | socklen_t optlen; |
114 | int filter_prio = fib_prio; |
115 | int filter_flags = RTF_LLINFO0x400 | RTF_BROADCAST0x400000; |
116 | |
117 | kr_state.fib_sync = fs; |
118 | kr_state.rdomain = rdomain; |
119 | kr_state.fib_prio = fib_prio; |
120 | |
121 | if ((kr_state.fd = socket(AF_ROUTE17, |
122 | SOCK_RAW3 | SOCK_CLOEXEC0x8000 | SOCK_NONBLOCK0x4000, AF_INET624)) == -1) { |
123 | log_warn("kr_init: socket"); |
124 | return (-1); |
125 | } |
126 | |
127 | /* not interested in my own messages */ |
128 | if (setsockopt(kr_state.fd, SOL_SOCKET0xffff, SO_USELOOPBACK0x0040, |
129 | &opt, sizeof(opt)) == -1) |
130 | log_warn("kr_init: setsockopt"); /* not fatal */ |
131 | |
132 | if (redis_label_or_prefix) { |
133 | filter_prio = 0; |
134 | log_info("%s: priority filter disabled", __func__); |
135 | } else |
136 | log_debug("%s: priority filter enabled", __func__); |
137 | |
138 | if (setsockopt(kr_state.fd, AF_ROUTE17, ROUTE_PRIOFILTER3, &filter_prio, |
139 | sizeof(filter_prio)) == -1) { |
140 | log_warn("%s: setsockopt AF_ROUTE ROUTE_PRIOFILTER", __func__); |
141 | /* not fatal */ |
142 | } |
143 | |
144 | if (setsockopt(kr_state.fd, AF_ROUTE17, ROUTE_FLAGFILTER4, &filter_flags, |
145 | sizeof(filter_flags)) == -1) { |
146 | log_warn("%s: setsockopt AF_ROUTE ROUTE_FLAGFILTER", __func__); |
147 | /* not fatal */ |
148 | } |
149 | |
150 | /* grow receive buffer, don't wanna miss messages */ |
151 | optlen = sizeof(default_rcvbuf); |
152 | if (getsockopt(kr_state.fd, SOL_SOCKET0xffff, SO_RCVBUF0x1002, |
153 | &default_rcvbuf, &optlen) == -1) |
154 | log_warn("kr_init getsockopt SOL_SOCKET SO_RCVBUF"); |
155 | else |
156 | for (rcvbuf = MAX_RTSOCK_BUF(2 * 1024 * 1024); |
157 | rcvbuf > default_rcvbuf && |
158 | setsockopt(kr_state.fd, SOL_SOCKET0xffff, SO_RCVBUF0x1002, |
159 | &rcvbuf, sizeof(rcvbuf)) == -1 && errno(*__errno()) == ENOBUFS55; |
160 | rcvbuf /= 2) |
161 | ; /* nothing */ |
162 | |
163 | kr_state.pid = getpid(); |
164 | kr_state.rtseq = 1; |
165 | |
166 | RB_INIT(&krt)do { (&krt)->rbh_root = ((void *)0); } while (0); |
167 | |
168 | if (fetchtable() == -1) |
169 | return (-1); |
170 | |
171 | if (protect_lo() == -1) |
172 | return (-1); |
173 | |
174 | event_set(&kr_state.ev, kr_state.fd, EV_READ0x02 | EV_PERSIST0x10, |
175 | kr_dispatch_msg, NULL((void *)0)); |
176 | event_add(&kr_state.ev, NULL((void *)0)); |
177 | |
178 | kr_state.reload_state = KR_RELOAD_IDLE0; |
179 | evtimer_set(&kr_state.reload, kr_fib_reload_timer, NULL)event_set(&kr_state.reload, -1, 0, kr_fib_reload_timer, ( (void *)0)); |
180 | |
181 | return (0); |
182 | } |
183 | |
184 | int |
185 | kr_change_fib(struct kroute_node *kr, struct kroute *kroute, int krcount, |
186 | int action) |
187 | { |
188 | int i; |
189 | struct kroute_node *kn, *nkn; |
190 | |
191 | if (action == RTM_ADD0x1) { |
192 | /* |
193 | * First remove all stale multipath routes. |
194 | * This step must be skipped when the action is RTM_CHANGE |
195 | * because it is already a single path route that will be |
196 | * changed. |
197 | */ |
198 | for (kn = kr; kn != NULL((void *)0); kn = nkn) { |
199 | for (i = 0; i < krcount; i++) { |
200 | if (kn->r.scope == kroute[i].scope && |
201 | IN6_ARE_ADDR_EQUAL(&kn->r.nexthop,(memcmp(&(&kn->r.nexthop)->__u6_addr.__u6_addr8 [0], &(&kroute[i].nexthop)->__u6_addr.__u6_addr8[0 ], sizeof(struct in6_addr)) == 0) |
202 | &kroute[i].nexthop)(memcmp(&(&kn->r.nexthop)->__u6_addr.__u6_addr8 [0], &(&kroute[i].nexthop)->__u6_addr.__u6_addr8[0 ], sizeof(struct in6_addr)) == 0)) |
203 | break; |
204 | } |
205 | nkn = kn->next; |
206 | if (i == krcount) { |
207 | /* stale route */ |
208 | if (kr_delete_fib(kn) == -1) |
209 | log_warnx("kr_delete_fib failed"); |
210 | /* |
211 | * if head element was removed we need to adjust |
212 | * the head |
213 | */ |
214 | if (kr == kn) |
215 | kr = nkn; |
216 | } |
217 | } |
218 | } |
219 | |
220 | /* |
221 | * now add or change the route |
222 | */ |
223 | for (i = 0; i < krcount; i++) { |
224 | /* nexthop ::1 -> ignore silently */ |
225 | if (IN6_IS_ADDR_LOOPBACK(&kroute[i].nexthop)((*(const u_int32_t *)(const void *)(&(&kroute[i].nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute[i].nexthop)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute[i].nexthop)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &kroute[i].nexthop)->__u6_addr.__u6_addr8[12]) == (__uint32_t )(__builtin_constant_p(1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t )(1) & 0xff000000) >> 24) : __swap32md(1))))) |
226 | continue; |
227 | |
228 | if (action == RTM_ADD0x1 && kr) { |
229 | for (kn = kr; kn != NULL((void *)0); kn = kn->next) { |
230 | if (kn->r.scope == kroute[i].scope && |
231 | IN6_ARE_ADDR_EQUAL(&kn->r.nexthop,(memcmp(&(&kn->r.nexthop)->__u6_addr.__u6_addr8 [0], &(&kroute[i].nexthop)->__u6_addr.__u6_addr8[0 ], sizeof(struct in6_addr)) == 0) |
232 | &kroute[i].nexthop)(memcmp(&(&kn->r.nexthop)->__u6_addr.__u6_addr8 [0], &(&kroute[i].nexthop)->__u6_addr.__u6_addr8[0 ], sizeof(struct in6_addr)) == 0)) |
233 | break; |
234 | } |
235 | |
236 | if (kn != NULL((void *)0)) |
237 | /* nexthop already present, skip it */ |
238 | continue; |
239 | } else |
240 | /* modify first entry */ |
241 | kn = kr; |
242 | |
243 | /* send update */ |
244 | if (send_rtmsg(kr_state.fd, action, &kroute[i]) == -1) |
245 | return (-1); |
246 | |
247 | /* create new entry unless we are changing the first entry */ |
248 | if (action == RTM_ADD0x1) |
249 | if ((kn = calloc(1, sizeof(*kn))) == NULL((void *)0)) |
250 | fatal(NULL((void *)0)); |
251 | |
252 | kn->r.prefix = kroute[i].prefix; |
253 | kn->r.prefixlen = kroute[i].prefixlen; |
254 | kn->r.nexthop = kroute[i].nexthop; |
255 | kn->r.scope = kroute[i].scope; |
256 | kn->r.flags = kroute[i].flags | F_OSPFD_INSERTED0x0001; |
257 | kn->r.priority = kr_state.fib_prio; |
258 | kn->r.ext_tag = kroute[i].ext_tag; |
259 | rtlabel_unref(kn->r.rtlabel); /* for RTM_CHANGE */ |
260 | kn->r.rtlabel = kroute[i].rtlabel; |
261 | |
262 | if (action == RTM_ADD0x1) |
263 | if (kroute_insert(kn) == -1) { |
264 | log_debug("kr_update_fib: cannot insert %s", |
265 | log_in6addr(&kn->r.nexthop)); |
266 | free(kn); |
267 | } |
268 | action = RTM_ADD0x1; |
269 | } |
270 | return (0); |
271 | } |
272 | |
273 | int |
274 | kr_change(struct kroute *kroute, int krcount) |
275 | { |
276 | struct kroute_node *kr; |
277 | int action = RTM_ADD0x1; |
278 | |
279 | kroute->rtlabel = rtlabel_tag2id(kroute->ext_tag); |
280 | |
281 | kr = kroute_find(&kroute->prefix, kroute->prefixlen, kr_state.fib_prio); |
282 | if (kr != NULL((void *)0) && kr->next == NULL((void *)0) && krcount == 1) { |
283 | /* |
284 | * single path OSPF route. |
285 | * The kernel does not allow to change a gateway route to a |
286 | * cloning route or contrary. In this case remove and add the |
287 | * route, otherwise change the existing one. |
288 | */ |
289 | if ((IN6_IS_ADDR_UNSPECIFIED(&kroute->nexthop)((*(const u_int32_t *)(const void *)(&(&kroute->nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute->nexthop)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute->nexthop)->__u6_addr.__u6_addr8[8 ]) == 0) && (*(const u_int32_t *)(const void *)(& (&kroute->nexthop)->__u6_addr.__u6_addr8[12]) == 0) ) && |
290 | !IN6_IS_ADDR_UNSPECIFIED(&kr->r.nexthop)((*(const u_int32_t *)(const void *)(&(&kr->r.nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->r.nexthop)->__u6_addr. __u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->r.nexthop)->__u6_addr.__u6_addr8[8]) == 0) && (*(const u_int32_t *)(const void *)(&(& kr->r.nexthop)->__u6_addr.__u6_addr8[12]) == 0))) || |
291 | (!IN6_IS_ADDR_UNSPECIFIED(&kroute->nexthop)((*(const u_int32_t *)(const void *)(&(&kroute->nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute->nexthop)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute->nexthop)->__u6_addr.__u6_addr8[8 ]) == 0) && (*(const u_int32_t *)(const void *)(& (&kroute->nexthop)->__u6_addr.__u6_addr8[12]) == 0) ) && |
292 | IN6_IS_ADDR_UNSPECIFIED(&kr->r.nexthop)((*(const u_int32_t *)(const void *)(&(&kr->r.nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->r.nexthop)->__u6_addr. __u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->r.nexthop)->__u6_addr.__u6_addr8[8]) == 0) && (*(const u_int32_t *)(const void *)(&(& kr->r.nexthop)->__u6_addr.__u6_addr8[12]) == 0)))) { |
293 | if (kr_delete_fib(kr) == 0) |
294 | kr = NULL((void *)0); |
295 | else { |
296 | log_warn("kr_change: failed to remove route: " |
297 | "%s/%d", log_in6addr(&kr->r.prefix), |
298 | kr->r.prefixlen); |
299 | return (-1); |
300 | } |
301 | } else |
302 | action = RTM_CHANGE0x3; |
303 | } |
304 | |
305 | return (kr_change_fib(kr, kroute, krcount, action)); |
306 | } |
307 | |
308 | int |
309 | kr_delete_fib(struct kroute_node *kr) |
310 | { |
311 | if (kr->r.priority != kr_state.fib_prio) |
312 | log_warn("kr_delete_fib: %s/%d has wrong priority %d", |
313 | log_in6addr(&kr->r.prefix), kr->r.prefixlen, |
314 | kr->r.priority); |
315 | |
316 | if (send_rtmsg(kr_state.fd, RTM_DELETE0x2, &kr->r) == -1) |
317 | return (-1); |
318 | |
319 | if (kroute_remove(kr) == -1) |
320 | return (-1); |
321 | |
322 | return (0); |
323 | } |
324 | |
325 | int |
326 | kr_delete(struct kroute *kroute) |
327 | { |
328 | struct kroute_node *kr, *nkr; |
329 | |
330 | if ((kr = kroute_find(&kroute->prefix, kroute->prefixlen, |
331 | kr_state.fib_prio)) == NULL((void *)0)) |
332 | return (0); |
333 | |
334 | while (kr != NULL((void *)0)) { |
335 | nkr = kr->next; |
336 | if (kr_delete_fib(kr) == -1) |
337 | return (-1); |
338 | kr = nkr; |
339 | } |
340 | |
341 | return (0); |
342 | } |
343 | |
344 | void |
345 | kr_shutdown(void) |
346 | { |
347 | kr_fib_decouple(); |
348 | kroute_clear(); |
349 | } |
350 | |
351 | void |
352 | kr_fib_couple(void) |
353 | { |
354 | struct kroute_node *kr; |
355 | struct kroute_node *kn; |
356 | |
357 | if (kr_state.fib_sync == 1) /* already coupled */ |
358 | return; |
359 | |
360 | kr_state.fib_sync = 1; |
361 | |
362 | RB_FOREACH(kr, kroute_tree, &krt)for ((kr) = kroute_tree_RB_MINMAX(&krt, -1); (kr) != ((void *)0); (kr) = kroute_tree_RB_NEXT(kr)) |
363 | if (kr->r.priority == kr_state.fib_prio) |
364 | for (kn = kr; kn != NULL((void *)0); kn = kn->next) |
365 | send_rtmsg(kr_state.fd, RTM_ADD0x1, &kn->r); |
366 | |
367 | log_info("kernel routing table coupled"); |
368 | } |
369 | |
370 | void |
371 | kr_fib_decouple(void) |
372 | { |
373 | struct kroute_node *kr; |
374 | struct kroute_node *kn; |
375 | |
376 | if (kr_state.fib_sync == 0) /* already decoupled */ |
377 | return; |
378 | |
379 | RB_FOREACH(kr, kroute_tree, &krt)for ((kr) = kroute_tree_RB_MINMAX(&krt, -1); (kr) != ((void *)0); (kr) = kroute_tree_RB_NEXT(kr)) |
380 | if (kr->r.priority == kr_state.fib_prio) |
381 | for (kn = kr; kn != NULL((void *)0); kn = kn->next) |
382 | send_rtmsg(kr_state.fd, RTM_DELETE0x2, &kn->r); |
383 | |
384 | kr_state.fib_sync = 0; |
385 | |
386 | log_info("kernel routing table decoupled"); |
387 | } |
388 | |
389 | void |
390 | kr_fib_reload_timer(int fd, short event, void *bula) |
391 | { |
392 | if (kr_state.reload_state == KR_RELOAD_FETCH1) { |
393 | kr_fib_reload(); |
394 | kr_state.reload_state = KR_RELOAD_HOLD2; |
395 | kr_fib_reload_arm_timer(KR_RELOAD_HOLD_TIMER5000); |
396 | } else { |
397 | kr_state.reload_state = KR_RELOAD_IDLE0; |
398 | } |
399 | } |
400 | |
401 | void |
402 | kr_fib_reload_arm_timer(int delay) |
403 | { |
404 | struct timeval tv; |
405 | |
406 | timerclear(&tv)(&tv)->tv_sec = (&tv)->tv_usec = 0; |
407 | tv.tv_sec = delay / 1000; |
408 | tv.tv_usec = (delay % 1000) * 1000; |
409 | |
410 | if (evtimer_add(&kr_state.reload, &tv)event_add(&kr_state.reload, &tv) == -1) |
411 | fatal("add_reload_timer"); |
412 | } |
413 | |
414 | void |
415 | kr_fib_reload(void) |
416 | { |
417 | struct kroute_node *krn, *kr, *kn; |
418 | |
419 | log_info("reloading interface list and routing table"); |
420 | |
421 | kr_state.fib_serial++; |
422 | |
423 | if (fetchifs(0) != 0 || fetchtable() != 0) |
424 | return; |
425 | |
426 | for (kr = RB_MIN(kroute_tree, &krt)kroute_tree_RB_MINMAX(&krt, -1); kr != NULL((void *)0); kr = krn) { |
427 | krn = RB_NEXT(kroute_tree, &krt, kr)kroute_tree_RB_NEXT(kr); |
428 | |
429 | do { |
430 | kn = kr->next; |
431 | |
432 | if (kr->serial != kr_state.fib_serial) { |
433 | |
434 | if (kr->r.priority == kr_state.fib_prio) { |
435 | kr->serial = kr_state.fib_serial; |
436 | if (send_rtmsg(kr_state.fd, |
437 | RTM_ADD0x1, &kr->r) != 0) |
438 | break; |
439 | } else |
440 | kroute_remove(kr); |
441 | } |
442 | |
443 | } while ((kr = kn) != NULL((void *)0)); |
444 | } |
445 | } |
446 | |
447 | void |
448 | kr_fib_update_prio(u_int8_t fib_prio) |
449 | { |
450 | struct kroute_node *kr; |
451 | |
452 | RB_FOREACH(kr, kroute_tree, &krt)for ((kr) = kroute_tree_RB_MINMAX(&krt, -1); (kr) != ((void *)0); (kr) = kroute_tree_RB_NEXT(kr)) |
453 | if ((kr->r.flags & F_OSPFD_INSERTED0x0001)) |
454 | kr->r.priority = fib_prio; |
455 | |
456 | log_info("fib priority changed from %hhu to %hhu", kr_state.fib_prio, |
457 | fib_prio); |
458 | |
459 | kr_state.fib_prio = fib_prio; |
460 | } |
461 | |
462 | void |
463 | kr_dispatch_msg(int fd, short event, void *bula) |
464 | { |
465 | /* XXX this is stupid */ |
466 | dispatch_rtmsg(); |
467 | } |
468 | |
469 | void |
470 | kr_show_route(struct imsg *imsg) |
471 | { |
472 | struct kroute_node *kr; |
473 | struct kroute_node *kn; |
474 | int flags; |
475 | struct in6_addr addr; |
476 | |
477 | switch (imsg->hdr.type) { |
478 | case IMSG_CTL_KROUTE: |
479 | if (imsg->hdr.len != IMSG_HEADER_SIZEsizeof(struct imsg_hdr) + sizeof(flags)) { |
480 | log_warnx("kr_show_route: wrong imsg len"); |
481 | return; |
482 | } |
483 | memcpy(&flags, imsg->data, sizeof(flags)); |
484 | RB_FOREACH(kr, kroute_tree, &krt)for ((kr) = kroute_tree_RB_MINMAX(&krt, -1); (kr) != ((void *)0); (kr) = kroute_tree_RB_NEXT(kr)) |
485 | if (!flags || kr->r.flags & flags) { |
486 | kn = kr; |
487 | do { |
488 | main_imsg_compose_ospfe(IMSG_CTL_KROUTE, |
489 | imsg->hdr.pid, |
490 | &kn->r, sizeof(kn->r)); |
491 | } while ((kn = kn->next) != NULL((void *)0)); |
492 | } |
493 | break; |
494 | case IMSG_CTL_KROUTE_ADDR: |
495 | if (imsg->hdr.len != IMSG_HEADER_SIZEsizeof(struct imsg_hdr) + |
496 | sizeof(struct in6_addr)) { |
497 | log_warnx("kr_show_route: wrong imsg len"); |
498 | return; |
499 | } |
500 | memcpy(&addr, imsg->data, sizeof(addr)); |
501 | kr = kroute_match(&addr); |
502 | if (kr != NULL((void *)0)) |
503 | main_imsg_compose_ospfe(IMSG_CTL_KROUTE, imsg->hdr.pid, |
504 | &kr->r, sizeof(kr->r)); |
505 | break; |
506 | default: |
507 | log_debug("kr_show_route: error handling imsg"); |
508 | break; |
509 | } |
510 | |
511 | main_imsg_compose_ospfe(IMSG_CTL_END, imsg->hdr.pid, NULL((void *)0), 0); |
512 | } |
513 | |
514 | void |
515 | kr_redist_remove(struct kroute_node *kh, struct kroute_node *kn) |
516 | { |
517 | struct kroute *kr; |
518 | |
519 | /* was the route redistributed? */ |
520 | if ((kn->r.flags & F_REDISTRIBUTED0x0200) == 0) |
521 | return; |
522 | |
523 | /* remove redistributed flag */ |
524 | kn->r.flags &= ~F_REDISTRIBUTED0x0200; |
525 | kr = &kn->r; |
526 | |
527 | /* probably inform the RDE (check if no other path is redistributed) */ |
528 | for (kn = kh; kn; kn = kn->next) |
529 | if (kn->r.flags & F_REDISTRIBUTED0x0200) |
530 | break; |
531 | |
532 | if (kn == NULL((void *)0)) |
533 | main_imsg_compose_rde(IMSG_NETWORK_DEL, 0, kr, |
534 | sizeof(struct kroute)); |
535 | } |
536 | |
537 | int |
538 | kr_redist_eval(struct kroute *kr, struct kroute *new_kr) |
539 | { |
540 | u_int32_t metric = 0; |
541 | |
542 | /* Only non-ospfd routes are considered for redistribution. */ |
543 | if (!(kr->flags & F_KERNEL0x0002)) |
544 | goto dont_redistribute; |
545 | |
546 | /* Dynamic routes are not redistributable. */ |
547 | if (kr->flags & F_DYNAMIC0x0040) |
548 | goto dont_redistribute; |
549 | |
550 | /* interface is not up and running so don't announce */ |
551 | if (kr->flags & F_DOWN0x0010) |
552 | goto dont_redistribute; |
553 | |
554 | /* |
555 | * We consider loopback, multicast, link- and site-local, |
556 | * IPv4 mapped and IPv4 compatible addresses as not redistributable. |
557 | */ |
558 | if (IN6_IS_ADDR_LOOPBACK(&kr->prefix)((*(const u_int32_t *)(const void *)(&(&kr->prefix )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->prefix)->__u6_addr.__u6_addr8 [4]) == 0) && (*(const u_int32_t *)(const void *)(& (&kr->prefix)->__u6_addr.__u6_addr8[8]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->prefix )->__u6_addr.__u6_addr8[12]) == (__uint32_t)(__builtin_constant_p (1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t)(1) & 0xff000000 ) >> 24) : __swap32md(1)))) || |
559 | IN6_IS_ADDR_MULTICAST(&kr->prefix)((&kr->prefix)->__u6_addr.__u6_addr8[0] == 0xff) || |
560 | IN6_IS_ADDR_LINKLOCAL(&kr->prefix)(((&kr->prefix)->__u6_addr.__u6_addr8[0] == 0xfe) && (((&kr->prefix)->__u6_addr.__u6_addr8[1] & 0xc0 ) == 0x80)) || |
561 | IN6_IS_ADDR_SITELOCAL(&kr->prefix)(((&kr->prefix)->__u6_addr.__u6_addr8[0] == 0xfe) && (((&kr->prefix)->__u6_addr.__u6_addr8[1] & 0xc0 ) == 0xc0)) || |
562 | IN6_IS_ADDR_V4MAPPED(&kr->prefix)((*(const u_int32_t *)(const void *)(&(&kr->prefix )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->prefix)->__u6_addr.__u6_addr8 [4]) == 0) && (*(const u_int32_t *)(const void *)(& (&kr->prefix)->__u6_addr.__u6_addr8[8]) == (__uint32_t )(__builtin_constant_p(0x0000ffff) ? (__uint32_t)(((__uint32_t )(0x0000ffff) & 0xff) << 24 | ((__uint32_t)(0x0000ffff ) & 0xff00) << 8 | ((__uint32_t)(0x0000ffff) & 0xff0000 ) >> 8 | ((__uint32_t)(0x0000ffff) & 0xff000000) >> 24) : __swap32md(0x0000ffff)))) || |
563 | IN6_IS_ADDR_V4COMPAT(&kr->prefix)((*(const u_int32_t *)(const void *)(&(&kr->prefix )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->prefix)->__u6_addr.__u6_addr8 [4]) == 0) && (*(const u_int32_t *)(const void *)(& (&kr->prefix)->__u6_addr.__u6_addr8[8]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->prefix )->__u6_addr.__u6_addr8[12]) != 0) && (*(const u_int32_t *)(const void *)(&(&kr->prefix)->__u6_addr.__u6_addr8 [12]) != (__uint32_t)(__builtin_constant_p(1) ? (__uint32_t)( ((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t)(1) & 0xff000000) >> 24) : __swap32md (1))))) |
564 | goto dont_redistribute; |
565 | /* |
566 | * Consider networks with nexthop loopback as not redistributable |
567 | * unless it is a reject or blackhole route. |
568 | */ |
569 | if (IN6_IS_ADDR_LOOPBACK(&kr->nexthop)((*(const u_int32_t *)(const void *)(&(&kr->nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->nexthop)->__u6_addr.__u6_addr8 [4]) == 0) && (*(const u_int32_t *)(const void *)(& (&kr->nexthop)->__u6_addr.__u6_addr8[8]) == 0) && (*(const u_int32_t *)(const void *)(&(&kr->nexthop )->__u6_addr.__u6_addr8[12]) == (__uint32_t)(__builtin_constant_p (1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t)(1) & 0xff000000 ) >> 24) : __swap32md(1)))) && |
570 | !(kr->flags & (F_BLACKHOLE0x0100|F_REJECT0x0080))) |
571 | goto dont_redistribute; |
572 | |
573 | /* Should we redistribute this route? */ |
574 | if (!ospf_redistribute(kr, &metric)) |
575 | goto dont_redistribute; |
576 | |
577 | /* prefix should be redistributed */ |
578 | kr->flags |= F_REDISTRIBUTED0x0200; |
579 | /* |
580 | * only one of all multipath routes can be redistributed so |
581 | * redistribute the best one. |
582 | */ |
583 | if (new_kr->metric > metric) { |
584 | *new_kr = *kr; |
585 | new_kr->metric = metric; |
586 | } |
587 | |
588 | return (1); |
589 | |
590 | dont_redistribute: |
591 | /* was the route redistributed? */ |
592 | if ((kr->flags & F_REDISTRIBUTED0x0200) == 0) |
593 | return (0); |
594 | |
595 | kr->flags &= ~F_REDISTRIBUTED0x0200; |
596 | return (1); |
597 | } |
598 | |
599 | void |
600 | kr_redistribute(struct kroute_node *kh) |
601 | { |
602 | struct kroute_node *kn; |
603 | struct kroute kr; |
604 | int redistribute = 0; |
605 | |
606 | /* only the highest prio route can be redistributed */ |
607 | if (kroute_find(&kh->r.prefix, kh->r.prefixlen, RTP_ANY64) != kh) |
608 | return; |
609 | |
610 | bzero(&kr, sizeof(kr)); |
611 | kr.metric = UINT_MAX0xffffffffU; |
612 | for (kn = kh; kn; kn = kn->next) |
613 | if (kr_redist_eval(&kn->r, &kr)) |
614 | redistribute = 1; |
615 | |
616 | if (!redistribute) |
617 | return; |
618 | |
619 | if (kr.flags & F_REDISTRIBUTED0x0200) { |
620 | main_imsg_compose_rde(IMSG_NETWORK_ADD, 0, &kr, |
621 | sizeof(struct kroute)); |
622 | } else { |
623 | kr = kh->r; |
624 | main_imsg_compose_rde(IMSG_NETWORK_DEL, 0, &kr, |
625 | sizeof(struct kroute)); |
626 | } |
627 | } |
628 | |
629 | void |
630 | kr_reload(int redis_label_or_prefix) |
631 | { |
632 | struct kroute_node *kr, *kn; |
633 | u_int32_t dummy; |
634 | int r; |
635 | int filter_prio = kr_state.fib_prio; |
636 | |
637 | /* update the priority filter */ |
638 | if (redis_label_or_prefix) { |
639 | filter_prio = 0; |
640 | log_info("%s: priority filter disabled", __func__); |
641 | } else |
642 | log_debug("%s: priority filter enabled", __func__); |
643 | |
644 | if (setsockopt(kr_state.fd, AF_ROUTE17, ROUTE_PRIOFILTER3, &filter_prio, |
645 | sizeof(filter_prio)) == -1) { |
646 | log_warn("%s: setsockopt AF_ROUTE ROUTE_PRIOFILTER", __func__); |
647 | /* not fatal */ |
648 | } |
649 | |
650 | RB_FOREACH(kr, kroute_tree, &krt)for ((kr) = kroute_tree_RB_MINMAX(&krt, -1); (kr) != ((void *)0); (kr) = kroute_tree_RB_NEXT(kr)) { |
651 | for (kn = kr; kn; kn = kn->next) { |
652 | r = ospf_redistribute(&kn->r, &dummy); |
653 | /* |
654 | * if it is redistributed, redistribute again metric |
655 | * may have changed. |
656 | */ |
657 | if ((kn->r.flags & F_REDISTRIBUTED0x0200 && !r) || r) |
658 | break; |
659 | } |
660 | if (kn) { |
661 | /* |
662 | * kr_redistribute copes with removes and RDE with |
663 | * duplicates |
664 | */ |
665 | kr_redistribute(kr); |
666 | } |
667 | } |
668 | } |
669 | |
670 | /* rb-tree compare */ |
671 | int |
672 | kroute_compare(struct kroute_node *a, struct kroute_node *b) |
673 | { |
674 | int i; |
675 | |
676 | /* XXX maybe switch a & b */ |
677 | i = memcmp(&a->r.prefix, &b->r.prefix, sizeof(a->r.prefix)); |
678 | if (i) |
679 | return (i); |
680 | if (a->r.prefixlen < b->r.prefixlen) |
681 | return (-1); |
682 | if (a->r.prefixlen > b->r.prefixlen) |
683 | return (1); |
684 | |
685 | /* if the priority is RTP_ANY finish on the first address hit */ |
686 | if (a->r.priority == RTP_ANY64 || b->r.priority == RTP_ANY64) |
687 | return (0); |
688 | if (a->r.priority < b->r.priority) |
689 | return (-1); |
690 | if (a->r.priority > b->r.priority) |
691 | return (1); |
692 | return (0); |
693 | } |
694 | |
695 | /* tree management */ |
696 | struct kroute_node * |
697 | kroute_find(const struct in6_addr *prefix, u_int8_t prefixlen, u_int8_t prio) |
698 | { |
699 | struct kroute_node s; |
700 | struct kroute_node *kn, *tmp; |
701 | |
702 | s.r.prefix = *prefix; |
703 | s.r.prefixlen = prefixlen; |
704 | s.r.priority = prio; |
705 | |
706 | kn = RB_FIND(kroute_tree, &krt, &s)kroute_tree_RB_FIND(&krt, &s); |
707 | if (kn && prio == RTP_ANY64) { |
708 | tmp = RB_PREV(kroute_tree, &krt, kn)kroute_tree_RB_PREV(kn); |
709 | while (tmp) { |
710 | if (kroute_compare(&s, tmp) == 0) |
711 | kn = tmp; |
712 | else |
713 | break; |
714 | tmp = RB_PREV(kroute_tree, &krt, kn)kroute_tree_RB_PREV(kn); |
715 | } |
716 | } |
717 | return (kn); |
718 | } |
719 | |
720 | struct kroute_node * |
721 | kroute_matchgw(struct kroute_node *kr, struct in6_addr *nh, unsigned int scope) |
722 | { |
723 | while (kr) { |
724 | if (scope == kr->r.scope && |
725 | IN6_ARE_ADDR_EQUAL(&kr->r.nexthop, nh)(memcmp(&(&kr->r.nexthop)->__u6_addr.__u6_addr8 [0], &(nh)->__u6_addr.__u6_addr8[0], sizeof(struct in6_addr )) == 0)) |
726 | return (kr); |
727 | kr = kr->next; |
728 | } |
729 | |
730 | return (NULL((void *)0)); |
731 | } |
732 | |
733 | int |
734 | kroute_insert(struct kroute_node *kr) |
735 | { |
736 | struct kroute_node *krm, *krh; |
737 | |
738 | kr->serial = kr_state.fib_serial; |
739 | |
740 | if ((krh = RB_INSERT(kroute_tree, &krt, kr)kroute_tree_RB_INSERT(&krt, kr)) != NULL((void *)0)) { |
741 | /* |
742 | * Multipath route, add at end of list. |
743 | */ |
744 | krm = krh; |
745 | while (krm->next != NULL((void *)0)) |
746 | krm = krm->next; |
747 | krm->next = kr; |
748 | kr->next = NULL((void *)0); /* to be sure */ |
749 | } else |
750 | krh = kr; |
751 | |
752 | if (!(kr->r.flags & F_KERNEL0x0002)) { |
753 | /* don't validate or redistribute ospf route */ |
754 | kr->r.flags &= ~F_DOWN0x0010; |
755 | return (0); |
756 | } |
757 | |
758 | if (kif_validate(kr->r.ifindex)) |
759 | kr->r.flags &= ~F_DOWN0x0010; |
760 | else |
761 | kr->r.flags |= F_DOWN0x0010; |
762 | |
763 | kr_redistribute(krh); |
764 | return (0); |
765 | } |
766 | |
767 | int |
768 | kroute_remove(struct kroute_node *kr) |
769 | { |
770 | struct kroute_node *krm; |
771 | |
772 | if ((krm = RB_FIND(kroute_tree, &krt, kr)kroute_tree_RB_FIND(&krt, kr)) == NULL((void *)0)) { |
773 | log_warnx("kroute_remove failed to find %s/%u", |
774 | log_in6addr(&kr->r.prefix), kr->r.prefixlen); |
775 | return (-1); |
776 | } |
777 | |
778 | if (krm == kr) { |
779 | /* head element */ |
780 | if (RB_REMOVE(kroute_tree, &krt, kr)kroute_tree_RB_REMOVE(&krt, kr) == NULL((void *)0)) { |
781 | log_warnx("kroute_remove failed for %s/%u", |
782 | log_in6addr(&kr->r.prefix), kr->r.prefixlen); |
783 | return (-1); |
784 | } |
785 | if (kr->next != NULL((void *)0)) { |
786 | if (RB_INSERT(kroute_tree, &krt, kr->next)kroute_tree_RB_INSERT(&krt, kr->next) != NULL((void *)0)) { |
787 | log_warnx("kroute_remove failed to add %s/%u", |
788 | log_in6addr(&kr->r.prefix), |
789 | kr->r.prefixlen); |
790 | return (-1); |
791 | } |
792 | } |
793 | } else { |
794 | /* somewhere in the list */ |
795 | while (krm->next != kr && krm->next != NULL((void *)0)) |
796 | krm = krm->next; |
797 | if (krm->next == NULL((void *)0)) { |
798 | log_warnx("kroute_remove multipath list corrupted " |
799 | "for %s/%u", log_in6addr(&kr->r.prefix), |
800 | kr->r.prefixlen); |
801 | return (-1); |
802 | } |
803 | krm->next = kr->next; |
804 | } |
805 | |
806 | kr_redist_remove(krm, kr); |
807 | rtlabel_unref(kr->r.rtlabel); |
808 | |
809 | free(kr); |
810 | return (0); |
811 | } |
812 | |
813 | void |
814 | kroute_clear(void) |
815 | { |
816 | struct kroute_node *kr; |
817 | |
818 | while ((kr = RB_MIN(kroute_tree, &krt)kroute_tree_RB_MINMAX(&krt, -1)) != NULL((void *)0)) |
819 | kroute_remove(kr); |
820 | } |
821 | |
822 | struct iface * |
823 | kif_update(u_short ifindex, int flags, struct if_data *ifd, |
824 | struct sockaddr_dl *sdl) |
825 | { |
826 | struct iface *iface; |
827 | char ifname[IF_NAMESIZE16]; |
828 | |
829 | if ((iface = if_find(ifindex)) == NULL((void *)0)) { |
830 | bzero(ifname, sizeof(ifname)); |
831 | if (sdl && sdl->sdl_family == AF_LINK18) { |
832 | if (sdl->sdl_nlen >= sizeof(ifname)) |
833 | memcpy(ifname, sdl->sdl_data, |
834 | sizeof(ifname) - 1); |
835 | else if (sdl->sdl_nlen > 0) |
836 | memcpy(ifname, sdl->sdl_data, sdl->sdl_nlen); |
837 | else |
838 | return (NULL((void *)0)); |
839 | } else |
840 | return (NULL((void *)0)); |
841 | if ((iface = if_new(ifindex, ifname)) == NULL((void *)0)) |
842 | return (NULL((void *)0)); |
843 | } |
844 | |
845 | if_update(iface, ifd->ifi_mtu, flags, ifd->ifi_type, |
846 | ifd->ifi_link_state, ifd->ifi_baudrate, ifd->ifi_rdomain); |
847 | |
848 | return (iface); |
849 | } |
850 | |
851 | int |
852 | kif_validate(u_short ifindex) |
853 | { |
854 | struct iface *iface; |
855 | |
856 | if ((iface = if_find(ifindex)) == NULL((void *)0)) { |
857 | log_warnx("interface with index %u not found", ifindex); |
858 | return (-1); |
859 | } |
860 | |
861 | return ((iface->flags & IFF_UP0x1) && LINK_STATE_IS_UP(iface->linkstate)((iface->linkstate) >= 4 || (iface->linkstate) == 0)); |
862 | } |
863 | |
864 | struct kroute_node * |
865 | kroute_match(struct in6_addr *key) |
866 | { |
867 | int i; |
868 | struct kroute_node *kr; |
869 | struct in6_addr ina; |
870 | |
871 | /* we will never match the default route */ |
872 | for (i = 128; i > 0; i--) { |
873 | inet6applymask(&ina, key, i); |
874 | if ((kr = kroute_find(&ina, i, RTP_ANY64)) != NULL((void *)0)) |
875 | return (kr); |
876 | } |
877 | |
878 | /* if we don't have a match yet, try to find a default route */ |
879 | if ((kr = kroute_find(&in6addr_any, 0, RTP_ANY64)) != NULL((void *)0)) |
880 | return (kr); |
881 | |
882 | return (NULL((void *)0)); |
883 | } |
884 | |
885 | /* misc */ |
886 | int |
887 | protect_lo(void) |
888 | { |
889 | struct kroute_node *kr; |
890 | |
891 | /* special protection for loopback */ |
892 | if ((kr = calloc(1, sizeof(struct kroute_node))) == NULL((void *)0)) { |
893 | log_warn("protect_lo"); |
894 | return (-1); |
895 | } |
896 | memcpy(&kr->r.prefix, &in6addr_loopback, sizeof(kr->r.prefix)); |
897 | kr->r.prefixlen = 128; |
898 | kr->r.flags = F_KERNEL0x0002|F_CONNECTED0x0008; |
899 | |
900 | if (RB_INSERT(kroute_tree, &krt, kr)kroute_tree_RB_INSERT(&krt, kr) != NULL((void *)0)) |
901 | free(kr); /* kernel route already there, no problem */ |
902 | |
903 | return (0); |
904 | } |
905 | |
906 | #define ROUNDUP(a)((a) > 0 ? (1 + (((a) - 1) | (sizeof(long) - 1))) : sizeof (long)) \ |
907 | ((a) > 0 ? (1 + (((a) - 1) | (sizeof(long) - 1))) : sizeof(long)) |
908 | |
909 | void |
910 | get_rtaddrs(int addrs, struct sockaddr *sa, struct sockaddr **rti_info) |
911 | { |
912 | int i; |
913 | |
914 | for (i = 0; i < RTAX_MAX15; i++) { |
915 | if (addrs & (1 << i)) { |
916 | rti_info[i] = sa; |
917 | sa = (struct sockaddr *)((char *)(sa) + |
918 | ROUNDUP(sa->sa_len)((sa->sa_len) > 0 ? (1 + (((sa->sa_len) - 1) | (sizeof (long) - 1))) : sizeof(long))); |
919 | } else |
920 | rti_info[i] = NULL((void *)0); |
921 | } |
922 | } |
923 | |
924 | void |
925 | if_change(u_short ifindex, int flags, struct if_data *ifd, |
926 | struct sockaddr_dl *sdl) |
927 | { |
928 | struct kroute_node *kr, *tkr; |
929 | struct iface *iface; |
930 | u_int8_t wasvalid, isvalid; |
931 | |
932 | wasvalid = kif_validate(ifindex); |
933 | |
934 | if ((iface = kif_update(ifindex, flags, ifd, sdl)) == NULL((void *)0)) { |
935 | log_warn("if_change: kif_update(%u)", ifindex); |
936 | return; |
937 | } |
938 | |
939 | /* inform engine and rde about state change */ |
940 | main_imsg_compose_rde(IMSG_IFINFO, 0, iface, sizeof(struct iface)); |
941 | main_imsg_compose_ospfe(IMSG_IFINFO, 0, iface, sizeof(struct iface)); |
942 | |
943 | isvalid = (iface->flags & IFF_UP0x1) && |
944 | LINK_STATE_IS_UP(iface->linkstate)((iface->linkstate) >= 4 || (iface->linkstate) == 0); |
945 | |
946 | if (wasvalid == isvalid) |
947 | return; /* nothing changed wrt validity */ |
948 | |
949 | /* update redistribute list */ |
950 | RB_FOREACH(kr, kroute_tree, &krt)for ((kr) = kroute_tree_RB_MINMAX(&krt, -1); (kr) != ((void *)0); (kr) = kroute_tree_RB_NEXT(kr)) { |
951 | for (tkr = kr; tkr != NULL((void *)0); tkr = tkr->next) { |
952 | if (tkr->r.ifindex == ifindex) { |
953 | if (isvalid) |
954 | tkr->r.flags &= ~F_DOWN0x0010; |
955 | else |
956 | tkr->r.flags |= F_DOWN0x0010; |
957 | |
958 | } |
959 | } |
960 | kr_redistribute(kr); |
961 | } |
962 | } |
963 | |
964 | void |
965 | if_newaddr(u_short ifindex, struct sockaddr_in6 *ifa, struct sockaddr_in6 *mask, |
966 | struct sockaddr_in6 *brd) |
967 | { |
968 | struct iface *iface; |
969 | struct iface_addr *ia; |
970 | struct ifaddrchange ifc; |
971 | |
972 | if (ifa == NULL((void *)0) || ifa->sin6_family != AF_INET624) |
973 | return; |
974 | if ((iface = if_find(ifindex)) == NULL((void *)0)) { |
975 | log_warnx("if_newaddr: corresponding if %d not found", ifindex); |
976 | return; |
977 | } |
978 | |
979 | /* We only care about link-local and global-scope. */ |
980 | if (IN6_IS_ADDR_UNSPECIFIED(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &ifa->sin6_addr)->__u6_addr.__u6_addr8[12]) == 0)) || |
981 | IN6_IS_ADDR_LOOPBACK(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &ifa->sin6_addr)->__u6_addr.__u6_addr8[12]) == (__uint32_t )(__builtin_constant_p(1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t )(1) & 0xff000000) >> 24) : __swap32md(1)))) || |
982 | IN6_IS_ADDR_MULTICAST(&ifa->sin6_addr)((&ifa->sin6_addr)->__u6_addr.__u6_addr8[0] == 0xff ) || |
983 | IN6_IS_ADDR_SITELOCAL(&ifa->sin6_addr)(((&ifa->sin6_addr)->__u6_addr.__u6_addr8[0] == 0xfe ) && (((&ifa->sin6_addr)->__u6_addr.__u6_addr8 [1] & 0xc0) == 0xc0)) || |
984 | IN6_IS_ADDR_V4MAPPED(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == (__uint32_t)(__builtin_constant_p(0x0000ffff) ? (__uint32_t )(((__uint32_t)(0x0000ffff) & 0xff) << 24 | ((__uint32_t )(0x0000ffff) & 0xff00) << 8 | ((__uint32_t)(0x0000ffff ) & 0xff0000) >> 8 | ((__uint32_t)(0x0000ffff) & 0xff000000) >> 24) : __swap32md(0x0000ffff)))) || |
985 | IN6_IS_ADDR_V4COMPAT(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &ifa->sin6_addr)->__u6_addr.__u6_addr8[12]) != 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[12]) != (__uint32_t)(__builtin_constant_p (1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t)(1) & 0xff000000 ) >> 24) : __swap32md(1))))) |
986 | return; |
987 | |
988 | clearscope(&ifa->sin6_addr); |
989 | |
990 | if (IN6_IS_ADDR_LINKLOCAL(&ifa->sin6_addr)(((&ifa->sin6_addr)->__u6_addr.__u6_addr8[0] == 0xfe ) && (((&ifa->sin6_addr)->__u6_addr.__u6_addr8 [1] & 0xc0) == 0x80)) || |
991 | iface->flags & IFF_LOOPBACK0x8) |
992 | iface->addr = ifa->sin6_addr; |
993 | |
994 | if ((ia = calloc(1, sizeof(struct iface_addr))) == NULL((void *)0)) |
995 | fatal("if_newaddr"); |
996 | |
997 | ia->addr = ifa->sin6_addr; |
998 | |
999 | if (mask) |
1000 | ia->prefixlen = mask2prefixlen(mask); |
1001 | else |
1002 | ia->prefixlen = 0; |
1003 | if (brd && brd->sin6_family == AF_INET624) |
1004 | ia->dstbrd = brd->sin6_addr; |
1005 | else |
1006 | bzero(&ia->dstbrd, sizeof(ia->dstbrd)); |
1007 | |
1008 | switch (iface->type) { |
1009 | case IF_TYPE_BROADCAST: |
1010 | case IF_TYPE_NBMA: |
1011 | log_debug("if_newaddr: ifindex %u, addr %s/%d", |
1012 | ifindex, log_in6addr(&ia->addr), ia->prefixlen); |
1013 | break; |
1014 | case IF_TYPE_VIRTUALLINK: /* FIXME */ |
1015 | break; |
1016 | case IF_TYPE_POINTOPOINT: |
1017 | case IF_TYPE_POINTOMULTIPOINT: |
1018 | log_debug("if_newaddr: ifindex %u, addr %s/%d, " |
1019 | "dest %s", ifindex, log_in6addr(&ia->addr), |
1020 | ia->prefixlen, log_in6addr(&ia->dstbrd)); |
1021 | break; |
1022 | default: |
1023 | fatalx("if_newaddr: unknown interface type"); |
1024 | } |
1025 | |
1026 | TAILQ_INSERT_TAIL(&iface->ifa_list, ia, entry)do { (ia)->entry.tqe_next = ((void *)0); (ia)->entry.tqe_prev = (&iface->ifa_list)->tqh_last; *(&iface->ifa_list )->tqh_last = (ia); (&iface->ifa_list)->tqh_last = &(ia)->entry.tqe_next; } while (0); |
1027 | /* inform engine and rde if interface is used */ |
1028 | if (iface->cflags & F_IFACE_CONFIGURED0x02) { |
1029 | ifc.addr = ia->addr; |
1030 | ifc.dstbrd = ia->dstbrd; |
1031 | ifc.prefixlen = ia->prefixlen; |
1032 | ifc.ifindex = ifindex; |
1033 | main_imsg_compose_ospfe(IMSG_IFADDRNEW, 0, &ifc, sizeof(ifc)); |
1034 | main_imsg_compose_rde(IMSG_IFADDRNEW, 0, &ifc, sizeof(ifc)); |
1035 | } |
1036 | } |
1037 | |
1038 | void |
1039 | if_deladdr(u_short ifindex, struct sockaddr_in6 *ifa, struct sockaddr_in6 *mask, |
1040 | struct sockaddr_in6 *brd) |
1041 | { |
1042 | struct iface *iface; |
1043 | struct iface_addr *ia, *nia; |
1044 | struct ifaddrchange ifc; |
1045 | |
1046 | if (ifa == NULL((void *)0) || ifa->sin6_family != AF_INET624) |
1047 | return; |
1048 | if ((iface = if_find(ifindex)) == NULL((void *)0)) { |
1049 | log_warnx("if_deladdr: corresponding if %d not found", ifindex); |
1050 | return; |
1051 | } |
1052 | |
1053 | /* We only care about link-local and global-scope. */ |
1054 | if (IN6_IS_ADDR_UNSPECIFIED(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &ifa->sin6_addr)->__u6_addr.__u6_addr8[12]) == 0)) || |
1055 | IN6_IS_ADDR_LOOPBACK(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &ifa->sin6_addr)->__u6_addr.__u6_addr8[12]) == (__uint32_t )(__builtin_constant_p(1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t )(1) & 0xff000000) >> 24) : __swap32md(1)))) || |
1056 | IN6_IS_ADDR_MULTICAST(&ifa->sin6_addr)((&ifa->sin6_addr)->__u6_addr.__u6_addr8[0] == 0xff ) || |
1057 | IN6_IS_ADDR_SITELOCAL(&ifa->sin6_addr)(((&ifa->sin6_addr)->__u6_addr.__u6_addr8[0] == 0xfe ) && (((&ifa->sin6_addr)->__u6_addr.__u6_addr8 [1] & 0xc0) == 0xc0)) || |
1058 | IN6_IS_ADDR_V4MAPPED(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == (__uint32_t)(__builtin_constant_p(0x0000ffff) ? (__uint32_t )(((__uint32_t)(0x0000ffff) & 0xff) << 24 | ((__uint32_t )(0x0000ffff) & 0xff00) << 8 | ((__uint32_t)(0x0000ffff ) & 0xff0000) >> 8 | ((__uint32_t)(0x0000ffff) & 0xff000000) >> 24) : __swap32md(0x0000ffff)))) || |
1059 | IN6_IS_ADDR_V4COMPAT(&ifa->sin6_addr)((*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr)->__u6_addr.__u6_addr8[8] ) == 0) && (*(const u_int32_t *)(const void *)(&( &ifa->sin6_addr)->__u6_addr.__u6_addr8[12]) != 0) && (*(const u_int32_t *)(const void *)(&(&ifa->sin6_addr )->__u6_addr.__u6_addr8[12]) != (__uint32_t)(__builtin_constant_p (1) ? (__uint32_t)(((__uint32_t)(1) & 0xff) << 24 | ((__uint32_t)(1) & 0xff00) << 8 | ((__uint32_t)(1) & 0xff0000) >> 8 | ((__uint32_t)(1) & 0xff000000 ) >> 24) : __swap32md(1))))) |
1060 | return; |
1061 | |
1062 | clearscope(&ifa->sin6_addr); |
1063 | |
1064 | for (ia = TAILQ_FIRST(&iface->ifa_list)((&iface->ifa_list)->tqh_first); ia != NULL((void *)0); ia = nia) { |
1065 | nia = TAILQ_NEXT(ia, entry)((ia)->entry.tqe_next); |
1066 | |
1067 | if (IN6_ARE_ADDR_EQUAL(&ia->addr, &ifa->sin6_addr)(memcmp(&(&ia->addr)->__u6_addr.__u6_addr8[0], & (&ifa->sin6_addr)->__u6_addr.__u6_addr8[0], sizeof( struct in6_addr)) == 0)) { |
1068 | log_debug("if_deladdr: ifindex %u, addr %s/%d", |
1069 | ifindex, log_in6addr(&ia->addr), ia->prefixlen); |
1070 | TAILQ_REMOVE(&iface->ifa_list, ia, entry)do { if (((ia)->entry.tqe_next) != ((void *)0)) (ia)->entry .tqe_next->entry.tqe_prev = (ia)->entry.tqe_prev; else ( &iface->ifa_list)->tqh_last = (ia)->entry.tqe_prev ; *(ia)->entry.tqe_prev = (ia)->entry.tqe_next; ; ; } while (0); |
1071 | /* inform engine and rde if interface is used */ |
1072 | if (iface->cflags & F_IFACE_CONFIGURED0x02) { |
1073 | ifc.addr = ia->addr; |
1074 | ifc.dstbrd = ia->dstbrd; |
1075 | ifc.prefixlen = ia->prefixlen; |
1076 | ifc.ifindex = ifindex; |
1077 | main_imsg_compose_ospfe(IMSG_IFADDRDEL, 0, &ifc, |
1078 | sizeof(ifc)); |
1079 | main_imsg_compose_rde(IMSG_IFADDRDEL, 0, &ifc, |
1080 | sizeof(ifc)); |
1081 | } |
1082 | free(ia); |
1083 | return; |
1084 | } |
1085 | } |
1086 | } |
1087 | |
1088 | void |
1089 | if_announce(void *msg) |
1090 | { |
1091 | struct if_announcemsghdr *ifan; |
1092 | struct iface *iface; |
1093 | |
1094 | ifan = msg; |
1095 | |
1096 | switch (ifan->ifan_what) { |
1097 | case IFAN_ARRIVAL0: |
1098 | if ((iface = if_new(ifan->ifan_index, ifan->ifan_name)) == NULL((void *)0)) |
Although the value stored to 'iface' is used in the enclosing expression, the value is never actually read from 'iface' | |
1099 | fatal("if_announce failed"); |
1100 | break; |
1101 | case IFAN_DEPARTURE1: |
1102 | iface = if_find(ifan->ifan_index); |
1103 | if_del(iface); |
1104 | break; |
1105 | } |
1106 | } |
1107 | |
1108 | /* rtsock */ |
1109 | int |
1110 | send_rtmsg(int fd, int action, struct kroute *kroute) |
1111 | { |
1112 | struct iovec iov[5]; |
1113 | struct rt_msghdr hdr; |
1114 | struct pad { |
1115 | struct sockaddr_in6 addr; |
1116 | char pad[sizeof(long)]; /* thank you IPv6 */ |
1117 | } prefix, nexthop, mask; |
1118 | struct { |
1119 | struct sockaddr_dl addr; |
1120 | char pad[sizeof(long)]; |
1121 | } ifp; |
1122 | struct sockaddr_rtlabel sa_rl; |
1123 | int iovcnt = 0; |
1124 | const char *label; |
1125 | |
1126 | if (kr_state.fib_sync == 0) |
1127 | return (0); |
1128 | |
1129 | /* initialize header */ |
1130 | bzero(&hdr, sizeof(hdr)); |
1131 | hdr.rtm_version = RTM_VERSION5; |
1132 | hdr.rtm_type = action; |
1133 | hdr.rtm_priority = kr_state.fib_prio; |
1134 | hdr.rtm_tableid = kr_state.rdomain; /* rtableid */ |
1135 | if (action == RTM_CHANGE0x3) |
1136 | hdr.rtm_fmask = RTF_REJECT0x8|RTF_BLACKHOLE0x1000; |
1137 | else |
1138 | hdr.rtm_flags = RTF_MPATH0x40000; |
1139 | hdr.rtm_seq = kr_state.rtseq++; /* overflow doesn't matter */ |
1140 | hdr.rtm_hdrlen = sizeof(hdr); |
1141 | hdr.rtm_msglen = sizeof(hdr); |
1142 | /* adjust iovec */ |
1143 | iov[iovcnt].iov_base = &hdr; |
1144 | iov[iovcnt++].iov_len = sizeof(hdr); |
1145 | |
1146 | bzero(&prefix, sizeof(prefix)); |
1147 | prefix.addr.sin6_len = sizeof(struct sockaddr_in6); |
1148 | prefix.addr.sin6_family = AF_INET624; |
1149 | prefix.addr.sin6_addr = kroute->prefix; |
1150 | /* adjust header */ |
1151 | hdr.rtm_addrs |= RTA_DST0x1; |
1152 | hdr.rtm_msglen += ROUNDUP(sizeof(struct sockaddr_in6))((sizeof(struct sockaddr_in6)) > 0 ? (1 + (((sizeof(struct sockaddr_in6)) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1153 | /* adjust iovec */ |
1154 | iov[iovcnt].iov_base = &prefix; |
1155 | iov[iovcnt++].iov_len = ROUNDUP(sizeof(struct sockaddr_in6))((sizeof(struct sockaddr_in6)) > 0 ? (1 + (((sizeof(struct sockaddr_in6)) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1156 | |
1157 | if (!IN6_IS_ADDR_UNSPECIFIED(&kroute->nexthop)((*(const u_int32_t *)(const void *)(&(&kroute->nexthop )->__u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute->nexthop)->__u6_addr .__u6_addr8[4]) == 0) && (*(const u_int32_t *)(const void *)(&(&kroute->nexthop)->__u6_addr.__u6_addr8[8 ]) == 0) && (*(const u_int32_t *)(const void *)(& (&kroute->nexthop)->__u6_addr.__u6_addr8[12]) == 0) )) { |
1158 | bzero(&nexthop, sizeof(nexthop)); |
1159 | nexthop.addr.sin6_len = sizeof(struct sockaddr_in6); |
1160 | nexthop.addr.sin6_family = AF_INET624; |
1161 | nexthop.addr.sin6_addr = kroute->nexthop; |
1162 | nexthop.addr.sin6_scope_id = kroute->scope; |
1163 | /* |
1164 | * XXX we should set the sin6_scope_id but the kernel |
1165 | * XXX does not expect it that way. It must be fiddled |
1166 | * XXX into the sin6_addr. Welcome to the typical |
1167 | * XXX IPv6 insanity and all without wine bottles. |
1168 | */ |
1169 | embedscope(&nexthop.addr); |
1170 | |
1171 | /* adjust header */ |
1172 | hdr.rtm_flags |= RTF_GATEWAY0x2; |
1173 | hdr.rtm_addrs |= RTA_GATEWAY0x2; |
1174 | hdr.rtm_msglen += ROUNDUP(sizeof(struct sockaddr_in6))((sizeof(struct sockaddr_in6)) > 0 ? (1 + (((sizeof(struct sockaddr_in6)) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1175 | /* adjust iovec */ |
1176 | iov[iovcnt].iov_base = &nexthop; |
1177 | iov[iovcnt++].iov_len = ROUNDUP(sizeof(struct sockaddr_in6))((sizeof(struct sockaddr_in6)) > 0 ? (1 + (((sizeof(struct sockaddr_in6)) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1178 | } else if (kroute->ifindex) { |
1179 | /* |
1180 | * We don't have an interface address in that network, |
1181 | * so we install a cloning route. The kernel will then |
1182 | * do neighbor discovery. |
1183 | */ |
1184 | bzero(&ifp, sizeof(ifp)); |
1185 | ifp.addr.sdl_len = sizeof(struct sockaddr_dl); |
1186 | ifp.addr.sdl_family = AF_LINK18; |
1187 | |
1188 | ifp.addr.sdl_index = kroute->ifindex; |
1189 | /* adjust header */ |
1190 | hdr.rtm_flags |= RTF_CLONING0x100; |
1191 | hdr.rtm_addrs |= RTA_GATEWAY0x2; |
1192 | hdr.rtm_msglen += ROUNDUP(sizeof(struct sockaddr_dl))((sizeof(struct sockaddr_dl)) > 0 ? (1 + (((sizeof(struct sockaddr_dl )) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1193 | /* adjust iovec */ |
1194 | iov[iovcnt].iov_base = &ifp; |
1195 | iov[iovcnt++].iov_len = ROUNDUP(sizeof(struct sockaddr_dl))((sizeof(struct sockaddr_dl)) > 0 ? (1 + (((sizeof(struct sockaddr_dl )) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1196 | } |
1197 | |
1198 | bzero(&mask, sizeof(mask)); |
1199 | mask.addr.sin6_len = sizeof(struct sockaddr_in6); |
1200 | mask.addr.sin6_family = AF_INET624; |
1201 | mask.addr.sin6_addr = *prefixlen2mask(kroute->prefixlen); |
1202 | /* adjust header */ |
1203 | if (kroute->prefixlen == 128) |
1204 | hdr.rtm_flags |= RTF_HOST0x4; |
1205 | hdr.rtm_addrs |= RTA_NETMASK0x4; |
1206 | hdr.rtm_msglen += ROUNDUP(sizeof(struct sockaddr_in6))((sizeof(struct sockaddr_in6)) > 0 ? (1 + (((sizeof(struct sockaddr_in6)) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1207 | /* adjust iovec */ |
1208 | iov[iovcnt].iov_base = &mask; |
1209 | iov[iovcnt++].iov_len = ROUNDUP(sizeof(struct sockaddr_in6))((sizeof(struct sockaddr_in6)) > 0 ? (1 + (((sizeof(struct sockaddr_in6)) - 1) | (sizeof(long) - 1))) : sizeof(long)); |
1210 | |
1211 | if (kroute->rtlabel != 0) { |
1212 | sa_rl.sr_len = sizeof(sa_rl); |
1213 | sa_rl.sr_family = AF_UNSPEC0; |
1214 | label = rtlabel_id2name(kroute->rtlabel); |
1215 | if (strlcpy(sa_rl.sr_label, label, |
1216 | sizeof(sa_rl.sr_label)) >= sizeof(sa_rl.sr_label)) { |
1217 | log_warnx("send_rtmsg: invalid rtlabel"); |
1218 | return (-1); |
1219 | } |
1220 | /* adjust header */ |
1221 | hdr.rtm_addrs |= RTA_LABEL0x400; |
1222 | hdr.rtm_msglen += sizeof(sa_rl); |
1223 | /* adjust iovec */ |
1224 | iov[iovcnt].iov_base = &sa_rl; |
1225 | iov[iovcnt++].iov_len = sizeof(sa_rl); |
1226 | } |
1227 | |
1228 | retry: |
1229 | if (writev(fd, iov, iovcnt) == -1) { |
1230 | if (errno(*__errno()) == ESRCH3) { |
1231 | if (hdr.rtm_type == RTM_CHANGE0x3) { |
1232 | hdr.rtm_type = RTM_ADD0x1; |
1233 | goto retry; |
1234 | } else if (hdr.rtm_type == RTM_DELETE0x2) { |
1235 | log_info("route %s/%u vanished before delete", |
1236 | log_sockaddr(&prefix), kroute->prefixlen); |
1237 | return (0); |
1238 | } |
1239 | } |
1240 | log_warn("send_rtmsg: action %u, prefix %s/%u", hdr.rtm_type, |
1241 | log_sockaddr(&prefix), kroute->prefixlen); |
1242 | return (0); |
1243 | } |
1244 | |
1245 | return (0); |
1246 | } |
1247 | |
1248 | int |
1249 | fetchtable(void) |
1250 | { |
1251 | size_t len; |
1252 | int mib[7]; |
1253 | char *buf; |
1254 | int rv; |
1255 | |
1256 | mib[0] = CTL_NET4; |
1257 | mib[1] = PF_ROUTE17; |
1258 | mib[2] = 0; |
1259 | mib[3] = AF_INET624; |
1260 | mib[4] = NET_RT_DUMP1; |
1261 | mib[5] = 0; |
1262 | mib[6] = kr_state.rdomain; /* rtableid */ |
1263 | |
1264 | if (sysctl(mib, 7, NULL((void *)0), &len, NULL((void *)0), 0) == -1) { |
1265 | log_warn("sysctl"); |
1266 | return (-1); |
1267 | } |
1268 | if ((buf = malloc(len)) == NULL((void *)0)) { |
1269 | log_warn("fetchtable"); |
1270 | return (-1); |
1271 | } |
1272 | if (sysctl(mib, 7, buf, &len, NULL((void *)0), 0) == -1) { |
1273 | log_warn("sysctl"); |
1274 | free(buf); |
1275 | return (-1); |
1276 | } |
1277 | |
1278 | rv = rtmsg_process(buf, len); |
1279 | free(buf); |
1280 | |
1281 | return (rv); |
1282 | } |
1283 | |
1284 | int |
1285 | fetchifs(u_short ifindex) |
1286 | { |
1287 | size_t len; |
1288 | int mib[6]; |
1289 | char *buf; |
1290 | int rv; |
1291 | |
1292 | mib[0] = CTL_NET4; |
1293 | mib[1] = PF_ROUTE17; |
1294 | mib[2] = 0; |
1295 | mib[3] = AF_INET624; |
1296 | mib[4] = NET_RT_IFLIST3; |
1297 | mib[5] = ifindex; |
1298 | |
1299 | if (sysctl(mib, 6, NULL((void *)0), &len, NULL((void *)0), 0) == -1) { |
1300 | log_warn("sysctl"); |
1301 | return (-1); |
1302 | } |
1303 | if ((buf = malloc(len)) == NULL((void *)0)) { |
1304 | log_warn("fetchifs"); |
1305 | return (-1); |
1306 | } |
1307 | if (sysctl(mib, 6, buf, &len, NULL((void *)0), 0) == -1) { |
1308 | log_warn("sysctl"); |
1309 | free(buf); |
1310 | return (-1); |
1311 | } |
1312 | |
1313 | rv = rtmsg_process(buf, len); |
1314 | free(buf); |
1315 | |
1316 | return (rv); |
1317 | } |
1318 | |
1319 | int |
1320 | dispatch_rtmsg(void) |
1321 | { |
1322 | char buf[RT_BUF_SIZE16384]; |
1323 | ssize_t n; |
1324 | |
1325 | if ((n = read(kr_state.fd, &buf, sizeof(buf))) == -1) { |
1326 | if (errno(*__errno()) == EAGAIN35 || errno(*__errno()) == EINTR4) |
1327 | return (0); |
1328 | log_warn("dispatch_rtmsg: read error"); |
1329 | return (-1); |
1330 | } |
1331 | |
1332 | if (n == 0) { |
1333 | log_warnx("routing socket closed"); |
1334 | return (-1); |
1335 | } |
1336 | |
1337 | return (rtmsg_process(buf, n)); |
1338 | } |
1339 | |
1340 | int |
1341 | rtmsg_process(char *buf, size_t len) |
1342 | { |
1343 | struct rt_msghdr *rtm; |
1344 | struct if_msghdr ifm; |
1345 | struct ifa_msghdr *ifam; |
1346 | struct sockaddr *sa, *rti_info[RTAX_MAX15]; |
1347 | struct sockaddr_in6 *sa_in6; |
1348 | struct sockaddr_rtlabel *label; |
1349 | struct kroute_node *kr, *okr; |
1350 | struct in6_addr prefix, nexthop; |
1351 | u_int8_t prefixlen, prio; |
1352 | int flags, mpath; |
1353 | unsigned int scope; |
1354 | u_short ifindex = 0; |
1355 | int rv, delay; |
1356 | size_t offset; |
1357 | char *next; |
1358 | |
1359 | for (offset = 0; offset < len; offset += rtm->rtm_msglen) { |
1360 | next = buf + offset; |
1361 | rtm = (struct rt_msghdr *)next; |
1362 | if (len < offset + sizeof(u_short) || |
1363 | len < offset + rtm->rtm_msglen) |
1364 | fatalx("rtmsg_process: partial rtm in buffer"); |
1365 | if (rtm->rtm_version != RTM_VERSION5) |
1366 | continue; |
1367 | |
1368 | bzero(&prefix, sizeof(prefix)); |
1369 | bzero(&nexthop, sizeof(nexthop)); |
1370 | scope = 0; |
1371 | prefixlen = 0; |
1372 | flags = F_KERNEL0x0002; |
1373 | mpath = 0; |
1374 | prio = 0; |
1375 | |
1376 | sa = (struct sockaddr *)(next + rtm->rtm_hdrlen); |
1377 | get_rtaddrs(rtm->rtm_addrs, sa, rti_info); |
1378 | |
1379 | switch (rtm->rtm_type) { |
1380 | case RTM_ADD0x1: |
1381 | case RTM_GET0x4: |
1382 | case RTM_CHANGE0x3: |
1383 | case RTM_DELETE0x2: |
1384 | if (rtm->rtm_errno) /* failed attempts... */ |
1385 | continue; |
1386 | |
1387 | if (rtm->rtm_tableid != kr_state.rdomain) |
1388 | continue; |
1389 | |
1390 | if (rtm->rtm_type == RTM_GET0x4 && |
1391 | rtm->rtm_pid != kr_state.pid) /* caused by us */ |
1392 | continue; |
1393 | |
1394 | if ((sa = rti_info[RTAX_DST0]) == NULL((void *)0)) |
1395 | continue; |
1396 | |
1397 | /* Skip ARP/ND cache and broadcast routes. */ |
1398 | if (rtm->rtm_flags & (RTF_LLINFO0x400|RTF_BROADCAST0x400000)) |
1399 | continue; |
1400 | |
1401 | if (rtm->rtm_flags & RTF_MPATH0x40000) |
1402 | mpath = 1; |
1403 | prio = rtm->rtm_priority; |
1404 | flags = (prio == kr_state.fib_prio) ? |
1405 | F_OSPFD_INSERTED0x0001 : F_KERNEL0x0002; |
1406 | |
1407 | switch (sa->sa_family) { |
1408 | case AF_INET624: |
1409 | prefix = |
1410 | ((struct sockaddr_in6 *)sa)->sin6_addr; |
1411 | sa_in6 = (struct sockaddr_in6 *) |
1412 | rti_info[RTAX_NETMASK2]; |
1413 | if (sa_in6 != NULL((void *)0)) { |
1414 | if (sa_in6->sin6_len != 0) |
1415 | prefixlen = mask2prefixlen( |
1416 | sa_in6); |
1417 | } else if (rtm->rtm_flags & RTF_HOST0x4) |
1418 | prefixlen = 128; |
1419 | else |
1420 | fatalx("classful IPv6 address?!!"); |
1421 | if (rtm->rtm_flags & RTF_STATIC0x800) |
1422 | flags |= F_STATIC0x0020; |
1423 | if (rtm->rtm_flags & RTF_BLACKHOLE0x1000) |
1424 | flags |= F_BLACKHOLE0x0100; |
1425 | if (rtm->rtm_flags & RTF_REJECT0x8) |
1426 | flags |= F_REJECT0x0080; |
1427 | if (rtm->rtm_flags & RTF_DYNAMIC0x10) |
1428 | flags |= F_DYNAMIC0x0040; |
1429 | break; |
1430 | default: |
1431 | continue; |
1432 | } |
1433 | |
1434 | ifindex = rtm->rtm_index; |
1435 | if ((sa = rti_info[RTAX_GATEWAY1]) != NULL((void *)0)) { |
1436 | switch (sa->sa_family) { |
1437 | case AF_INET624: |
1438 | if (rtm->rtm_flags & RTF_CONNECTED0x800000) |
1439 | flags |= F_CONNECTED0x0008; |
1440 | |
1441 | sa_in6 = (struct sockaddr_in6 *)sa; |
1442 | /* |
1443 | * XXX The kernel provides the scope |
1444 | * XXX via the kame hack instead of |
1445 | * XXX the scope_id field. |
1446 | */ |
1447 | recoverscope(sa_in6); |
1448 | nexthop = sa_in6->sin6_addr; |
1449 | scope = sa_in6->sin6_scope_id; |
1450 | break; |
1451 | case AF_LINK18: |
1452 | flags |= F_CONNECTED0x0008; |
1453 | break; |
1454 | } |
1455 | } |
1456 | } |
1457 | |
1458 | switch (rtm->rtm_type) { |
1459 | case RTM_ADD0x1: |
1460 | case RTM_GET0x4: |
1461 | case RTM_CHANGE0x3: |
1462 | if (IN6_IS_ADDR_UNSPECIFIED(&nexthop)((*(const u_int32_t *)(const void *)(&(&nexthop)-> __u6_addr.__u6_addr8[0]) == 0) && (*(const u_int32_t * )(const void *)(&(&nexthop)->__u6_addr.__u6_addr8[ 4]) == 0) && (*(const u_int32_t *)(const void *)(& (&nexthop)->__u6_addr.__u6_addr8[8]) == 0) && ( *(const u_int32_t *)(const void *)(&(&nexthop)->__u6_addr .__u6_addr8[12]) == 0)) && |
1463 | !(flags & F_CONNECTED0x0008)) { |
1464 | log_warnx("rtmsg_process no nexthop for %s/%u", |
1465 | log_in6addr(&prefix), prefixlen); |
1466 | continue; |
1467 | } |
1468 | |
1469 | if ((okr = kroute_find(&prefix, prefixlen, prio)) |
1470 | != NULL((void *)0)) { |
1471 | kr = okr; |
1472 | if ((mpath || prio == kr_state.fib_prio) && |
1473 | (kr = kroute_matchgw(okr, &nexthop, scope)) == |
1474 | NULL((void *)0)) { |
1475 | log_warnx("rtmsg_process: mpath route" |
1476 | " not found"); |
1477 | /* add routes we missed out earlier */ |
1478 | goto add; |
1479 | } |
1480 | |
1481 | if (kr->r.flags & F_REDISTRIBUTED0x0200) |
1482 | flags |= F_REDISTRIBUTED0x0200; |
1483 | kr->r.nexthop = nexthop; |
1484 | kr->r.scope = scope; |
1485 | kr->r.flags = flags; |
1486 | kr->r.ifindex = ifindex; |
1487 | |
1488 | rtlabel_unref(kr->r.rtlabel); |
1489 | kr->r.rtlabel = 0; |
1490 | kr->r.ext_tag = 0; |
1491 | if ((label = (struct sockaddr_rtlabel *) |
1492 | rti_info[RTAX_LABEL10]) != NULL((void *)0)) { |
1493 | kr->r.rtlabel = |
1494 | rtlabel_name2id(label->sr_label); |
1495 | kr->r.ext_tag = |
1496 | rtlabel_id2tag(kr->r.rtlabel); |
1497 | } |
1498 | |
1499 | if (kif_validate(kr->r.ifindex)) |
1500 | kr->r.flags &= ~F_DOWN0x0010; |
1501 | else |
1502 | kr->r.flags |= F_DOWN0x0010; |
1503 | |
1504 | /* just readd, the RDE will care */ |
1505 | kr->serial = kr_state.fib_serial; |
1506 | kr_redistribute(kr); |
1507 | } else { |
1508 | add: |
1509 | if ((kr = calloc(1, |
1510 | sizeof(struct kroute_node))) == NULL((void *)0)) { |
1511 | log_warn("rtmsg_process calloc"); |
1512 | return (-1); |
1513 | } |
1514 | kr->r.prefix = prefix; |
1515 | kr->r.prefixlen = prefixlen; |
1516 | kr->r.nexthop = nexthop; |
1517 | kr->r.scope = scope; |
1518 | kr->r.flags = flags; |
1519 | kr->r.ifindex = ifindex; |
1520 | kr->r.priority = prio; |
1521 | |
1522 | if (rtm->rtm_priority == kr_state.fib_prio) { |
1523 | log_warnx("alien OSPF route %s/%d", |
1524 | log_in6addr(&prefix), prefixlen); |
1525 | rv = send_rtmsg(kr_state.fd, |
1526 | RTM_DELETE0x2, &kr->r); |
1527 | free(kr); |
1528 | if (rv == -1) |
1529 | return (-1); |
1530 | } else { |
1531 | if ((label = (struct sockaddr_rtlabel *) |
1532 | rti_info[RTAX_LABEL10]) != NULL((void *)0)) { |
1533 | kr->r.rtlabel = |
1534 | rtlabel_name2id( |
1535 | label->sr_label); |
1536 | kr->r.ext_tag = |
1537 | rtlabel_id2tag( |
1538 | kr->r.rtlabel); |
1539 | } |
1540 | |
1541 | kroute_insert(kr); |
1542 | } |
1543 | } |
1544 | break; |
1545 | case RTM_DELETE0x2: |
1546 | if ((kr = kroute_find(&prefix, prefixlen, prio)) == |
1547 | NULL((void *)0)) |
1548 | continue; |
1549 | if (!(kr->r.flags & F_KERNEL0x0002)) |
1550 | continue; |
1551 | /* get the correct route */ |
1552 | okr = kr; |
1553 | if (mpath && (kr = kroute_matchgw(kr, &nexthop, |
1554 | scope)) == NULL((void *)0)) { |
1555 | log_warnx("rtmsg_process mpath route" |
1556 | " not found"); |
1557 | return (-1); |
1558 | } |
1559 | if (kroute_remove(kr) == -1) |
1560 | return (-1); |
1561 | break; |
1562 | case RTM_IFINFO0xe: |
1563 | memcpy(&ifm, next, sizeof(ifm)); |
1564 | if_change(ifm.ifm_index, ifm.ifm_flags, &ifm.ifm_data, |
1565 | (struct sockaddr_dl *)rti_info[RTAX_IFP4]); |
1566 | break; |
1567 | case RTM_NEWADDR0xc: |
1568 | ifam = (struct ifa_msghdr *)rtm; |
1569 | if ((ifam->ifam_addrs & (RTA_NETMASK0x4 | RTA_IFA0x20 | |
1570 | RTA_BRD0x80)) == 0) |
1571 | break; |
1572 | |
1573 | if_newaddr(ifam->ifam_index, |
1574 | (struct sockaddr_in6 *)rti_info[RTAX_IFA5], |
1575 | (struct sockaddr_in6 *)rti_info[RTAX_NETMASK2], |
1576 | (struct sockaddr_in6 *)rti_info[RTAX_BRD7]); |
1577 | break; |
1578 | case RTM_DELADDR0xd: |
1579 | ifam = (struct ifa_msghdr *)rtm; |
1580 | if ((ifam->ifam_addrs & (RTA_NETMASK0x4 | RTA_IFA0x20 | |
1581 | RTA_BRD0x80)) == 0) |
1582 | break; |
1583 | |
1584 | if_deladdr(ifam->ifam_index, |
1585 | (struct sockaddr_in6 *)rti_info[RTAX_IFA5], |
1586 | (struct sockaddr_in6 *)rti_info[RTAX_NETMASK2], |
1587 | (struct sockaddr_in6 *)rti_info[RTAX_BRD7]); |
1588 | break; |
1589 | case RTM_IFANNOUNCE0xf: |
1590 | if_announce(next); |
1591 | break; |
1592 | case RTM_DESYNC0x10: |
1593 | /* |
1594 | * We lost some routing packets. Schedule a reload |
1595 | * of the kernel route/interface information. |
1596 | */ |
1597 | if (kr_state.reload_state == KR_RELOAD_IDLE0) { |
1598 | delay = KR_RELOAD_TIMER250; |
1599 | log_info("desync; scheduling fib reload"); |
1600 | } else { |
1601 | delay = KR_RELOAD_HOLD_TIMER5000; |
1602 | log_debug("desync during KR_RELOAD_%s", |
1603 | kr_state.reload_state == |
1604 | KR_RELOAD_FETCH1 ? "FETCH" : "HOLD"); |
1605 | } |
1606 | kr_state.reload_state = KR_RELOAD_FETCH1; |
1607 | kr_fib_reload_arm_timer(delay); |
1608 | break; |
1609 | default: |
1610 | /* ignore for now */ |
1611 | break; |
1612 | } |
1613 | } |
1614 | return (offset); |
1615 | } |