@@ -46,36 +46,107 @@ static inline struct fib_xid_lpm_local *fxid_llpm(struct fib_xid *fxid)
46
46
*/
47
47
const struct xia_ppal_rt_iops * lpm_rt_iops = & xia_ppal_tree_rt_iops ;
48
48
49
+ /* Only call this function after an RCU synchronization,
50
+ * such as by calling free_fxid.
51
+ */
52
+ static void local_free_lpm (struct fib_xid_table * xtbl , struct fib_xid * fxid )
53
+ {
54
+ struct fib_xid_lpm_local * llpm = fxid_llpm (fxid );
55
+
56
+ xdst_free_anchor (& llpm -> anchor );
57
+ kfree (llpm );
58
+ }
59
+
49
60
/* Assuming the FIB is locked, find the appropriate anchor,
50
- * flush it, and unlock the FIB.
61
+ * flush it, and unlock the FIB. To do this, we create a copy of
62
+ * the predecessor so that some readers can continue using the tree while
63
+ * we wait for other readers to finish to flush the anchor.
51
64
*/
52
- static void newroute_flush_anchor_locked (struct fib_xid_table * xtbl ,
53
- struct fib_xid * new_fxid ,
54
- struct xip_deferred_negdep_flush * dnf )
65
+ static int newroute_flush_anchor_locked (struct fib_xid_table * xtbl ,
66
+ struct fib_xid * new_fxid ,
67
+ struct xip_deferred_negdep_flush * dnf )
55
68
{
56
- struct fib_xid * pred_fxid = tree_fib_get_pred_locked (new_fxid );
69
+ /* At most one of @dup_llpm and @dup_mrd should be used. */
70
+ struct fib_xid_lpm_local * dup_llpm = NULL ;
71
+ struct fib_xid_redirect_main * dup_mrd = NULL ;
57
72
73
+ /* Find the predecessor. If it doesn't exist, we're done. */
74
+ struct fib_xid * pred_fxid = tree_fib_get_pred_locked (new_fxid );
58
75
if (!pred_fxid ) {
59
76
lpm_rt_iops -> fib_unlock (xtbl , NULL );
60
77
fib_defer_dnf (dnf , xtbl_net (xtbl ), xtbl_ppalty (xtbl ));
61
- return ;
78
+ return 0 ;
62
79
}
63
80
64
- fib_free_dnf (dnf );
65
- synchronize_rcu ();
81
+ /* Flush the predecessor's anchor by first making a copy,
82
+ * replacing the old entry, waiting an RCU synchronization,
83
+ * and then freeing the old entry when done.
84
+ */
66
85
switch (pred_fxid -> fx_table_id ) {
67
86
case XRTABLE_LOCAL_INDEX :
68
- xdst_free_anchor (& fxid_llpm (pred_fxid )-> anchor );
87
+ /* Allocate a duplicate of the predecessor entry. */
88
+ dup_llpm = lpm_rt_iops -> fxid_ppal_alloc (sizeof (* dup_llpm ),
89
+ GFP_ATOMIC );
90
+ if (!dup_llpm ) {
91
+ /* Can't add this entry now due to lack of memory. */
92
+ lpm_rt_iops -> fxid_rm_locked (NULL , xtbl , new_fxid );
93
+ lpm_rt_iops -> fib_unlock (xtbl , NULL );
94
+ return - ENOMEM ;
95
+ }
96
+
97
+ /* Replace the old predecessor with the new predecessor by
98
+ * copying the generic struct fib_xid and replacing the old
99
+ * node with the new one in the tree.
100
+ */
101
+ xdst_init_anchor (& dup_llpm -> anchor );
102
+ dup_llpm -> common = * pred_fxid ;
103
+ lpm_rt_iops -> fxid_replace_locked (xtbl , pred_fxid ,
104
+ & dup_llpm -> common );
105
+
106
+ /* Release write lock to let tree readers that get a write
107
+ * lock (such as in lpm_deliver()) continue, avoiding deadlock.
108
+ */
109
+ lpm_rt_iops -> fib_unlock (xtbl , NULL );
110
+
111
+ /* Wait for existing RCU readers in routing mechanism to
112
+ * finish, and then flush the anchor.
113
+ *
114
+ * The old predecessor is no longer accessible by the tree and
115
+ * existing readers on its anchor have finished, so we can
116
+ * release the old predecessor. Since we just called
117
+ * synchronize_rcu(), we can directly call local_free_lpm().
118
+ */
119
+ synchronize_rcu ();
120
+ local_free_lpm (xtbl , pred_fxid );
121
+ BUG_ON (dup_mrd );
69
122
break ;
70
123
case XRTABLE_MAIN_INDEX :
71
- xdst_invalidate_redirect (xtbl_net (xtbl ), xtbl_ppalty (xtbl ),
72
- pred_fxid -> fx_xid ,
73
- & fxid_mrd (pred_fxid )-> gw );
124
+ /* Same algorithm as above for main predecessor entries. */
125
+ dup_mrd = lpm_rt_iops -> fxid_ppal_alloc (sizeof (* dup_mrd ),
126
+ GFP_ATOMIC );
127
+ if (!dup_mrd ) {
128
+ lpm_rt_iops -> fxid_rm_locked (NULL , xtbl , new_fxid );
129
+ lpm_rt_iops -> fib_unlock (xtbl , NULL );
130
+ return - ENOMEM ;
131
+ }
132
+
133
+ dup_mrd -> gw = fxid_mrd (pred_fxid )-> gw ;
134
+ dup_mrd -> common = * pred_fxid ;
135
+ lpm_rt_iops -> fxid_replace_locked (xtbl , pred_fxid ,
136
+ & dup_mrd -> common );
137
+
138
+ lpm_rt_iops -> fib_unlock (xtbl , NULL );
139
+
140
+ synchronize_rcu ();
141
+ fib_mrd_free (xtbl , pred_fxid );
142
+ BUG_ON (dup_llpm );
74
143
break ;
75
144
default :
76
145
BUG ();
77
146
}
78
- lpm_rt_iops -> fib_unlock (xtbl , NULL );
147
+
148
+ fib_free_dnf (dnf );
149
+ return 0 ;
79
150
}
80
151
81
152
static int local_newroute (struct xip_ppal_ctx * ctx ,
@@ -111,15 +182,22 @@ static int local_newroute(struct xip_ppal_ctx *ctx,
111
182
* atomically to flush the appropriate anchor.
112
183
*/
113
184
rc = tree_fib_newroute_lock (& new_llpm -> common , xtbl , cfg , NULL );
114
- if (rc ) {
115
- fib_free_dnf (dnf );
116
- fxid_free_norcu (xtbl , & new_llpm -> common );
117
- return rc ;
118
- }
185
+ if (rc )
186
+ goto unlock_and_free ;
119
187
120
188
/* Flush appropriate anchor and release lock. */
121
- newroute_flush_anchor_locked (xtbl , & new_llpm -> common , dnf );
189
+ rc = newroute_flush_anchor_locked (xtbl , & new_llpm -> common , dnf );
190
+ if (rc )
191
+ goto free ;
192
+
122
193
return 0 ;
194
+
195
+ unlock_and_free :
196
+ lpm_rt_iops -> fib_unlock (NULL , xtbl );
197
+ free :
198
+ fib_free_dnf (dnf );
199
+ fxid_free_norcu (xtbl , & new_llpm -> common );
200
+ return rc ;
123
201
}
124
202
125
203
static int local_dump_lpm (struct fib_xid * fxid , struct fib_xid_table * xtbl ,
@@ -169,15 +247,6 @@ static int local_dump_lpm(struct fib_xid *fxid, struct fib_xid_table *xtbl,
169
247
return - EMSGSIZE ;
170
248
}
171
249
172
- /* Don't call this function! Use free_fxid instead. */
173
- static void local_free_lpm (struct fib_xid_table * xtbl , struct fib_xid * fxid )
174
- {
175
- struct fib_xid_lpm_local * llpm = fxid_llpm (fxid );
176
-
177
- xdst_free_anchor (& llpm -> anchor );
178
- kfree (llpm );
179
- }
180
-
181
250
static int main_newroute (struct xip_ppal_ctx * ctx , struct fib_xid_table * xtbl ,
182
251
struct xia_fib_config * cfg )
183
252
{
@@ -208,15 +277,21 @@ static int main_newroute(struct xip_ppal_ctx *ctx, struct fib_xid_table *xtbl,
208
277
* atomically to flush the appropriate anchor.
209
278
*/
210
279
rc = tree_fib_newroute_lock (& new_mrd -> common , xtbl , cfg , NULL );
211
- if (rc ) {
212
- fib_free_dnf (dnf );
213
- fxid_free_norcu (xtbl , & new_mrd -> common );
214
- return rc ;
215
- }
280
+ if (rc )
281
+ goto dnf ;
216
282
217
283
/* Flush appropriate anchor and release lock. */
218
- newroute_flush_anchor_locked (xtbl , & new_mrd -> common , dnf );
284
+ rc = newroute_flush_anchor_locked (xtbl , & new_mrd -> common , dnf );
285
+ if (rc )
286
+ goto fxid ;
287
+
219
288
return 0 ;
289
+
290
+ dnf :
291
+ fib_free_dnf (dnf );
292
+ fxid :
293
+ fxid_free_norcu (xtbl , & new_mrd -> common );
294
+ return rc ;
220
295
}
221
296
222
297
static const xia_ppal_all_rt_eops_t lpm_all_rt_eops = {
0 commit comments