Commit | Line | Data |
---|---|---|
0c83ed8e KH |
1 | /* -*- mode: c; c-basic-offset: 8; -*- |
2 | * vim: noexpandtab sw=8 ts=8 sts=0: | |
3 | * | |
4 | * Copyright (C) 2004, 2005 Oracle. All rights reserved. | |
5 | * | |
6 | * This program is free software; you can redistribute it and/or | |
7 | * modify it under the terms of the GNU General Public | |
8 | * License as published by the Free Software Foundation; either | |
9 | * version 2 of the License, or (at your option) any later version. | |
10 | * | |
11 | * This program is distributed in the hope that it will be useful, | |
12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
14 | * General Public License for more details. | |
15 | * | |
16 | * You should have received a copy of the GNU General Public | |
17 | * License along with this program; if not, write to the | |
18 | * Free Software Foundation, Inc., 59 Temple Place - Suite 330, | |
19 | * Boston, MA 021110-1307, USA. | |
20 | */ | |
21 | ||
22 | #include <linux/kernel.h> | |
23 | #include <linux/module.h> | |
24 | #include <linux/sysctl.h> | |
25 | #include <linux/configfs.h> | |
26 | ||
27 | #include "endian.h" | |
28 | #include "tcp.h" | |
29 | #include "nodemanager.h" | |
30 | #include "heartbeat.h" | |
31 | #include "masklog.h" | |
32 | #include "sys.h" | |
33 | #include "ver.h" | |
34 | ||
35 | /* for now we operate under the assertion that there can be only one | |
36 | * cluster active at a time. Changing this will require trickling | |
37 | * cluster references throughout where nodes are looked up */ | |
296b75ed | 38 | struct o2nm_cluster *o2nm_single_cluster = NULL; |
0c83ed8e KH |
39 | |
40 | #define OCFS2_MAX_HB_CTL_PATH 256 | |
41 | static char ocfs2_hb_ctl_path[OCFS2_MAX_HB_CTL_PATH] = "/sbin/ocfs2_hb_ctl"; | |
42 | ||
43 | static ctl_table ocfs2_nm_table[] = { | |
44 | { | |
45 | .ctl_name = 1, | |
46 | .procname = "hb_ctl_path", | |
47 | .data = ocfs2_hb_ctl_path, | |
48 | .maxlen = OCFS2_MAX_HB_CTL_PATH, | |
49 | .mode = 0644, | |
50 | .proc_handler = &proc_dostring, | |
51 | .strategy = &sysctl_string, | |
52 | }, | |
53 | { .ctl_name = 0 } | |
54 | }; | |
55 | ||
56 | static ctl_table ocfs2_mod_table[] = { | |
57 | { | |
58 | .ctl_name = KERN_OCFS2_NM, | |
59 | .procname = "nm", | |
60 | .data = NULL, | |
61 | .maxlen = 0, | |
62 | .mode = 0555, | |
63 | .child = ocfs2_nm_table | |
64 | }, | |
65 | { .ctl_name = 0} | |
66 | }; | |
67 | ||
68 | static ctl_table ocfs2_kern_table[] = { | |
69 | { | |
70 | .ctl_name = KERN_OCFS2, | |
71 | .procname = "ocfs2", | |
72 | .data = NULL, | |
73 | .maxlen = 0, | |
74 | .mode = 0555, | |
75 | .child = ocfs2_mod_table | |
76 | }, | |
77 | { .ctl_name = 0} | |
78 | }; | |
79 | ||
80 | static ctl_table ocfs2_root_table[] = { | |
81 | { | |
82 | .ctl_name = CTL_FS, | |
83 | .procname = "fs", | |
84 | .data = NULL, | |
85 | .maxlen = 0, | |
86 | .mode = 0555, | |
87 | .child = ocfs2_kern_table | |
88 | }, | |
89 | { .ctl_name = 0 } | |
90 | }; | |
91 | ||
92 | static struct ctl_table_header *ocfs2_table_header = NULL; | |
93 | ||
94 | const char *o2nm_get_hb_ctl_path(void) | |
95 | { | |
96 | return ocfs2_hb_ctl_path; | |
97 | } | |
98 | EXPORT_SYMBOL_GPL(o2nm_get_hb_ctl_path); | |
99 | ||
0c83ed8e KH |
100 | struct o2nm_node *o2nm_get_node_by_num(u8 node_num) |
101 | { | |
102 | struct o2nm_node *node = NULL; | |
103 | ||
104 | if (node_num >= O2NM_MAX_NODES || o2nm_single_cluster == NULL) | |
105 | goto out; | |
106 | ||
107 | read_lock(&o2nm_single_cluster->cl_nodes_lock); | |
108 | node = o2nm_single_cluster->cl_nodes[node_num]; | |
109 | if (node) | |
110 | config_item_get(&node->nd_item); | |
111 | read_unlock(&o2nm_single_cluster->cl_nodes_lock); | |
112 | out: | |
113 | return node; | |
114 | } | |
115 | EXPORT_SYMBOL_GPL(o2nm_get_node_by_num); | |
116 | ||
117 | int o2nm_configured_node_map(unsigned long *map, unsigned bytes) | |
118 | { | |
119 | struct o2nm_cluster *cluster = o2nm_single_cluster; | |
120 | ||
121 | BUG_ON(bytes < (sizeof(cluster->cl_nodes_bitmap))); | |
122 | ||
123 | if (cluster == NULL) | |
124 | return -EINVAL; | |
125 | ||
126 | read_lock(&cluster->cl_nodes_lock); | |
127 | memcpy(map, cluster->cl_nodes_bitmap, sizeof(cluster->cl_nodes_bitmap)); | |
128 | read_unlock(&cluster->cl_nodes_lock); | |
129 | ||
130 | return 0; | |
131 | } | |
132 | EXPORT_SYMBOL_GPL(o2nm_configured_node_map); | |
133 | ||
134 | static struct o2nm_node *o2nm_node_ip_tree_lookup(struct o2nm_cluster *cluster, | |
135 | __be32 ip_needle, | |
136 | struct rb_node ***ret_p, | |
137 | struct rb_node **ret_parent) | |
138 | { | |
139 | struct rb_node **p = &cluster->cl_node_ip_tree.rb_node; | |
140 | struct rb_node *parent = NULL; | |
141 | struct o2nm_node *node, *ret = NULL; | |
142 | ||
143 | while (*p) { | |
79cd22d3 AM |
144 | int cmp; |
145 | ||
0c83ed8e KH |
146 | parent = *p; |
147 | node = rb_entry(parent, struct o2nm_node, nd_ip_node); | |
148 | ||
79cd22d3 AM |
149 | cmp = memcmp(&ip_needle, &node->nd_ipv4_address, |
150 | sizeof(ip_needle)); | |
151 | if (cmp < 0) | |
0c83ed8e | 152 | p = &(*p)->rb_left; |
79cd22d3 | 153 | else if (cmp > 0) |
0c83ed8e KH |
154 | p = &(*p)->rb_right; |
155 | else { | |
156 | ret = node; | |
157 | break; | |
158 | } | |
159 | } | |
160 | ||
161 | if (ret_p != NULL) | |
162 | *ret_p = p; | |
163 | if (ret_parent != NULL) | |
164 | *ret_parent = parent; | |
165 | ||
166 | return ret; | |
167 | } | |
168 | ||
169 | struct o2nm_node *o2nm_get_node_by_ip(__be32 addr) | |
170 | { | |
171 | struct o2nm_node *node = NULL; | |
172 | struct o2nm_cluster *cluster = o2nm_single_cluster; | |
173 | ||
174 | if (cluster == NULL) | |
175 | goto out; | |
176 | ||
177 | read_lock(&cluster->cl_nodes_lock); | |
178 | node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL); | |
179 | if (node) | |
180 | config_item_get(&node->nd_item); | |
181 | read_unlock(&cluster->cl_nodes_lock); | |
182 | ||
183 | out: | |
184 | return node; | |
185 | } | |
186 | EXPORT_SYMBOL_GPL(o2nm_get_node_by_ip); | |
187 | ||
188 | void o2nm_node_put(struct o2nm_node *node) | |
189 | { | |
190 | config_item_put(&node->nd_item); | |
191 | } | |
192 | EXPORT_SYMBOL_GPL(o2nm_node_put); | |
193 | ||
194 | void o2nm_node_get(struct o2nm_node *node) | |
195 | { | |
196 | config_item_get(&node->nd_item); | |
197 | } | |
198 | EXPORT_SYMBOL_GPL(o2nm_node_get); | |
199 | ||
200 | u8 o2nm_this_node(void) | |
201 | { | |
202 | u8 node_num = O2NM_MAX_NODES; | |
203 | ||
204 | if (o2nm_single_cluster && o2nm_single_cluster->cl_has_local) | |
205 | node_num = o2nm_single_cluster->cl_local_node; | |
206 | ||
207 | return node_num; | |
208 | } | |
209 | EXPORT_SYMBOL_GPL(o2nm_this_node); | |
210 | ||
211 | /* node configfs bits */ | |
212 | ||
213 | static struct o2nm_cluster *to_o2nm_cluster(struct config_item *item) | |
214 | { | |
215 | return item ? | |
216 | container_of(to_config_group(item), struct o2nm_cluster, | |
217 | cl_group) | |
218 | : NULL; | |
219 | } | |
220 | ||
221 | static struct o2nm_node *to_o2nm_node(struct config_item *item) | |
222 | { | |
223 | return item ? container_of(item, struct o2nm_node, nd_item) : NULL; | |
224 | } | |
225 | ||
226 | static void o2nm_node_release(struct config_item *item) | |
227 | { | |
228 | struct o2nm_node *node = to_o2nm_node(item); | |
229 | kfree(node); | |
230 | } | |
231 | ||
232 | static ssize_t o2nm_node_num_read(struct o2nm_node *node, char *page) | |
233 | { | |
234 | return sprintf(page, "%d\n", node->nd_num); | |
235 | } | |
236 | ||
237 | static struct o2nm_cluster *to_o2nm_cluster_from_node(struct o2nm_node *node) | |
238 | { | |
239 | /* through the first node_set .parent | |
240 | * mycluster/nodes/mynode == o2nm_cluster->o2nm_node_group->o2nm_node */ | |
241 | return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent); | |
242 | } | |
243 | ||
244 | enum { | |
245 | O2NM_NODE_ATTR_NUM = 0, | |
246 | O2NM_NODE_ATTR_PORT, | |
247 | O2NM_NODE_ATTR_ADDRESS, | |
248 | O2NM_NODE_ATTR_LOCAL, | |
249 | }; | |
250 | ||
251 | static ssize_t o2nm_node_num_write(struct o2nm_node *node, const char *page, | |
252 | size_t count) | |
253 | { | |
254 | struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node); | |
255 | unsigned long tmp; | |
256 | char *p = (char *)page; | |
257 | ||
258 | tmp = simple_strtoul(p, &p, 0); | |
259 | if (!p || (*p && (*p != '\n'))) | |
260 | return -EINVAL; | |
261 | ||
262 | if (tmp >= O2NM_MAX_NODES) | |
263 | return -ERANGE; | |
264 | ||
265 | /* once we're in the cl_nodes tree networking can look us up by | |
266 | * node number and try to use our address and port attributes | |
267 | * to connect to this node.. make sure that they've been set | |
268 | * before writing the node attribute? */ | |
269 | if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) || | |
270 | !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) | |
271 | return -EINVAL; /* XXX */ | |
272 | ||
273 | write_lock(&cluster->cl_nodes_lock); | |
274 | if (cluster->cl_nodes[tmp]) | |
275 | p = NULL; | |
276 | else { | |
277 | cluster->cl_nodes[tmp] = node; | |
278 | node->nd_num = tmp; | |
279 | set_bit(tmp, cluster->cl_nodes_bitmap); | |
280 | } | |
281 | write_unlock(&cluster->cl_nodes_lock); | |
282 | if (p == NULL) | |
283 | return -EEXIST; | |
284 | ||
285 | return count; | |
286 | } | |
287 | static ssize_t o2nm_node_ipv4_port_read(struct o2nm_node *node, char *page) | |
288 | { | |
289 | return sprintf(page, "%u\n", ntohs(node->nd_ipv4_port)); | |
290 | } | |
291 | ||
292 | static ssize_t o2nm_node_ipv4_port_write(struct o2nm_node *node, | |
293 | const char *page, size_t count) | |
294 | { | |
295 | unsigned long tmp; | |
296 | char *p = (char *)page; | |
297 | ||
298 | tmp = simple_strtoul(p, &p, 0); | |
299 | if (!p || (*p && (*p != '\n'))) | |
300 | return -EINVAL; | |
301 | ||
302 | if (tmp == 0) | |
303 | return -EINVAL; | |
304 | if (tmp >= (u16)-1) | |
305 | return -ERANGE; | |
306 | ||
307 | node->nd_ipv4_port = htons(tmp); | |
308 | ||
309 | return count; | |
310 | } | |
311 | ||
312 | static ssize_t o2nm_node_ipv4_address_read(struct o2nm_node *node, char *page) | |
313 | { | |
314 | return sprintf(page, "%u.%u.%u.%u\n", NIPQUAD(node->nd_ipv4_address)); | |
315 | } | |
316 | ||
317 | static ssize_t o2nm_node_ipv4_address_write(struct o2nm_node *node, | |
318 | const char *page, | |
319 | size_t count) | |
320 | { | |
321 | struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node); | |
322 | int ret, i; | |
323 | struct rb_node **p, *parent; | |
324 | unsigned int octets[4]; | |
325 | __be32 ipv4_addr = 0; | |
326 | ||
327 | ret = sscanf(page, "%3u.%3u.%3u.%3u", &octets[3], &octets[2], | |
328 | &octets[1], &octets[0]); | |
329 | if (ret != 4) | |
330 | return -EINVAL; | |
331 | ||
332 | for (i = 0; i < ARRAY_SIZE(octets); i++) { | |
333 | if (octets[i] > 255) | |
334 | return -ERANGE; | |
335 | be32_add_cpu(&ipv4_addr, octets[i] << (i * 8)); | |
336 | } | |
337 | ||
338 | ret = 0; | |
339 | write_lock(&cluster->cl_nodes_lock); | |
340 | if (o2nm_node_ip_tree_lookup(cluster, ipv4_addr, &p, &parent)) | |
341 | ret = -EEXIST; | |
342 | else { | |
343 | rb_link_node(&node->nd_ip_node, parent, p); | |
344 | rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree); | |
345 | } | |
346 | write_unlock(&cluster->cl_nodes_lock); | |
347 | if (ret) | |
348 | return ret; | |
349 | ||
350 | memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr)); | |
351 | ||
352 | return count; | |
353 | } | |
354 | ||
355 | static ssize_t o2nm_node_local_read(struct o2nm_node *node, char *page) | |
356 | { | |
357 | return sprintf(page, "%d\n", node->nd_local); | |
358 | } | |
359 | ||
360 | static ssize_t o2nm_node_local_write(struct o2nm_node *node, const char *page, | |
361 | size_t count) | |
362 | { | |
363 | struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node); | |
364 | unsigned long tmp; | |
365 | char *p = (char *)page; | |
366 | ssize_t ret; | |
367 | ||
368 | tmp = simple_strtoul(p, &p, 0); | |
369 | if (!p || (*p && (*p != '\n'))) | |
370 | return -EINVAL; | |
371 | ||
372 | tmp = !!tmp; /* boolean of whether this node wants to be local */ | |
373 | ||
374 | /* setting local turns on networking rx for now so we require having | |
375 | * set everything else first */ | |
376 | if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) || | |
377 | !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) || | |
378 | !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) | |
379 | return -EINVAL; /* XXX */ | |
380 | ||
381 | /* the only failure case is trying to set a new local node | |
382 | * when a different one is already set */ | |
383 | if (tmp && tmp == cluster->cl_has_local && | |
384 | cluster->cl_local_node != node->nd_num) | |
385 | return -EBUSY; | |
386 | ||
387 | /* bring up the rx thread if we're setting the new local node. */ | |
388 | if (tmp && !cluster->cl_has_local) { | |
389 | ret = o2net_start_listening(node); | |
390 | if (ret) | |
391 | return ret; | |
392 | } | |
393 | ||
394 | if (!tmp && cluster->cl_has_local && | |
395 | cluster->cl_local_node == node->nd_num) { | |
396 | o2net_stop_listening(node); | |
397 | cluster->cl_local_node = O2NM_INVALID_NODE_NUM; | |
398 | } | |
399 | ||
400 | node->nd_local = tmp; | |
401 | if (node->nd_local) { | |
402 | cluster->cl_has_local = tmp; | |
403 | cluster->cl_local_node = node->nd_num; | |
404 | } | |
405 | ||
406 | return count; | |
407 | } | |
408 | ||
409 | struct o2nm_node_attribute { | |
410 | struct configfs_attribute attr; | |
411 | ssize_t (*show)(struct o2nm_node *, char *); | |
412 | ssize_t (*store)(struct o2nm_node *, const char *, size_t); | |
413 | }; | |
414 | ||
415 | static struct o2nm_node_attribute o2nm_node_attr_num = { | |
416 | .attr = { .ca_owner = THIS_MODULE, | |
417 | .ca_name = "num", | |
418 | .ca_mode = S_IRUGO | S_IWUSR }, | |
419 | .show = o2nm_node_num_read, | |
420 | .store = o2nm_node_num_write, | |
421 | }; | |
422 | ||
423 | static struct o2nm_node_attribute o2nm_node_attr_ipv4_port = { | |
424 | .attr = { .ca_owner = THIS_MODULE, | |
425 | .ca_name = "ipv4_port", | |
426 | .ca_mode = S_IRUGO | S_IWUSR }, | |
427 | .show = o2nm_node_ipv4_port_read, | |
428 | .store = o2nm_node_ipv4_port_write, | |
429 | }; | |
430 | ||
431 | static struct o2nm_node_attribute o2nm_node_attr_ipv4_address = { | |
432 | .attr = { .ca_owner = THIS_MODULE, | |
433 | .ca_name = "ipv4_address", | |
434 | .ca_mode = S_IRUGO | S_IWUSR }, | |
435 | .show = o2nm_node_ipv4_address_read, | |
436 | .store = o2nm_node_ipv4_address_write, | |
437 | }; | |
438 | ||
439 | static struct o2nm_node_attribute o2nm_node_attr_local = { | |
440 | .attr = { .ca_owner = THIS_MODULE, | |
441 | .ca_name = "local", | |
442 | .ca_mode = S_IRUGO | S_IWUSR }, | |
443 | .show = o2nm_node_local_read, | |
444 | .store = o2nm_node_local_write, | |
445 | }; | |
446 | ||
447 | static struct configfs_attribute *o2nm_node_attrs[] = { | |
448 | [O2NM_NODE_ATTR_NUM] = &o2nm_node_attr_num.attr, | |
449 | [O2NM_NODE_ATTR_PORT] = &o2nm_node_attr_ipv4_port.attr, | |
450 | [O2NM_NODE_ATTR_ADDRESS] = &o2nm_node_attr_ipv4_address.attr, | |
451 | [O2NM_NODE_ATTR_LOCAL] = &o2nm_node_attr_local.attr, | |
452 | NULL, | |
453 | }; | |
454 | ||
455 | static int o2nm_attr_index(struct configfs_attribute *attr) | |
456 | { | |
457 | int i; | |
458 | for (i = 0; i < ARRAY_SIZE(o2nm_node_attrs); i++) { | |
459 | if (attr == o2nm_node_attrs[i]) | |
460 | return i; | |
461 | } | |
462 | BUG(); | |
463 | return 0; | |
464 | } | |
465 | ||
466 | static ssize_t o2nm_node_show(struct config_item *item, | |
467 | struct configfs_attribute *attr, | |
468 | char *page) | |
469 | { | |
470 | struct o2nm_node *node = to_o2nm_node(item); | |
471 | struct o2nm_node_attribute *o2nm_node_attr = | |
472 | container_of(attr, struct o2nm_node_attribute, attr); | |
473 | ssize_t ret = 0; | |
474 | ||
475 | if (o2nm_node_attr->show) | |
476 | ret = o2nm_node_attr->show(node, page); | |
477 | return ret; | |
478 | } | |
479 | ||
480 | static ssize_t o2nm_node_store(struct config_item *item, | |
481 | struct configfs_attribute *attr, | |
482 | const char *page, size_t count) | |
483 | { | |
484 | struct o2nm_node *node = to_o2nm_node(item); | |
485 | struct o2nm_node_attribute *o2nm_node_attr = | |
486 | container_of(attr, struct o2nm_node_attribute, attr); | |
487 | ssize_t ret; | |
488 | int attr_index = o2nm_attr_index(attr); | |
489 | ||
490 | if (o2nm_node_attr->store == NULL) { | |
491 | ret = -EINVAL; | |
492 | goto out; | |
493 | } | |
494 | ||
495 | if (test_bit(attr_index, &node->nd_set_attributes)) | |
496 | return -EBUSY; | |
497 | ||
498 | ret = o2nm_node_attr->store(node, page, count); | |
499 | if (ret < count) | |
500 | goto out; | |
501 | ||
502 | set_bit(attr_index, &node->nd_set_attributes); | |
503 | out: | |
504 | return ret; | |
505 | } | |
506 | ||
507 | static struct configfs_item_operations o2nm_node_item_ops = { | |
508 | .release = o2nm_node_release, | |
509 | .show_attribute = o2nm_node_show, | |
510 | .store_attribute = o2nm_node_store, | |
511 | }; | |
512 | ||
513 | static struct config_item_type o2nm_node_type = { | |
514 | .ct_item_ops = &o2nm_node_item_ops, | |
515 | .ct_attrs = o2nm_node_attrs, | |
516 | .ct_owner = THIS_MODULE, | |
517 | }; | |
518 | ||
519 | /* node set */ | |
520 | ||
521 | struct o2nm_node_group { | |
522 | struct config_group ns_group; | |
523 | /* some stuff? */ | |
524 | }; | |
525 | ||
526 | #if 0 | |
527 | static struct o2nm_node_group *to_o2nm_node_group(struct config_group *group) | |
528 | { | |
529 | return group ? | |
530 | container_of(group, struct o2nm_node_group, ns_group) | |
531 | : NULL; | |
532 | } | |
533 | #endif | |
534 | ||
535 | static struct config_item *o2nm_node_group_make_item(struct config_group *group, | |
536 | const char *name) | |
537 | { | |
538 | struct o2nm_node *node = NULL; | |
539 | struct config_item *ret = NULL; | |
540 | ||
541 | if (strlen(name) > O2NM_MAX_NAME_LEN) | |
542 | goto out; /* ENAMETOOLONG */ | |
543 | ||
544 | node = kcalloc(1, sizeof(struct o2nm_node), GFP_KERNEL); | |
545 | if (node == NULL) | |
546 | goto out; /* ENOMEM */ | |
547 | ||
548 | strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */ | |
549 | config_item_init_type_name(&node->nd_item, name, &o2nm_node_type); | |
550 | spin_lock_init(&node->nd_lock); | |
551 | ||
552 | ret = &node->nd_item; | |
553 | ||
554 | out: | |
555 | if (ret == NULL) | |
556 | kfree(node); | |
557 | ||
558 | return ret; | |
559 | } | |
560 | ||
561 | static void o2nm_node_group_drop_item(struct config_group *group, | |
562 | struct config_item *item) | |
563 | { | |
564 | struct o2nm_node *node = to_o2nm_node(item); | |
565 | struct o2nm_cluster *cluster = to_o2nm_cluster(group->cg_item.ci_parent); | |
566 | ||
567 | o2net_disconnect_node(node); | |
568 | ||
569 | if (cluster->cl_has_local && | |
570 | (cluster->cl_local_node == node->nd_num)) { | |
571 | cluster->cl_has_local = 0; | |
572 | cluster->cl_local_node = O2NM_INVALID_NODE_NUM; | |
573 | o2net_stop_listening(node); | |
574 | } | |
575 | ||
576 | /* XXX call into net to stop this node from trading messages */ | |
577 | ||
578 | write_lock(&cluster->cl_nodes_lock); | |
579 | ||
580 | /* XXX sloppy */ | |
581 | if (node->nd_ipv4_address) | |
582 | rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree); | |
583 | ||
584 | /* nd_num might be 0 if the node number hasn't been set.. */ | |
585 | if (cluster->cl_nodes[node->nd_num] == node) { | |
586 | cluster->cl_nodes[node->nd_num] = NULL; | |
587 | clear_bit(node->nd_num, cluster->cl_nodes_bitmap); | |
588 | } | |
589 | write_unlock(&cluster->cl_nodes_lock); | |
590 | ||
591 | config_item_put(item); | |
592 | } | |
593 | ||
594 | static struct configfs_group_operations o2nm_node_group_group_ops = { | |
595 | .make_item = o2nm_node_group_make_item, | |
596 | .drop_item = o2nm_node_group_drop_item, | |
597 | }; | |
598 | ||
599 | static struct config_item_type o2nm_node_group_type = { | |
600 | .ct_group_ops = &o2nm_node_group_group_ops, | |
601 | .ct_owner = THIS_MODULE, | |
602 | }; | |
603 | ||
604 | /* cluster */ | |
605 | ||
606 | static void o2nm_cluster_release(struct config_item *item) | |
607 | { | |
608 | struct o2nm_cluster *cluster = to_o2nm_cluster(item); | |
609 | ||
610 | kfree(cluster->cl_group.default_groups); | |
611 | kfree(cluster); | |
612 | } | |
613 | ||
614 | static struct configfs_item_operations o2nm_cluster_item_ops = { | |
615 | .release = o2nm_cluster_release, | |
616 | }; | |
617 | ||
618 | static struct config_item_type o2nm_cluster_type = { | |
619 | .ct_item_ops = &o2nm_cluster_item_ops, | |
620 | .ct_owner = THIS_MODULE, | |
621 | }; | |
622 | ||
623 | /* cluster set */ | |
624 | ||
625 | struct o2nm_cluster_group { | |
626 | struct configfs_subsystem cs_subsys; | |
627 | /* some stuff? */ | |
628 | }; | |
629 | ||
630 | #if 0 | |
631 | static struct o2nm_cluster_group *to_o2nm_cluster_group(struct config_group *group) | |
632 | { | |
633 | return group ? | |
634 | container_of(to_configfs_subsystem(group), struct o2nm_cluster_group, cs_subsys) | |
635 | : NULL; | |
636 | } | |
637 | #endif | |
638 | ||
639 | static struct config_group *o2nm_cluster_group_make_group(struct config_group *group, | |
640 | const char *name) | |
641 | { | |
642 | struct o2nm_cluster *cluster = NULL; | |
643 | struct o2nm_node_group *ns = NULL; | |
644 | struct config_group *o2hb_group = NULL, *ret = NULL; | |
645 | void *defs = NULL; | |
646 | ||
1b1dcc1b | 647 | /* this runs under the parent dir's i_mutex; there can be only |
0c83ed8e KH |
648 | * one caller in here at a time */ |
649 | if (o2nm_single_cluster) | |
650 | goto out; /* ENOSPC */ | |
651 | ||
652 | cluster = kcalloc(1, sizeof(struct o2nm_cluster), GFP_KERNEL); | |
653 | ns = kcalloc(1, sizeof(struct o2nm_node_group), GFP_KERNEL); | |
654 | defs = kcalloc(3, sizeof(struct config_group *), GFP_KERNEL); | |
655 | o2hb_group = o2hb_alloc_hb_set(); | |
656 | if (cluster == NULL || ns == NULL || o2hb_group == NULL || defs == NULL) | |
657 | goto out; | |
658 | ||
659 | config_group_init_type_name(&cluster->cl_group, name, | |
660 | &o2nm_cluster_type); | |
661 | config_group_init_type_name(&ns->ns_group, "node", | |
662 | &o2nm_node_group_type); | |
663 | ||
664 | cluster->cl_group.default_groups = defs; | |
665 | cluster->cl_group.default_groups[0] = &ns->ns_group; | |
666 | cluster->cl_group.default_groups[1] = o2hb_group; | |
667 | cluster->cl_group.default_groups[2] = NULL; | |
668 | rwlock_init(&cluster->cl_nodes_lock); | |
669 | cluster->cl_node_ip_tree = RB_ROOT; | |
670 | ||
671 | ret = &cluster->cl_group; | |
672 | o2nm_single_cluster = cluster; | |
673 | ||
674 | out: | |
675 | if (ret == NULL) { | |
676 | kfree(cluster); | |
677 | kfree(ns); | |
678 | o2hb_free_hb_set(o2hb_group); | |
679 | kfree(defs); | |
680 | } | |
681 | ||
682 | return ret; | |
683 | } | |
684 | ||
685 | static void o2nm_cluster_group_drop_item(struct config_group *group, struct config_item *item) | |
686 | { | |
687 | struct o2nm_cluster *cluster = to_o2nm_cluster(item); | |
688 | int i; | |
689 | struct config_item *killme; | |
690 | ||
691 | BUG_ON(o2nm_single_cluster != cluster); | |
692 | o2nm_single_cluster = NULL; | |
693 | ||
694 | for (i = 0; cluster->cl_group.default_groups[i]; i++) { | |
695 | killme = &cluster->cl_group.default_groups[i]->cg_item; | |
696 | cluster->cl_group.default_groups[i] = NULL; | |
697 | config_item_put(killme); | |
698 | } | |
699 | ||
700 | config_item_put(item); | |
701 | } | |
702 | ||
703 | static struct configfs_group_operations o2nm_cluster_group_group_ops = { | |
704 | .make_group = o2nm_cluster_group_make_group, | |
705 | .drop_item = o2nm_cluster_group_drop_item, | |
706 | }; | |
707 | ||
708 | static struct config_item_type o2nm_cluster_group_type = { | |
709 | .ct_group_ops = &o2nm_cluster_group_group_ops, | |
710 | .ct_owner = THIS_MODULE, | |
711 | }; | |
712 | ||
713 | static struct o2nm_cluster_group o2nm_cluster_group = { | |
714 | .cs_subsys = { | |
715 | .su_group = { | |
716 | .cg_item = { | |
717 | .ci_namebuf = "cluster", | |
718 | .ci_type = &o2nm_cluster_group_type, | |
719 | }, | |
720 | }, | |
721 | }, | |
722 | }; | |
723 | ||
724 | static void __exit exit_o2nm(void) | |
725 | { | |
726 | if (ocfs2_table_header) | |
727 | unregister_sysctl_table(ocfs2_table_header); | |
728 | ||
729 | /* XXX sync with hb callbacks and shut down hb? */ | |
730 | o2net_unregister_hb_callbacks(); | |
731 | configfs_unregister_subsystem(&o2nm_cluster_group.cs_subsys); | |
732 | o2cb_sys_shutdown(); | |
733 | ||
734 | o2net_exit(); | |
735 | } | |
736 | ||
737 | static int __init init_o2nm(void) | |
738 | { | |
739 | int ret = -1; | |
740 | ||
741 | cluster_print_version(); | |
742 | ||
743 | o2hb_init(); | |
744 | o2net_init(); | |
745 | ||
746 | ocfs2_table_header = register_sysctl_table(ocfs2_root_table, 0); | |
747 | if (!ocfs2_table_header) { | |
748 | printk(KERN_ERR "nodemanager: unable to register sysctl\n"); | |
749 | ret = -ENOMEM; /* or something. */ | |
895928b8 | 750 | goto out_o2net; |
0c83ed8e KH |
751 | } |
752 | ||
753 | ret = o2net_register_hb_callbacks(); | |
754 | if (ret) | |
755 | goto out_sysctl; | |
756 | ||
757 | config_group_init(&o2nm_cluster_group.cs_subsys.su_group); | |
758 | init_MUTEX(&o2nm_cluster_group.cs_subsys.su_sem); | |
759 | ret = configfs_register_subsystem(&o2nm_cluster_group.cs_subsys); | |
760 | if (ret) { | |
761 | printk(KERN_ERR "nodemanager: Registration returned %d\n", ret); | |
762 | goto out_callbacks; | |
763 | } | |
764 | ||
765 | ret = o2cb_sys_init(); | |
766 | if (!ret) | |
767 | goto out; | |
768 | ||
769 | configfs_unregister_subsystem(&o2nm_cluster_group.cs_subsys); | |
770 | out_callbacks: | |
771 | o2net_unregister_hb_callbacks(); | |
772 | out_sysctl: | |
773 | unregister_sysctl_table(ocfs2_table_header); | |
895928b8 JM |
774 | out_o2net: |
775 | o2net_exit(); | |
0c83ed8e KH |
776 | out: |
777 | return ret; | |
778 | } | |
779 | ||
780 | MODULE_AUTHOR("Oracle"); | |
781 | MODULE_LICENSE("GPL"); | |
782 | ||
783 | module_init(init_o2nm) | |
784 | module_exit(exit_o2nm) |