1 // SPDX-License-Identifier: GPL-2.0-only 2 /****************************************************************************** 3 ******************************************************************************* 4 ** 5 ** Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. 6 ** Copyright (C) 2004-2021 Red Hat, Inc. All rights reserved. 7 ** 8 ** 9 ******************************************************************************* 10 ******************************************************************************/ 11 12 /* 13 * midcomms.c 14 * 15 * This is the appallingly named "mid-level" comms layer. It takes care about 16 * deliver an on application layer "reliable" communication above the used 17 * lowcomms transport layer. 18 * 19 * How it works: 20 * 21 * Each nodes keeps track of all send DLM messages in send_queue with a sequence 22 * number. The receive will send an DLM_ACK message back for every DLM message 23 * received at the other side. If a reconnect happens in lowcomms we will send 24 * all unacknowledged dlm messages again. The receiving side might drop any already 25 * received message by comparing sequence numbers. 26 * 27 * How version detection works: 28 * 29 * Due the fact that dlm has pre-configured node addresses on every side 30 * it is in it's nature that every side connects at starts to transmit 31 * dlm messages which ends in a race. However DLM_RCOM_NAMES, DLM_RCOM_STATUS 32 * and their replies are the first messages which are exchanges. Due backwards 33 * compatibility these messages are not covered by the midcomms re-transmission 34 * layer. These messages have their own re-transmission handling in the dlm 35 * application layer. The version field of every node will be set on these RCOM 36 * messages as soon as they arrived and the node isn't yet part of the nodes 37 * hash. There exists also logic to detect version mismatched if something weird 38 * going on or the first messages isn't an expected one. 39 * 40 * Termination: 41 * 42 * The midcomms layer does a 4 way handshake for termination on DLM protocol 43 * like TCP supports it with half-closed socket support. SCTP doesn't support 44 * half-closed socket, so we do it on DLM layer. Also socket shutdown() can be 45 * interrupted by .e.g. tcp reset itself. Additional there exists the othercon 46 * paradigm in lowcomms which cannot be easily without breaking backwards 47 * compatibility. A node cannot send anything to another node when a DLM_FIN 48 * message was send. There exists additional logic to print a warning if 49 * DLM wants to do it. There exists a state handling like RFC 793 but reduced 50 * to termination only. The event "member removal event" describes the cluster 51 * manager removed the node from internal lists, at this point DLM does not 52 * send any message to the other node. There exists two cases: 53 * 54 * 1. The cluster member was removed and we received a FIN 55 * OR 56 * 2. We received a FIN but the member was not removed yet 57 * 58 * One of these cases will do the CLOSE_WAIT to LAST_ACK change. 59 * 60 * 61 * +---------+ 62 * | CLOSED | 63 * +---------+ 64 * | add member/receive RCOM version 65 * | detection msg 66 * V 67 * +---------+ 68 * | ESTAB | 69 * +---------+ 70 * CLOSE | | rcv FIN 71 * ------- | | ------- 72 * +---------+ snd FIN / \ snd ACK +---------+ 73 * | FIN |<----------------- ------------------>| CLOSE | 74 * | WAIT-1 |------------------ | WAIT | 75 * +---------+ rcv FIN \ +---------+ 76 * | rcv ACK of FIN ------- | CLOSE | member 77 * | -------------- snd ACK | ------- | removal 78 * V x V snd FIN V event 79 * +---------+ +---------+ +---------+ 80 * |FINWAIT-2| | CLOSING | | LAST-ACK| 81 * +---------+ +---------+ +---------+ 82 * | rcv ACK of FIN | rcv ACK of FIN | 83 * | rcv FIN -------------- | -------------- | 84 * | ------- x V x V 85 * \ snd ACK +---------+ +---------+ 86 * ------------------------>| CLOSED | | CLOSED | 87 * +---------+ +---------+ 88 * 89 * NOTE: any state can interrupted by midcomms_close() and state will be 90 * switched to CLOSED in case of fencing. There exists also some timeout 91 * handling when we receive the version detection RCOM messages which is 92 * made by observation. 93 * 94 * Future improvements: 95 * 96 * There exists some known issues/improvements of the dlm handling. Some 97 * of them should be done in a next major dlm version bump which makes 98 * it incompatible with previous versions. 99 * 100 * Unaligned memory access: 101 * 102 * There exists cases when the dlm message buffer length is not aligned 103 * to 8 byte. However seems nobody detected any problem with it. This 104 * can be fixed in the next major version bump of dlm. 105 * 106 * Version detection: 107 * 108 * The version detection and how it's done is related to backwards 109 * compatibility. There exists better ways to make a better handling. 110 * However this should be changed in the next major version bump of dlm. 111 * 112 * Tail Size checking: 113 * 114 * There exists a message tail payload in e.g. DLM_MSG however we don't 115 * check it against the message length yet regarding to the receive buffer 116 * length. That need to be validated. 117 * 118 * Fencing bad nodes: 119 * 120 * At timeout places or weird sequence number behaviours we should send 121 * a fencing request to the cluster manager. 122 */ 123 124 /* Debug switch to enable a 5 seconds sleep waiting of a termination. 125 * This can be useful to test fencing while termination is running. 126 * This requires a setup with only gfs2 as dlm user, so that the 127 * last umount will terminate the connection. 128 * 129 * However it became useful to test, while the 5 seconds block in umount 130 * just press the reset button. In a lot of dropping the termination 131 * process can could take several seconds. 132 */ 133 #define DLM_DEBUG_FENCE_TERMINATION 0 134 135 #include <trace/events/dlm.h> 136 #include <net/tcp.h> 137 138 #include "dlm_internal.h" 139 #include "lowcomms.h" 140 #include "config.h" 141 #include "memory.h" 142 #include "lock.h" 143 #include "util.h" 144 #include "midcomms.h" 145 146 /* init value for sequence numbers for testing purpose only e.g. overflows */ 147 #define DLM_SEQ_INIT 0 148 /* 5 seconds wait to sync ending of dlm */ 149 #define DLM_SHUTDOWN_TIMEOUT msecs_to_jiffies(5000) 150 #define DLM_VERSION_NOT_SET 0 151 #define DLM_SEND_ACK_BACK_MSG_THRESHOLD 32 152 #define DLM_RECV_ACK_BACK_MSG_THRESHOLD (DLM_SEND_ACK_BACK_MSG_THRESHOLD * 8) 153 154 struct midcomms_node { 155 int nodeid; 156 uint32_t version; 157 atomic_t seq_send; 158 atomic_t seq_next; 159 /* These queues are unbound because we cannot drop any message in dlm. 160 * We could send a fence signal for a specific node to the cluster 161 * manager if queues hits some maximum value, however this handling 162 * not supported yet. 163 */ 164 struct list_head send_queue; 165 spinlock_t send_queue_lock; 166 atomic_t send_queue_cnt; 167 #define DLM_NODE_FLAG_CLOSE 1 168 #define DLM_NODE_FLAG_STOP_TX 2 169 #define DLM_NODE_FLAG_STOP_RX 3 170 atomic_t ulp_delivered; 171 unsigned long flags; 172 wait_queue_head_t shutdown_wait; 173 174 /* dlm tcp termination state */ 175 #define DLM_CLOSED 1 176 #define DLM_ESTABLISHED 2 177 #define DLM_FIN_WAIT1 3 178 #define DLM_FIN_WAIT2 4 179 #define DLM_CLOSE_WAIT 5 180 #define DLM_LAST_ACK 6 181 #define DLM_CLOSING 7 182 int state; 183 spinlock_t state_lock; 184 185 /* counts how many lockspaces are using this node 186 * this refcount is necessary to determine if the 187 * node wants to disconnect. 188 */ 189 int users; 190 191 /* not protected by srcu, node_hash lifetime */ 192 void *debugfs; 193 194 struct hlist_node hlist; 195 struct rcu_head rcu; 196 }; 197 198 struct dlm_mhandle { 199 const union dlm_packet *inner_p; 200 struct midcomms_node *node; 201 struct dlm_opts *opts; 202 struct dlm_msg *msg; 203 bool committed; 204 uint32_t seq; 205 206 void (*ack_rcv)(struct midcomms_node *node); 207 208 /* get_mhandle/commit srcu idx exchange */ 209 int idx; 210 211 struct list_head list; 212 struct rcu_head rcu; 213 }; 214 215 static struct hlist_head node_hash[CONN_HASH_SIZE]; 216 static DEFINE_SPINLOCK(nodes_lock); 217 DEFINE_STATIC_SRCU(nodes_srcu); 218 219 /* This mutex prevents that midcomms_close() is running while 220 * stop() or remove(). As I experienced invalid memory access 221 * behaviours when DLM_DEBUG_FENCE_TERMINATION is enabled and 222 * resetting machines. I will end in some double deletion in nodes 223 * datastructure. 224 */ 225 static DEFINE_MUTEX(close_lock); 226 227 struct kmem_cache *dlm_midcomms_cache_create(void) 228 { 229 return KMEM_CACHE(dlm_mhandle, 0); 230 } 231 232 static inline const char *dlm_state_str(int state) 233 { 234 switch (state) { 235 case DLM_CLOSED: 236 return "CLOSED"; 237 case DLM_ESTABLISHED: 238 return "ESTABLISHED"; 239 case DLM_FIN_WAIT1: 240 return "FIN_WAIT1"; 241 case DLM_FIN_WAIT2: 242 return "FIN_WAIT2"; 243 case DLM_CLOSE_WAIT: 244 return "CLOSE_WAIT"; 245 case DLM_LAST_ACK: 246 return "LAST_ACK"; 247 case DLM_CLOSING: 248 return "CLOSING"; 249 default: 250 return "UNKNOWN"; 251 } 252 } 253 254 const char *dlm_midcomms_state(struct midcomms_node *node) 255 { 256 return dlm_state_str(node->state); 257 } 258 259 unsigned long dlm_midcomms_flags(struct midcomms_node *node) 260 { 261 return node->flags; 262 } 263 264 int dlm_midcomms_send_queue_cnt(struct midcomms_node *node) 265 { 266 return atomic_read(&node->send_queue_cnt); 267 } 268 269 uint32_t dlm_midcomms_version(struct midcomms_node *node) 270 { 271 return node->version; 272 } 273 274 static struct midcomms_node *__find_node(int nodeid, int r) 275 { 276 struct midcomms_node *node; 277 278 hlist_for_each_entry_rcu(node, &node_hash[r], hlist) { 279 if (node->nodeid == nodeid) 280 return node; 281 } 282 283 return NULL; 284 } 285 286 static void dlm_mhandle_release(struct rcu_head *rcu) 287 { 288 struct dlm_mhandle *mh = container_of(rcu, struct dlm_mhandle, rcu); 289 290 dlm_lowcomms_put_msg(mh->msg); 291 dlm_free_mhandle(mh); 292 } 293 294 static void dlm_mhandle_delete(struct midcomms_node *node, 295 struct dlm_mhandle *mh) 296 { 297 list_del_rcu(&mh->list); 298 atomic_dec(&node->send_queue_cnt); 299 call_rcu(&mh->rcu, dlm_mhandle_release); 300 } 301 302 static void dlm_send_queue_flush(struct midcomms_node *node) 303 { 304 struct dlm_mhandle *mh; 305 306 pr_debug("flush midcomms send queue of node %d\n", node->nodeid); 307 308 rcu_read_lock(); 309 spin_lock_bh(&node->send_queue_lock); 310 list_for_each_entry_rcu(mh, &node->send_queue, list) { 311 dlm_mhandle_delete(node, mh); 312 } 313 spin_unlock_bh(&node->send_queue_lock); 314 rcu_read_unlock(); 315 } 316 317 static void midcomms_node_reset(struct midcomms_node *node) 318 { 319 pr_debug("reset node %d\n", node->nodeid); 320 321 atomic_set(&node->seq_next, DLM_SEQ_INIT); 322 atomic_set(&node->seq_send, DLM_SEQ_INIT); 323 atomic_set(&node->ulp_delivered, 0); 324 node->version = DLM_VERSION_NOT_SET; 325 node->flags = 0; 326 327 dlm_send_queue_flush(node); 328 node->state = DLM_CLOSED; 329 wake_up(&node->shutdown_wait); 330 } 331 332 static struct midcomms_node *nodeid2node(int nodeid) 333 { 334 return __find_node(nodeid, nodeid_hash(nodeid)); 335 } 336 337 int dlm_midcomms_addr(int nodeid, struct sockaddr_storage *addr) 338 { 339 int ret, idx, r = nodeid_hash(nodeid); 340 struct midcomms_node *node; 341 342 ret = dlm_lowcomms_addr(nodeid, addr); 343 if (ret) 344 return ret; 345 346 idx = srcu_read_lock(&nodes_srcu); 347 node = __find_node(nodeid, r); 348 if (node) { 349 srcu_read_unlock(&nodes_srcu, idx); 350 return 0; 351 } 352 srcu_read_unlock(&nodes_srcu, idx); 353 354 node = kmalloc(sizeof(*node), GFP_NOFS); 355 if (!node) 356 return -ENOMEM; 357 358 node->nodeid = nodeid; 359 spin_lock_init(&node->state_lock); 360 spin_lock_init(&node->send_queue_lock); 361 atomic_set(&node->send_queue_cnt, 0); 362 INIT_LIST_HEAD(&node->send_queue); 363 init_waitqueue_head(&node->shutdown_wait); 364 node->users = 0; 365 midcomms_node_reset(node); 366 367 spin_lock_bh(&nodes_lock); 368 hlist_add_head_rcu(&node->hlist, &node_hash[r]); 369 spin_unlock_bh(&nodes_lock); 370 371 node->debugfs = dlm_create_debug_comms_file(nodeid, node); 372 return 0; 373 } 374 375 static int dlm_send_ack(int nodeid, uint32_t seq) 376 { 377 int mb_len = sizeof(struct dlm_header); 378 struct dlm_header *m_header; 379 struct dlm_msg *msg; 380 char *ppc; 381 382 msg = dlm_lowcomms_new_msg(nodeid, mb_len, &ppc, NULL, NULL); 383 if (!msg) 384 return -ENOMEM; 385 386 m_header = (struct dlm_header *)ppc; 387 388 m_header->h_version = cpu_to_le32(DLM_HEADER_MAJOR | DLM_HEADER_MINOR); 389 m_header->h_nodeid = cpu_to_le32(dlm_our_nodeid()); 390 m_header->h_length = cpu_to_le16(mb_len); 391 m_header->h_cmd = DLM_ACK; 392 m_header->u.h_seq = cpu_to_le32(seq); 393 394 dlm_lowcomms_commit_msg(msg); 395 dlm_lowcomms_put_msg(msg); 396 397 return 0; 398 } 399 400 static void dlm_send_ack_threshold(struct midcomms_node *node, 401 uint32_t threshold) 402 { 403 uint32_t oval, nval; 404 bool send_ack; 405 406 /* let only send one user trigger threshold to send ack back */ 407 do { 408 oval = atomic_read(&node->ulp_delivered); 409 send_ack = (oval > threshold); 410 /* abort if threshold is not reached */ 411 if (!send_ack) 412 break; 413 414 nval = 0; 415 /* try to reset ulp_delivered counter */ 416 } while (atomic_cmpxchg(&node->ulp_delivered, oval, nval) != oval); 417 418 if (send_ack) 419 dlm_send_ack(node->nodeid, atomic_read(&node->seq_next)); 420 } 421 422 static int dlm_send_fin(struct midcomms_node *node, 423 void (*ack_rcv)(struct midcomms_node *node)) 424 { 425 int mb_len = sizeof(struct dlm_header); 426 struct dlm_header *m_header; 427 struct dlm_mhandle *mh; 428 char *ppc; 429 430 mh = dlm_midcomms_get_mhandle(node->nodeid, mb_len, &ppc); 431 if (!mh) 432 return -ENOMEM; 433 434 set_bit(DLM_NODE_FLAG_STOP_TX, &node->flags); 435 mh->ack_rcv = ack_rcv; 436 437 m_header = (struct dlm_header *)ppc; 438 439 m_header->h_version = cpu_to_le32(DLM_HEADER_MAJOR | DLM_HEADER_MINOR); 440 m_header->h_nodeid = cpu_to_le32(dlm_our_nodeid()); 441 m_header->h_length = cpu_to_le16(mb_len); 442 m_header->h_cmd = DLM_FIN; 443 444 pr_debug("sending fin msg to node %d\n", node->nodeid); 445 dlm_midcomms_commit_mhandle(mh, NULL, 0); 446 447 return 0; 448 } 449 450 static void dlm_receive_ack(struct midcomms_node *node, uint32_t seq) 451 { 452 struct dlm_mhandle *mh; 453 454 rcu_read_lock(); 455 list_for_each_entry_rcu(mh, &node->send_queue, list) { 456 if (before(mh->seq, seq)) { 457 if (mh->ack_rcv) 458 mh->ack_rcv(node); 459 } else { 460 /* send queue should be ordered */ 461 break; 462 } 463 } 464 465 spin_lock_bh(&node->send_queue_lock); 466 list_for_each_entry_rcu(mh, &node->send_queue, list) { 467 if (before(mh->seq, seq)) { 468 dlm_mhandle_delete(node, mh); 469 } else { 470 /* send queue should be ordered */ 471 break; 472 } 473 } 474 spin_unlock_bh(&node->send_queue_lock); 475 rcu_read_unlock(); 476 } 477 478 static void dlm_pas_fin_ack_rcv(struct midcomms_node *node) 479 { 480 spin_lock_bh(&node->state_lock); 481 pr_debug("receive passive fin ack from node %d with state %s\n", 482 node->nodeid, dlm_state_str(node->state)); 483 484 switch (node->state) { 485 case DLM_LAST_ACK: 486 /* DLM_CLOSED */ 487 midcomms_node_reset(node); 488 break; 489 case DLM_CLOSED: 490 /* not valid but somehow we got what we want */ 491 wake_up(&node->shutdown_wait); 492 break; 493 default: 494 spin_unlock_bh(&node->state_lock); 495 log_print("%s: unexpected state: %d", 496 __func__, node->state); 497 WARN_ON_ONCE(1); 498 return; 499 } 500 spin_unlock_bh(&node->state_lock); 501 } 502 503 static void dlm_receive_buffer_3_2_trace(uint32_t seq, 504 const union dlm_packet *p) 505 { 506 switch (p->header.h_cmd) { 507 case DLM_MSG: 508 trace_dlm_recv_message(dlm_our_nodeid(), seq, &p->message); 509 break; 510 case DLM_RCOM: 511 trace_dlm_recv_rcom(dlm_our_nodeid(), seq, &p->rcom); 512 break; 513 default: 514 break; 515 } 516 } 517 518 static void dlm_midcomms_receive_buffer(const union dlm_packet *p, 519 struct midcomms_node *node, 520 uint32_t seq) 521 { 522 bool is_expected_seq; 523 uint32_t oval, nval; 524 525 do { 526 oval = atomic_read(&node->seq_next); 527 is_expected_seq = (oval == seq); 528 if (!is_expected_seq) 529 break; 530 531 nval = oval + 1; 532 } while (atomic_cmpxchg(&node->seq_next, oval, nval) != oval); 533 534 if (is_expected_seq) { 535 switch (p->header.h_cmd) { 536 case DLM_FIN: 537 spin_lock_bh(&node->state_lock); 538 pr_debug("receive fin msg from node %d with state %s\n", 539 node->nodeid, dlm_state_str(node->state)); 540 541 switch (node->state) { 542 case DLM_ESTABLISHED: 543 dlm_send_ack(node->nodeid, nval); 544 545 /* passive shutdown DLM_LAST_ACK case 1 546 * additional we check if the node is used by 547 * cluster manager events at all. 548 */ 549 if (node->users == 0) { 550 node->state = DLM_LAST_ACK; 551 pr_debug("switch node %d to state %s case 1\n", 552 node->nodeid, dlm_state_str(node->state)); 553 set_bit(DLM_NODE_FLAG_STOP_RX, &node->flags); 554 dlm_send_fin(node, dlm_pas_fin_ack_rcv); 555 } else { 556 node->state = DLM_CLOSE_WAIT; 557 pr_debug("switch node %d to state %s\n", 558 node->nodeid, dlm_state_str(node->state)); 559 } 560 break; 561 case DLM_FIN_WAIT1: 562 dlm_send_ack(node->nodeid, nval); 563 node->state = DLM_CLOSING; 564 set_bit(DLM_NODE_FLAG_STOP_RX, &node->flags); 565 pr_debug("switch node %d to state %s\n", 566 node->nodeid, dlm_state_str(node->state)); 567 break; 568 case DLM_FIN_WAIT2: 569 dlm_send_ack(node->nodeid, nval); 570 midcomms_node_reset(node); 571 pr_debug("switch node %d to state %s\n", 572 node->nodeid, dlm_state_str(node->state)); 573 break; 574 case DLM_LAST_ACK: 575 /* probably remove_member caught it, do nothing */ 576 break; 577 default: 578 spin_unlock_bh(&node->state_lock); 579 log_print("%s: unexpected state: %d", 580 __func__, node->state); 581 WARN_ON_ONCE(1); 582 return; 583 } 584 spin_unlock_bh(&node->state_lock); 585 break; 586 default: 587 WARN_ON_ONCE(test_bit(DLM_NODE_FLAG_STOP_RX, &node->flags)); 588 dlm_receive_buffer_3_2_trace(seq, p); 589 dlm_receive_buffer(p, node->nodeid); 590 atomic_inc(&node->ulp_delivered); 591 /* unlikely case to send ack back when we don't transmit */ 592 dlm_send_ack_threshold(node, DLM_RECV_ACK_BACK_MSG_THRESHOLD); 593 break; 594 } 595 } else { 596 /* retry to ack message which we already have by sending back 597 * current node->seq_next number as ack. 598 */ 599 if (seq < oval) 600 dlm_send_ack(node->nodeid, oval); 601 602 log_print_ratelimited("ignore dlm msg because seq mismatch, seq: %u, expected: %u, nodeid: %d", 603 seq, oval, node->nodeid); 604 } 605 } 606 607 static int dlm_opts_check_msglen(const union dlm_packet *p, uint16_t msglen, 608 int nodeid) 609 { 610 int len = msglen; 611 612 /* we only trust outer header msglen because 613 * it's checked against receive buffer length. 614 */ 615 if (len < sizeof(struct dlm_opts)) 616 return -1; 617 len -= sizeof(struct dlm_opts); 618 619 if (len < le16_to_cpu(p->opts.o_optlen)) 620 return -1; 621 len -= le16_to_cpu(p->opts.o_optlen); 622 623 switch (p->opts.o_nextcmd) { 624 case DLM_FIN: 625 if (len < sizeof(struct dlm_header)) { 626 log_print("fin too small: %d, will skip this message from node %d", 627 len, nodeid); 628 return -1; 629 } 630 631 break; 632 case DLM_MSG: 633 if (len < sizeof(struct dlm_message)) { 634 log_print("msg too small: %d, will skip this message from node %d", 635 msglen, nodeid); 636 return -1; 637 } 638 639 break; 640 case DLM_RCOM: 641 if (len < sizeof(struct dlm_rcom)) { 642 log_print("rcom msg too small: %d, will skip this message from node %d", 643 len, nodeid); 644 return -1; 645 } 646 647 break; 648 default: 649 log_print("unsupported o_nextcmd received: %u, will skip this message from node %d", 650 p->opts.o_nextcmd, nodeid); 651 return -1; 652 } 653 654 return 0; 655 } 656 657 static void dlm_midcomms_receive_buffer_3_2(const union dlm_packet *p, int nodeid) 658 { 659 uint16_t msglen = le16_to_cpu(p->header.h_length); 660 struct midcomms_node *node; 661 uint32_t seq; 662 int ret, idx; 663 664 idx = srcu_read_lock(&nodes_srcu); 665 node = nodeid2node(nodeid); 666 if (WARN_ON_ONCE(!node)) 667 goto out; 668 669 switch (node->version) { 670 case DLM_VERSION_NOT_SET: 671 node->version = DLM_VERSION_3_2; 672 wake_up(&node->shutdown_wait); 673 log_print("version 0x%08x for node %d detected", DLM_VERSION_3_2, 674 node->nodeid); 675 676 spin_lock(&node->state_lock); 677 switch (node->state) { 678 case DLM_CLOSED: 679 node->state = DLM_ESTABLISHED; 680 pr_debug("switch node %d to state %s\n", 681 node->nodeid, dlm_state_str(node->state)); 682 break; 683 default: 684 break; 685 } 686 spin_unlock(&node->state_lock); 687 688 break; 689 case DLM_VERSION_3_2: 690 break; 691 default: 692 log_print_ratelimited("version mismatch detected, assumed 0x%08x but node %d has 0x%08x", 693 DLM_VERSION_3_2, node->nodeid, node->version); 694 goto out; 695 } 696 697 switch (p->header.h_cmd) { 698 case DLM_RCOM: 699 /* these rcom message we use to determine version. 700 * they have their own retransmission handling and 701 * are the first messages of dlm. 702 * 703 * length already checked. 704 */ 705 switch (p->rcom.rc_type) { 706 case cpu_to_le32(DLM_RCOM_NAMES): 707 fallthrough; 708 case cpu_to_le32(DLM_RCOM_NAMES_REPLY): 709 fallthrough; 710 case cpu_to_le32(DLM_RCOM_STATUS): 711 fallthrough; 712 case cpu_to_le32(DLM_RCOM_STATUS_REPLY): 713 break; 714 default: 715 log_print("unsupported rcom type received: %u, will skip this message from node %d", 716 le32_to_cpu(p->rcom.rc_type), nodeid); 717 goto out; 718 } 719 720 WARN_ON_ONCE(test_bit(DLM_NODE_FLAG_STOP_RX, &node->flags)); 721 dlm_receive_buffer(p, nodeid); 722 break; 723 case DLM_OPTS: 724 seq = le32_to_cpu(p->header.u.h_seq); 725 726 ret = dlm_opts_check_msglen(p, msglen, nodeid); 727 if (ret < 0) { 728 log_print("opts msg too small: %u, will skip this message from node %d", 729 msglen, nodeid); 730 goto out; 731 } 732 733 p = (union dlm_packet *)((unsigned char *)p->opts.o_opts + 734 le16_to_cpu(p->opts.o_optlen)); 735 736 /* recheck inner msglen just if it's not garbage */ 737 msglen = le16_to_cpu(p->header.h_length); 738 switch (p->header.h_cmd) { 739 case DLM_RCOM: 740 if (msglen < sizeof(struct dlm_rcom)) { 741 log_print("inner rcom msg too small: %u, will skip this message from node %d", 742 msglen, nodeid); 743 goto out; 744 } 745 746 break; 747 case DLM_MSG: 748 if (msglen < sizeof(struct dlm_message)) { 749 log_print("inner msg too small: %u, will skip this message from node %d", 750 msglen, nodeid); 751 goto out; 752 } 753 754 break; 755 case DLM_FIN: 756 if (msglen < sizeof(struct dlm_header)) { 757 log_print("inner fin too small: %u, will skip this message from node %d", 758 msglen, nodeid); 759 goto out; 760 } 761 762 break; 763 default: 764 log_print("unsupported inner h_cmd received: %u, will skip this message from node %d", 765 msglen, nodeid); 766 goto out; 767 } 768 769 dlm_midcomms_receive_buffer(p, node, seq); 770 break; 771 case DLM_ACK: 772 seq = le32_to_cpu(p->header.u.h_seq); 773 dlm_receive_ack(node, seq); 774 break; 775 default: 776 log_print("unsupported h_cmd received: %u, will skip this message from node %d", 777 p->header.h_cmd, nodeid); 778 break; 779 } 780 781 out: 782 srcu_read_unlock(&nodes_srcu, idx); 783 } 784 785 static void dlm_midcomms_receive_buffer_3_1(const union dlm_packet *p, int nodeid) 786 { 787 uint16_t msglen = le16_to_cpu(p->header.h_length); 788 struct midcomms_node *node; 789 int idx; 790 791 idx = srcu_read_lock(&nodes_srcu); 792 node = nodeid2node(nodeid); 793 if (WARN_ON_ONCE(!node)) { 794 srcu_read_unlock(&nodes_srcu, idx); 795 return; 796 } 797 798 switch (node->version) { 799 case DLM_VERSION_NOT_SET: 800 node->version = DLM_VERSION_3_1; 801 wake_up(&node->shutdown_wait); 802 log_print("version 0x%08x for node %d detected", DLM_VERSION_3_1, 803 node->nodeid); 804 break; 805 case DLM_VERSION_3_1: 806 break; 807 default: 808 log_print_ratelimited("version mismatch detected, assumed 0x%08x but node %d has 0x%08x", 809 DLM_VERSION_3_1, node->nodeid, node->version); 810 srcu_read_unlock(&nodes_srcu, idx); 811 return; 812 } 813 srcu_read_unlock(&nodes_srcu, idx); 814 815 switch (p->header.h_cmd) { 816 case DLM_RCOM: 817 /* length already checked */ 818 break; 819 case DLM_MSG: 820 if (msglen < sizeof(struct dlm_message)) { 821 log_print("msg too small: %u, will skip this message from node %d", 822 msglen, nodeid); 823 return; 824 } 825 826 break; 827 default: 828 log_print("unsupported h_cmd received: %u, will skip this message from node %d", 829 p->header.h_cmd, nodeid); 830 return; 831 } 832 833 dlm_receive_buffer(p, nodeid); 834 } 835 836 int dlm_validate_incoming_buffer(int nodeid, unsigned char *buf, int len) 837 { 838 const unsigned char *ptr = buf; 839 const struct dlm_header *hd; 840 uint16_t msglen; 841 int ret = 0; 842 843 while (len >= sizeof(struct dlm_header)) { 844 hd = (struct dlm_header *)ptr; 845 846 /* no message should be more than DLM_MAX_SOCKET_BUFSIZE or 847 * less than dlm_header size. 848 * 849 * Some messages does not have a 8 byte length boundary yet 850 * which can occur in a unaligned memory access of some dlm 851 * messages. However this problem need to be fixed at the 852 * sending side, for now it seems nobody run into architecture 853 * related issues yet but it slows down some processing. 854 * Fixing this issue should be scheduled in future by doing 855 * the next major version bump. 856 */ 857 msglen = le16_to_cpu(hd->h_length); 858 if (msglen > DLM_MAX_SOCKET_BUFSIZE || 859 msglen < sizeof(struct dlm_header)) { 860 log_print("received invalid length header: %u from node %d, will abort message parsing", 861 msglen, nodeid); 862 return -EBADMSG; 863 } 864 865 /* caller will take care that leftover 866 * will be parsed next call with more data 867 */ 868 if (msglen > len) 869 break; 870 871 ret += msglen; 872 len -= msglen; 873 ptr += msglen; 874 } 875 876 return ret; 877 } 878 879 /* 880 * Called from the low-level comms layer to process a buffer of 881 * commands. 882 */ 883 int dlm_process_incoming_buffer(int nodeid, unsigned char *buf, int len) 884 { 885 const unsigned char *ptr = buf; 886 const struct dlm_header *hd; 887 uint16_t msglen; 888 int ret = 0; 889 890 while (len >= sizeof(struct dlm_header)) { 891 hd = (struct dlm_header *)ptr; 892 893 msglen = le16_to_cpu(hd->h_length); 894 if (msglen > len) 895 break; 896 897 switch (hd->h_version) { 898 case cpu_to_le32(DLM_VERSION_3_1): 899 dlm_midcomms_receive_buffer_3_1((const union dlm_packet *)ptr, nodeid); 900 break; 901 case cpu_to_le32(DLM_VERSION_3_2): 902 dlm_midcomms_receive_buffer_3_2((const union dlm_packet *)ptr, nodeid); 903 break; 904 default: 905 log_print("received invalid version header: %u from node %d, will skip this message", 906 le32_to_cpu(hd->h_version), nodeid); 907 break; 908 } 909 910 ret += msglen; 911 len -= msglen; 912 ptr += msglen; 913 } 914 915 return ret; 916 } 917 918 void dlm_midcomms_unack_msg_resend(int nodeid) 919 { 920 struct midcomms_node *node; 921 struct dlm_mhandle *mh; 922 int idx, ret; 923 924 idx = srcu_read_lock(&nodes_srcu); 925 node = nodeid2node(nodeid); 926 if (WARN_ON_ONCE(!node)) { 927 srcu_read_unlock(&nodes_srcu, idx); 928 return; 929 } 930 931 /* old protocol, we don't support to retransmit on failure */ 932 switch (node->version) { 933 case DLM_VERSION_3_2: 934 break; 935 default: 936 srcu_read_unlock(&nodes_srcu, idx); 937 return; 938 } 939 940 rcu_read_lock(); 941 list_for_each_entry_rcu(mh, &node->send_queue, list) { 942 if (!mh->committed) 943 continue; 944 945 ret = dlm_lowcomms_resend_msg(mh->msg); 946 if (!ret) 947 log_print_ratelimited("retransmit dlm msg, seq %u, nodeid %d", 948 mh->seq, node->nodeid); 949 } 950 rcu_read_unlock(); 951 srcu_read_unlock(&nodes_srcu, idx); 952 } 953 954 static void dlm_fill_opts_header(struct dlm_opts *opts, uint16_t inner_len, 955 uint32_t seq) 956 { 957 opts->o_header.h_cmd = DLM_OPTS; 958 opts->o_header.h_version = cpu_to_le32(DLM_HEADER_MAJOR | DLM_HEADER_MINOR); 959 opts->o_header.h_nodeid = cpu_to_le32(dlm_our_nodeid()); 960 opts->o_header.h_length = cpu_to_le16(DLM_MIDCOMMS_OPT_LEN + inner_len); 961 opts->o_header.u.h_seq = cpu_to_le32(seq); 962 } 963 964 static void midcomms_new_msg_cb(void *data) 965 { 966 struct dlm_mhandle *mh = data; 967 968 atomic_inc(&mh->node->send_queue_cnt); 969 970 spin_lock_bh(&mh->node->send_queue_lock); 971 list_add_tail_rcu(&mh->list, &mh->node->send_queue); 972 spin_unlock_bh(&mh->node->send_queue_lock); 973 974 mh->seq = atomic_fetch_inc(&mh->node->seq_send); 975 } 976 977 static struct dlm_msg *dlm_midcomms_get_msg_3_2(struct dlm_mhandle *mh, int nodeid, 978 int len, char **ppc) 979 { 980 struct dlm_opts *opts; 981 struct dlm_msg *msg; 982 983 msg = dlm_lowcomms_new_msg(nodeid, len + DLM_MIDCOMMS_OPT_LEN, 984 ppc, midcomms_new_msg_cb, mh); 985 if (!msg) 986 return NULL; 987 988 opts = (struct dlm_opts *)*ppc; 989 mh->opts = opts; 990 991 /* add possible options here */ 992 dlm_fill_opts_header(opts, len, mh->seq); 993 994 *ppc += sizeof(*opts); 995 mh->inner_p = (const union dlm_packet *)*ppc; 996 return msg; 997 } 998 999 /* avoid false positive for nodes_srcu, unlock happens in 1000 * dlm_midcomms_commit_mhandle which is a must call if success 1001 */ 1002 #ifndef __CHECKER__ 1003 struct dlm_mhandle *dlm_midcomms_get_mhandle(int nodeid, int len, char **ppc) 1004 { 1005 struct midcomms_node *node; 1006 struct dlm_mhandle *mh; 1007 struct dlm_msg *msg; 1008 int idx; 1009 1010 idx = srcu_read_lock(&nodes_srcu); 1011 node = nodeid2node(nodeid); 1012 if (WARN_ON_ONCE(!node)) 1013 goto err; 1014 1015 /* this is a bug, however we going on and hope it will be resolved */ 1016 WARN_ON_ONCE(test_bit(DLM_NODE_FLAG_STOP_TX, &node->flags)); 1017 1018 mh = dlm_allocate_mhandle(); 1019 if (!mh) 1020 goto err; 1021 1022 mh->committed = false; 1023 mh->ack_rcv = NULL; 1024 mh->idx = idx; 1025 mh->node = node; 1026 1027 switch (node->version) { 1028 case DLM_VERSION_3_1: 1029 msg = dlm_lowcomms_new_msg(nodeid, len, ppc, NULL, NULL); 1030 if (!msg) { 1031 dlm_free_mhandle(mh); 1032 goto err; 1033 } 1034 1035 break; 1036 case DLM_VERSION_3_2: 1037 /* send ack back if necessary */ 1038 dlm_send_ack_threshold(node, DLM_SEND_ACK_BACK_MSG_THRESHOLD); 1039 1040 msg = dlm_midcomms_get_msg_3_2(mh, nodeid, len, ppc); 1041 if (!msg) { 1042 dlm_free_mhandle(mh); 1043 goto err; 1044 } 1045 break; 1046 default: 1047 dlm_free_mhandle(mh); 1048 WARN_ON_ONCE(1); 1049 goto err; 1050 } 1051 1052 mh->msg = msg; 1053 1054 /* keep in mind that is a must to call 1055 * dlm_midcomms_commit_msg() which releases 1056 * nodes_srcu using mh->idx which is assumed 1057 * here that the application will call it. 1058 */ 1059 return mh; 1060 1061 err: 1062 srcu_read_unlock(&nodes_srcu, idx); 1063 return NULL; 1064 } 1065 #endif 1066 1067 static void dlm_midcomms_commit_msg_3_2_trace(const struct dlm_mhandle *mh, 1068 const void *name, int namelen) 1069 { 1070 switch (mh->inner_p->header.h_cmd) { 1071 case DLM_MSG: 1072 trace_dlm_send_message(mh->node->nodeid, mh->seq, 1073 &mh->inner_p->message, 1074 name, namelen); 1075 break; 1076 case DLM_RCOM: 1077 trace_dlm_send_rcom(mh->node->nodeid, mh->seq, 1078 &mh->inner_p->rcom); 1079 break; 1080 default: 1081 /* nothing to trace */ 1082 break; 1083 } 1084 } 1085 1086 static void dlm_midcomms_commit_msg_3_2(struct dlm_mhandle *mh, 1087 const void *name, int namelen) 1088 { 1089 /* nexthdr chain for fast lookup */ 1090 mh->opts->o_nextcmd = mh->inner_p->header.h_cmd; 1091 mh->committed = true; 1092 dlm_midcomms_commit_msg_3_2_trace(mh, name, namelen); 1093 dlm_lowcomms_commit_msg(mh->msg); 1094 } 1095 1096 /* avoid false positive for nodes_srcu, lock was happen in 1097 * dlm_midcomms_get_mhandle 1098 */ 1099 #ifndef __CHECKER__ 1100 void dlm_midcomms_commit_mhandle(struct dlm_mhandle *mh, 1101 const void *name, int namelen) 1102 { 1103 1104 switch (mh->node->version) { 1105 case DLM_VERSION_3_1: 1106 srcu_read_unlock(&nodes_srcu, mh->idx); 1107 1108 dlm_lowcomms_commit_msg(mh->msg); 1109 dlm_lowcomms_put_msg(mh->msg); 1110 /* mh is not part of rcu list in this case */ 1111 dlm_free_mhandle(mh); 1112 break; 1113 case DLM_VERSION_3_2: 1114 /* held rcu read lock here, because we sending the 1115 * dlm message out, when we do that we could receive 1116 * an ack back which releases the mhandle and we 1117 * get a use after free. 1118 */ 1119 rcu_read_lock(); 1120 dlm_midcomms_commit_msg_3_2(mh, name, namelen); 1121 srcu_read_unlock(&nodes_srcu, mh->idx); 1122 rcu_read_unlock(); 1123 break; 1124 default: 1125 srcu_read_unlock(&nodes_srcu, mh->idx); 1126 WARN_ON_ONCE(1); 1127 break; 1128 } 1129 } 1130 #endif 1131 1132 int dlm_midcomms_start(void) 1133 { 1134 return dlm_lowcomms_start(); 1135 } 1136 1137 void dlm_midcomms_stop(void) 1138 { 1139 dlm_lowcomms_stop(); 1140 } 1141 1142 void dlm_midcomms_init(void) 1143 { 1144 int i; 1145 1146 for (i = 0; i < CONN_HASH_SIZE; i++) 1147 INIT_HLIST_HEAD(&node_hash[i]); 1148 1149 dlm_lowcomms_init(); 1150 } 1151 1152 static void midcomms_node_release(struct rcu_head *rcu) 1153 { 1154 struct midcomms_node *node = container_of(rcu, struct midcomms_node, rcu); 1155 1156 WARN_ON_ONCE(atomic_read(&node->send_queue_cnt)); 1157 dlm_send_queue_flush(node); 1158 kfree(node); 1159 } 1160 1161 void dlm_midcomms_exit(void) 1162 { 1163 struct midcomms_node *node; 1164 int i, idx; 1165 1166 idx = srcu_read_lock(&nodes_srcu); 1167 for (i = 0; i < CONN_HASH_SIZE; i++) { 1168 hlist_for_each_entry_rcu(node, &node_hash[i], hlist) { 1169 dlm_delete_debug_comms_file(node->debugfs); 1170 1171 spin_lock(&nodes_lock); 1172 hlist_del_rcu(&node->hlist); 1173 spin_unlock(&nodes_lock); 1174 1175 call_srcu(&nodes_srcu, &node->rcu, midcomms_node_release); 1176 } 1177 } 1178 srcu_read_unlock(&nodes_srcu, idx); 1179 1180 dlm_lowcomms_exit(); 1181 } 1182 1183 static void dlm_act_fin_ack_rcv(struct midcomms_node *node) 1184 { 1185 spin_lock_bh(&node->state_lock); 1186 pr_debug("receive active fin ack from node %d with state %s\n", 1187 node->nodeid, dlm_state_str(node->state)); 1188 1189 switch (node->state) { 1190 case DLM_FIN_WAIT1: 1191 node->state = DLM_FIN_WAIT2; 1192 pr_debug("switch node %d to state %s\n", 1193 node->nodeid, dlm_state_str(node->state)); 1194 break; 1195 case DLM_CLOSING: 1196 midcomms_node_reset(node); 1197 pr_debug("switch node %d to state %s\n", 1198 node->nodeid, dlm_state_str(node->state)); 1199 break; 1200 case DLM_CLOSED: 1201 /* not valid but somehow we got what we want */ 1202 wake_up(&node->shutdown_wait); 1203 break; 1204 default: 1205 spin_unlock_bh(&node->state_lock); 1206 log_print("%s: unexpected state: %d", 1207 __func__, node->state); 1208 WARN_ON_ONCE(1); 1209 return; 1210 } 1211 spin_unlock_bh(&node->state_lock); 1212 } 1213 1214 void dlm_midcomms_add_member(int nodeid) 1215 { 1216 struct midcomms_node *node; 1217 int idx; 1218 1219 idx = srcu_read_lock(&nodes_srcu); 1220 node = nodeid2node(nodeid); 1221 if (WARN_ON_ONCE(!node)) { 1222 srcu_read_unlock(&nodes_srcu, idx); 1223 return; 1224 } 1225 1226 spin_lock_bh(&node->state_lock); 1227 if (!node->users) { 1228 pr_debug("receive add member from node %d with state %s\n", 1229 node->nodeid, dlm_state_str(node->state)); 1230 switch (node->state) { 1231 case DLM_ESTABLISHED: 1232 break; 1233 case DLM_CLOSED: 1234 node->state = DLM_ESTABLISHED; 1235 pr_debug("switch node %d to state %s\n", 1236 node->nodeid, dlm_state_str(node->state)); 1237 break; 1238 default: 1239 /* some invalid state passive shutdown 1240 * was failed, we try to reset and 1241 * hope it will go on. 1242 */ 1243 log_print("reset node %d because shutdown stuck", 1244 node->nodeid); 1245 1246 midcomms_node_reset(node); 1247 node->state = DLM_ESTABLISHED; 1248 break; 1249 } 1250 } 1251 1252 node->users++; 1253 pr_debug("node %d users inc count %d\n", nodeid, node->users); 1254 spin_unlock_bh(&node->state_lock); 1255 1256 srcu_read_unlock(&nodes_srcu, idx); 1257 } 1258 1259 void dlm_midcomms_remove_member(int nodeid) 1260 { 1261 struct midcomms_node *node; 1262 int idx; 1263 1264 idx = srcu_read_lock(&nodes_srcu); 1265 node = nodeid2node(nodeid); 1266 /* in case of dlm_midcomms_close() removes node */ 1267 if (!node) { 1268 srcu_read_unlock(&nodes_srcu, idx); 1269 return; 1270 } 1271 1272 spin_lock_bh(&node->state_lock); 1273 /* case of dlm_midcomms_addr() created node but 1274 * was not added before because dlm_midcomms_close() 1275 * removed the node 1276 */ 1277 if (!node->users) { 1278 spin_unlock_bh(&node->state_lock); 1279 srcu_read_unlock(&nodes_srcu, idx); 1280 return; 1281 } 1282 1283 node->users--; 1284 pr_debug("node %d users dec count %d\n", nodeid, node->users); 1285 1286 /* hitting users count to zero means the 1287 * other side is running dlm_midcomms_stop() 1288 * we meet us to have a clean disconnect. 1289 */ 1290 if (node->users == 0) { 1291 pr_debug("receive remove member from node %d with state %s\n", 1292 node->nodeid, dlm_state_str(node->state)); 1293 switch (node->state) { 1294 case DLM_ESTABLISHED: 1295 break; 1296 case DLM_CLOSE_WAIT: 1297 /* passive shutdown DLM_LAST_ACK case 2 */ 1298 node->state = DLM_LAST_ACK; 1299 pr_debug("switch node %d to state %s case 2\n", 1300 node->nodeid, dlm_state_str(node->state)); 1301 set_bit(DLM_NODE_FLAG_STOP_RX, &node->flags); 1302 dlm_send_fin(node, dlm_pas_fin_ack_rcv); 1303 break; 1304 case DLM_LAST_ACK: 1305 /* probably receive fin caught it, do nothing */ 1306 break; 1307 case DLM_CLOSED: 1308 /* already gone, do nothing */ 1309 break; 1310 default: 1311 log_print("%s: unexpected state: %d", 1312 __func__, node->state); 1313 break; 1314 } 1315 } 1316 spin_unlock_bh(&node->state_lock); 1317 1318 srcu_read_unlock(&nodes_srcu, idx); 1319 } 1320 1321 void dlm_midcomms_version_wait(void) 1322 { 1323 struct midcomms_node *node; 1324 int i, idx, ret; 1325 1326 idx = srcu_read_lock(&nodes_srcu); 1327 for (i = 0; i < CONN_HASH_SIZE; i++) { 1328 hlist_for_each_entry_rcu(node, &node_hash[i], hlist) { 1329 ret = wait_event_timeout(node->shutdown_wait, 1330 node->version != DLM_VERSION_NOT_SET || 1331 node->state == DLM_CLOSED || 1332 test_bit(DLM_NODE_FLAG_CLOSE, &node->flags), 1333 DLM_SHUTDOWN_TIMEOUT); 1334 if (!ret || test_bit(DLM_NODE_FLAG_CLOSE, &node->flags)) 1335 pr_debug("version wait timed out for node %d with state %s\n", 1336 node->nodeid, dlm_state_str(node->state)); 1337 } 1338 } 1339 srcu_read_unlock(&nodes_srcu, idx); 1340 } 1341 1342 static void midcomms_shutdown(struct midcomms_node *node) 1343 { 1344 int ret; 1345 1346 /* old protocol, we don't wait for pending operations */ 1347 switch (node->version) { 1348 case DLM_VERSION_3_2: 1349 break; 1350 default: 1351 return; 1352 } 1353 1354 spin_lock_bh(&node->state_lock); 1355 pr_debug("receive active shutdown for node %d with state %s\n", 1356 node->nodeid, dlm_state_str(node->state)); 1357 switch (node->state) { 1358 case DLM_ESTABLISHED: 1359 node->state = DLM_FIN_WAIT1; 1360 pr_debug("switch node %d to state %s case 2\n", 1361 node->nodeid, dlm_state_str(node->state)); 1362 dlm_send_fin(node, dlm_act_fin_ack_rcv); 1363 break; 1364 case DLM_CLOSED: 1365 /* we have what we want */ 1366 break; 1367 default: 1368 /* busy to enter DLM_FIN_WAIT1, wait until passive 1369 * done in shutdown_wait to enter DLM_CLOSED. 1370 */ 1371 break; 1372 } 1373 spin_unlock_bh(&node->state_lock); 1374 1375 if (DLM_DEBUG_FENCE_TERMINATION) 1376 msleep(5000); 1377 1378 /* wait for other side dlm + fin */ 1379 ret = wait_event_timeout(node->shutdown_wait, 1380 node->state == DLM_CLOSED || 1381 test_bit(DLM_NODE_FLAG_CLOSE, &node->flags), 1382 DLM_SHUTDOWN_TIMEOUT); 1383 if (!ret) 1384 pr_debug("active shutdown timed out for node %d with state %s\n", 1385 node->nodeid, dlm_state_str(node->state)); 1386 else 1387 pr_debug("active shutdown done for node %d with state %s\n", 1388 node->nodeid, dlm_state_str(node->state)); 1389 } 1390 1391 void dlm_midcomms_shutdown(void) 1392 { 1393 struct midcomms_node *node; 1394 int i, idx; 1395 1396 mutex_lock(&close_lock); 1397 idx = srcu_read_lock(&nodes_srcu); 1398 for (i = 0; i < CONN_HASH_SIZE; i++) { 1399 hlist_for_each_entry_rcu(node, &node_hash[i], hlist) { 1400 midcomms_shutdown(node); 1401 } 1402 } 1403 1404 dlm_lowcomms_shutdown(); 1405 1406 for (i = 0; i < CONN_HASH_SIZE; i++) { 1407 hlist_for_each_entry_rcu(node, &node_hash[i], hlist) { 1408 midcomms_node_reset(node); 1409 } 1410 } 1411 srcu_read_unlock(&nodes_srcu, idx); 1412 mutex_unlock(&close_lock); 1413 } 1414 1415 int dlm_midcomms_close(int nodeid) 1416 { 1417 struct midcomms_node *node; 1418 int idx, ret; 1419 1420 idx = srcu_read_lock(&nodes_srcu); 1421 /* Abort pending close/remove operation */ 1422 node = nodeid2node(nodeid); 1423 if (node) { 1424 /* let shutdown waiters leave */ 1425 set_bit(DLM_NODE_FLAG_CLOSE, &node->flags); 1426 wake_up(&node->shutdown_wait); 1427 } 1428 srcu_read_unlock(&nodes_srcu, idx); 1429 1430 synchronize_srcu(&nodes_srcu); 1431 1432 mutex_lock(&close_lock); 1433 idx = srcu_read_lock(&nodes_srcu); 1434 node = nodeid2node(nodeid); 1435 if (!node) { 1436 srcu_read_unlock(&nodes_srcu, idx); 1437 mutex_unlock(&close_lock); 1438 return dlm_lowcomms_close(nodeid); 1439 } 1440 1441 ret = dlm_lowcomms_close(nodeid); 1442 dlm_delete_debug_comms_file(node->debugfs); 1443 1444 spin_lock_bh(&nodes_lock); 1445 hlist_del_rcu(&node->hlist); 1446 spin_unlock_bh(&nodes_lock); 1447 srcu_read_unlock(&nodes_srcu, idx); 1448 1449 /* wait that all readers left until flush send queue */ 1450 synchronize_srcu(&nodes_srcu); 1451 1452 /* drop all pending dlm messages, this is fine as 1453 * this function get called when the node is fenced 1454 */ 1455 dlm_send_queue_flush(node); 1456 1457 call_srcu(&nodes_srcu, &node->rcu, midcomms_node_release); 1458 mutex_unlock(&close_lock); 1459 1460 return ret; 1461 } 1462 1463 /* debug functionality to send raw dlm msg from user space */ 1464 struct dlm_rawmsg_data { 1465 struct midcomms_node *node; 1466 void *buf; 1467 }; 1468 1469 static void midcomms_new_rawmsg_cb(void *data) 1470 { 1471 struct dlm_rawmsg_data *rd = data; 1472 struct dlm_header *h = rd->buf; 1473 1474 switch (h->h_version) { 1475 case cpu_to_le32(DLM_VERSION_3_1): 1476 break; 1477 default: 1478 switch (h->h_cmd) { 1479 case DLM_OPTS: 1480 if (!h->u.h_seq) 1481 h->u.h_seq = cpu_to_le32(atomic_fetch_inc(&rd->node->seq_send)); 1482 break; 1483 default: 1484 break; 1485 } 1486 break; 1487 } 1488 } 1489 1490 int dlm_midcomms_rawmsg_send(struct midcomms_node *node, void *buf, 1491 int buflen) 1492 { 1493 struct dlm_rawmsg_data rd; 1494 struct dlm_msg *msg; 1495 char *msgbuf; 1496 1497 rd.node = node; 1498 rd.buf = buf; 1499 1500 msg = dlm_lowcomms_new_msg(node->nodeid, buflen, &msgbuf, 1501 midcomms_new_rawmsg_cb, &rd); 1502 if (!msg) 1503 return -ENOMEM; 1504 1505 memcpy(msgbuf, buf, buflen); 1506 dlm_lowcomms_commit_msg(msg); 1507 return 0; 1508 } 1509 1510
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.