~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/fs/hfsplus/bnode.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0
  2 /*
  3  *  linux/fs/hfsplus/bnode.c
  4  *
  5  * Copyright (C) 2001
  6  * Brad Boyer (flar@allandria.com)
  7  * (C) 2003 Ardis Technologies <roman@ardistech.com>
  8  *
  9  * Handle basic btree node operations
 10  */
 11 
 12 #include <linux/string.h>
 13 #include <linux/slab.h>
 14 #include <linux/pagemap.h>
 15 #include <linux/fs.h>
 16 #include <linux/swap.h>
 17 
 18 #include "hfsplus_fs.h"
 19 #include "hfsplus_raw.h"
 20 
 21 /* Copy a specified range of bytes from the raw data of a node */
 22 void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len)
 23 {
 24         struct page **pagep;
 25         int l;
 26 
 27         off += node->page_offset;
 28         pagep = node->page + (off >> PAGE_SHIFT);
 29         off &= ~PAGE_MASK;
 30 
 31         l = min_t(int, len, PAGE_SIZE - off);
 32         memcpy_from_page(buf, *pagep, off, l);
 33 
 34         while ((len -= l) != 0) {
 35                 buf += l;
 36                 l = min_t(int, len, PAGE_SIZE);
 37                 memcpy_from_page(buf, *++pagep, 0, l);
 38         }
 39 }
 40 
 41 u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off)
 42 {
 43         __be16 data;
 44         /* TODO: optimize later... */
 45         hfs_bnode_read(node, &data, off, 2);
 46         return be16_to_cpu(data);
 47 }
 48 
 49 u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off)
 50 {
 51         u8 data;
 52         /* TODO: optimize later... */
 53         hfs_bnode_read(node, &data, off, 1);
 54         return data;
 55 }
 56 
 57 void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off)
 58 {
 59         struct hfs_btree *tree;
 60         int key_len;
 61 
 62         tree = node->tree;
 63         if (node->type == HFS_NODE_LEAF ||
 64             tree->attributes & HFS_TREE_VARIDXKEYS ||
 65             node->tree->cnid == HFSPLUS_ATTR_CNID)
 66                 key_len = hfs_bnode_read_u16(node, off) + 2;
 67         else
 68                 key_len = tree->max_key_len + 2;
 69 
 70         hfs_bnode_read(node, key, off, key_len);
 71 }
 72 
 73 void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len)
 74 {
 75         struct page **pagep;
 76         int l;
 77 
 78         off += node->page_offset;
 79         pagep = node->page + (off >> PAGE_SHIFT);
 80         off &= ~PAGE_MASK;
 81 
 82         l = min_t(int, len, PAGE_SIZE - off);
 83         memcpy_to_page(*pagep, off, buf, l);
 84         set_page_dirty(*pagep);
 85 
 86         while ((len -= l) != 0) {
 87                 buf += l;
 88                 l = min_t(int, len, PAGE_SIZE);
 89                 memcpy_to_page(*++pagep, 0, buf, l);
 90                 set_page_dirty(*pagep);
 91         }
 92 }
 93 
 94 void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data)
 95 {
 96         __be16 v = cpu_to_be16(data);
 97         /* TODO: optimize later... */
 98         hfs_bnode_write(node, &v, off, 2);
 99 }
100 
101 void hfs_bnode_clear(struct hfs_bnode *node, int off, int len)
102 {
103         struct page **pagep;
104         int l;
105 
106         off += node->page_offset;
107         pagep = node->page + (off >> PAGE_SHIFT);
108         off &= ~PAGE_MASK;
109 
110         l = min_t(int, len, PAGE_SIZE - off);
111         memzero_page(*pagep, off, l);
112         set_page_dirty(*pagep);
113 
114         while ((len -= l) != 0) {
115                 l = min_t(int, len, PAGE_SIZE);
116                 memzero_page(*++pagep, 0, l);
117                 set_page_dirty(*pagep);
118         }
119 }
120 
121 void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
122                     struct hfs_bnode *src_node, int src, int len)
123 {
124         struct page **src_page, **dst_page;
125         int l;
126 
127         hfs_dbg(BNODE_MOD, "copybytes: %u,%u,%u\n", dst, src, len);
128         if (!len)
129                 return;
130         src += src_node->page_offset;
131         dst += dst_node->page_offset;
132         src_page = src_node->page + (src >> PAGE_SHIFT);
133         src &= ~PAGE_MASK;
134         dst_page = dst_node->page + (dst >> PAGE_SHIFT);
135         dst &= ~PAGE_MASK;
136 
137         if (src == dst) {
138                 l = min_t(int, len, PAGE_SIZE - src);
139                 memcpy_page(*dst_page, src, *src_page, src, l);
140                 set_page_dirty(*dst_page);
141 
142                 while ((len -= l) != 0) {
143                         l = min_t(int, len, PAGE_SIZE);
144                         memcpy_page(*++dst_page, 0, *++src_page, 0, l);
145                         set_page_dirty(*dst_page);
146                 }
147         } else {
148                 void *src_ptr, *dst_ptr;
149 
150                 do {
151                         dst_ptr = kmap_local_page(*dst_page) + dst;
152                         src_ptr = kmap_local_page(*src_page) + src;
153                         if (PAGE_SIZE - src < PAGE_SIZE - dst) {
154                                 l = PAGE_SIZE - src;
155                                 src = 0;
156                                 dst += l;
157                         } else {
158                                 l = PAGE_SIZE - dst;
159                                 src += l;
160                                 dst = 0;
161                         }
162                         l = min(len, l);
163                         memcpy(dst_ptr, src_ptr, l);
164                         kunmap_local(src_ptr);
165                         set_page_dirty(*dst_page);
166                         kunmap_local(dst_ptr);
167                         if (!dst)
168                                 dst_page++;
169                         else
170                                 src_page++;
171                 } while ((len -= l));
172         }
173 }
174 
175 void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
176 {
177         struct page **src_page, **dst_page;
178         void *src_ptr, *dst_ptr;
179         int l;
180 
181         hfs_dbg(BNODE_MOD, "movebytes: %u,%u,%u\n", dst, src, len);
182         if (!len)
183                 return;
184         src += node->page_offset;
185         dst += node->page_offset;
186         if (dst > src) {
187                 src += len - 1;
188                 src_page = node->page + (src >> PAGE_SHIFT);
189                 src = (src & ~PAGE_MASK) + 1;
190                 dst += len - 1;
191                 dst_page = node->page + (dst >> PAGE_SHIFT);
192                 dst = (dst & ~PAGE_MASK) + 1;
193 
194                 if (src == dst) {
195                         while (src < len) {
196                                 dst_ptr = kmap_local_page(*dst_page);
197                                 src_ptr = kmap_local_page(*src_page);
198                                 memmove(dst_ptr, src_ptr, src);
199                                 kunmap_local(src_ptr);
200                                 set_page_dirty(*dst_page);
201                                 kunmap_local(dst_ptr);
202                                 len -= src;
203                                 src = PAGE_SIZE;
204                                 src_page--;
205                                 dst_page--;
206                         }
207                         src -= len;
208                         dst_ptr = kmap_local_page(*dst_page);
209                         src_ptr = kmap_local_page(*src_page);
210                         memmove(dst_ptr + src, src_ptr + src, len);
211                         kunmap_local(src_ptr);
212                         set_page_dirty(*dst_page);
213                         kunmap_local(dst_ptr);
214                 } else {
215                         do {
216                                 dst_ptr = kmap_local_page(*dst_page) + dst;
217                                 src_ptr = kmap_local_page(*src_page) + src;
218                                 if (src < dst) {
219                                         l = src;
220                                         src = PAGE_SIZE;
221                                         dst -= l;
222                                 } else {
223                                         l = dst;
224                                         src -= l;
225                                         dst = PAGE_SIZE;
226                                 }
227                                 l = min(len, l);
228                                 memmove(dst_ptr - l, src_ptr - l, l);
229                                 kunmap_local(src_ptr);
230                                 set_page_dirty(*dst_page);
231                                 kunmap_local(dst_ptr);
232                                 if (dst == PAGE_SIZE)
233                                         dst_page--;
234                                 else
235                                         src_page--;
236                         } while ((len -= l));
237                 }
238         } else {
239                 src_page = node->page + (src >> PAGE_SHIFT);
240                 src &= ~PAGE_MASK;
241                 dst_page = node->page + (dst >> PAGE_SHIFT);
242                 dst &= ~PAGE_MASK;
243 
244                 if (src == dst) {
245                         l = min_t(int, len, PAGE_SIZE - src);
246 
247                         dst_ptr = kmap_local_page(*dst_page) + src;
248                         src_ptr = kmap_local_page(*src_page) + src;
249                         memmove(dst_ptr, src_ptr, l);
250                         kunmap_local(src_ptr);
251                         set_page_dirty(*dst_page);
252                         kunmap_local(dst_ptr);
253 
254                         while ((len -= l) != 0) {
255                                 l = min_t(int, len, PAGE_SIZE);
256                                 dst_ptr = kmap_local_page(*++dst_page);
257                                 src_ptr = kmap_local_page(*++src_page);
258                                 memmove(dst_ptr, src_ptr, l);
259                                 kunmap_local(src_ptr);
260                                 set_page_dirty(*dst_page);
261                                 kunmap_local(dst_ptr);
262                         }
263                 } else {
264                         do {
265                                 dst_ptr = kmap_local_page(*dst_page) + dst;
266                                 src_ptr = kmap_local_page(*src_page) + src;
267                                 if (PAGE_SIZE - src <
268                                                 PAGE_SIZE - dst) {
269                                         l = PAGE_SIZE - src;
270                                         src = 0;
271                                         dst += l;
272                                 } else {
273                                         l = PAGE_SIZE - dst;
274                                         src += l;
275                                         dst = 0;
276                                 }
277                                 l = min(len, l);
278                                 memmove(dst_ptr, src_ptr, l);
279                                 kunmap_local(src_ptr);
280                                 set_page_dirty(*dst_page);
281                                 kunmap_local(dst_ptr);
282                                 if (!dst)
283                                         dst_page++;
284                                 else
285                                         src_page++;
286                         } while ((len -= l));
287                 }
288         }
289 }
290 
291 void hfs_bnode_dump(struct hfs_bnode *node)
292 {
293         struct hfs_bnode_desc desc;
294         __be32 cnid;
295         int i, off, key_off;
296 
297         hfs_dbg(BNODE_MOD, "bnode: %d\n", node->this);
298         hfs_bnode_read(node, &desc, 0, sizeof(desc));
299         hfs_dbg(BNODE_MOD, "%d, %d, %d, %d, %d\n",
300                 be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
301                 desc.type, desc.height, be16_to_cpu(desc.num_recs));
302 
303         off = node->tree->node_size - 2;
304         for (i = be16_to_cpu(desc.num_recs); i >= 0; off -= 2, i--) {
305                 key_off = hfs_bnode_read_u16(node, off);
306                 hfs_dbg(BNODE_MOD, " %d", key_off);
307                 if (i && node->type == HFS_NODE_INDEX) {
308                         int tmp;
309 
310                         if (node->tree->attributes & HFS_TREE_VARIDXKEYS ||
311                                         node->tree->cnid == HFSPLUS_ATTR_CNID)
312                                 tmp = hfs_bnode_read_u16(node, key_off) + 2;
313                         else
314                                 tmp = node->tree->max_key_len + 2;
315                         hfs_dbg_cont(BNODE_MOD, " (%d", tmp);
316                         hfs_bnode_read(node, &cnid, key_off + tmp, 4);
317                         hfs_dbg_cont(BNODE_MOD, ",%d)", be32_to_cpu(cnid));
318                 } else if (i && node->type == HFS_NODE_LEAF) {
319                         int tmp;
320 
321                         tmp = hfs_bnode_read_u16(node, key_off);
322                         hfs_dbg_cont(BNODE_MOD, " (%d)", tmp);
323                 }
324         }
325         hfs_dbg_cont(BNODE_MOD, "\n");
326 }
327 
328 void hfs_bnode_unlink(struct hfs_bnode *node)
329 {
330         struct hfs_btree *tree;
331         struct hfs_bnode *tmp;
332         __be32 cnid;
333 
334         tree = node->tree;
335         if (node->prev) {
336                 tmp = hfs_bnode_find(tree, node->prev);
337                 if (IS_ERR(tmp))
338                         return;
339                 tmp->next = node->next;
340                 cnid = cpu_to_be32(tmp->next);
341                 hfs_bnode_write(tmp, &cnid,
342                         offsetof(struct hfs_bnode_desc, next), 4);
343                 hfs_bnode_put(tmp);
344         } else if (node->type == HFS_NODE_LEAF)
345                 tree->leaf_head = node->next;
346 
347         if (node->next) {
348                 tmp = hfs_bnode_find(tree, node->next);
349                 if (IS_ERR(tmp))
350                         return;
351                 tmp->prev = node->prev;
352                 cnid = cpu_to_be32(tmp->prev);
353                 hfs_bnode_write(tmp, &cnid,
354                         offsetof(struct hfs_bnode_desc, prev), 4);
355                 hfs_bnode_put(tmp);
356         } else if (node->type == HFS_NODE_LEAF)
357                 tree->leaf_tail = node->prev;
358 
359         /* move down? */
360         if (!node->prev && !node->next)
361                 hfs_dbg(BNODE_MOD, "hfs_btree_del_level\n");
362         if (!node->parent) {
363                 tree->root = 0;
364                 tree->depth = 0;
365         }
366         set_bit(HFS_BNODE_DELETED, &node->flags);
367 }
368 
369 static inline int hfs_bnode_hash(u32 num)
370 {
371         num = (num >> 16) + num;
372         num += num >> 8;
373         return num & (NODE_HASH_SIZE - 1);
374 }
375 
376 struct hfs_bnode *hfs_bnode_findhash(struct hfs_btree *tree, u32 cnid)
377 {
378         struct hfs_bnode *node;
379 
380         if (cnid >= tree->node_count) {
381                 pr_err("request for non-existent node %d in B*Tree\n",
382                        cnid);
383                 return NULL;
384         }
385 
386         for (node = tree->node_hash[hfs_bnode_hash(cnid)];
387                         node; node = node->next_hash)
388                 if (node->this == cnid)
389                         return node;
390         return NULL;
391 }
392 
393 static struct hfs_bnode *__hfs_bnode_create(struct hfs_btree *tree, u32 cnid)
394 {
395         struct hfs_bnode *node, *node2;
396         struct address_space *mapping;
397         struct page *page;
398         int size, block, i, hash;
399         loff_t off;
400 
401         if (cnid >= tree->node_count) {
402                 pr_err("request for non-existent node %d in B*Tree\n",
403                        cnid);
404                 return NULL;
405         }
406 
407         size = sizeof(struct hfs_bnode) + tree->pages_per_bnode *
408                 sizeof(struct page *);
409         node = kzalloc(size, GFP_KERNEL);
410         if (!node)
411                 return NULL;
412         node->tree = tree;
413         node->this = cnid;
414         set_bit(HFS_BNODE_NEW, &node->flags);
415         atomic_set(&node->refcnt, 1);
416         hfs_dbg(BNODE_REFS, "new_node(%d:%d): 1\n",
417                 node->tree->cnid, node->this);
418         init_waitqueue_head(&node->lock_wq);
419         spin_lock(&tree->hash_lock);
420         node2 = hfs_bnode_findhash(tree, cnid);
421         if (!node2) {
422                 hash = hfs_bnode_hash(cnid);
423                 node->next_hash = tree->node_hash[hash];
424                 tree->node_hash[hash] = node;
425                 tree->node_hash_cnt++;
426         } else {
427                 spin_unlock(&tree->hash_lock);
428                 kfree(node);
429                 wait_event(node2->lock_wq,
430                         !test_bit(HFS_BNODE_NEW, &node2->flags));
431                 return node2;
432         }
433         spin_unlock(&tree->hash_lock);
434 
435         mapping = tree->inode->i_mapping;
436         off = (loff_t)cnid << tree->node_size_shift;
437         block = off >> PAGE_SHIFT;
438         node->page_offset = off & ~PAGE_MASK;
439         for (i = 0; i < tree->pages_per_bnode; block++, i++) {
440                 page = read_mapping_page(mapping, block, NULL);
441                 if (IS_ERR(page))
442                         goto fail;
443                 node->page[i] = page;
444         }
445 
446         return node;
447 fail:
448         set_bit(HFS_BNODE_ERROR, &node->flags);
449         return node;
450 }
451 
452 void hfs_bnode_unhash(struct hfs_bnode *node)
453 {
454         struct hfs_bnode **p;
455 
456         hfs_dbg(BNODE_REFS, "remove_node(%d:%d): %d\n",
457                 node->tree->cnid, node->this, atomic_read(&node->refcnt));
458         for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
459              *p && *p != node; p = &(*p)->next_hash)
460                 ;
461         BUG_ON(!*p);
462         *p = node->next_hash;
463         node->tree->node_hash_cnt--;
464 }
465 
466 /* Load a particular node out of a tree */
467 struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
468 {
469         struct hfs_bnode *node;
470         struct hfs_bnode_desc *desc;
471         int i, rec_off, off, next_off;
472         int entry_size, key_size;
473 
474         spin_lock(&tree->hash_lock);
475         node = hfs_bnode_findhash(tree, num);
476         if (node) {
477                 hfs_bnode_get(node);
478                 spin_unlock(&tree->hash_lock);
479                 wait_event(node->lock_wq,
480                         !test_bit(HFS_BNODE_NEW, &node->flags));
481                 if (test_bit(HFS_BNODE_ERROR, &node->flags))
482                         goto node_error;
483                 return node;
484         }
485         spin_unlock(&tree->hash_lock);
486         node = __hfs_bnode_create(tree, num);
487         if (!node)
488                 return ERR_PTR(-ENOMEM);
489         if (test_bit(HFS_BNODE_ERROR, &node->flags))
490                 goto node_error;
491         if (!test_bit(HFS_BNODE_NEW, &node->flags))
492                 return node;
493 
494         desc = (struct hfs_bnode_desc *)(kmap_local_page(node->page[0]) +
495                                                          node->page_offset);
496         node->prev = be32_to_cpu(desc->prev);
497         node->next = be32_to_cpu(desc->next);
498         node->num_recs = be16_to_cpu(desc->num_recs);
499         node->type = desc->type;
500         node->height = desc->height;
501         kunmap_local(desc);
502 
503         switch (node->type) {
504         case HFS_NODE_HEADER:
505         case HFS_NODE_MAP:
506                 if (node->height != 0)
507                         goto node_error;
508                 break;
509         case HFS_NODE_LEAF:
510                 if (node->height != 1)
511                         goto node_error;
512                 break;
513         case HFS_NODE_INDEX:
514                 if (node->height <= 1 || node->height > tree->depth)
515                         goto node_error;
516                 break;
517         default:
518                 goto node_error;
519         }
520 
521         rec_off = tree->node_size - 2;
522         off = hfs_bnode_read_u16(node, rec_off);
523         if (off != sizeof(struct hfs_bnode_desc))
524                 goto node_error;
525         for (i = 1; i <= node->num_recs; off = next_off, i++) {
526                 rec_off -= 2;
527                 next_off = hfs_bnode_read_u16(node, rec_off);
528                 if (next_off <= off ||
529                     next_off > tree->node_size ||
530                     next_off & 1)
531                         goto node_error;
532                 entry_size = next_off - off;
533                 if (node->type != HFS_NODE_INDEX &&
534                     node->type != HFS_NODE_LEAF)
535                         continue;
536                 key_size = hfs_bnode_read_u16(node, off) + 2;
537                 if (key_size >= entry_size || key_size & 1)
538                         goto node_error;
539         }
540         clear_bit(HFS_BNODE_NEW, &node->flags);
541         wake_up(&node->lock_wq);
542         return node;
543 
544 node_error:
545         set_bit(HFS_BNODE_ERROR, &node->flags);
546         clear_bit(HFS_BNODE_NEW, &node->flags);
547         wake_up(&node->lock_wq);
548         hfs_bnode_put(node);
549         return ERR_PTR(-EIO);
550 }
551 
552 void hfs_bnode_free(struct hfs_bnode *node)
553 {
554         int i;
555 
556         for (i = 0; i < node->tree->pages_per_bnode; i++)
557                 if (node->page[i])
558                         put_page(node->page[i]);
559         kfree(node);
560 }
561 
562 struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num)
563 {
564         struct hfs_bnode *node;
565         struct page **pagep;
566         int i;
567 
568         spin_lock(&tree->hash_lock);
569         node = hfs_bnode_findhash(tree, num);
570         spin_unlock(&tree->hash_lock);
571         if (node) {
572                 pr_crit("new node %u already hashed?\n", num);
573                 WARN_ON(1);
574                 return node;
575         }
576         node = __hfs_bnode_create(tree, num);
577         if (!node)
578                 return ERR_PTR(-ENOMEM);
579         if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
580                 hfs_bnode_put(node);
581                 return ERR_PTR(-EIO);
582         }
583 
584         pagep = node->page;
585         memzero_page(*pagep, node->page_offset,
586                      min_t(int, PAGE_SIZE, tree->node_size));
587         set_page_dirty(*pagep);
588         for (i = 1; i < tree->pages_per_bnode; i++) {
589                 memzero_page(*++pagep, 0, PAGE_SIZE);
590                 set_page_dirty(*pagep);
591         }
592         clear_bit(HFS_BNODE_NEW, &node->flags);
593         wake_up(&node->lock_wq);
594 
595         return node;
596 }
597 
598 void hfs_bnode_get(struct hfs_bnode *node)
599 {
600         if (node) {
601                 atomic_inc(&node->refcnt);
602                 hfs_dbg(BNODE_REFS, "get_node(%d:%d): %d\n",
603                         node->tree->cnid, node->this,
604                         atomic_read(&node->refcnt));
605         }
606 }
607 
608 /* Dispose of resources used by a node */
609 void hfs_bnode_put(struct hfs_bnode *node)
610 {
611         if (node) {
612                 struct hfs_btree *tree = node->tree;
613                 int i;
614 
615                 hfs_dbg(BNODE_REFS, "put_node(%d:%d): %d\n",
616                         node->tree->cnid, node->this,
617                         atomic_read(&node->refcnt));
618                 BUG_ON(!atomic_read(&node->refcnt));
619                 if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
620                         return;
621                 for (i = 0; i < tree->pages_per_bnode; i++) {
622                         if (!node->page[i])
623                                 continue;
624                         mark_page_accessed(node->page[i]);
625                 }
626 
627                 if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
628                         hfs_bnode_unhash(node);
629                         spin_unlock(&tree->hash_lock);
630                         if (hfs_bnode_need_zeroout(tree))
631                                 hfs_bnode_clear(node, 0, tree->node_size);
632                         hfs_bmap_free(node);
633                         hfs_bnode_free(node);
634                         return;
635                 }
636                 spin_unlock(&tree->hash_lock);
637         }
638 }
639 
640 /*
641  * Unused nodes have to be zeroed if this is the catalog tree and
642  * a corresponding flag in the volume header is set.
643  */
644 bool hfs_bnode_need_zeroout(struct hfs_btree *tree)
645 {
646         struct super_block *sb = tree->inode->i_sb;
647         struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
648         const u32 volume_attr = be32_to_cpu(sbi->s_vhdr->attributes);
649 
650         return tree->cnid == HFSPLUS_CAT_CNID &&
651                 volume_attr & HFSPLUS_VOL_UNUSED_NODE_FIX;
652 }
653 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php