11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * JFFS2 -- Journalling Flash File System, Version 2. 31da177e4SLinus Torvalds * 4c00c310eSDavid Woodhouse * Copyright © 2001-2007 Red Hat, Inc. 51da177e4SLinus Torvalds * 61da177e4SLinus Torvalds * Created by David Woodhouse <dwmw2@infradead.org> 71da177e4SLinus Torvalds * 81da177e4SLinus Torvalds * For licensing information, see the file 'LICENCE' in this directory. 91da177e4SLinus Torvalds * 101da177e4SLinus Torvalds */ 111da177e4SLinus Torvalds 121da177e4SLinus Torvalds #include <linux/kernel.h> 131da177e4SLinus Torvalds #include <linux/sched.h> 141da177e4SLinus Torvalds #include <linux/fs.h> 151da177e4SLinus Torvalds #include <linux/mtd/mtd.h> 161da177e4SLinus Torvalds #include <linux/rbtree.h> 171da177e4SLinus Torvalds #include <linux/crc32.h> 181da177e4SLinus Torvalds #include <linux/slab.h> 191da177e4SLinus Torvalds #include <linux/pagemap.h> 201da177e4SLinus Torvalds #include "nodelist.h" 211da177e4SLinus Torvalds 2290a18fabSAdrian Bunk static void jffs2_obsolete_node_frag(struct jffs2_sb_info *c, 2390a18fabSAdrian Bunk struct jffs2_node_frag *this); 2490a18fabSAdrian Bunk 251da177e4SLinus Torvalds void jffs2_add_fd_to_list(struct jffs2_sb_info *c, struct jffs2_full_dirent *new, struct jffs2_full_dirent **list) 261da177e4SLinus Torvalds { 271da177e4SLinus Torvalds struct jffs2_full_dirent **prev = list; 28e0d60137SArtem B. Bityutskiy 29733802d9SArtem B. Bityutskiy dbg_dentlist("add dirent \"%s\", ino #%u\n", new->name, new->ino); 301da177e4SLinus Torvalds 311da177e4SLinus Torvalds while ((*prev) && (*prev)->nhash <= new->nhash) { 321da177e4SLinus Torvalds if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { 331da177e4SLinus Torvalds /* Duplicate. Free one */ 341da177e4SLinus Torvalds if (new->version < (*prev)->version) { 35733802d9SArtem B. Bityutskiy dbg_dentlist("Eep! Marking new dirent node is obsolete, old is \"%s\", ino #%u\n", 36e0d60137SArtem B. Bityutskiy (*prev)->name, (*prev)->ino); 371da177e4SLinus Torvalds jffs2_mark_node_obsolete(c, new->raw); 381da177e4SLinus Torvalds jffs2_free_full_dirent(new); 391da177e4SLinus Torvalds } else { 40733802d9SArtem B. Bityutskiy dbg_dentlist("marking old dirent \"%s\", ino #%u bsolete\n", 41e0d60137SArtem B. Bityutskiy (*prev)->name, (*prev)->ino); 421da177e4SLinus Torvalds new->next = (*prev)->next; 431da177e4SLinus Torvalds jffs2_mark_node_obsolete(c, ((*prev)->raw)); 441da177e4SLinus Torvalds jffs2_free_full_dirent(*prev); 451da177e4SLinus Torvalds *prev = new; 461da177e4SLinus Torvalds } 47e0d60137SArtem B. Bityutskiy return; 481da177e4SLinus Torvalds } 491da177e4SLinus Torvalds prev = &((*prev)->next); 501da177e4SLinus Torvalds } 511da177e4SLinus Torvalds new->next = *prev; 521da177e4SLinus Torvalds *prev = new; 531da177e4SLinus Torvalds } 541da177e4SLinus Torvalds 551e900979SArtem B. Bityutskiy void jffs2_truncate_fragtree(struct jffs2_sb_info *c, struct rb_root *list, uint32_t size) 561e900979SArtem B. Bityutskiy { 571e900979SArtem B. Bityutskiy struct jffs2_node_frag *frag = jffs2_lookup_node_frag(list, size); 581e900979SArtem B. Bityutskiy 59733802d9SArtem B. Bityutskiy dbg_fragtree("truncating fragtree to 0x%08x bytes\n", size); 601e900979SArtem B. Bityutskiy 611e900979SArtem B. Bityutskiy /* We know frag->ofs <= size. That's what lookup does for us */ 621e900979SArtem B. Bityutskiy if (frag && frag->ofs != size) { 631e0da3cbSArtem B. Bityutskiy if (frag->ofs+frag->size > size) { 641e900979SArtem B. Bityutskiy frag->size = size - frag->ofs; 651e900979SArtem B. Bityutskiy } 661e900979SArtem B. Bityutskiy frag = frag_next(frag); 671e900979SArtem B. Bityutskiy } 681e900979SArtem B. Bityutskiy while (frag && frag->ofs >= size) { 691e900979SArtem B. Bityutskiy struct jffs2_node_frag *next = frag_next(frag); 701e900979SArtem B. Bityutskiy 711e900979SArtem B. Bityutskiy frag_erase(frag, list); 721e900979SArtem B. Bityutskiy jffs2_obsolete_node_frag(c, frag); 731e900979SArtem B. Bityutskiy frag = next; 741e900979SArtem B. Bityutskiy } 751e0da3cbSArtem B. Bityutskiy 761e0da3cbSArtem B. Bityutskiy if (size == 0) 771e0da3cbSArtem B. Bityutskiy return; 781e0da3cbSArtem B. Bityutskiy 791e0da3cbSArtem B. Bityutskiy /* 801e0da3cbSArtem B. Bityutskiy * If the last fragment starts at the RAM page boundary, it is 811e0da3cbSArtem B. Bityutskiy * REF_PRISTINE irrespective of its size. 821e0da3cbSArtem B. Bityutskiy */ 831e0da3cbSArtem B. Bityutskiy frag = frag_last(list); 84f0507530SArtem B. Bityutskiy if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) { 85733802d9SArtem B. Bityutskiy dbg_fragtree2("marking the last fragment 0x%08x-0x%08x REF_PRISTINE.\n", 861e0da3cbSArtem B. Bityutskiy frag->ofs, frag->ofs + frag->size); 871e0da3cbSArtem B. Bityutskiy frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE; 881e0da3cbSArtem B. Bityutskiy } 891e900979SArtem B. Bityutskiy } 901e900979SArtem B. Bityutskiy 9190a18fabSAdrian Bunk static void jffs2_obsolete_node_frag(struct jffs2_sb_info *c, 9290a18fabSAdrian Bunk struct jffs2_node_frag *this) 931da177e4SLinus Torvalds { 94f97117d1SArtem B. Bityutskiy if (this->node) { 95f97117d1SArtem B. Bityutskiy this->node->frags--; 96f97117d1SArtem B. Bityutskiy if (!this->node->frags) { 97f97117d1SArtem B. Bityutskiy /* The node has no valid frags left. It's totally obsoleted */ 98733802d9SArtem B. Bityutskiy dbg_fragtree2("marking old node @0x%08x (0x%04x-0x%04x) obsolete\n", 99e0d60137SArtem B. Bityutskiy ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size); 100f97117d1SArtem B. Bityutskiy jffs2_mark_node_obsolete(c, this->node->raw); 101f97117d1SArtem B. Bityutskiy jffs2_free_full_dnode(this->node); 102f97117d1SArtem B. Bityutskiy } else { 103733802d9SArtem B. Bityutskiy dbg_fragtree2("marking old node @0x%08x (0x%04x-0x%04x) REF_NORMAL. frags is %d\n", 104e0d60137SArtem B. Bityutskiy ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); 105f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 1061da177e4SLinus Torvalds } 1071da177e4SLinus Torvalds 108f97117d1SArtem B. Bityutskiy } 109f97117d1SArtem B. Bityutskiy jffs2_free_node_frag(this); 1109dee7503SDavid Woodhouse } 1119dee7503SDavid Woodhouse 112f97117d1SArtem B. Bityutskiy static void jffs2_fragtree_insert(struct jffs2_node_frag *newfrag, struct jffs2_node_frag *base) 1131da177e4SLinus Torvalds { 114f97117d1SArtem B. Bityutskiy struct rb_node *parent = &base->rb; 115f97117d1SArtem B. Bityutskiy struct rb_node **link = &parent; 1161da177e4SLinus Torvalds 117733802d9SArtem B. Bityutskiy dbg_fragtree2("insert frag (0x%04x-0x%04x)\n", newfrag->ofs, newfrag->ofs + newfrag->size); 1189dee7503SDavid Woodhouse 119f97117d1SArtem B. Bityutskiy while (*link) { 120f97117d1SArtem B. Bityutskiy parent = *link; 121f97117d1SArtem B. Bityutskiy base = rb_entry(parent, struct jffs2_node_frag, rb); 122f97117d1SArtem B. Bityutskiy 123f97117d1SArtem B. Bityutskiy if (newfrag->ofs > base->ofs) 124f97117d1SArtem B. Bityutskiy link = &base->rb.rb_right; 125f97117d1SArtem B. Bityutskiy else if (newfrag->ofs < base->ofs) 126f97117d1SArtem B. Bityutskiy link = &base->rb.rb_left; 1279dee7503SDavid Woodhouse else { 128e0d60137SArtem B. Bityutskiy JFFS2_ERROR("duplicate frag at %08x (%p,%p)\n", newfrag->ofs, newfrag, base); 129dae6227fSArtem B. Bityutskiy BUG(); 130dae6227fSArtem B. Bityutskiy } 131dae6227fSArtem B. Bityutskiy } 132dae6227fSArtem B. Bityutskiy 133f97117d1SArtem B. Bityutskiy rb_link_node(&newfrag->rb, &base->rb, link); 134dae6227fSArtem B. Bityutskiy } 135dae6227fSArtem B. Bityutskiy 1361e0da3cbSArtem B. Bityutskiy /* 1371e0da3cbSArtem B. Bityutskiy * Allocate and initializes a new fragment. 1381e0da3cbSArtem B. Bityutskiy */ 139858119e1SArjan van de Ven static struct jffs2_node_frag * new_fragment(struct jffs2_full_dnode *fn, uint32_t ofs, uint32_t size) 1401e0da3cbSArtem B. Bityutskiy { 1411e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag; 1421e0da3cbSArtem B. Bityutskiy 1431e0da3cbSArtem B. Bityutskiy newfrag = jffs2_alloc_node_frag(); 1441e0da3cbSArtem B. Bityutskiy if (likely(newfrag)) { 1451e0da3cbSArtem B. Bityutskiy newfrag->ofs = ofs; 1461e0da3cbSArtem B. Bityutskiy newfrag->size = size; 1471e0da3cbSArtem B. Bityutskiy newfrag->node = fn; 1481e0da3cbSArtem B. Bityutskiy } else { 1491e0da3cbSArtem B. Bityutskiy JFFS2_ERROR("cannot allocate a jffs2_node_frag object\n"); 1501e0da3cbSArtem B. Bityutskiy } 1511e0da3cbSArtem B. Bityutskiy 1521e0da3cbSArtem B. Bityutskiy return newfrag; 1531e0da3cbSArtem B. Bityutskiy } 1541e0da3cbSArtem B. Bityutskiy 1551e0da3cbSArtem B. Bityutskiy /* 1561e0da3cbSArtem B. Bityutskiy * Called when there is no overlapping fragment exist. Inserts a hole before the new 1571e0da3cbSArtem B. Bityutskiy * fragment and inserts the new fragment to the fragtree. 1581e0da3cbSArtem B. Bityutskiy */ 1591e0da3cbSArtem B. Bityutskiy static int no_overlapping_node(struct jffs2_sb_info *c, struct rb_root *root, 1601e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag, 1611e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *this, uint32_t lastend) 1621e0da3cbSArtem B. Bityutskiy { 1631e0da3cbSArtem B. Bityutskiy if (lastend < newfrag->node->ofs) { 1641e0da3cbSArtem B. Bityutskiy /* put a hole in before the new fragment */ 1651e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *holefrag; 1661e0da3cbSArtem B. Bityutskiy 1671e0da3cbSArtem B. Bityutskiy holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend); 1681e0da3cbSArtem B. Bityutskiy if (unlikely(!holefrag)) { 1691e0da3cbSArtem B. Bityutskiy jffs2_free_node_frag(newfrag); 1701e0da3cbSArtem B. Bityutskiy return -ENOMEM; 1711e0da3cbSArtem B. Bityutskiy } 1721e0da3cbSArtem B. Bityutskiy 1731e0da3cbSArtem B. Bityutskiy if (this) { 1741e0da3cbSArtem B. Bityutskiy /* By definition, the 'this' node has no right-hand child, 1751e0da3cbSArtem B. Bityutskiy because there are no frags with offset greater than it. 1761e0da3cbSArtem B. Bityutskiy So that's where we want to put the hole */ 177733802d9SArtem B. Bityutskiy dbg_fragtree2("add hole frag %#04x-%#04x on the right of the new frag.\n", 1781e0da3cbSArtem B. Bityutskiy holefrag->ofs, holefrag->ofs + holefrag->size); 1791e0da3cbSArtem B. Bityutskiy rb_link_node(&holefrag->rb, &this->rb, &this->rb.rb_right); 1801e0da3cbSArtem B. Bityutskiy } else { 181733802d9SArtem B. Bityutskiy dbg_fragtree2("Add hole frag %#04x-%#04x to the root of the tree.\n", 1821e0da3cbSArtem B. Bityutskiy holefrag->ofs, holefrag->ofs + holefrag->size); 1831e0da3cbSArtem B. Bityutskiy rb_link_node(&holefrag->rb, NULL, &root->rb_node); 1841e0da3cbSArtem B. Bityutskiy } 1851e0da3cbSArtem B. Bityutskiy rb_insert_color(&holefrag->rb, root); 1861e0da3cbSArtem B. Bityutskiy this = holefrag; 1871e0da3cbSArtem B. Bityutskiy } 1881e0da3cbSArtem B. Bityutskiy 1891e0da3cbSArtem B. Bityutskiy if (this) { 1901e0da3cbSArtem B. Bityutskiy /* By definition, the 'this' node has no right-hand child, 1911e0da3cbSArtem B. Bityutskiy because there are no frags with offset greater than it. 1921e0da3cbSArtem B. Bityutskiy So that's where we want to put new fragment */ 193733802d9SArtem B. Bityutskiy dbg_fragtree2("add the new node at the right\n"); 1941e0da3cbSArtem B. Bityutskiy rb_link_node(&newfrag->rb, &this->rb, &this->rb.rb_right); 1951e0da3cbSArtem B. Bityutskiy } else { 196733802d9SArtem B. Bityutskiy dbg_fragtree2("insert the new node at the root of the tree\n"); 1971e0da3cbSArtem B. Bityutskiy rb_link_node(&newfrag->rb, NULL, &root->rb_node); 1981e0da3cbSArtem B. Bityutskiy } 1991e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 2001e0da3cbSArtem B. Bityutskiy 2011e0da3cbSArtem B. Bityutskiy return 0; 2021e0da3cbSArtem B. Bityutskiy } 2031e0da3cbSArtem B. Bityutskiy 204f97117d1SArtem B. Bityutskiy /* Doesn't set inode->i_size */ 2051e0da3cbSArtem B. Bityutskiy static int jffs2_add_frag_to_fragtree(struct jffs2_sb_info *c, struct rb_root *root, struct jffs2_node_frag *newfrag) 2061da177e4SLinus Torvalds { 207f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *this; 208f97117d1SArtem B. Bityutskiy uint32_t lastend; 2091da177e4SLinus Torvalds 210f97117d1SArtem B. Bityutskiy /* Skip all the nodes which are completed before this one starts */ 2111e0da3cbSArtem B. Bityutskiy this = jffs2_lookup_node_frag(root, newfrag->node->ofs); 2121da177e4SLinus Torvalds 213f97117d1SArtem B. Bityutskiy if (this) { 214733802d9SArtem B. Bityutskiy dbg_fragtree2("lookup gave frag 0x%04x-0x%04x; phys 0x%08x (*%p)\n", 215e0d60137SArtem B. Bityutskiy this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this); 216f97117d1SArtem B. Bityutskiy lastend = this->ofs + this->size; 217f97117d1SArtem B. Bityutskiy } else { 218733802d9SArtem B. Bityutskiy dbg_fragtree2("lookup gave no frag\n"); 219f97117d1SArtem B. Bityutskiy lastend = 0; 220f97117d1SArtem B. Bityutskiy } 2211da177e4SLinus Torvalds 2221e0da3cbSArtem B. Bityutskiy /* See if we ran off the end of the fragtree */ 223f97117d1SArtem B. Bityutskiy if (lastend <= newfrag->ofs) { 224f97117d1SArtem B. Bityutskiy /* We did */ 2251da177e4SLinus Torvalds 226f97117d1SArtem B. Bityutskiy /* Check if 'this' node was on the same page as the new node. 227f97117d1SArtem B. Bityutskiy If so, both 'this' and the new node get marked REF_NORMAL so 228f97117d1SArtem B. Bityutskiy the GC can take a look. 2291da177e4SLinus Torvalds */ 230f97117d1SArtem B. Bityutskiy if (lastend && (lastend-1) >> PAGE_CACHE_SHIFT == newfrag->ofs >> PAGE_CACHE_SHIFT) { 231f97117d1SArtem B. Bityutskiy if (this->node) 232f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 233f97117d1SArtem B. Bityutskiy mark_ref_normal(newfrag->node->raw); 2341da177e4SLinus Torvalds } 2351da177e4SLinus Torvalds 2361e0da3cbSArtem B. Bityutskiy return no_overlapping_node(c, root, newfrag, this, lastend); 2371da177e4SLinus Torvalds } 238dae6227fSArtem B. Bityutskiy 2391e0da3cbSArtem B. Bityutskiy if (this->node) 240733802d9SArtem B. Bityutskiy dbg_fragtree2("dealing with frag %u-%u, phys %#08x(%d).\n", 2411e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size, 2421e0da3cbSArtem B. Bityutskiy ref_offset(this->node->raw), ref_flags(this->node->raw)); 2431e0da3cbSArtem B. Bityutskiy else 244733802d9SArtem B. Bityutskiy dbg_fragtree2("dealing with hole frag %u-%u.\n", 2451e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size); 246dae6227fSArtem B. Bityutskiy 247f97117d1SArtem B. Bityutskiy /* OK. 'this' is pointing at the first frag that newfrag->ofs at least partially obsoletes, 248f97117d1SArtem B. Bityutskiy * - i.e. newfrag->ofs < this->ofs+this->size && newfrag->ofs >= this->ofs 249f97117d1SArtem B. Bityutskiy */ 250f97117d1SArtem B. Bityutskiy if (newfrag->ofs > this->ofs) { 251f97117d1SArtem B. Bityutskiy /* This node isn't completely obsoleted. The start of it remains valid */ 2521da177e4SLinus Torvalds 253f97117d1SArtem B. Bityutskiy /* Mark the new node and the partially covered node REF_NORMAL -- let 254f97117d1SArtem B. Bityutskiy the GC take a look at them */ 255f97117d1SArtem B. Bityutskiy mark_ref_normal(newfrag->node->raw); 256f97117d1SArtem B. Bityutskiy if (this->node) 257f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 2581da177e4SLinus Torvalds 259f97117d1SArtem B. Bityutskiy if (this->ofs + this->size > newfrag->ofs + newfrag->size) { 260f97117d1SArtem B. Bityutskiy /* The new node splits 'this' frag into two */ 2611e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag2; 2621e0da3cbSArtem B. Bityutskiy 263f97117d1SArtem B. Bityutskiy if (this->node) 264733802d9SArtem B. Bityutskiy dbg_fragtree2("split old frag 0x%04x-0x%04x, phys 0x%08x\n", 265e0d60137SArtem B. Bityutskiy this->ofs, this->ofs+this->size, ref_offset(this->node->raw)); 266f97117d1SArtem B. Bityutskiy else 267733802d9SArtem B. Bityutskiy dbg_fragtree2("split old hole frag 0x%04x-0x%04x\n", 2688d5df409SArtem B. Bityutskiy this->ofs, this->ofs+this->size); 269dae6227fSArtem B. Bityutskiy 270f97117d1SArtem B. Bityutskiy /* New second frag pointing to this's node */ 2711e0da3cbSArtem B. Bityutskiy newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size, 2721e0da3cbSArtem B. Bityutskiy this->ofs + this->size - newfrag->ofs - newfrag->size); 2731e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag2)) 2741e0da3cbSArtem B. Bityutskiy return -ENOMEM; 275f97117d1SArtem B. Bityutskiy if (this->node) 276f97117d1SArtem B. Bityutskiy this->node->frags++; 277dae6227fSArtem B. Bityutskiy 278f97117d1SArtem B. Bityutskiy /* Adjust size of original 'this' */ 279f97117d1SArtem B. Bityutskiy this->size = newfrag->ofs - this->ofs; 2801da177e4SLinus Torvalds 281f97117d1SArtem B. Bityutskiy /* Now, we know there's no node with offset 282f97117d1SArtem B. Bityutskiy greater than this->ofs but smaller than 283f97117d1SArtem B. Bityutskiy newfrag2->ofs or newfrag->ofs, for obvious 284f97117d1SArtem B. Bityutskiy reasons. So we can do a tree insert from 285f97117d1SArtem B. Bityutskiy 'this' to insert newfrag, and a tree insert 286f97117d1SArtem B. Bityutskiy from newfrag to insert newfrag2. */ 287f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(newfrag, this); 2881e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 2891da177e4SLinus Torvalds 290f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(newfrag2, newfrag); 2911e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag2->rb, root); 2921da177e4SLinus Torvalds 2931da177e4SLinus Torvalds return 0; 2941da177e4SLinus Torvalds } 295f97117d1SArtem B. Bityutskiy /* New node just reduces 'this' frag in size, doesn't split it */ 296f97117d1SArtem B. Bityutskiy this->size = newfrag->ofs - this->ofs; 297f97117d1SArtem B. Bityutskiy 298f97117d1SArtem B. Bityutskiy /* Again, we know it lives down here in the tree */ 299f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(newfrag, this); 3001e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 301f97117d1SArtem B. Bityutskiy } else { 302f97117d1SArtem B. Bityutskiy /* New frag starts at the same point as 'this' used to. Replace 303f97117d1SArtem B. Bityutskiy it in the tree without doing a delete and insertion */ 304733802d9SArtem B. Bityutskiy dbg_fragtree2("inserting newfrag (*%p),%d-%d in before 'this' (*%p),%d-%d\n", 305e0d60137SArtem B. Bityutskiy newfrag, newfrag->ofs, newfrag->ofs+newfrag->size, this, this->ofs, this->ofs+this->size); 306f97117d1SArtem B. Bityutskiy 3071e0da3cbSArtem B. Bityutskiy rb_replace_node(&this->rb, &newfrag->rb, root); 308f97117d1SArtem B. Bityutskiy 309f97117d1SArtem B. Bityutskiy if (newfrag->ofs + newfrag->size >= this->ofs+this->size) { 310733802d9SArtem B. Bityutskiy dbg_fragtree2("obsoleting node frag %p (%x-%x)\n", this, this->ofs, this->ofs+this->size); 311f97117d1SArtem B. Bityutskiy jffs2_obsolete_node_frag(c, this); 312f97117d1SArtem B. Bityutskiy } else { 313f97117d1SArtem B. Bityutskiy this->ofs += newfrag->size; 314f97117d1SArtem B. Bityutskiy this->size -= newfrag->size; 315f97117d1SArtem B. Bityutskiy 316f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(this, newfrag); 3171e0da3cbSArtem B. Bityutskiy rb_insert_color(&this->rb, root); 318f97117d1SArtem B. Bityutskiy return 0; 319f97117d1SArtem B. Bityutskiy } 320f97117d1SArtem B. Bityutskiy } 321f97117d1SArtem B. Bityutskiy /* OK, now we have newfrag added in the correct place in the tree, but 322f97117d1SArtem B. Bityutskiy frag_next(newfrag) may be a fragment which is overlapped by it 323f97117d1SArtem B. Bityutskiy */ 324f97117d1SArtem B. Bityutskiy while ((this = frag_next(newfrag)) && newfrag->ofs + newfrag->size >= this->ofs + this->size) { 325f97117d1SArtem B. Bityutskiy /* 'this' frag is obsoleted completely. */ 326733802d9SArtem B. Bityutskiy dbg_fragtree2("obsoleting node frag %p (%x-%x) and removing from tree\n", 327e0d60137SArtem B. Bityutskiy this, this->ofs, this->ofs+this->size); 3281e0da3cbSArtem B. Bityutskiy rb_erase(&this->rb, root); 329f97117d1SArtem B. Bityutskiy jffs2_obsolete_node_frag(c, this); 330f97117d1SArtem B. Bityutskiy } 331f97117d1SArtem B. Bityutskiy /* Now we're pointing at the first frag which isn't totally obsoleted by 332f97117d1SArtem B. Bityutskiy the new frag */ 333f97117d1SArtem B. Bityutskiy 3341e0da3cbSArtem B. Bityutskiy if (!this || newfrag->ofs + newfrag->size == this->ofs) 335f97117d1SArtem B. Bityutskiy return 0; 3361e0da3cbSArtem B. Bityutskiy 337f97117d1SArtem B. Bityutskiy /* Still some overlap but we don't need to move it in the tree */ 338f97117d1SArtem B. Bityutskiy this->size = (this->ofs + this->size) - (newfrag->ofs + newfrag->size); 339f97117d1SArtem B. Bityutskiy this->ofs = newfrag->ofs + newfrag->size; 340f97117d1SArtem B. Bityutskiy 341f97117d1SArtem B. Bityutskiy /* And mark them REF_NORMAL so the GC takes a look at them */ 342f97117d1SArtem B. Bityutskiy if (this->node) 343f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 344f97117d1SArtem B. Bityutskiy mark_ref_normal(newfrag->node->raw); 345f97117d1SArtem B. Bityutskiy 346f97117d1SArtem B. Bityutskiy return 0; 347f97117d1SArtem B. Bityutskiy } 348f97117d1SArtem B. Bityutskiy 3491e0da3cbSArtem B. Bityutskiy /* 3501e0da3cbSArtem B. Bityutskiy * Given an inode, probably with existing tree of fragments, add the new node 3511e0da3cbSArtem B. Bityutskiy * to the fragment tree. 352f97117d1SArtem B. Bityutskiy */ 353f97117d1SArtem B. Bityutskiy int jffs2_add_full_dnode_to_inode(struct jffs2_sb_info *c, struct jffs2_inode_info *f, struct jffs2_full_dnode *fn) 354f97117d1SArtem B. Bityutskiy { 355f97117d1SArtem B. Bityutskiy int ret; 356f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *newfrag; 357f97117d1SArtem B. Bityutskiy 358f97117d1SArtem B. Bityutskiy if (unlikely(!fn->size)) 359f97117d1SArtem B. Bityutskiy return 0; 360f97117d1SArtem B. Bityutskiy 3611e0da3cbSArtem B. Bityutskiy newfrag = new_fragment(fn, fn->ofs, fn->size); 362f97117d1SArtem B. Bityutskiy if (unlikely(!newfrag)) 363f97117d1SArtem B. Bityutskiy return -ENOMEM; 3641e0da3cbSArtem B. Bityutskiy newfrag->node->frags = 1; 365f97117d1SArtem B. Bityutskiy 366733802d9SArtem B. Bityutskiy dbg_fragtree("adding node %#04x-%#04x @0x%08x on flash, newfrag *%p\n", 367e0d60137SArtem B. Bityutskiy fn->ofs, fn->ofs+fn->size, ref_offset(fn->raw), newfrag); 368f97117d1SArtem B. Bityutskiy 369f97117d1SArtem B. Bityutskiy ret = jffs2_add_frag_to_fragtree(c, &f->fragtree, newfrag); 370f97117d1SArtem B. Bityutskiy if (unlikely(ret)) 371f97117d1SArtem B. Bityutskiy return ret; 372f97117d1SArtem B. Bityutskiy 373f97117d1SArtem B. Bityutskiy /* If we now share a page with other nodes, mark either previous 374f97117d1SArtem B. Bityutskiy or next node REF_NORMAL, as appropriate. */ 375f97117d1SArtem B. Bityutskiy if (newfrag->ofs & (PAGE_CACHE_SIZE-1)) { 376f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *prev = frag_prev(newfrag); 377f97117d1SArtem B. Bityutskiy 378f97117d1SArtem B. Bityutskiy mark_ref_normal(fn->raw); 379f97117d1SArtem B. Bityutskiy /* If we don't start at zero there's _always_ a previous */ 380f97117d1SArtem B. Bityutskiy if (prev->node) 381f97117d1SArtem B. Bityutskiy mark_ref_normal(prev->node->raw); 382f97117d1SArtem B. Bityutskiy } 383f97117d1SArtem B. Bityutskiy 384f97117d1SArtem B. Bityutskiy if ((newfrag->ofs+newfrag->size) & (PAGE_CACHE_SIZE-1)) { 385f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *next = frag_next(newfrag); 386f97117d1SArtem B. Bityutskiy 387f97117d1SArtem B. Bityutskiy if (next) { 388f97117d1SArtem B. Bityutskiy mark_ref_normal(fn->raw); 389f97117d1SArtem B. Bityutskiy if (next->node) 390f97117d1SArtem B. Bityutskiy mark_ref_normal(next->node->raw); 391f97117d1SArtem B. Bityutskiy } 392f97117d1SArtem B. Bityutskiy } 393f97117d1SArtem B. Bityutskiy jffs2_dbg_fragtree_paranoia_check_nolock(f); 3941e0da3cbSArtem B. Bityutskiy 395f97117d1SArtem B. Bityutskiy return 0; 396f97117d1SArtem B. Bityutskiy } 397f97117d1SArtem B. Bityutskiy 3981da177e4SLinus Torvalds void jffs2_set_inocache_state(struct jffs2_sb_info *c, struct jffs2_inode_cache *ic, int state) 3991da177e4SLinus Torvalds { 4001da177e4SLinus Torvalds spin_lock(&c->inocache_lock); 4011da177e4SLinus Torvalds ic->state = state; 4021da177e4SLinus Torvalds wake_up(&c->inocache_wq); 4031da177e4SLinus Torvalds spin_unlock(&c->inocache_lock); 4041da177e4SLinus Torvalds } 4051da177e4SLinus Torvalds 4061da177e4SLinus Torvalds /* During mount, this needs no locking. During normal operation, its 4071da177e4SLinus Torvalds callers want to do other stuff while still holding the inocache_lock. 4081da177e4SLinus Torvalds Rather than introducing special case get_ino_cache functions or 4091da177e4SLinus Torvalds callbacks, we just let the caller do the locking itself. */ 4101da177e4SLinus Torvalds 4111da177e4SLinus Torvalds struct jffs2_inode_cache *jffs2_get_ino_cache(struct jffs2_sb_info *c, uint32_t ino) 4121da177e4SLinus Torvalds { 4131da177e4SLinus Torvalds struct jffs2_inode_cache *ret; 4141da177e4SLinus Torvalds 4151da177e4SLinus Torvalds ret = c->inocache_list[ino % INOCACHE_HASHSIZE]; 4161da177e4SLinus Torvalds while (ret && ret->ino < ino) { 4171da177e4SLinus Torvalds ret = ret->next; 4181da177e4SLinus Torvalds } 4191da177e4SLinus Torvalds 4201da177e4SLinus Torvalds if (ret && ret->ino != ino) 4211da177e4SLinus Torvalds ret = NULL; 4221da177e4SLinus Torvalds 4231da177e4SLinus Torvalds return ret; 4241da177e4SLinus Torvalds } 4251da177e4SLinus Torvalds 4261da177e4SLinus Torvalds void jffs2_add_ino_cache (struct jffs2_sb_info *c, struct jffs2_inode_cache *new) 4271da177e4SLinus Torvalds { 4281da177e4SLinus Torvalds struct jffs2_inode_cache **prev; 4297d27c814SThomas Gleixner 4301da177e4SLinus Torvalds spin_lock(&c->inocache_lock); 4317d200960SDavid Woodhouse if (!new->ino) 4327d200960SDavid Woodhouse new->ino = ++c->highest_ino; 4337d200960SDavid Woodhouse 434733802d9SArtem B. Bityutskiy dbg_inocache("add %p (ino #%u)\n", new, new->ino); 4351da177e4SLinus Torvalds 4361da177e4SLinus Torvalds prev = &c->inocache_list[new->ino % INOCACHE_HASHSIZE]; 4371da177e4SLinus Torvalds 4381da177e4SLinus Torvalds while ((*prev) && (*prev)->ino < new->ino) { 4391da177e4SLinus Torvalds prev = &(*prev)->next; 4401da177e4SLinus Torvalds } 4411da177e4SLinus Torvalds new->next = *prev; 4421da177e4SLinus Torvalds *prev = new; 4431da177e4SLinus Torvalds 4441da177e4SLinus Torvalds spin_unlock(&c->inocache_lock); 4451da177e4SLinus Torvalds } 4461da177e4SLinus Torvalds 4471da177e4SLinus Torvalds void jffs2_del_ino_cache(struct jffs2_sb_info *c, struct jffs2_inode_cache *old) 4481da177e4SLinus Torvalds { 4491da177e4SLinus Torvalds struct jffs2_inode_cache **prev; 450e0d60137SArtem B. Bityutskiy 451355ed4e1SKaiGai Kohei #ifdef CONFIG_JFFS2_FS_XATTR 452355ed4e1SKaiGai Kohei BUG_ON(old->xref); 453355ed4e1SKaiGai Kohei #endif 454733802d9SArtem B. Bityutskiy dbg_inocache("del %p (ino #%u)\n", old, old->ino); 4551da177e4SLinus Torvalds spin_lock(&c->inocache_lock); 4561da177e4SLinus Torvalds 4571da177e4SLinus Torvalds prev = &c->inocache_list[old->ino % INOCACHE_HASHSIZE]; 4581da177e4SLinus Torvalds 4591da177e4SLinus Torvalds while ((*prev) && (*prev)->ino < old->ino) { 4601da177e4SLinus Torvalds prev = &(*prev)->next; 4611da177e4SLinus Torvalds } 4621da177e4SLinus Torvalds if ((*prev) == old) { 4631da177e4SLinus Torvalds *prev = old->next; 4641da177e4SLinus Torvalds } 4651da177e4SLinus Torvalds 46667e345d1SDavid Woodhouse /* Free it now unless it's in READING or CLEARING state, which 46767e345d1SDavid Woodhouse are the transitions upon read_inode() and clear_inode(). The 46867e345d1SDavid Woodhouse rest of the time we know nobody else is looking at it, and 46967e345d1SDavid Woodhouse if it's held by read_inode() or clear_inode() they'll free it 47067e345d1SDavid Woodhouse for themselves. */ 47167e345d1SDavid Woodhouse if (old->state != INO_STATE_READING && old->state != INO_STATE_CLEARING) 47267e345d1SDavid Woodhouse jffs2_free_inode_cache(old); 47367e345d1SDavid Woodhouse 4741da177e4SLinus Torvalds spin_unlock(&c->inocache_lock); 4751da177e4SLinus Torvalds } 4761da177e4SLinus Torvalds 4771da177e4SLinus Torvalds void jffs2_free_ino_caches(struct jffs2_sb_info *c) 4781da177e4SLinus Torvalds { 4791da177e4SLinus Torvalds int i; 4801da177e4SLinus Torvalds struct jffs2_inode_cache *this, *next; 4811da177e4SLinus Torvalds 4821da177e4SLinus Torvalds for (i=0; i<INOCACHE_HASHSIZE; i++) { 4831da177e4SLinus Torvalds this = c->inocache_list[i]; 4841da177e4SLinus Torvalds while (this) { 4851da177e4SLinus Torvalds next = this->next; 486aa98d7cfSKaiGai Kohei jffs2_xattr_free_inode(c, this); 4871da177e4SLinus Torvalds jffs2_free_inode_cache(this); 4881da177e4SLinus Torvalds this = next; 4891da177e4SLinus Torvalds } 4901da177e4SLinus Torvalds c->inocache_list[i] = NULL; 4911da177e4SLinus Torvalds } 4921da177e4SLinus Torvalds } 4931da177e4SLinus Torvalds 4941da177e4SLinus Torvalds void jffs2_free_raw_node_refs(struct jffs2_sb_info *c) 4951da177e4SLinus Torvalds { 4961da177e4SLinus Torvalds int i; 4971da177e4SLinus Torvalds struct jffs2_raw_node_ref *this, *next; 4981da177e4SLinus Torvalds 4991da177e4SLinus Torvalds for (i=0; i<c->nr_blocks; i++) { 5001da177e4SLinus Torvalds this = c->blocks[i].first_node; 5011da177e4SLinus Torvalds while (this) { 5029bfeb691SDavid Woodhouse if (this[REFS_PER_BLOCK].flash_offset == REF_LINK_NODE) 5039bfeb691SDavid Woodhouse next = this[REFS_PER_BLOCK].next_in_ino; 5049bfeb691SDavid Woodhouse else 5059bfeb691SDavid Woodhouse next = NULL; 5069bfeb691SDavid Woodhouse 5079bfeb691SDavid Woodhouse jffs2_free_refblock(this); 5081da177e4SLinus Torvalds this = next; 5091da177e4SLinus Torvalds } 5101da177e4SLinus Torvalds c->blocks[i].first_node = c->blocks[i].last_node = NULL; 5111da177e4SLinus Torvalds } 5121da177e4SLinus Torvalds } 5131da177e4SLinus Torvalds 5141da177e4SLinus Torvalds struct jffs2_node_frag *jffs2_lookup_node_frag(struct rb_root *fragtree, uint32_t offset) 5151da177e4SLinus Torvalds { 5161da177e4SLinus Torvalds /* The common case in lookup is that there will be a node 5171da177e4SLinus Torvalds which precisely matches. So we go looking for that first */ 5181da177e4SLinus Torvalds struct rb_node *next; 5191da177e4SLinus Torvalds struct jffs2_node_frag *prev = NULL; 5201da177e4SLinus Torvalds struct jffs2_node_frag *frag = NULL; 5211da177e4SLinus Torvalds 522733802d9SArtem B. Bityutskiy dbg_fragtree2("root %p, offset %d\n", fragtree, offset); 5231da177e4SLinus Torvalds 5241da177e4SLinus Torvalds next = fragtree->rb_node; 5251da177e4SLinus Torvalds 5261da177e4SLinus Torvalds while(next) { 5271da177e4SLinus Torvalds frag = rb_entry(next, struct jffs2_node_frag, rb); 5281da177e4SLinus Torvalds 5291da177e4SLinus Torvalds if (frag->ofs + frag->size <= offset) { 5301da177e4SLinus Torvalds /* Remember the closest smaller match on the way down */ 5311da177e4SLinus Torvalds if (!prev || frag->ofs > prev->ofs) 5321da177e4SLinus Torvalds prev = frag; 5331da177e4SLinus Torvalds next = frag->rb.rb_right; 5341da177e4SLinus Torvalds } else if (frag->ofs > offset) { 5351da177e4SLinus Torvalds next = frag->rb.rb_left; 5361da177e4SLinus Torvalds } else { 5371da177e4SLinus Torvalds return frag; 5381da177e4SLinus Torvalds } 5391da177e4SLinus Torvalds } 5401da177e4SLinus Torvalds 5411da177e4SLinus Torvalds /* Exact match not found. Go back up looking at each parent, 5421da177e4SLinus Torvalds and return the closest smaller one */ 5431da177e4SLinus Torvalds 5441da177e4SLinus Torvalds if (prev) 545733802d9SArtem B. Bityutskiy dbg_fragtree2("no match. Returning frag %#04x-%#04x, closest previous\n", 546e0d60137SArtem B. Bityutskiy prev->ofs, prev->ofs+prev->size); 5471da177e4SLinus Torvalds else 548733802d9SArtem B. Bityutskiy dbg_fragtree2("returning NULL, empty fragtree\n"); 5491da177e4SLinus Torvalds 5501da177e4SLinus Torvalds return prev; 5511da177e4SLinus Torvalds } 5521da177e4SLinus Torvalds 5531da177e4SLinus Torvalds /* Pass 'c' argument to indicate that nodes should be marked obsolete as 5541da177e4SLinus Torvalds they're killed. */ 5551da177e4SLinus Torvalds void jffs2_kill_fragtree(struct rb_root *root, struct jffs2_sb_info *c) 5561da177e4SLinus Torvalds { 5571da177e4SLinus Torvalds struct jffs2_node_frag *frag; 5581da177e4SLinus Torvalds struct jffs2_node_frag *parent; 5591da177e4SLinus Torvalds 5601da177e4SLinus Torvalds if (!root->rb_node) 5611da177e4SLinus Torvalds return; 5621da177e4SLinus Torvalds 563733802d9SArtem B. Bityutskiy dbg_fragtree("killing\n"); 5641da177e4SLinus Torvalds 565e0d60137SArtem B. Bityutskiy frag = (rb_entry(root->rb_node, struct jffs2_node_frag, rb)); 5661da177e4SLinus Torvalds while(frag) { 5671da177e4SLinus Torvalds if (frag->rb.rb_left) { 5681da177e4SLinus Torvalds frag = frag_left(frag); 5691da177e4SLinus Torvalds continue; 5701da177e4SLinus Torvalds } 5711da177e4SLinus Torvalds if (frag->rb.rb_right) { 5721da177e4SLinus Torvalds frag = frag_right(frag); 5731da177e4SLinus Torvalds continue; 5741da177e4SLinus Torvalds } 5751da177e4SLinus Torvalds 5761da177e4SLinus Torvalds if (frag->node && !(--frag->node->frags)) { 5771da177e4SLinus Torvalds /* Not a hole, and it's the final remaining frag 5781da177e4SLinus Torvalds of this node. Free the node */ 5791da177e4SLinus Torvalds if (c) 5801da177e4SLinus Torvalds jffs2_mark_node_obsolete(c, frag->node->raw); 5811da177e4SLinus Torvalds 5821da177e4SLinus Torvalds jffs2_free_full_dnode(frag->node); 5831da177e4SLinus Torvalds } 5841da177e4SLinus Torvalds parent = frag_parent(frag); 5851da177e4SLinus Torvalds if (parent) { 5861da177e4SLinus Torvalds if (frag_left(parent) == frag) 5871da177e4SLinus Torvalds parent->rb.rb_left = NULL; 5881da177e4SLinus Torvalds else 5891da177e4SLinus Torvalds parent->rb.rb_right = NULL; 5901da177e4SLinus Torvalds } 5911da177e4SLinus Torvalds 5921da177e4SLinus Torvalds jffs2_free_node_frag(frag); 5931da177e4SLinus Torvalds frag = parent; 5941da177e4SLinus Torvalds 5951da177e4SLinus Torvalds cond_resched(); 5961da177e4SLinus Torvalds } 5971da177e4SLinus Torvalds } 598f1f9671bSDavid Woodhouse 5992f785402SDavid Woodhouse struct jffs2_raw_node_ref *jffs2_link_node_ref(struct jffs2_sb_info *c, 6002f785402SDavid Woodhouse struct jffs2_eraseblock *jeb, 6012f785402SDavid Woodhouse uint32_t ofs, uint32_t len, 602fcb75787SDavid Woodhouse struct jffs2_inode_cache *ic) 603f1f9671bSDavid Woodhouse { 6042f785402SDavid Woodhouse struct jffs2_raw_node_ref *ref; 6052f785402SDavid Woodhouse 6069bfeb691SDavid Woodhouse BUG_ON(!jeb->allocated_refs); 6079bfeb691SDavid Woodhouse jeb->allocated_refs--; 6089bfeb691SDavid Woodhouse 6099bfeb691SDavid Woodhouse ref = jeb->last_node; 6109bfeb691SDavid Woodhouse 6119bfeb691SDavid Woodhouse dbg_noderef("Last node at %p is (%08x,%p)\n", ref, ref->flash_offset, 6129bfeb691SDavid Woodhouse ref->next_in_ino); 6139bfeb691SDavid Woodhouse 6149bfeb691SDavid Woodhouse while (ref->flash_offset != REF_EMPTY_NODE) { 6159bfeb691SDavid Woodhouse if (ref->flash_offset == REF_LINK_NODE) 6169bfeb691SDavid Woodhouse ref = ref->next_in_ino; 6179bfeb691SDavid Woodhouse else 6189bfeb691SDavid Woodhouse ref++; 6192f785402SDavid Woodhouse } 6202f785402SDavid Woodhouse 6219bfeb691SDavid Woodhouse dbg_noderef("New ref is %p (%08x becomes %08x,%p) len 0x%x\n", ref, 6229bfeb691SDavid Woodhouse ref->flash_offset, ofs, ref->next_in_ino, len); 6239bfeb691SDavid Woodhouse 6242f785402SDavid Woodhouse ref->flash_offset = ofs; 6252f785402SDavid Woodhouse 6269bfeb691SDavid Woodhouse if (!jeb->first_node) { 627f1f9671bSDavid Woodhouse jeb->first_node = ref; 6289bfeb691SDavid Woodhouse BUG_ON(ref_offset(ref) != jeb->offset); 6299bfeb691SDavid Woodhouse } else if (unlikely(ref_offset(ref) != jeb->offset + c->sector_size - jeb->free_size)) { 6309bfeb691SDavid Woodhouse uint32_t last_len = ref_totlen(c, jeb, jeb->last_node); 6319bfeb691SDavid Woodhouse 6329bfeb691SDavid Woodhouse JFFS2_ERROR("Adding new ref %p at (0x%08x-0x%08x) not immediately after previous (0x%08x-0x%08x)\n", 6339bfeb691SDavid Woodhouse ref, ref_offset(ref), ref_offset(ref)+len, 6349bfeb691SDavid Woodhouse ref_offset(jeb->last_node), 6359bfeb691SDavid Woodhouse ref_offset(jeb->last_node)+last_len); 6369bfeb691SDavid Woodhouse BUG(); 637ca89a517SDavid Woodhouse } 638f1f9671bSDavid Woodhouse jeb->last_node = ref; 639f1f9671bSDavid Woodhouse 640fcb75787SDavid Woodhouse if (ic) { 641fcb75787SDavid Woodhouse ref->next_in_ino = ic->nodes; 642fcb75787SDavid Woodhouse ic->nodes = ref; 643fcb75787SDavid Woodhouse } else { 644fcb75787SDavid Woodhouse ref->next_in_ino = NULL; 645fcb75787SDavid Woodhouse } 646fcb75787SDavid Woodhouse 647f1f9671bSDavid Woodhouse switch(ref_flags(ref)) { 648f1f9671bSDavid Woodhouse case REF_UNCHECKED: 649f1f9671bSDavid Woodhouse c->unchecked_size += len; 650f1f9671bSDavid Woodhouse jeb->unchecked_size += len; 651f1f9671bSDavid Woodhouse break; 652f1f9671bSDavid Woodhouse 653f1f9671bSDavid Woodhouse case REF_NORMAL: 654f1f9671bSDavid Woodhouse case REF_PRISTINE: 655f1f9671bSDavid Woodhouse c->used_size += len; 656f1f9671bSDavid Woodhouse jeb->used_size += len; 657f1f9671bSDavid Woodhouse break; 658f1f9671bSDavid Woodhouse 659f1f9671bSDavid Woodhouse case REF_OBSOLETE: 660f1f9671bSDavid Woodhouse c->dirty_size += len; 6613b79673cSDavid Woodhouse jeb->dirty_size += len; 662f1f9671bSDavid Woodhouse break; 663f1f9671bSDavid Woodhouse } 664f1f9671bSDavid Woodhouse c->free_size -= len; 665f1f9671bSDavid Woodhouse jeb->free_size -= len; 666f1f9671bSDavid Woodhouse 667ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 668ca89a517SDavid Woodhouse /* Set (and test) __totlen field... for now */ 669ca89a517SDavid Woodhouse ref->__totlen = len; 670ca89a517SDavid Woodhouse ref_totlen(c, jeb, ref); 671ca89a517SDavid Woodhouse #endif 6722f785402SDavid Woodhouse return ref; 673f1f9671bSDavid Woodhouse } 67468270995SDavid Woodhouse 6752f785402SDavid Woodhouse /* No locking, no reservation of 'ref'. Do not use on a live file system */ 67668270995SDavid Woodhouse int jffs2_scan_dirty_space(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, 67768270995SDavid Woodhouse uint32_t size) 67868270995SDavid Woodhouse { 679ca89a517SDavid Woodhouse if (!size) 680ca89a517SDavid Woodhouse return 0; 6819bfeb691SDavid Woodhouse if (unlikely(size > jeb->free_size)) { 6829bfeb691SDavid Woodhouse printk(KERN_CRIT "Dirty space 0x%x larger then free_size 0x%x (wasted 0x%x)\n", 6839bfeb691SDavid Woodhouse size, jeb->free_size, jeb->wasted_size); 684ca89a517SDavid Woodhouse BUG(); 685ca89a517SDavid Woodhouse } 6869bfeb691SDavid Woodhouse /* REF_EMPTY_NODE is !obsolete, so that works OK */ 6872ebf09c2SDavid Woodhouse if (jeb->last_node && ref_obsolete(jeb->last_node)) { 688ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 689ca89a517SDavid Woodhouse jeb->last_node->__totlen += size; 690ca89a517SDavid Woodhouse #endif 69168270995SDavid Woodhouse c->dirty_size += size; 69268270995SDavid Woodhouse c->free_size -= size; 69368270995SDavid Woodhouse jeb->dirty_size += size; 69468270995SDavid Woodhouse jeb->free_size -= size; 695ca89a517SDavid Woodhouse } else { 6962f785402SDavid Woodhouse uint32_t ofs = jeb->offset + c->sector_size - jeb->free_size; 6972f785402SDavid Woodhouse ofs |= REF_OBSOLETE; 698ca89a517SDavid Woodhouse 6992f785402SDavid Woodhouse jffs2_link_node_ref(c, jeb, ofs, size, NULL); 700ca89a517SDavid Woodhouse } 70168270995SDavid Woodhouse 70268270995SDavid Woodhouse return 0; 70368270995SDavid Woodhouse } 704ca89a517SDavid Woodhouse 705ca89a517SDavid Woodhouse /* Calculate totlen from surrounding nodes or eraseblock */ 706ca89a517SDavid Woodhouse static inline uint32_t __ref_totlen(struct jffs2_sb_info *c, 707ca89a517SDavid Woodhouse struct jffs2_eraseblock *jeb, 708ca89a517SDavid Woodhouse struct jffs2_raw_node_ref *ref) 709ca89a517SDavid Woodhouse { 710ca89a517SDavid Woodhouse uint32_t ref_end; 71199988f7bSDavid Woodhouse struct jffs2_raw_node_ref *next_ref = ref_next(ref); 712ca89a517SDavid Woodhouse 71399988f7bSDavid Woodhouse if (next_ref) 71499988f7bSDavid Woodhouse ref_end = ref_offset(next_ref); 715ca89a517SDavid Woodhouse else { 716ca89a517SDavid Woodhouse if (!jeb) 717ca89a517SDavid Woodhouse jeb = &c->blocks[ref->flash_offset / c->sector_size]; 718ca89a517SDavid Woodhouse 719ca89a517SDavid Woodhouse /* Last node in block. Use free_space */ 7209bfeb691SDavid Woodhouse if (unlikely(ref != jeb->last_node)) { 721ca89a517SDavid Woodhouse printk(KERN_CRIT "ref %p @0x%08x is not jeb->last_node (%p @0x%08x)\n", 722ca89a517SDavid Woodhouse ref, ref_offset(ref), jeb->last_node, jeb->last_node?ref_offset(jeb->last_node):0); 723ca89a517SDavid Woodhouse BUG(); 724ca89a517SDavid Woodhouse } 725ca89a517SDavid Woodhouse ref_end = jeb->offset + c->sector_size - jeb->free_size; 726ca89a517SDavid Woodhouse } 727ca89a517SDavid Woodhouse return ref_end - ref_offset(ref); 728ca89a517SDavid Woodhouse } 729ca89a517SDavid Woodhouse 730ca89a517SDavid Woodhouse uint32_t __jffs2_ref_totlen(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, 731ca89a517SDavid Woodhouse struct jffs2_raw_node_ref *ref) 732ca89a517SDavid Woodhouse { 733ca89a517SDavid Woodhouse uint32_t ret; 734ca89a517SDavid Woodhouse 735ca89a517SDavid Woodhouse ret = __ref_totlen(c, jeb, ref); 7369bfeb691SDavid Woodhouse 737ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 7389bfeb691SDavid Woodhouse if (unlikely(ret != ref->__totlen)) { 7399bfeb691SDavid Woodhouse if (!jeb) 7409bfeb691SDavid Woodhouse jeb = &c->blocks[ref->flash_offset / c->sector_size]; 7419bfeb691SDavid Woodhouse 742ca89a517SDavid Woodhouse printk(KERN_CRIT "Totlen for ref at %p (0x%08x-0x%08x) miscalculated as 0x%x instead of %x\n", 743ca89a517SDavid Woodhouse ref, ref_offset(ref), ref_offset(ref)+ref->__totlen, 744ca89a517SDavid Woodhouse ret, ref->__totlen); 74599988f7bSDavid Woodhouse if (ref_next(ref)) { 74699988f7bSDavid Woodhouse printk(KERN_CRIT "next %p (0x%08x-0x%08x)\n", ref_next(ref), ref_offset(ref_next(ref)), 74799988f7bSDavid Woodhouse ref_offset(ref_next(ref))+ref->__totlen); 748ca89a517SDavid Woodhouse } else 74999988f7bSDavid Woodhouse printk(KERN_CRIT "No next ref. jeb->last_node is %p\n", jeb->last_node); 750ca89a517SDavid Woodhouse 751ca89a517SDavid Woodhouse printk(KERN_CRIT "jeb->wasted_size %x, dirty_size %x, used_size %x, free_size %x\n", jeb->wasted_size, jeb->dirty_size, jeb->used_size, jeb->free_size); 7529bfeb691SDavid Woodhouse 753ca89a517SDavid Woodhouse #if defined(JFFS2_DBG_DUMPS) || defined(JFFS2_DBG_PARANOIA_CHECKS) 754ca89a517SDavid Woodhouse __jffs2_dbg_dump_node_refs_nolock(c, jeb); 755ca89a517SDavid Woodhouse #endif 7569bfeb691SDavid Woodhouse 757ca89a517SDavid Woodhouse WARN_ON(1); 7589bfeb691SDavid Woodhouse 7599bfeb691SDavid Woodhouse ret = ref->__totlen; 760ca89a517SDavid Woodhouse } 761ca89a517SDavid Woodhouse #endif /* TEST_TOTLEN */ 762ca89a517SDavid Woodhouse return ret; 763ca89a517SDavid Woodhouse } 764