11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * JFFS2 -- Journalling Flash File System, Version 2. 31da177e4SLinus Torvalds * 41da177e4SLinus Torvalds * Copyright (C) 2001-2003 Red Hat, Inc. 51da177e4SLinus Torvalds * 61da177e4SLinus Torvalds * Created by David Woodhouse <dwmw2@infradead.org> 71da177e4SLinus Torvalds * 81da177e4SLinus Torvalds * For licensing information, see the file 'LICENCE' in this directory. 91da177e4SLinus Torvalds * 10182ec4eeSThomas Gleixner * $Id: nodelist.c,v 1.115 2005/11/07 11:14:40 gleixner Exp $ 111da177e4SLinus Torvalds * 121da177e4SLinus Torvalds */ 131da177e4SLinus Torvalds 141da177e4SLinus Torvalds #include <linux/kernel.h> 151da177e4SLinus Torvalds #include <linux/sched.h> 161da177e4SLinus Torvalds #include <linux/fs.h> 171da177e4SLinus Torvalds #include <linux/mtd/mtd.h> 181da177e4SLinus Torvalds #include <linux/rbtree.h> 191da177e4SLinus Torvalds #include <linux/crc32.h> 201da177e4SLinus Torvalds #include <linux/slab.h> 211da177e4SLinus Torvalds #include <linux/pagemap.h> 221da177e4SLinus Torvalds #include "nodelist.h" 231da177e4SLinus Torvalds 241da177e4SLinus Torvalds void jffs2_add_fd_to_list(struct jffs2_sb_info *c, struct jffs2_full_dirent *new, struct jffs2_full_dirent **list) 251da177e4SLinus Torvalds { 261da177e4SLinus Torvalds struct jffs2_full_dirent **prev = list; 27e0d60137SArtem B. Bityutskiy 28733802d9SArtem B. Bityutskiy dbg_dentlist("add dirent \"%s\", ino #%u\n", new->name, new->ino); 291da177e4SLinus Torvalds 301da177e4SLinus Torvalds while ((*prev) && (*prev)->nhash <= new->nhash) { 311da177e4SLinus Torvalds if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { 321da177e4SLinus Torvalds /* Duplicate. Free one */ 331da177e4SLinus Torvalds if (new->version < (*prev)->version) { 34733802d9SArtem B. Bityutskiy dbg_dentlist("Eep! Marking new dirent node is obsolete, old is \"%s\", ino #%u\n", 35e0d60137SArtem B. Bityutskiy (*prev)->name, (*prev)->ino); 361da177e4SLinus Torvalds jffs2_mark_node_obsolete(c, new->raw); 371da177e4SLinus Torvalds jffs2_free_full_dirent(new); 381da177e4SLinus Torvalds } else { 39733802d9SArtem B. Bityutskiy dbg_dentlist("marking old dirent \"%s\", ino #%u bsolete\n", 40e0d60137SArtem B. Bityutskiy (*prev)->name, (*prev)->ino); 411da177e4SLinus Torvalds new->next = (*prev)->next; 421da177e4SLinus Torvalds jffs2_mark_node_obsolete(c, ((*prev)->raw)); 431da177e4SLinus Torvalds jffs2_free_full_dirent(*prev); 441da177e4SLinus Torvalds *prev = new; 451da177e4SLinus Torvalds } 46e0d60137SArtem B. Bityutskiy return; 471da177e4SLinus Torvalds } 481da177e4SLinus Torvalds prev = &((*prev)->next); 491da177e4SLinus Torvalds } 501da177e4SLinus Torvalds new->next = *prev; 511da177e4SLinus Torvalds *prev = new; 521da177e4SLinus Torvalds } 531da177e4SLinus Torvalds 541e900979SArtem B. Bityutskiy void jffs2_truncate_fragtree(struct jffs2_sb_info *c, struct rb_root *list, uint32_t size) 551e900979SArtem B. Bityutskiy { 561e900979SArtem B. Bityutskiy struct jffs2_node_frag *frag = jffs2_lookup_node_frag(list, size); 571e900979SArtem B. Bityutskiy 58733802d9SArtem B. Bityutskiy dbg_fragtree("truncating fragtree to 0x%08x bytes\n", size); 591e900979SArtem B. Bityutskiy 601e900979SArtem B. Bityutskiy /* We know frag->ofs <= size. That's what lookup does for us */ 611e900979SArtem B. Bityutskiy if (frag && frag->ofs != size) { 621e0da3cbSArtem B. Bityutskiy if (frag->ofs+frag->size > size) { 631e900979SArtem B. Bityutskiy frag->size = size - frag->ofs; 641e900979SArtem B. Bityutskiy } 651e900979SArtem B. Bityutskiy frag = frag_next(frag); 661e900979SArtem B. Bityutskiy } 671e900979SArtem B. Bityutskiy while (frag && frag->ofs >= size) { 681e900979SArtem B. Bityutskiy struct jffs2_node_frag *next = frag_next(frag); 691e900979SArtem B. Bityutskiy 701e900979SArtem B. Bityutskiy frag_erase(frag, list); 711e900979SArtem B. Bityutskiy jffs2_obsolete_node_frag(c, frag); 721e900979SArtem B. Bityutskiy frag = next; 731e900979SArtem B. Bityutskiy } 741e0da3cbSArtem B. Bityutskiy 751e0da3cbSArtem B. Bityutskiy if (size == 0) 761e0da3cbSArtem B. Bityutskiy return; 771e0da3cbSArtem B. Bityutskiy 781e0da3cbSArtem B. Bityutskiy /* 791e0da3cbSArtem B. Bityutskiy * If the last fragment starts at the RAM page boundary, it is 801e0da3cbSArtem B. Bityutskiy * REF_PRISTINE irrespective of its size. 811e0da3cbSArtem B. Bityutskiy */ 821e0da3cbSArtem B. Bityutskiy frag = frag_last(list); 83f0507530SArtem B. Bityutskiy if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) { 84733802d9SArtem B. Bityutskiy dbg_fragtree2("marking the last fragment 0x%08x-0x%08x REF_PRISTINE.\n", 851e0da3cbSArtem B. Bityutskiy frag->ofs, frag->ofs + frag->size); 861e0da3cbSArtem B. Bityutskiy frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE; 871e0da3cbSArtem B. Bityutskiy } 881e900979SArtem B. Bityutskiy } 891e900979SArtem B. Bityutskiy 90f97117d1SArtem B. Bityutskiy void jffs2_obsolete_node_frag(struct jffs2_sb_info *c, struct jffs2_node_frag *this) 911da177e4SLinus Torvalds { 92f97117d1SArtem B. Bityutskiy if (this->node) { 93f97117d1SArtem B. Bityutskiy this->node->frags--; 94f97117d1SArtem B. Bityutskiy if (!this->node->frags) { 95f97117d1SArtem B. Bityutskiy /* The node has no valid frags left. It's totally obsoleted */ 96733802d9SArtem B. Bityutskiy dbg_fragtree2("marking old node @0x%08x (0x%04x-0x%04x) obsolete\n", 97e0d60137SArtem B. Bityutskiy ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size); 98f97117d1SArtem B. Bityutskiy jffs2_mark_node_obsolete(c, this->node->raw); 99f97117d1SArtem B. Bityutskiy jffs2_free_full_dnode(this->node); 100f97117d1SArtem B. Bityutskiy } else { 101733802d9SArtem B. Bityutskiy dbg_fragtree2("marking old node @0x%08x (0x%04x-0x%04x) REF_NORMAL. frags is %d\n", 102e0d60137SArtem B. Bityutskiy ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); 103f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 1041da177e4SLinus Torvalds } 1051da177e4SLinus Torvalds 106f97117d1SArtem B. Bityutskiy } 107f97117d1SArtem B. Bityutskiy jffs2_free_node_frag(this); 1089dee7503SDavid Woodhouse } 1099dee7503SDavid Woodhouse 110f97117d1SArtem B. Bityutskiy static void jffs2_fragtree_insert(struct jffs2_node_frag *newfrag, struct jffs2_node_frag *base) 1111da177e4SLinus Torvalds { 112f97117d1SArtem B. Bityutskiy struct rb_node *parent = &base->rb; 113f97117d1SArtem B. Bityutskiy struct rb_node **link = &parent; 1141da177e4SLinus Torvalds 115733802d9SArtem B. Bityutskiy dbg_fragtree2("insert frag (0x%04x-0x%04x)\n", newfrag->ofs, newfrag->ofs + newfrag->size); 1169dee7503SDavid Woodhouse 117f97117d1SArtem B. Bityutskiy while (*link) { 118f97117d1SArtem B. Bityutskiy parent = *link; 119f97117d1SArtem B. Bityutskiy base = rb_entry(parent, struct jffs2_node_frag, rb); 120f97117d1SArtem B. Bityutskiy 121f97117d1SArtem B. Bityutskiy if (newfrag->ofs > base->ofs) 122f97117d1SArtem B. Bityutskiy link = &base->rb.rb_right; 123f97117d1SArtem B. Bityutskiy else if (newfrag->ofs < base->ofs) 124f97117d1SArtem B. Bityutskiy link = &base->rb.rb_left; 1259dee7503SDavid Woodhouse else { 126e0d60137SArtem B. Bityutskiy JFFS2_ERROR("duplicate frag at %08x (%p,%p)\n", newfrag->ofs, newfrag, base); 127dae6227fSArtem B. Bityutskiy BUG(); 128dae6227fSArtem B. Bityutskiy } 129dae6227fSArtem B. Bityutskiy } 130dae6227fSArtem B. Bityutskiy 131f97117d1SArtem B. Bityutskiy rb_link_node(&newfrag->rb, &base->rb, link); 132dae6227fSArtem B. Bityutskiy } 133dae6227fSArtem B. Bityutskiy 1341e0da3cbSArtem B. Bityutskiy /* 1351e0da3cbSArtem B. Bityutskiy * Allocate and initializes a new fragment. 1361e0da3cbSArtem B. Bityutskiy */ 137858119e1SArjan van de Ven static struct jffs2_node_frag * new_fragment(struct jffs2_full_dnode *fn, uint32_t ofs, uint32_t size) 1381e0da3cbSArtem B. Bityutskiy { 1391e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag; 1401e0da3cbSArtem B. Bityutskiy 1411e0da3cbSArtem B. Bityutskiy newfrag = jffs2_alloc_node_frag(); 1421e0da3cbSArtem B. Bityutskiy if (likely(newfrag)) { 1431e0da3cbSArtem B. Bityutskiy newfrag->ofs = ofs; 1441e0da3cbSArtem B. Bityutskiy newfrag->size = size; 1451e0da3cbSArtem B. Bityutskiy newfrag->node = fn; 1461e0da3cbSArtem B. Bityutskiy } else { 1471e0da3cbSArtem B. Bityutskiy JFFS2_ERROR("cannot allocate a jffs2_node_frag object\n"); 1481e0da3cbSArtem B. Bityutskiy } 1491e0da3cbSArtem B. Bityutskiy 1501e0da3cbSArtem B. Bityutskiy return newfrag; 1511e0da3cbSArtem B. Bityutskiy } 1521e0da3cbSArtem B. Bityutskiy 1531e0da3cbSArtem B. Bityutskiy /* 1541e0da3cbSArtem B. Bityutskiy * Called when there is no overlapping fragment exist. Inserts a hole before the new 1551e0da3cbSArtem B. Bityutskiy * fragment and inserts the new fragment to the fragtree. 1561e0da3cbSArtem B. Bityutskiy */ 1571e0da3cbSArtem B. Bityutskiy static int no_overlapping_node(struct jffs2_sb_info *c, struct rb_root *root, 1581e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag, 1591e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *this, uint32_t lastend) 1601e0da3cbSArtem B. Bityutskiy { 1611e0da3cbSArtem B. Bityutskiy if (lastend < newfrag->node->ofs) { 1621e0da3cbSArtem B. Bityutskiy /* put a hole in before the new fragment */ 1631e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *holefrag; 1641e0da3cbSArtem B. Bityutskiy 1651e0da3cbSArtem B. Bityutskiy holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend); 1661e0da3cbSArtem B. Bityutskiy if (unlikely(!holefrag)) { 1671e0da3cbSArtem B. Bityutskiy jffs2_free_node_frag(newfrag); 1681e0da3cbSArtem B. Bityutskiy return -ENOMEM; 1691e0da3cbSArtem B. Bityutskiy } 1701e0da3cbSArtem B. Bityutskiy 1711e0da3cbSArtem B. Bityutskiy if (this) { 1721e0da3cbSArtem B. Bityutskiy /* By definition, the 'this' node has no right-hand child, 1731e0da3cbSArtem B. Bityutskiy because there are no frags with offset greater than it. 1741e0da3cbSArtem B. Bityutskiy So that's where we want to put the hole */ 175733802d9SArtem B. Bityutskiy dbg_fragtree2("add hole frag %#04x-%#04x on the right of the new frag.\n", 1761e0da3cbSArtem B. Bityutskiy holefrag->ofs, holefrag->ofs + holefrag->size); 1771e0da3cbSArtem B. Bityutskiy rb_link_node(&holefrag->rb, &this->rb, &this->rb.rb_right); 1781e0da3cbSArtem B. Bityutskiy } else { 179733802d9SArtem B. Bityutskiy dbg_fragtree2("Add hole frag %#04x-%#04x to the root of the tree.\n", 1801e0da3cbSArtem B. Bityutskiy holefrag->ofs, holefrag->ofs + holefrag->size); 1811e0da3cbSArtem B. Bityutskiy rb_link_node(&holefrag->rb, NULL, &root->rb_node); 1821e0da3cbSArtem B. Bityutskiy } 1831e0da3cbSArtem B. Bityutskiy rb_insert_color(&holefrag->rb, root); 1841e0da3cbSArtem B. Bityutskiy this = holefrag; 1851e0da3cbSArtem B. Bityutskiy } 1861e0da3cbSArtem B. Bityutskiy 1871e0da3cbSArtem B. Bityutskiy if (this) { 1881e0da3cbSArtem B. Bityutskiy /* By definition, the 'this' node has no right-hand child, 1891e0da3cbSArtem B. Bityutskiy because there are no frags with offset greater than it. 1901e0da3cbSArtem B. Bityutskiy So that's where we want to put new fragment */ 191733802d9SArtem B. Bityutskiy dbg_fragtree2("add the new node at the right\n"); 1921e0da3cbSArtem B. Bityutskiy rb_link_node(&newfrag->rb, &this->rb, &this->rb.rb_right); 1931e0da3cbSArtem B. Bityutskiy } else { 194733802d9SArtem B. Bityutskiy dbg_fragtree2("insert the new node at the root of the tree\n"); 1951e0da3cbSArtem B. Bityutskiy rb_link_node(&newfrag->rb, NULL, &root->rb_node); 1961e0da3cbSArtem B. Bityutskiy } 1971e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 1981e0da3cbSArtem B. Bityutskiy 1991e0da3cbSArtem B. Bityutskiy return 0; 2001e0da3cbSArtem B. Bityutskiy } 2011e0da3cbSArtem B. Bityutskiy 202f97117d1SArtem B. Bityutskiy /* Doesn't set inode->i_size */ 2031e0da3cbSArtem B. Bityutskiy static int jffs2_add_frag_to_fragtree(struct jffs2_sb_info *c, struct rb_root *root, struct jffs2_node_frag *newfrag) 2041da177e4SLinus Torvalds { 205f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *this; 206f97117d1SArtem B. Bityutskiy uint32_t lastend; 2071da177e4SLinus Torvalds 208f97117d1SArtem B. Bityutskiy /* Skip all the nodes which are completed before this one starts */ 2091e0da3cbSArtem B. Bityutskiy this = jffs2_lookup_node_frag(root, newfrag->node->ofs); 2101da177e4SLinus Torvalds 211f97117d1SArtem B. Bityutskiy if (this) { 212733802d9SArtem B. Bityutskiy dbg_fragtree2("lookup gave frag 0x%04x-0x%04x; phys 0x%08x (*%p)\n", 213e0d60137SArtem B. Bityutskiy this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this); 214f97117d1SArtem B. Bityutskiy lastend = this->ofs + this->size; 215f97117d1SArtem B. Bityutskiy } else { 216733802d9SArtem B. Bityutskiy dbg_fragtree2("lookup gave no frag\n"); 217f97117d1SArtem B. Bityutskiy lastend = 0; 218f97117d1SArtem B. Bityutskiy } 2191da177e4SLinus Torvalds 2201e0da3cbSArtem B. Bityutskiy /* See if we ran off the end of the fragtree */ 221f97117d1SArtem B. Bityutskiy if (lastend <= newfrag->ofs) { 222f97117d1SArtem B. Bityutskiy /* We did */ 2231da177e4SLinus Torvalds 224f97117d1SArtem B. Bityutskiy /* Check if 'this' node was on the same page as the new node. 225f97117d1SArtem B. Bityutskiy If so, both 'this' and the new node get marked REF_NORMAL so 226f97117d1SArtem B. Bityutskiy the GC can take a look. 2271da177e4SLinus Torvalds */ 228f97117d1SArtem B. Bityutskiy if (lastend && (lastend-1) >> PAGE_CACHE_SHIFT == newfrag->ofs >> PAGE_CACHE_SHIFT) { 229f97117d1SArtem B. Bityutskiy if (this->node) 230f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 231f97117d1SArtem B. Bityutskiy mark_ref_normal(newfrag->node->raw); 2321da177e4SLinus Torvalds } 2331da177e4SLinus Torvalds 2341e0da3cbSArtem B. Bityutskiy return no_overlapping_node(c, root, newfrag, this, lastend); 2351da177e4SLinus Torvalds } 236dae6227fSArtem B. Bityutskiy 2371e0da3cbSArtem B. Bityutskiy if (this->node) 238733802d9SArtem B. Bityutskiy dbg_fragtree2("dealing with frag %u-%u, phys %#08x(%d).\n", 2391e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size, 2401e0da3cbSArtem B. Bityutskiy ref_offset(this->node->raw), ref_flags(this->node->raw)); 2411e0da3cbSArtem B. Bityutskiy else 242733802d9SArtem B. Bityutskiy dbg_fragtree2("dealing with hole frag %u-%u.\n", 2431e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size); 244dae6227fSArtem B. Bityutskiy 245f97117d1SArtem B. Bityutskiy /* OK. 'this' is pointing at the first frag that newfrag->ofs at least partially obsoletes, 246f97117d1SArtem B. Bityutskiy * - i.e. newfrag->ofs < this->ofs+this->size && newfrag->ofs >= this->ofs 247f97117d1SArtem B. Bityutskiy */ 248f97117d1SArtem B. Bityutskiy if (newfrag->ofs > this->ofs) { 249f97117d1SArtem B. Bityutskiy /* This node isn't completely obsoleted. The start of it remains valid */ 2501da177e4SLinus Torvalds 251f97117d1SArtem B. Bityutskiy /* Mark the new node and the partially covered node REF_NORMAL -- let 252f97117d1SArtem B. Bityutskiy the GC take a look at them */ 253f97117d1SArtem B. Bityutskiy mark_ref_normal(newfrag->node->raw); 254f97117d1SArtem B. Bityutskiy if (this->node) 255f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 2561da177e4SLinus Torvalds 257f97117d1SArtem B. Bityutskiy if (this->ofs + this->size > newfrag->ofs + newfrag->size) { 258f97117d1SArtem B. Bityutskiy /* The new node splits 'this' frag into two */ 2591e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag2; 2601e0da3cbSArtem B. Bityutskiy 261f97117d1SArtem B. Bityutskiy if (this->node) 262733802d9SArtem B. Bityutskiy dbg_fragtree2("split old frag 0x%04x-0x%04x, phys 0x%08x\n", 263e0d60137SArtem B. Bityutskiy this->ofs, this->ofs+this->size, ref_offset(this->node->raw)); 264f97117d1SArtem B. Bityutskiy else 265733802d9SArtem B. Bityutskiy dbg_fragtree2("split old hole frag 0x%04x-0x%04x\n", 2668d5df409SArtem B. Bityutskiy this->ofs, this->ofs+this->size); 267dae6227fSArtem B. Bityutskiy 268f97117d1SArtem B. Bityutskiy /* New second frag pointing to this's node */ 2691e0da3cbSArtem B. Bityutskiy newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size, 2701e0da3cbSArtem B. Bityutskiy this->ofs + this->size - newfrag->ofs - newfrag->size); 2711e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag2)) 2721e0da3cbSArtem B. Bityutskiy return -ENOMEM; 273f97117d1SArtem B. Bityutskiy if (this->node) 274f97117d1SArtem B. Bityutskiy this->node->frags++; 275dae6227fSArtem B. Bityutskiy 276f97117d1SArtem B. Bityutskiy /* Adjust size of original 'this' */ 277f97117d1SArtem B. Bityutskiy this->size = newfrag->ofs - this->ofs; 2781da177e4SLinus Torvalds 279f97117d1SArtem B. Bityutskiy /* Now, we know there's no node with offset 280f97117d1SArtem B. Bityutskiy greater than this->ofs but smaller than 281f97117d1SArtem B. Bityutskiy newfrag2->ofs or newfrag->ofs, for obvious 282f97117d1SArtem B. Bityutskiy reasons. So we can do a tree insert from 283f97117d1SArtem B. Bityutskiy 'this' to insert newfrag, and a tree insert 284f97117d1SArtem B. Bityutskiy from newfrag to insert newfrag2. */ 285f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(newfrag, this); 2861e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 2871da177e4SLinus Torvalds 288f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(newfrag2, newfrag); 2891e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag2->rb, root); 2901da177e4SLinus Torvalds 2911da177e4SLinus Torvalds return 0; 2921da177e4SLinus Torvalds } 293f97117d1SArtem B. Bityutskiy /* New node just reduces 'this' frag in size, doesn't split it */ 294f97117d1SArtem B. Bityutskiy this->size = newfrag->ofs - this->ofs; 295f97117d1SArtem B. Bityutskiy 296f97117d1SArtem B. Bityutskiy /* Again, we know it lives down here in the tree */ 297f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(newfrag, this); 2981e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 299f97117d1SArtem B. Bityutskiy } else { 300f97117d1SArtem B. Bityutskiy /* New frag starts at the same point as 'this' used to. Replace 301f97117d1SArtem B. Bityutskiy it in the tree without doing a delete and insertion */ 302733802d9SArtem B. Bityutskiy dbg_fragtree2("inserting newfrag (*%p),%d-%d in before 'this' (*%p),%d-%d\n", 303e0d60137SArtem B. Bityutskiy newfrag, newfrag->ofs, newfrag->ofs+newfrag->size, this, this->ofs, this->ofs+this->size); 304f97117d1SArtem B. Bityutskiy 3051e0da3cbSArtem B. Bityutskiy rb_replace_node(&this->rb, &newfrag->rb, root); 306f97117d1SArtem B. Bityutskiy 307f97117d1SArtem B. Bityutskiy if (newfrag->ofs + newfrag->size >= this->ofs+this->size) { 308733802d9SArtem B. Bityutskiy dbg_fragtree2("obsoleting node frag %p (%x-%x)\n", this, this->ofs, this->ofs+this->size); 309f97117d1SArtem B. Bityutskiy jffs2_obsolete_node_frag(c, this); 310f97117d1SArtem B. Bityutskiy } else { 311f97117d1SArtem B. Bityutskiy this->ofs += newfrag->size; 312f97117d1SArtem B. Bityutskiy this->size -= newfrag->size; 313f97117d1SArtem B. Bityutskiy 314f97117d1SArtem B. Bityutskiy jffs2_fragtree_insert(this, newfrag); 3151e0da3cbSArtem B. Bityutskiy rb_insert_color(&this->rb, root); 316f97117d1SArtem B. Bityutskiy return 0; 317f97117d1SArtem B. Bityutskiy } 318f97117d1SArtem B. Bityutskiy } 319f97117d1SArtem B. Bityutskiy /* OK, now we have newfrag added in the correct place in the tree, but 320f97117d1SArtem B. Bityutskiy frag_next(newfrag) may be a fragment which is overlapped by it 321f97117d1SArtem B. Bityutskiy */ 322f97117d1SArtem B. Bityutskiy while ((this = frag_next(newfrag)) && newfrag->ofs + newfrag->size >= this->ofs + this->size) { 323f97117d1SArtem B. Bityutskiy /* 'this' frag is obsoleted completely. */ 324733802d9SArtem B. Bityutskiy dbg_fragtree2("obsoleting node frag %p (%x-%x) and removing from tree\n", 325e0d60137SArtem B. Bityutskiy this, this->ofs, this->ofs+this->size); 3261e0da3cbSArtem B. Bityutskiy rb_erase(&this->rb, root); 327f97117d1SArtem B. Bityutskiy jffs2_obsolete_node_frag(c, this); 328f97117d1SArtem B. Bityutskiy } 329f97117d1SArtem B. Bityutskiy /* Now we're pointing at the first frag which isn't totally obsoleted by 330f97117d1SArtem B. Bityutskiy the new frag */ 331f97117d1SArtem B. Bityutskiy 3321e0da3cbSArtem B. Bityutskiy if (!this || newfrag->ofs + newfrag->size == this->ofs) 333f97117d1SArtem B. Bityutskiy return 0; 3341e0da3cbSArtem B. Bityutskiy 335f97117d1SArtem B. Bityutskiy /* Still some overlap but we don't need to move it in the tree */ 336f97117d1SArtem B. Bityutskiy this->size = (this->ofs + this->size) - (newfrag->ofs + newfrag->size); 337f97117d1SArtem B. Bityutskiy this->ofs = newfrag->ofs + newfrag->size; 338f97117d1SArtem B. Bityutskiy 339f97117d1SArtem B. Bityutskiy /* And mark them REF_NORMAL so the GC takes a look at them */ 340f97117d1SArtem B. Bityutskiy if (this->node) 341f97117d1SArtem B. Bityutskiy mark_ref_normal(this->node->raw); 342f97117d1SArtem B. Bityutskiy mark_ref_normal(newfrag->node->raw); 343f97117d1SArtem B. Bityutskiy 344f97117d1SArtem B. Bityutskiy return 0; 345f97117d1SArtem B. Bityutskiy } 346f97117d1SArtem B. Bityutskiy 3471e0da3cbSArtem B. Bityutskiy /* 3481e0da3cbSArtem B. Bityutskiy * Given an inode, probably with existing tree of fragments, add the new node 3491e0da3cbSArtem B. Bityutskiy * to the fragment tree. 350f97117d1SArtem B. Bityutskiy */ 351f97117d1SArtem B. Bityutskiy int jffs2_add_full_dnode_to_inode(struct jffs2_sb_info *c, struct jffs2_inode_info *f, struct jffs2_full_dnode *fn) 352f97117d1SArtem B. Bityutskiy { 353f97117d1SArtem B. Bityutskiy int ret; 354f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *newfrag; 355f97117d1SArtem B. Bityutskiy 356f97117d1SArtem B. Bityutskiy if (unlikely(!fn->size)) 357f97117d1SArtem B. Bityutskiy return 0; 358f97117d1SArtem B. Bityutskiy 3591e0da3cbSArtem B. Bityutskiy newfrag = new_fragment(fn, fn->ofs, fn->size); 360f97117d1SArtem B. Bityutskiy if (unlikely(!newfrag)) 361f97117d1SArtem B. Bityutskiy return -ENOMEM; 3621e0da3cbSArtem B. Bityutskiy newfrag->node->frags = 1; 363f97117d1SArtem B. Bityutskiy 364733802d9SArtem B. Bityutskiy dbg_fragtree("adding node %#04x-%#04x @0x%08x on flash, newfrag *%p\n", 365e0d60137SArtem B. Bityutskiy fn->ofs, fn->ofs+fn->size, ref_offset(fn->raw), newfrag); 366f97117d1SArtem B. Bityutskiy 367f97117d1SArtem B. Bityutskiy ret = jffs2_add_frag_to_fragtree(c, &f->fragtree, newfrag); 368f97117d1SArtem B. Bityutskiy if (unlikely(ret)) 369f97117d1SArtem B. Bityutskiy return ret; 370f97117d1SArtem B. Bityutskiy 371f97117d1SArtem B. Bityutskiy /* If we now share a page with other nodes, mark either previous 372f97117d1SArtem B. Bityutskiy or next node REF_NORMAL, as appropriate. */ 373f97117d1SArtem B. Bityutskiy if (newfrag->ofs & (PAGE_CACHE_SIZE-1)) { 374f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *prev = frag_prev(newfrag); 375f97117d1SArtem B. Bityutskiy 376f97117d1SArtem B. Bityutskiy mark_ref_normal(fn->raw); 377f97117d1SArtem B. Bityutskiy /* If we don't start at zero there's _always_ a previous */ 378f97117d1SArtem B. Bityutskiy if (prev->node) 379f97117d1SArtem B. Bityutskiy mark_ref_normal(prev->node->raw); 380f97117d1SArtem B. Bityutskiy } 381f97117d1SArtem B. Bityutskiy 382f97117d1SArtem B. Bityutskiy if ((newfrag->ofs+newfrag->size) & (PAGE_CACHE_SIZE-1)) { 383f97117d1SArtem B. Bityutskiy struct jffs2_node_frag *next = frag_next(newfrag); 384f97117d1SArtem B. Bityutskiy 385f97117d1SArtem B. Bityutskiy if (next) { 386f97117d1SArtem B. Bityutskiy mark_ref_normal(fn->raw); 387f97117d1SArtem B. Bityutskiy if (next->node) 388f97117d1SArtem B. Bityutskiy mark_ref_normal(next->node->raw); 389f97117d1SArtem B. Bityutskiy } 390f97117d1SArtem B. Bityutskiy } 391f97117d1SArtem B. Bityutskiy jffs2_dbg_fragtree_paranoia_check_nolock(f); 3921e0da3cbSArtem B. Bityutskiy 393f97117d1SArtem B. Bityutskiy return 0; 394f97117d1SArtem B. Bityutskiy } 395f97117d1SArtem B. Bityutskiy 3961e0da3cbSArtem B. Bityutskiy /* 3971e0da3cbSArtem B. Bityutskiy * Check the data CRC of the node. 3981e0da3cbSArtem B. Bityutskiy * 3991e0da3cbSArtem B. Bityutskiy * Returns: 0 if the data CRC is correct; 4001e0da3cbSArtem B. Bityutskiy * 1 - if incorrect; 4011e0da3cbSArtem B. Bityutskiy * error code if an error occured. 4021e0da3cbSArtem B. Bityutskiy */ 4031e0da3cbSArtem B. Bityutskiy static int check_node_data(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn) 4041e0da3cbSArtem B. Bityutskiy { 4051e0da3cbSArtem B. Bityutskiy struct jffs2_raw_node_ref *ref = tn->fn->raw; 4061e0da3cbSArtem B. Bityutskiy int err = 0, pointed = 0; 4071e0da3cbSArtem B. Bityutskiy struct jffs2_eraseblock *jeb; 4081e0da3cbSArtem B. Bityutskiy unsigned char *buffer; 4090ef675d4SAtsushi Nemoto uint32_t crc, ofs, len; 4100ef675d4SAtsushi Nemoto size_t retlen; 4111e0da3cbSArtem B. Bityutskiy 4121e0da3cbSArtem B. Bityutskiy BUG_ON(tn->csize == 0); 4131e0da3cbSArtem B. Bityutskiy 414733802d9SArtem B. Bityutskiy if (!jffs2_is_writebuffered(c)) 415733802d9SArtem B. Bityutskiy goto adj_acc; 416733802d9SArtem B. Bityutskiy 4171e0da3cbSArtem B. Bityutskiy /* Calculate how many bytes were already checked */ 4181e0da3cbSArtem B. Bityutskiy ofs = ref_offset(ref) + sizeof(struct jffs2_raw_inode); 419733802d9SArtem B. Bityutskiy len = ofs % c->wbuf_pagesize; 420280562b2SArtem B. Bityutskiy if (likely(len)) 4211e0da3cbSArtem B. Bityutskiy len = c->wbuf_pagesize - len; 4221e0da3cbSArtem B. Bityutskiy 4231e0da3cbSArtem B. Bityutskiy if (len >= tn->csize) { 424733802d9SArtem B. Bityutskiy dbg_readinode("no need to check node at %#08x, data length %u, data starts at %#08x - it has already been checked.\n", 4251e0da3cbSArtem B. Bityutskiy ref_offset(ref), tn->csize, ofs); 4261e0da3cbSArtem B. Bityutskiy goto adj_acc; 4271e0da3cbSArtem B. Bityutskiy } 4281e0da3cbSArtem B. Bityutskiy 4291e0da3cbSArtem B. Bityutskiy ofs += len; 4301e0da3cbSArtem B. Bityutskiy len = tn->csize - len; 4311e0da3cbSArtem B. Bityutskiy 432733802d9SArtem B. Bityutskiy dbg_readinode("check node at %#08x, data length %u, partial CRC %#08x, correct CRC %#08x, data starts at %#08x, start checking from %#08x - %u bytes.\n", 4331e0da3cbSArtem B. Bityutskiy ref_offset(ref), tn->csize, tn->partial_crc, tn->data_crc, ofs - len, ofs, len); 4341e0da3cbSArtem B. Bityutskiy 4351e0da3cbSArtem B. Bityutskiy #ifndef __ECOS 4361e0da3cbSArtem B. Bityutskiy /* TODO: instead, incapsulate point() stuff to jffs2_flash_read(), 4371e0da3cbSArtem B. Bityutskiy * adding and jffs2_flash_read_end() interface. */ 4381e0da3cbSArtem B. Bityutskiy if (c->mtd->point) { 4391e0da3cbSArtem B. Bityutskiy err = c->mtd->point(c->mtd, ofs, len, &retlen, &buffer); 4401e0da3cbSArtem B. Bityutskiy if (!err && retlen < tn->csize) { 441184f5652SAndrew Morton JFFS2_WARNING("MTD point returned len too short: %zu instead of %u.\n", retlen, tn->csize); 4421e0da3cbSArtem B. Bityutskiy c->mtd->unpoint(c->mtd, buffer, ofs, len); 4431e0da3cbSArtem B. Bityutskiy } else if (err) 4441e0da3cbSArtem B. Bityutskiy JFFS2_WARNING("MTD point failed: error code %d.\n", err); 4451e0da3cbSArtem B. Bityutskiy else 4461e0da3cbSArtem B. Bityutskiy pointed = 1; /* succefully pointed to device */ 4471e0da3cbSArtem B. Bityutskiy } 4481e0da3cbSArtem B. Bityutskiy #endif 4491e0da3cbSArtem B. Bityutskiy 4501e0da3cbSArtem B. Bityutskiy if (!pointed) { 4511e0da3cbSArtem B. Bityutskiy buffer = kmalloc(len, GFP_KERNEL); 4521e0da3cbSArtem B. Bityutskiy if (unlikely(!buffer)) 4531e0da3cbSArtem B. Bityutskiy return -ENOMEM; 4541e0da3cbSArtem B. Bityutskiy 4551e0da3cbSArtem B. Bityutskiy /* TODO: this is very frequent pattern, make it a separate 4561e0da3cbSArtem B. Bityutskiy * routine */ 4571e0da3cbSArtem B. Bityutskiy err = jffs2_flash_read(c, ofs, len, &retlen, buffer); 4581e0da3cbSArtem B. Bityutskiy if (err) { 4591e0da3cbSArtem B. Bityutskiy JFFS2_ERROR("can not read %d bytes from 0x%08x, error code: %d.\n", len, ofs, err); 4601e0da3cbSArtem B. Bityutskiy goto free_out; 4611e0da3cbSArtem B. Bityutskiy } 4621e0da3cbSArtem B. Bityutskiy 4631e0da3cbSArtem B. Bityutskiy if (retlen != len) { 464184f5652SAndrew Morton JFFS2_ERROR("short read at %#08x: %zd instead of %d.\n", ofs, retlen, len); 4651e0da3cbSArtem B. Bityutskiy err = -EIO; 4661e0da3cbSArtem B. Bityutskiy goto free_out; 4671e0da3cbSArtem B. Bityutskiy } 4681e0da3cbSArtem B. Bityutskiy } 4691e0da3cbSArtem B. Bityutskiy 4701e0da3cbSArtem B. Bityutskiy /* Continue calculating CRC */ 4711e0da3cbSArtem B. Bityutskiy crc = crc32(tn->partial_crc, buffer, len); 4721e0da3cbSArtem B. Bityutskiy if(!pointed) 4731e0da3cbSArtem B. Bityutskiy kfree(buffer); 4741e0da3cbSArtem B. Bityutskiy #ifndef __ECOS 4751e0da3cbSArtem B. Bityutskiy else 4761e0da3cbSArtem B. Bityutskiy c->mtd->unpoint(c->mtd, buffer, ofs, len); 4771e0da3cbSArtem B. Bityutskiy #endif 4781e0da3cbSArtem B. Bityutskiy 4791e0da3cbSArtem B. Bityutskiy if (crc != tn->data_crc) { 48039243508SArtem B. Bityutskiy JFFS2_NOTICE("wrong data CRC in data node at 0x%08x: read %#08x, calculated %#08x.\n", 4811e0da3cbSArtem B. Bityutskiy ofs, tn->data_crc, crc); 4821e0da3cbSArtem B. Bityutskiy return 1; 4831e0da3cbSArtem B. Bityutskiy } 4841e0da3cbSArtem B. Bityutskiy 4851e0da3cbSArtem B. Bityutskiy adj_acc: 4861e0da3cbSArtem B. Bityutskiy jeb = &c->blocks[ref->flash_offset / c->sector_size]; 4871e0da3cbSArtem B. Bityutskiy len = ref_totlen(c, jeb, ref); 4881e0da3cbSArtem B. Bityutskiy 4891e0da3cbSArtem B. Bityutskiy /* 4901e0da3cbSArtem B. Bityutskiy * Mark the node as having been checked and fix the 4911e0da3cbSArtem B. Bityutskiy * accounting accordingly. 4921e0da3cbSArtem B. Bityutskiy */ 4931e0da3cbSArtem B. Bityutskiy spin_lock(&c->erase_completion_lock); 4941e0da3cbSArtem B. Bityutskiy jeb->used_size += len; 4951e0da3cbSArtem B. Bityutskiy jeb->unchecked_size -= len; 4961e0da3cbSArtem B. Bityutskiy c->used_size += len; 4971e0da3cbSArtem B. Bityutskiy c->unchecked_size -= len; 4981e0da3cbSArtem B. Bityutskiy spin_unlock(&c->erase_completion_lock); 4991e0da3cbSArtem B. Bityutskiy 5001e0da3cbSArtem B. Bityutskiy return 0; 5011e0da3cbSArtem B. Bityutskiy 5021e0da3cbSArtem B. Bityutskiy free_out: 5031e0da3cbSArtem B. Bityutskiy if(!pointed) 5041e0da3cbSArtem B. Bityutskiy kfree(buffer); 5051e0da3cbSArtem B. Bityutskiy #ifndef __ECOS 5061e0da3cbSArtem B. Bityutskiy else 5071e0da3cbSArtem B. Bityutskiy c->mtd->unpoint(c->mtd, buffer, ofs, len); 5081e0da3cbSArtem B. Bityutskiy #endif 5091e0da3cbSArtem B. Bityutskiy return err; 5101e0da3cbSArtem B. Bityutskiy } 5111e0da3cbSArtem B. Bityutskiy 5121e0da3cbSArtem B. Bityutskiy /* 5131e0da3cbSArtem B. Bityutskiy * Helper function for jffs2_add_older_frag_to_fragtree(). 5141e0da3cbSArtem B. Bityutskiy * 5151e0da3cbSArtem B. Bityutskiy * Checks the node if we are in the checking stage. 5161e0da3cbSArtem B. Bityutskiy */ 517858119e1SArjan van de Ven static int check_node(struct jffs2_sb_info *c, struct jffs2_inode_info *f, struct jffs2_tmp_dnode_info *tn) 5181e0da3cbSArtem B. Bityutskiy { 5191e0da3cbSArtem B. Bityutskiy int ret; 5201e0da3cbSArtem B. Bityutskiy 5211e0da3cbSArtem B. Bityutskiy BUG_ON(ref_obsolete(tn->fn->raw)); 5221e0da3cbSArtem B. Bityutskiy 5231e0da3cbSArtem B. Bityutskiy /* We only check the data CRC of unchecked nodes */ 5241e0da3cbSArtem B. Bityutskiy if (ref_flags(tn->fn->raw) != REF_UNCHECKED) 5251e0da3cbSArtem B. Bityutskiy return 0; 5261e0da3cbSArtem B. Bityutskiy 527733802d9SArtem B. Bityutskiy dbg_fragtree2("check node %#04x-%#04x, phys offs %#08x.\n", 52839243508SArtem B. Bityutskiy tn->fn->ofs, tn->fn->ofs + tn->fn->size, ref_offset(tn->fn->raw)); 5291e0da3cbSArtem B. Bityutskiy 5301e0da3cbSArtem B. Bityutskiy ret = check_node_data(c, tn); 5311e0da3cbSArtem B. Bityutskiy if (unlikely(ret < 0)) { 5321e0da3cbSArtem B. Bityutskiy JFFS2_ERROR("check_node_data() returned error: %d.\n", 5331e0da3cbSArtem B. Bityutskiy ret); 5341e0da3cbSArtem B. Bityutskiy } else if (unlikely(ret > 0)) { 535733802d9SArtem B. Bityutskiy dbg_fragtree2("CRC error, mark it obsolete.\n"); 5361e0da3cbSArtem B. Bityutskiy jffs2_mark_node_obsolete(c, tn->fn->raw); 5371e0da3cbSArtem B. Bityutskiy } 5381e0da3cbSArtem B. Bityutskiy 5391e0da3cbSArtem B. Bityutskiy return ret; 5401e0da3cbSArtem B. Bityutskiy } 5411e0da3cbSArtem B. Bityutskiy 5421e0da3cbSArtem B. Bityutskiy /* 5431e0da3cbSArtem B. Bityutskiy * Helper function for jffs2_add_older_frag_to_fragtree(). 5441e0da3cbSArtem B. Bityutskiy * 5451e0da3cbSArtem B. Bityutskiy * Called when the new fragment that is being inserted 5461e0da3cbSArtem B. Bityutskiy * splits a hole fragment. 5471e0da3cbSArtem B. Bityutskiy */ 5481e0da3cbSArtem B. Bityutskiy static int split_hole(struct jffs2_sb_info *c, struct rb_root *root, 5491e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *newfrag, struct jffs2_node_frag *hole) 5501e0da3cbSArtem B. Bityutskiy { 551733802d9SArtem B. Bityutskiy dbg_fragtree2("fragment %#04x-%#04x splits the hole %#04x-%#04x\n", 5521e0da3cbSArtem B. Bityutskiy newfrag->ofs, newfrag->ofs + newfrag->size, hole->ofs, hole->ofs + hole->size); 5531e0da3cbSArtem B. Bityutskiy 5541e0da3cbSArtem B. Bityutskiy if (hole->ofs == newfrag->ofs) { 5551e0da3cbSArtem B. Bityutskiy /* 5561e0da3cbSArtem B. Bityutskiy * Well, the new fragment actually starts at the same offset as 5571e0da3cbSArtem B. Bityutskiy * the hole. 5581e0da3cbSArtem B. Bityutskiy */ 5591e0da3cbSArtem B. Bityutskiy if (hole->ofs + hole->size > newfrag->ofs + newfrag->size) { 5601e0da3cbSArtem B. Bityutskiy /* 5611e0da3cbSArtem B. Bityutskiy * We replace the overlapped left part of the hole by 5621e0da3cbSArtem B. Bityutskiy * the new node. 5631e0da3cbSArtem B. Bityutskiy */ 5641e0da3cbSArtem B. Bityutskiy 565733802d9SArtem B. Bityutskiy dbg_fragtree2("insert fragment %#04x-%#04x and cut the left part of the hole\n", 5661e0da3cbSArtem B. Bityutskiy newfrag->ofs, newfrag->ofs + newfrag->size); 5671e0da3cbSArtem B. Bityutskiy rb_replace_node(&hole->rb, &newfrag->rb, root); 5681e0da3cbSArtem B. Bityutskiy 5691e0da3cbSArtem B. Bityutskiy hole->ofs += newfrag->size; 5701e0da3cbSArtem B. Bityutskiy hole->size -= newfrag->size; 5711e0da3cbSArtem B. Bityutskiy 5721e0da3cbSArtem B. Bityutskiy /* 5731e0da3cbSArtem B. Bityutskiy * We know that 'hole' should be the right hand 5741e0da3cbSArtem B. Bityutskiy * fragment. 5751e0da3cbSArtem B. Bityutskiy */ 5761e0da3cbSArtem B. Bityutskiy jffs2_fragtree_insert(hole, newfrag); 5771e0da3cbSArtem B. Bityutskiy rb_insert_color(&hole->rb, root); 5781e0da3cbSArtem B. Bityutskiy } else { 5791e0da3cbSArtem B. Bityutskiy /* 5801e0da3cbSArtem B. Bityutskiy * Ah, the new fragment is of the same size as the hole. 5811e0da3cbSArtem B. Bityutskiy * Relace the hole by it. 5821e0da3cbSArtem B. Bityutskiy */ 583733802d9SArtem B. Bityutskiy dbg_fragtree2("insert fragment %#04x-%#04x and overwrite hole\n", 5841e0da3cbSArtem B. Bityutskiy newfrag->ofs, newfrag->ofs + newfrag->size); 5851e0da3cbSArtem B. Bityutskiy rb_replace_node(&hole->rb, &newfrag->rb, root); 5861e0da3cbSArtem B. Bityutskiy jffs2_free_node_frag(hole); 5871e0da3cbSArtem B. Bityutskiy } 5881e0da3cbSArtem B. Bityutskiy } else { 5891e0da3cbSArtem B. Bityutskiy /* The new fragment lefts some hole space at the left */ 5901e0da3cbSArtem B. Bityutskiy 5911e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag * newfrag2 = NULL; 5921e0da3cbSArtem B. Bityutskiy 5931e0da3cbSArtem B. Bityutskiy if (hole->ofs + hole->size > newfrag->ofs + newfrag->size) { 5941e0da3cbSArtem B. Bityutskiy /* The new frag also lefts some space at the right */ 5951e0da3cbSArtem B. Bityutskiy newfrag2 = new_fragment(NULL, newfrag->ofs + 5961e0da3cbSArtem B. Bityutskiy newfrag->size, hole->ofs + hole->size 5971e0da3cbSArtem B. Bityutskiy - newfrag->ofs - newfrag->size); 5981e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag2)) { 5991e0da3cbSArtem B. Bityutskiy jffs2_free_node_frag(newfrag); 6001e0da3cbSArtem B. Bityutskiy return -ENOMEM; 6011e0da3cbSArtem B. Bityutskiy } 6021e0da3cbSArtem B. Bityutskiy } 6031e0da3cbSArtem B. Bityutskiy 6041e0da3cbSArtem B. Bityutskiy hole->size = newfrag->ofs - hole->ofs; 605733802d9SArtem B. Bityutskiy dbg_fragtree2("left the hole %#04x-%#04x at the left and inserd fragment %#04x-%#04x\n", 6061e0da3cbSArtem B. Bityutskiy hole->ofs, hole->ofs + hole->size, newfrag->ofs, newfrag->ofs + newfrag->size); 6071e0da3cbSArtem B. Bityutskiy 6081e0da3cbSArtem B. Bityutskiy jffs2_fragtree_insert(newfrag, hole); 6091e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 6101e0da3cbSArtem B. Bityutskiy 6111e0da3cbSArtem B. Bityutskiy if (newfrag2) { 612733802d9SArtem B. Bityutskiy dbg_fragtree2("left the hole %#04x-%#04x at the right\n", 6131e0da3cbSArtem B. Bityutskiy newfrag2->ofs, newfrag2->ofs + newfrag2->size); 6141e0da3cbSArtem B. Bityutskiy jffs2_fragtree_insert(newfrag2, newfrag); 6151e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag2->rb, root); 6161e0da3cbSArtem B. Bityutskiy } 6171e0da3cbSArtem B. Bityutskiy } 6181e0da3cbSArtem B. Bityutskiy 6191e0da3cbSArtem B. Bityutskiy return 0; 6201e0da3cbSArtem B. Bityutskiy } 6211e0da3cbSArtem B. Bityutskiy 6221e0da3cbSArtem B. Bityutskiy /* 6231e0da3cbSArtem B. Bityutskiy * This function is used when we build inode. It expects the nodes are passed 6241e0da3cbSArtem B. Bityutskiy * in the decreasing version order. The whole point of this is to improve the 6251e0da3cbSArtem B. Bityutskiy * inodes checking on NAND: we check the nodes' data CRC only when they are not 6261e0da3cbSArtem B. Bityutskiy * obsoleted. Previously, add_frag_to_fragtree() function was used and 6271e0da3cbSArtem B. Bityutskiy * nodes were passed to it in the increasing version ordes and CRCs of all 6281e0da3cbSArtem B. Bityutskiy * nodes were checked. 6291e0da3cbSArtem B. Bityutskiy * 6301e0da3cbSArtem B. Bityutskiy * Note: tn->fn->size shouldn't be zero. 6311e0da3cbSArtem B. Bityutskiy * 6321e0da3cbSArtem B. Bityutskiy * Returns 0 if the node was inserted 6331e0da3cbSArtem B. Bityutskiy * 1 if it wasn't inserted (since it is obsolete) 6341e0da3cbSArtem B. Bityutskiy * < 0 an if error occured 6351e0da3cbSArtem B. Bityutskiy */ 6361e0da3cbSArtem B. Bityutskiy int jffs2_add_older_frag_to_fragtree(struct jffs2_sb_info *c, struct jffs2_inode_info *f, 6371e0da3cbSArtem B. Bityutskiy struct jffs2_tmp_dnode_info *tn) 6381e0da3cbSArtem B. Bityutskiy { 6391e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *this, *newfrag; 6401e0da3cbSArtem B. Bityutskiy uint32_t lastend; 6411e0da3cbSArtem B. Bityutskiy struct jffs2_full_dnode *fn = tn->fn; 6421e0da3cbSArtem B. Bityutskiy struct rb_root *root = &f->fragtree; 6431e0da3cbSArtem B. Bityutskiy uint32_t fn_size = fn->size, fn_ofs = fn->ofs; 6441e0da3cbSArtem B. Bityutskiy int err, checked = 0; 6451e0da3cbSArtem B. Bityutskiy int ref_flag; 6461e0da3cbSArtem B. Bityutskiy 647733802d9SArtem B. Bityutskiy dbg_fragtree("insert fragment %#04x-%#04x, ver %u\n", fn_ofs, fn_ofs + fn_size, tn->version); 6481e0da3cbSArtem B. Bityutskiy 6491e0da3cbSArtem B. Bityutskiy /* Skip all the nodes which are completed before this one starts */ 6501e0da3cbSArtem B. Bityutskiy this = jffs2_lookup_node_frag(root, fn_ofs); 6511e0da3cbSArtem B. Bityutskiy if (this) 652733802d9SArtem B. Bityutskiy dbg_fragtree2("'this' found %#04x-%#04x (%s)\n", this->ofs, this->ofs + this->size, this->node ? "data" : "hole"); 6531e0da3cbSArtem B. Bityutskiy 6541e0da3cbSArtem B. Bityutskiy if (this) 6551e0da3cbSArtem B. Bityutskiy lastend = this->ofs + this->size; 6561e0da3cbSArtem B. Bityutskiy else 6571e0da3cbSArtem B. Bityutskiy lastend = 0; 6581e0da3cbSArtem B. Bityutskiy 6591e0da3cbSArtem B. Bityutskiy /* Detect the preliminary type of node */ 6601e0da3cbSArtem B. Bityutskiy if (fn->size >= PAGE_CACHE_SIZE) 6611e0da3cbSArtem B. Bityutskiy ref_flag = REF_PRISTINE; 6621e0da3cbSArtem B. Bityutskiy else 6631e0da3cbSArtem B. Bityutskiy ref_flag = REF_NORMAL; 6641e0da3cbSArtem B. Bityutskiy 6651e0da3cbSArtem B. Bityutskiy /* See if we ran off the end of the root */ 6661e0da3cbSArtem B. Bityutskiy if (lastend <= fn_ofs) { 6671e0da3cbSArtem B. Bityutskiy /* We did */ 6681e0da3cbSArtem B. Bityutskiy 6691e0da3cbSArtem B. Bityutskiy /* 6701e0da3cbSArtem B. Bityutskiy * We are going to insert the new node into the 6711e0da3cbSArtem B. Bityutskiy * fragment tree, so check it. 6721e0da3cbSArtem B. Bityutskiy */ 6731e0da3cbSArtem B. Bityutskiy err = check_node(c, f, tn); 6741e0da3cbSArtem B. Bityutskiy if (err != 0) 6751e0da3cbSArtem B. Bityutskiy return err; 6761e0da3cbSArtem B. Bityutskiy 6771e0da3cbSArtem B. Bityutskiy fn->frags = 1; 6781e0da3cbSArtem B. Bityutskiy 6791e0da3cbSArtem B. Bityutskiy newfrag = new_fragment(fn, fn_ofs, fn_size); 6801e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag)) 6811e0da3cbSArtem B. Bityutskiy return -ENOMEM; 6821e0da3cbSArtem B. Bityutskiy 6831e0da3cbSArtem B. Bityutskiy err = no_overlapping_node(c, root, newfrag, this, lastend); 6841e0da3cbSArtem B. Bityutskiy if (unlikely(err != 0)) { 6851e0da3cbSArtem B. Bityutskiy jffs2_free_node_frag(newfrag); 6861e0da3cbSArtem B. Bityutskiy return err; 6871e0da3cbSArtem B. Bityutskiy } 6881e0da3cbSArtem B. Bityutskiy 6891e0da3cbSArtem B. Bityutskiy goto out_ok; 6901e0da3cbSArtem B. Bityutskiy } 6911e0da3cbSArtem B. Bityutskiy 6921e0da3cbSArtem B. Bityutskiy fn->frags = 0; 6931e0da3cbSArtem B. Bityutskiy 6941e0da3cbSArtem B. Bityutskiy while (1) { 6951e0da3cbSArtem B. Bityutskiy /* 6961e0da3cbSArtem B. Bityutskiy * Here we have: 6971e0da3cbSArtem B. Bityutskiy * fn_ofs < this->ofs + this->size && fn_ofs >= this->ofs. 6981e0da3cbSArtem B. Bityutskiy * 6991e0da3cbSArtem B. Bityutskiy * Remember, 'this' has higher version, any non-hole node 7001e0da3cbSArtem B. Bityutskiy * which is already in the fragtree is newer then the newly 7011e0da3cbSArtem B. Bityutskiy * inserted. 7021e0da3cbSArtem B. Bityutskiy */ 7031e0da3cbSArtem B. Bityutskiy if (!this->node) { 7041e0da3cbSArtem B. Bityutskiy /* 7051e0da3cbSArtem B. Bityutskiy * 'this' is the hole fragment, so at least the 7061e0da3cbSArtem B. Bityutskiy * beginning of the new fragment is valid. 7071e0da3cbSArtem B. Bityutskiy */ 7081e0da3cbSArtem B. Bityutskiy 7091e0da3cbSArtem B. Bityutskiy /* 7101e0da3cbSArtem B. Bityutskiy * We are going to insert the new node into the 7111e0da3cbSArtem B. Bityutskiy * fragment tree, so check it. 7121e0da3cbSArtem B. Bityutskiy */ 7131e0da3cbSArtem B. Bityutskiy if (!checked) { 7141e0da3cbSArtem B. Bityutskiy err = check_node(c, f, tn); 7151e0da3cbSArtem B. Bityutskiy if (unlikely(err != 0)) 7161e0da3cbSArtem B. Bityutskiy return err; 7171e0da3cbSArtem B. Bityutskiy checked = 1; 7181e0da3cbSArtem B. Bityutskiy } 7191e0da3cbSArtem B. Bityutskiy 7201e0da3cbSArtem B. Bityutskiy if (this->ofs + this->size >= fn_ofs + fn_size) { 7211e0da3cbSArtem B. Bityutskiy /* We split the hole on two parts */ 7221e0da3cbSArtem B. Bityutskiy 7231e0da3cbSArtem B. Bityutskiy fn->frags += 1; 7241e0da3cbSArtem B. Bityutskiy newfrag = new_fragment(fn, fn_ofs, fn_size); 7251e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag)) 7261e0da3cbSArtem B. Bityutskiy return -ENOMEM; 7271e0da3cbSArtem B. Bityutskiy 7281e0da3cbSArtem B. Bityutskiy err = split_hole(c, root, newfrag, this); 7291e0da3cbSArtem B. Bityutskiy if (unlikely(err)) 7301e0da3cbSArtem B. Bityutskiy return err; 7311e0da3cbSArtem B. Bityutskiy goto out_ok; 7321e0da3cbSArtem B. Bityutskiy } 7331e0da3cbSArtem B. Bityutskiy 7341e0da3cbSArtem B. Bityutskiy /* 7351e0da3cbSArtem B. Bityutskiy * The beginning of the new fragment is valid since it 7361e0da3cbSArtem B. Bityutskiy * overlaps the hole node. 7371e0da3cbSArtem B. Bityutskiy */ 7381e0da3cbSArtem B. Bityutskiy 7391e0da3cbSArtem B. Bityutskiy ref_flag = REF_NORMAL; 7401e0da3cbSArtem B. Bityutskiy 7411e0da3cbSArtem B. Bityutskiy fn->frags += 1; 7421e0da3cbSArtem B. Bityutskiy newfrag = new_fragment(fn, fn_ofs, 7431e0da3cbSArtem B. Bityutskiy this->ofs + this->size - fn_ofs); 7441e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag)) 7451e0da3cbSArtem B. Bityutskiy return -ENOMEM; 7461e0da3cbSArtem B. Bityutskiy 7471e0da3cbSArtem B. Bityutskiy if (fn_ofs == this->ofs) { 7481e0da3cbSArtem B. Bityutskiy /* 7491e0da3cbSArtem B. Bityutskiy * The new node starts at the same offset as 7501e0da3cbSArtem B. Bityutskiy * the hole and supersieds the hole. 7511e0da3cbSArtem B. Bityutskiy */ 752733802d9SArtem B. Bityutskiy dbg_fragtree2("add the new fragment instead of hole %#04x-%#04x, refcnt %d\n", 7531e0da3cbSArtem B. Bityutskiy fn_ofs, fn_ofs + this->ofs + this->size - fn_ofs, fn->frags); 7541e0da3cbSArtem B. Bityutskiy 7551e0da3cbSArtem B. Bityutskiy rb_replace_node(&this->rb, &newfrag->rb, root); 7561e0da3cbSArtem B. Bityutskiy jffs2_free_node_frag(this); 7571e0da3cbSArtem B. Bityutskiy } else { 7581e0da3cbSArtem B. Bityutskiy /* 7591e0da3cbSArtem B. Bityutskiy * The hole becomes shorter as its right part 7601e0da3cbSArtem B. Bityutskiy * is supersieded by the new fragment. 7611e0da3cbSArtem B. Bityutskiy */ 762733802d9SArtem B. Bityutskiy dbg_fragtree2("reduce size of hole %#04x-%#04x to %#04x-%#04x\n", 7631e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size, this->ofs, this->ofs + this->size - newfrag->size); 7641e0da3cbSArtem B. Bityutskiy 765733802d9SArtem B. Bityutskiy dbg_fragtree2("add new fragment %#04x-%#04x, refcnt %d\n", fn_ofs, 7661e0da3cbSArtem B. Bityutskiy fn_ofs + this->ofs + this->size - fn_ofs, fn->frags); 7671e0da3cbSArtem B. Bityutskiy 7681e0da3cbSArtem B. Bityutskiy this->size -= newfrag->size; 7691e0da3cbSArtem B. Bityutskiy jffs2_fragtree_insert(newfrag, this); 7701e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 7711e0da3cbSArtem B. Bityutskiy } 7721e0da3cbSArtem B. Bityutskiy 7731e0da3cbSArtem B. Bityutskiy fn_ofs += newfrag->size; 7741e0da3cbSArtem B. Bityutskiy fn_size -= newfrag->size; 7751e0da3cbSArtem B. Bityutskiy this = rb_entry(rb_next(&newfrag->rb), 7761e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag, rb); 7771e0da3cbSArtem B. Bityutskiy 778733802d9SArtem B. Bityutskiy dbg_fragtree2("switch to the next 'this' fragment: %#04x-%#04x %s\n", 7791e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size, this->node ? "(data)" : "(hole)"); 7801e0da3cbSArtem B. Bityutskiy } 7811e0da3cbSArtem B. Bityutskiy 7821e0da3cbSArtem B. Bityutskiy /* 7831e0da3cbSArtem B. Bityutskiy * 'This' node is not the hole so it obsoletes the new fragment 7841e0da3cbSArtem B. Bityutskiy * either fully or partially. 7851e0da3cbSArtem B. Bityutskiy */ 7861e0da3cbSArtem B. Bityutskiy if (this->ofs + this->size >= fn_ofs + fn_size) { 7871e0da3cbSArtem B. Bityutskiy /* The new node is obsolete, drop it */ 7881e0da3cbSArtem B. Bityutskiy if (fn->frags == 0) { 789733802d9SArtem B. Bityutskiy dbg_fragtree2("%#04x-%#04x is obsolete, mark it obsolete\n", fn_ofs, fn_ofs + fn_size); 7901e0da3cbSArtem B. Bityutskiy ref_flag = REF_OBSOLETE; 7911e0da3cbSArtem B. Bityutskiy } 7921e0da3cbSArtem B. Bityutskiy goto out_ok; 7931e0da3cbSArtem B. Bityutskiy } else { 7941e0da3cbSArtem B. Bityutskiy struct jffs2_node_frag *new_this; 7951e0da3cbSArtem B. Bityutskiy 7961e0da3cbSArtem B. Bityutskiy /* 'This' node obsoletes the beginning of the new node */ 797733802d9SArtem B. Bityutskiy dbg_fragtree2("the beginning %#04x-%#04x is obsolete\n", fn_ofs, this->ofs + this->size); 7981e0da3cbSArtem B. Bityutskiy 7991e0da3cbSArtem B. Bityutskiy ref_flag = REF_NORMAL; 8001e0da3cbSArtem B. Bityutskiy 8011e0da3cbSArtem B. Bityutskiy fn_size -= this->ofs + this->size - fn_ofs; 8021e0da3cbSArtem B. Bityutskiy fn_ofs = this->ofs + this->size; 803733802d9SArtem B. Bityutskiy dbg_fragtree2("now considering %#04x-%#04x\n", fn_ofs, fn_ofs + fn_size); 8041e0da3cbSArtem B. Bityutskiy 8051e0da3cbSArtem B. Bityutskiy new_this = rb_entry(rb_next(&this->rb), struct jffs2_node_frag, rb); 8061e0da3cbSArtem B. Bityutskiy if (!new_this) { 8071e0da3cbSArtem B. Bityutskiy /* 8081e0da3cbSArtem B. Bityutskiy * There is no next fragment. Add the rest of 8091e0da3cbSArtem B. Bityutskiy * the new node as the right-hand child. 8101e0da3cbSArtem B. Bityutskiy */ 8111e0da3cbSArtem B. Bityutskiy if (!checked) { 8121e0da3cbSArtem B. Bityutskiy err = check_node(c, f, tn); 8131e0da3cbSArtem B. Bityutskiy if (unlikely(err != 0)) 8141e0da3cbSArtem B. Bityutskiy return err; 8151e0da3cbSArtem B. Bityutskiy checked = 1; 8161e0da3cbSArtem B. Bityutskiy } 8171e0da3cbSArtem B. Bityutskiy 8181e0da3cbSArtem B. Bityutskiy fn->frags += 1; 8191e0da3cbSArtem B. Bityutskiy newfrag = new_fragment(fn, fn_ofs, fn_size); 8201e0da3cbSArtem B. Bityutskiy if (unlikely(!newfrag)) 8211e0da3cbSArtem B. Bityutskiy return -ENOMEM; 8221e0da3cbSArtem B. Bityutskiy 823733802d9SArtem B. Bityutskiy dbg_fragtree2("there are no more fragments, insert %#04x-%#04x\n", 8241e0da3cbSArtem B. Bityutskiy newfrag->ofs, newfrag->ofs + newfrag->size); 8251e0da3cbSArtem B. Bityutskiy rb_link_node(&newfrag->rb, &this->rb, &this->rb.rb_right); 8261e0da3cbSArtem B. Bityutskiy rb_insert_color(&newfrag->rb, root); 8271e0da3cbSArtem B. Bityutskiy goto out_ok; 8281e0da3cbSArtem B. Bityutskiy } else { 8291e0da3cbSArtem B. Bityutskiy this = new_this; 830733802d9SArtem B. Bityutskiy dbg_fragtree2("switch to the next 'this' fragment: %#04x-%#04x %s\n", 8311e0da3cbSArtem B. Bityutskiy this->ofs, this->ofs + this->size, this->node ? "(data)" : "(hole)"); 8321e0da3cbSArtem B. Bityutskiy } 8331e0da3cbSArtem B. Bityutskiy } 8341e0da3cbSArtem B. Bityutskiy } 8351e0da3cbSArtem B. Bityutskiy 8361e0da3cbSArtem B. Bityutskiy out_ok: 8371e0da3cbSArtem B. Bityutskiy BUG_ON(fn->size < PAGE_CACHE_SIZE && ref_flag == REF_PRISTINE); 8381e0da3cbSArtem B. Bityutskiy 8391e0da3cbSArtem B. Bityutskiy if (ref_flag == REF_OBSOLETE) { 840733802d9SArtem B. Bityutskiy dbg_fragtree2("the node is obsolete now\n"); 8411e0da3cbSArtem B. Bityutskiy /* jffs2_mark_node_obsolete() will adjust space accounting */ 8421e0da3cbSArtem B. Bityutskiy jffs2_mark_node_obsolete(c, fn->raw); 8431e0da3cbSArtem B. Bityutskiy return 1; 8441e0da3cbSArtem B. Bityutskiy } 8451e0da3cbSArtem B. Bityutskiy 846733802d9SArtem B. Bityutskiy dbg_fragtree2("the node is \"%s\" now\n", ref_flag == REF_NORMAL ? "REF_NORMAL" : "REF_PRISTINE"); 8471e0da3cbSArtem B. Bityutskiy 8481e0da3cbSArtem B. Bityutskiy /* Space accounting was adjusted at check_node_data() */ 8491e0da3cbSArtem B. Bityutskiy spin_lock(&c->erase_completion_lock); 8501e0da3cbSArtem B. Bityutskiy fn->raw->flash_offset = ref_offset(fn->raw) | ref_flag; 8511e0da3cbSArtem B. Bityutskiy spin_unlock(&c->erase_completion_lock); 8521e0da3cbSArtem B. Bityutskiy 8531e0da3cbSArtem B. Bityutskiy return 0; 8541e0da3cbSArtem B. Bityutskiy } 8551da177e4SLinus Torvalds 8561da177e4SLinus Torvalds void jffs2_set_inocache_state(struct jffs2_sb_info *c, struct jffs2_inode_cache *ic, int state) 8571da177e4SLinus Torvalds { 8581da177e4SLinus Torvalds spin_lock(&c->inocache_lock); 8591da177e4SLinus Torvalds ic->state = state; 8601da177e4SLinus Torvalds wake_up(&c->inocache_wq); 8611da177e4SLinus Torvalds spin_unlock(&c->inocache_lock); 8621da177e4SLinus Torvalds } 8631da177e4SLinus Torvalds 8641da177e4SLinus Torvalds /* During mount, this needs no locking. During normal operation, its 8651da177e4SLinus Torvalds callers want to do other stuff while still holding the inocache_lock. 8661da177e4SLinus Torvalds Rather than introducing special case get_ino_cache functions or 8671da177e4SLinus Torvalds callbacks, we just let the caller do the locking itself. */ 8681da177e4SLinus Torvalds 8691da177e4SLinus Torvalds struct jffs2_inode_cache *jffs2_get_ino_cache(struct jffs2_sb_info *c, uint32_t ino) 8701da177e4SLinus Torvalds { 8711da177e4SLinus Torvalds struct jffs2_inode_cache *ret; 8721da177e4SLinus Torvalds 8731da177e4SLinus Torvalds ret = c->inocache_list[ino % INOCACHE_HASHSIZE]; 8741da177e4SLinus Torvalds while (ret && ret->ino < ino) { 8751da177e4SLinus Torvalds ret = ret->next; 8761da177e4SLinus Torvalds } 8771da177e4SLinus Torvalds 8781da177e4SLinus Torvalds if (ret && ret->ino != ino) 8791da177e4SLinus Torvalds ret = NULL; 8801da177e4SLinus Torvalds 8811da177e4SLinus Torvalds return ret; 8821da177e4SLinus Torvalds } 8831da177e4SLinus Torvalds 8841da177e4SLinus Torvalds void jffs2_add_ino_cache (struct jffs2_sb_info *c, struct jffs2_inode_cache *new) 8851da177e4SLinus Torvalds { 8861da177e4SLinus Torvalds struct jffs2_inode_cache **prev; 8877d27c814SThomas Gleixner 8881da177e4SLinus Torvalds spin_lock(&c->inocache_lock); 8897d200960SDavid Woodhouse if (!new->ino) 8907d200960SDavid Woodhouse new->ino = ++c->highest_ino; 8917d200960SDavid Woodhouse 892733802d9SArtem B. Bityutskiy dbg_inocache("add %p (ino #%u)\n", new, new->ino); 8931da177e4SLinus Torvalds 8941da177e4SLinus Torvalds prev = &c->inocache_list[new->ino % INOCACHE_HASHSIZE]; 8951da177e4SLinus Torvalds 8961da177e4SLinus Torvalds while ((*prev) && (*prev)->ino < new->ino) { 8971da177e4SLinus Torvalds prev = &(*prev)->next; 8981da177e4SLinus Torvalds } 8991da177e4SLinus Torvalds new->next = *prev; 9001da177e4SLinus Torvalds *prev = new; 9011da177e4SLinus Torvalds 9021da177e4SLinus Torvalds spin_unlock(&c->inocache_lock); 9031da177e4SLinus Torvalds } 9041da177e4SLinus Torvalds 9051da177e4SLinus Torvalds void jffs2_del_ino_cache(struct jffs2_sb_info *c, struct jffs2_inode_cache *old) 9061da177e4SLinus Torvalds { 9071da177e4SLinus Torvalds struct jffs2_inode_cache **prev; 908e0d60137SArtem B. Bityutskiy 909733802d9SArtem B. Bityutskiy dbg_inocache("del %p (ino #%u)\n", old, old->ino); 9101da177e4SLinus Torvalds spin_lock(&c->inocache_lock); 9111da177e4SLinus Torvalds 9121da177e4SLinus Torvalds prev = &c->inocache_list[old->ino % INOCACHE_HASHSIZE]; 9131da177e4SLinus Torvalds 9141da177e4SLinus Torvalds while ((*prev) && (*prev)->ino < old->ino) { 9151da177e4SLinus Torvalds prev = &(*prev)->next; 9161da177e4SLinus Torvalds } 9171da177e4SLinus Torvalds if ((*prev) == old) { 9181da177e4SLinus Torvalds *prev = old->next; 9191da177e4SLinus Torvalds } 9201da177e4SLinus Torvalds 92167e345d1SDavid Woodhouse /* Free it now unless it's in READING or CLEARING state, which 92267e345d1SDavid Woodhouse are the transitions upon read_inode() and clear_inode(). The 92367e345d1SDavid Woodhouse rest of the time we know nobody else is looking at it, and 92467e345d1SDavid Woodhouse if it's held by read_inode() or clear_inode() they'll free it 92567e345d1SDavid Woodhouse for themselves. */ 92667e345d1SDavid Woodhouse if (old->state != INO_STATE_READING && old->state != INO_STATE_CLEARING) 92767e345d1SDavid Woodhouse jffs2_free_inode_cache(old); 92867e345d1SDavid Woodhouse 9291da177e4SLinus Torvalds spin_unlock(&c->inocache_lock); 9301da177e4SLinus Torvalds } 9311da177e4SLinus Torvalds 9321da177e4SLinus Torvalds void jffs2_free_ino_caches(struct jffs2_sb_info *c) 9331da177e4SLinus Torvalds { 9341da177e4SLinus Torvalds int i; 9351da177e4SLinus Torvalds struct jffs2_inode_cache *this, *next; 9361da177e4SLinus Torvalds 9371da177e4SLinus Torvalds for (i=0; i<INOCACHE_HASHSIZE; i++) { 9381da177e4SLinus Torvalds this = c->inocache_list[i]; 9391da177e4SLinus Torvalds while (this) { 9401da177e4SLinus Torvalds next = this->next; 941aa98d7cfSKaiGai Kohei jffs2_xattr_free_inode(c, this); 9421da177e4SLinus Torvalds jffs2_free_inode_cache(this); 9431da177e4SLinus Torvalds this = next; 9441da177e4SLinus Torvalds } 9451da177e4SLinus Torvalds c->inocache_list[i] = NULL; 9461da177e4SLinus Torvalds } 9471da177e4SLinus Torvalds } 9481da177e4SLinus Torvalds 9491da177e4SLinus Torvalds void jffs2_free_raw_node_refs(struct jffs2_sb_info *c) 9501da177e4SLinus Torvalds { 9511da177e4SLinus Torvalds int i; 9521da177e4SLinus Torvalds struct jffs2_raw_node_ref *this, *next; 9531da177e4SLinus Torvalds 9541da177e4SLinus Torvalds for (i=0; i<c->nr_blocks; i++) { 9551da177e4SLinus Torvalds this = c->blocks[i].first_node; 9561da177e4SLinus Torvalds while(this) { 9571da177e4SLinus Torvalds next = this->next_phys; 9581da177e4SLinus Torvalds jffs2_free_raw_node_ref(this); 9591da177e4SLinus Torvalds this = next; 9601da177e4SLinus Torvalds } 9611da177e4SLinus Torvalds c->blocks[i].first_node = c->blocks[i].last_node = NULL; 9621da177e4SLinus Torvalds } 9631da177e4SLinus Torvalds } 9641da177e4SLinus Torvalds 9651da177e4SLinus Torvalds struct jffs2_node_frag *jffs2_lookup_node_frag(struct rb_root *fragtree, uint32_t offset) 9661da177e4SLinus Torvalds { 9671da177e4SLinus Torvalds /* The common case in lookup is that there will be a node 9681da177e4SLinus Torvalds which precisely matches. So we go looking for that first */ 9691da177e4SLinus Torvalds struct rb_node *next; 9701da177e4SLinus Torvalds struct jffs2_node_frag *prev = NULL; 9711da177e4SLinus Torvalds struct jffs2_node_frag *frag = NULL; 9721da177e4SLinus Torvalds 973733802d9SArtem B. Bityutskiy dbg_fragtree2("root %p, offset %d\n", fragtree, offset); 9741da177e4SLinus Torvalds 9751da177e4SLinus Torvalds next = fragtree->rb_node; 9761da177e4SLinus Torvalds 9771da177e4SLinus Torvalds while(next) { 9781da177e4SLinus Torvalds frag = rb_entry(next, struct jffs2_node_frag, rb); 9791da177e4SLinus Torvalds 9801da177e4SLinus Torvalds if (frag->ofs + frag->size <= offset) { 9811da177e4SLinus Torvalds /* Remember the closest smaller match on the way down */ 9821da177e4SLinus Torvalds if (!prev || frag->ofs > prev->ofs) 9831da177e4SLinus Torvalds prev = frag; 9841da177e4SLinus Torvalds next = frag->rb.rb_right; 9851da177e4SLinus Torvalds } else if (frag->ofs > offset) { 9861da177e4SLinus Torvalds next = frag->rb.rb_left; 9871da177e4SLinus Torvalds } else { 9881da177e4SLinus Torvalds return frag; 9891da177e4SLinus Torvalds } 9901da177e4SLinus Torvalds } 9911da177e4SLinus Torvalds 9921da177e4SLinus Torvalds /* Exact match not found. Go back up looking at each parent, 9931da177e4SLinus Torvalds and return the closest smaller one */ 9941da177e4SLinus Torvalds 9951da177e4SLinus Torvalds if (prev) 996733802d9SArtem B. Bityutskiy dbg_fragtree2("no match. Returning frag %#04x-%#04x, closest previous\n", 997e0d60137SArtem B. Bityutskiy prev->ofs, prev->ofs+prev->size); 9981da177e4SLinus Torvalds else 999733802d9SArtem B. Bityutskiy dbg_fragtree2("returning NULL, empty fragtree\n"); 10001da177e4SLinus Torvalds 10011da177e4SLinus Torvalds return prev; 10021da177e4SLinus Torvalds } 10031da177e4SLinus Torvalds 10041da177e4SLinus Torvalds /* Pass 'c' argument to indicate that nodes should be marked obsolete as 10051da177e4SLinus Torvalds they're killed. */ 10061da177e4SLinus Torvalds void jffs2_kill_fragtree(struct rb_root *root, struct jffs2_sb_info *c) 10071da177e4SLinus Torvalds { 10081da177e4SLinus Torvalds struct jffs2_node_frag *frag; 10091da177e4SLinus Torvalds struct jffs2_node_frag *parent; 10101da177e4SLinus Torvalds 10111da177e4SLinus Torvalds if (!root->rb_node) 10121da177e4SLinus Torvalds return; 10131da177e4SLinus Torvalds 1014733802d9SArtem B. Bityutskiy dbg_fragtree("killing\n"); 10151da177e4SLinus Torvalds 1016e0d60137SArtem B. Bityutskiy frag = (rb_entry(root->rb_node, struct jffs2_node_frag, rb)); 10171da177e4SLinus Torvalds while(frag) { 10181da177e4SLinus Torvalds if (frag->rb.rb_left) { 10191da177e4SLinus Torvalds frag = frag_left(frag); 10201da177e4SLinus Torvalds continue; 10211da177e4SLinus Torvalds } 10221da177e4SLinus Torvalds if (frag->rb.rb_right) { 10231da177e4SLinus Torvalds frag = frag_right(frag); 10241da177e4SLinus Torvalds continue; 10251da177e4SLinus Torvalds } 10261da177e4SLinus Torvalds 10271da177e4SLinus Torvalds if (frag->node && !(--frag->node->frags)) { 10281da177e4SLinus Torvalds /* Not a hole, and it's the final remaining frag 10291da177e4SLinus Torvalds of this node. Free the node */ 10301da177e4SLinus Torvalds if (c) 10311da177e4SLinus Torvalds jffs2_mark_node_obsolete(c, frag->node->raw); 10321da177e4SLinus Torvalds 10331da177e4SLinus Torvalds jffs2_free_full_dnode(frag->node); 10341da177e4SLinus Torvalds } 10351da177e4SLinus Torvalds parent = frag_parent(frag); 10361da177e4SLinus Torvalds if (parent) { 10371da177e4SLinus Torvalds if (frag_left(parent) == frag) 10381da177e4SLinus Torvalds parent->rb.rb_left = NULL; 10391da177e4SLinus Torvalds else 10401da177e4SLinus Torvalds parent->rb.rb_right = NULL; 10411da177e4SLinus Torvalds } 10421da177e4SLinus Torvalds 10431da177e4SLinus Torvalds jffs2_free_node_frag(frag); 10441da177e4SLinus Torvalds frag = parent; 10451da177e4SLinus Torvalds 10461da177e4SLinus Torvalds cond_resched(); 10471da177e4SLinus Torvalds } 10481da177e4SLinus Torvalds } 1049f1f9671bSDavid Woodhouse 1050f1f9671bSDavid Woodhouse void jffs2_link_node_ref(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, 1051f1f9671bSDavid Woodhouse struct jffs2_raw_node_ref *ref, uint32_t len) 1052f1f9671bSDavid Woodhouse { 1053f1f9671bSDavid Woodhouse if (!jeb->first_node) 1054f1f9671bSDavid Woodhouse jeb->first_node = ref; 1055ca89a517SDavid Woodhouse if (jeb->last_node) { 1056f1f9671bSDavid Woodhouse jeb->last_node->next_phys = ref; 1057ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 1058ca89a517SDavid Woodhouse if (ref_offset(jeb->last_node) + jeb->last_node->__totlen != ref_offset(ref)) { 1059ca89a517SDavid Woodhouse printk(KERN_CRIT "Adding new ref %p at (0x%08x-0x%08x) not immediately after previous (0x%08x-0x%08x)\n", 1060ca89a517SDavid Woodhouse ref, ref_offset(ref), ref_offset(ref)+ref->__totlen, 1061ca89a517SDavid Woodhouse ref_offset(jeb->last_node), ref_offset(jeb->last_node)+jeb->last_node->__totlen); 1062ca89a517SDavid Woodhouse WARN_ON(1); 1063ca89a517SDavid Woodhouse } 1064ca89a517SDavid Woodhouse #endif 1065ca89a517SDavid Woodhouse } 1066f1f9671bSDavid Woodhouse jeb->last_node = ref; 1067f1f9671bSDavid Woodhouse 1068f1f9671bSDavid Woodhouse switch(ref_flags(ref)) { 1069f1f9671bSDavid Woodhouse case REF_UNCHECKED: 1070f1f9671bSDavid Woodhouse c->unchecked_size += len; 1071f1f9671bSDavid Woodhouse jeb->unchecked_size += len; 1072f1f9671bSDavid Woodhouse break; 1073f1f9671bSDavid Woodhouse 1074f1f9671bSDavid Woodhouse case REF_NORMAL: 1075f1f9671bSDavid Woodhouse case REF_PRISTINE: 1076f1f9671bSDavid Woodhouse c->used_size += len; 1077f1f9671bSDavid Woodhouse jeb->used_size += len; 1078f1f9671bSDavid Woodhouse break; 1079f1f9671bSDavid Woodhouse 1080f1f9671bSDavid Woodhouse case REF_OBSOLETE: 1081f1f9671bSDavid Woodhouse c->dirty_size += len; 1082f1f9671bSDavid Woodhouse jeb->used_size += len; 1083f1f9671bSDavid Woodhouse break; 1084f1f9671bSDavid Woodhouse } 1085f1f9671bSDavid Woodhouse c->free_size -= len; 1086f1f9671bSDavid Woodhouse jeb->free_size -= len; 1087f1f9671bSDavid Woodhouse 1088f1f9671bSDavid Woodhouse ref->next_phys = NULL; 1089ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 1090ca89a517SDavid Woodhouse /* Set (and test) __totlen field... for now */ 1091ca89a517SDavid Woodhouse ref->__totlen = len; 1092ca89a517SDavid Woodhouse ref_totlen(c, jeb, ref); 1093ca89a517SDavid Woodhouse #endif 1094f1f9671bSDavid Woodhouse } 109568270995SDavid Woodhouse 1096ca89a517SDavid Woodhouse /* No locking. Do not use on a live file system */ 109768270995SDavid Woodhouse int jffs2_scan_dirty_space(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, 109868270995SDavid Woodhouse uint32_t size) 109968270995SDavid Woodhouse { 1100ca89a517SDavid Woodhouse if (!size) 1101ca89a517SDavid Woodhouse return 0; 1102ca89a517SDavid Woodhouse if (size > c->sector_size - jeb->used_size) { 1103ca89a517SDavid Woodhouse printk(KERN_CRIT "Dirty space 0x%x larger then used_size 0x%x (wasted 0x%x)\n", 1104ca89a517SDavid Woodhouse size, jeb->used_size, jeb->wasted_size); 1105ca89a517SDavid Woodhouse BUG(); 1106ca89a517SDavid Woodhouse } 1107ca89a517SDavid Woodhouse if (jeb->last_node && ref_obsolete(jeb->last_node)) { 1108ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 1109ca89a517SDavid Woodhouse jeb->last_node->__totlen += size; 1110ca89a517SDavid Woodhouse #endif 111168270995SDavid Woodhouse c->dirty_size += size; 111268270995SDavid Woodhouse c->free_size -= size; 111368270995SDavid Woodhouse jeb->dirty_size += size; 111468270995SDavid Woodhouse jeb->free_size -= size; 1115ca89a517SDavid Woodhouse } else { 1116ca89a517SDavid Woodhouse struct jffs2_raw_node_ref *ref; 1117ca89a517SDavid Woodhouse ref = jffs2_alloc_raw_node_ref(); 1118ca89a517SDavid Woodhouse if (!ref) 1119ca89a517SDavid Woodhouse return -ENOMEM; 1120ca89a517SDavid Woodhouse 1121ca89a517SDavid Woodhouse ref->flash_offset = jeb->offset + c->sector_size - jeb->free_size; 1122ca89a517SDavid Woodhouse ref->flash_offset |= REF_OBSOLETE; 1123ca89a517SDavid Woodhouse ref->next_in_ino = 0; 1124ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 1125ca89a517SDavid Woodhouse ref->__totlen = size; 1126ca89a517SDavid Woodhouse #endif 1127ca89a517SDavid Woodhouse 1128ca89a517SDavid Woodhouse jffs2_link_node_ref(c, jeb, ref, size); 1129ca89a517SDavid Woodhouse } 113068270995SDavid Woodhouse 113168270995SDavid Woodhouse return 0; 113268270995SDavid Woodhouse } 1133ca89a517SDavid Woodhouse 1134ca89a517SDavid Woodhouse /* Calculate totlen from surrounding nodes or eraseblock */ 1135ca89a517SDavid Woodhouse static inline uint32_t __ref_totlen(struct jffs2_sb_info *c, 1136ca89a517SDavid Woodhouse struct jffs2_eraseblock *jeb, 1137ca89a517SDavid Woodhouse struct jffs2_raw_node_ref *ref) 1138ca89a517SDavid Woodhouse { 1139ca89a517SDavid Woodhouse uint32_t ref_end; 1140ca89a517SDavid Woodhouse 1141ca89a517SDavid Woodhouse if (ref->next_phys) 1142ca89a517SDavid Woodhouse ref_end = ref_offset(ref->next_phys); 1143ca89a517SDavid Woodhouse else { 1144ca89a517SDavid Woodhouse if (!jeb) 1145ca89a517SDavid Woodhouse jeb = &c->blocks[ref->flash_offset / c->sector_size]; 1146ca89a517SDavid Woodhouse 1147ca89a517SDavid Woodhouse /* Last node in block. Use free_space */ 1148ca89a517SDavid Woodhouse if (ref != jeb->last_node) { 1149ca89a517SDavid Woodhouse printk(KERN_CRIT "ref %p @0x%08x is not jeb->last_node (%p @0x%08x)\n", 1150ca89a517SDavid Woodhouse ref, ref_offset(ref), jeb->last_node, jeb->last_node?ref_offset(jeb->last_node):0); 1151ca89a517SDavid Woodhouse BUG(); 1152ca89a517SDavid Woodhouse } 1153ca89a517SDavid Woodhouse ref_end = jeb->offset + c->sector_size - jeb->free_size; 1154ca89a517SDavid Woodhouse } 1155ca89a517SDavid Woodhouse return ref_end - ref_offset(ref); 1156ca89a517SDavid Woodhouse } 1157ca89a517SDavid Woodhouse 1158ca89a517SDavid Woodhouse uint32_t __jffs2_ref_totlen(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, 1159ca89a517SDavid Woodhouse struct jffs2_raw_node_ref *ref) 1160ca89a517SDavid Woodhouse { 1161ca89a517SDavid Woodhouse uint32_t ret; 1162ca89a517SDavid Woodhouse 1163ca89a517SDavid Woodhouse #if CONFIG_JFFS2_FS_DEBUG > 0 1164ca89a517SDavid Woodhouse if (jeb && jeb != &c->blocks[ref->flash_offset / c->sector_size]) { 1165ca89a517SDavid Woodhouse printk(KERN_CRIT "ref_totlen called with wrong block -- at 0x%08x instead of 0x%08x; ref 0x%08x\n", 1166ca89a517SDavid Woodhouse jeb->offset, c->blocks[ref->flash_offset / c->sector_size].offset, ref_offset(ref)); 1167ca89a517SDavid Woodhouse BUG(); 1168ca89a517SDavid Woodhouse } 1169ca89a517SDavid Woodhouse #endif 1170ca89a517SDavid Woodhouse 1171ca89a517SDavid Woodhouse ret = __ref_totlen(c, jeb, ref); 1172ca89a517SDavid Woodhouse #ifdef TEST_TOTLEN 1173ca89a517SDavid Woodhouse if (ret != ref->__totlen) { 1174ca89a517SDavid Woodhouse printk(KERN_CRIT "Totlen for ref at %p (0x%08x-0x%08x) miscalculated as 0x%x instead of %x\n", 1175ca89a517SDavid Woodhouse ref, ref_offset(ref), ref_offset(ref)+ref->__totlen, 1176ca89a517SDavid Woodhouse ret, ref->__totlen); 1177ca89a517SDavid Woodhouse if (ref->next_phys) { 1178ca89a517SDavid Woodhouse printk(KERN_CRIT "next_phys %p (0x%08x-0x%08x)\n", ref->next_phys, ref_offset(ref->next_phys), 1179ca89a517SDavid Woodhouse ref_offset(ref->next_phys)+ref->__totlen); 1180ca89a517SDavid Woodhouse } else 1181ca89a517SDavid Woodhouse printk(KERN_CRIT "No next_phys. jeb->last_node is %p\n", jeb->last_node); 1182ca89a517SDavid Woodhouse 1183ca89a517SDavid Woodhouse printk(KERN_CRIT "jeb->wasted_size %x, dirty_size %x, used_size %x, free_size %x\n", jeb->wasted_size, jeb->dirty_size, jeb->used_size, jeb->free_size); 1184ca89a517SDavid Woodhouse ret = ref->__totlen; 1185ca89a517SDavid Woodhouse if (!jeb) 1186ca89a517SDavid Woodhouse jeb = &c->blocks[ref->flash_offset / c->sector_size]; 1187ca89a517SDavid Woodhouse #if defined(JFFS2_DBG_DUMPS) || defined(JFFS2_DBG_PARANOIA_CHECKS) 1188ca89a517SDavid Woodhouse __jffs2_dbg_dump_node_refs_nolock(c, jeb); 1189ca89a517SDavid Woodhouse #endif 1190ca89a517SDavid Woodhouse WARN_ON(1); 1191ca89a517SDavid Woodhouse } 1192ca89a517SDavid Woodhouse #endif /* TEST_TOTLEN */ 1193ca89a517SDavid Woodhouse return ret; 1194ca89a517SDavid Woodhouse } 1195