12b27bdccSThomas Gleixner // SPDX-License-Identifier: GPL-2.0-only 21e51764aSArtem Bityutskiy /* 31e51764aSArtem Bityutskiy * This file is part of UBIFS. 41e51764aSArtem Bityutskiy * 51e51764aSArtem Bityutskiy * Copyright (C) 2006-2008 Nokia Corporation. 61e51764aSArtem Bityutskiy * Copyright (C) 2006, 2007 University of Szeged, Hungary 71e51764aSArtem Bityutskiy * 81e51764aSArtem Bityutskiy * Authors: Artem Bityutskiy (Битюцкий Артём) 91e51764aSArtem Bityutskiy * Adrian Hunter 101e51764aSArtem Bityutskiy * Zoltan Sogor 111e51764aSArtem Bityutskiy */ 121e51764aSArtem Bityutskiy 131e51764aSArtem Bityutskiy /* 141e51764aSArtem Bityutskiy * This file implements UBIFS I/O subsystem which provides various I/O-related 151e51764aSArtem Bityutskiy * helper functions (reading/writing/checking/validating nodes) and implements 161e51764aSArtem Bityutskiy * write-buffering support. Write buffers help to save space which otherwise 171e51764aSArtem Bityutskiy * would have been wasted for padding to the nearest minimal I/O unit boundary. 181e51764aSArtem Bityutskiy * Instead, data first goes to the write-buffer and is flushed when the 191e51764aSArtem Bityutskiy * buffer is full or when it is not used for some time (by timer). This is 201e51764aSArtem Bityutskiy * similar to the mechanism is used by JFFS2. 211e51764aSArtem Bityutskiy * 226c7f74f7SArtem Bityutskiy * UBIFS distinguishes between minimum write size (@c->min_io_size) and maximum 236c7f74f7SArtem Bityutskiy * write size (@c->max_write_size). The latter is the maximum amount of bytes 246c7f74f7SArtem Bityutskiy * the underlying flash is able to program at a time, and writing in 256c7f74f7SArtem Bityutskiy * @c->max_write_size units should presumably be faster. Obviously, 266c7f74f7SArtem Bityutskiy * @c->min_io_size <= @c->max_write_size. Write-buffers are of 276c7f74f7SArtem Bityutskiy * @c->max_write_size bytes in size for maximum performance. However, when a 286c7f74f7SArtem Bityutskiy * write-buffer is flushed, only the portion of it (aligned to @c->min_io_size 296c7f74f7SArtem Bityutskiy * boundary) which contains data is written, not the whole write-buffer, 306c7f74f7SArtem Bityutskiy * because this is more space-efficient. 316c7f74f7SArtem Bityutskiy * 326c7f74f7SArtem Bityutskiy * This optimization adds few complications to the code. Indeed, on the one 336c7f74f7SArtem Bityutskiy * hand, we want to write in optimal @c->max_write_size bytes chunks, which 346c7f74f7SArtem Bityutskiy * also means aligning writes at the @c->max_write_size bytes offsets. On the 356c7f74f7SArtem Bityutskiy * other hand, we do not want to waste space when synchronizing the write 366c7f74f7SArtem Bityutskiy * buffer, so during synchronization we writes in smaller chunks. And this makes 376c7f74f7SArtem Bityutskiy * the next write offset to be not aligned to @c->max_write_size bytes. So the 386c7f74f7SArtem Bityutskiy * have to make sure that the write-buffer offset (@wbuf->offs) becomes aligned 396c7f74f7SArtem Bityutskiy * to @c->max_write_size bytes again. We do this by temporarily shrinking 406c7f74f7SArtem Bityutskiy * write-buffer size (@wbuf->size). 416c7f74f7SArtem Bityutskiy * 421e51764aSArtem Bityutskiy * Write-buffers are defined by 'struct ubifs_wbuf' objects and protected by 431e51764aSArtem Bityutskiy * mutexes defined inside these objects. Since sometimes upper-level code 441e51764aSArtem Bityutskiy * has to lock the write-buffer (e.g. journal space reservation code), many 451e51764aSArtem Bityutskiy * functions related to write-buffers have "nolock" suffix which means that the 461e51764aSArtem Bityutskiy * caller has to lock the write-buffer before calling this function. 471e51764aSArtem Bityutskiy * 481e51764aSArtem Bityutskiy * UBIFS stores nodes at 64 bit-aligned addresses. If the node length is not 491e51764aSArtem Bityutskiy * aligned, UBIFS starts the next node from the aligned address, and the padded 501e51764aSArtem Bityutskiy * bytes may contain any rubbish. In other words, UBIFS does not put padding 511e51764aSArtem Bityutskiy * bytes in those small gaps. Common headers of nodes store real node lengths, 521e51764aSArtem Bityutskiy * not aligned lengths. Indexing nodes also store real lengths in branches. 531e51764aSArtem Bityutskiy * 541e51764aSArtem Bityutskiy * UBIFS uses padding when it pads to the next min. I/O unit. In this case it 551e51764aSArtem Bityutskiy * uses padding nodes or padding bytes, if the padding node does not fit. 561e51764aSArtem Bityutskiy * 576c7f74f7SArtem Bityutskiy * All UBIFS nodes are protected by CRC checksums and UBIFS checks CRC when 586c7f74f7SArtem Bityutskiy * they are read from the flash media. 591e51764aSArtem Bityutskiy */ 601e51764aSArtem Bityutskiy 611e51764aSArtem Bityutskiy #include <linux/crc32.h> 625a0e3ad6STejun Heo #include <linux/slab.h> 631e51764aSArtem Bityutskiy #include "ubifs.h" 641e51764aSArtem Bityutskiy 651e51764aSArtem Bityutskiy /** 66ff46d7b3SAdrian Hunter * ubifs_ro_mode - switch UBIFS to read read-only mode. 67ff46d7b3SAdrian Hunter * @c: UBIFS file-system description object 68ff46d7b3SAdrian Hunter * @err: error code which is the reason of switching to R/O mode 69ff46d7b3SAdrian Hunter */ 70ff46d7b3SAdrian Hunter void ubifs_ro_mode(struct ubifs_info *c, int err) 71ff46d7b3SAdrian Hunter { 722680d722SArtem Bityutskiy if (!c->ro_error) { 732680d722SArtem Bityutskiy c->ro_error = 1; 74ccb3eba7SArtem Bityutskiy c->no_chk_data_crc = 0; 751751e8a6SLinus Torvalds c->vfs_sb->s_flags |= SB_RDONLY; 76235c362bSSheng Yong ubifs_warn(c, "switched to read-only mode, error %d", err); 77d033c98bSArtem Bityutskiy dump_stack(); 78ff46d7b3SAdrian Hunter } 79ff46d7b3SAdrian Hunter } 80ff46d7b3SAdrian Hunter 8183cef708SArtem Bityutskiy /* 8283cef708SArtem Bityutskiy * Below are simple wrappers over UBI I/O functions which include some 8383cef708SArtem Bityutskiy * additional checks and UBIFS debugging stuff. See corresponding UBI function 8483cef708SArtem Bityutskiy * for more information. 8583cef708SArtem Bityutskiy */ 8683cef708SArtem Bityutskiy 8783cef708SArtem Bityutskiy int ubifs_leb_read(const struct ubifs_info *c, int lnum, void *buf, int offs, 8883cef708SArtem Bityutskiy int len, int even_ebadmsg) 8983cef708SArtem Bityutskiy { 9083cef708SArtem Bityutskiy int err; 9183cef708SArtem Bityutskiy 9283cef708SArtem Bityutskiy err = ubi_read(c->ubi, lnum, buf, offs, len); 9383cef708SArtem Bityutskiy /* 9483cef708SArtem Bityutskiy * In case of %-EBADMSG print the error message only if the 9583cef708SArtem Bityutskiy * @even_ebadmsg is true. 9683cef708SArtem Bityutskiy */ 9783cef708SArtem Bityutskiy if (err && (err != -EBADMSG || even_ebadmsg)) { 98235c362bSSheng Yong ubifs_err(c, "reading %d bytes from LEB %d:%d failed, error %d", 9983cef708SArtem Bityutskiy len, lnum, offs, err); 1007c46d0aeSArtem Bityutskiy dump_stack(); 10183cef708SArtem Bityutskiy } 10283cef708SArtem Bityutskiy return err; 10383cef708SArtem Bityutskiy } 10483cef708SArtem Bityutskiy 10583cef708SArtem Bityutskiy int ubifs_leb_write(struct ubifs_info *c, int lnum, const void *buf, int offs, 106b36a261eSRichard Weinberger int len) 10783cef708SArtem Bityutskiy { 10883cef708SArtem Bityutskiy int err; 10983cef708SArtem Bityutskiy 1106eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 11183cef708SArtem Bityutskiy if (c->ro_error) 11283cef708SArtem Bityutskiy return -EROFS; 11383cef708SArtem Bityutskiy if (!dbg_is_tst_rcvry(c)) 114b36a261eSRichard Weinberger err = ubi_leb_write(c->ubi, lnum, buf, offs, len); 11583cef708SArtem Bityutskiy else 116b36a261eSRichard Weinberger err = dbg_leb_write(c, lnum, buf, offs, len); 11783cef708SArtem Bityutskiy if (err) { 118235c362bSSheng Yong ubifs_err(c, "writing %d bytes to LEB %d:%d failed, error %d", 11983cef708SArtem Bityutskiy len, lnum, offs, err); 12083cef708SArtem Bityutskiy ubifs_ro_mode(c, err); 1217c46d0aeSArtem Bityutskiy dump_stack(); 12283cef708SArtem Bityutskiy } 12383cef708SArtem Bityutskiy return err; 12483cef708SArtem Bityutskiy } 12583cef708SArtem Bityutskiy 126b36a261eSRichard Weinberger int ubifs_leb_change(struct ubifs_info *c, int lnum, const void *buf, int len) 12783cef708SArtem Bityutskiy { 12883cef708SArtem Bityutskiy int err; 12983cef708SArtem Bityutskiy 1306eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 13183cef708SArtem Bityutskiy if (c->ro_error) 13283cef708SArtem Bityutskiy return -EROFS; 13383cef708SArtem Bityutskiy if (!dbg_is_tst_rcvry(c)) 134b36a261eSRichard Weinberger err = ubi_leb_change(c->ubi, lnum, buf, len); 13583cef708SArtem Bityutskiy else 136b36a261eSRichard Weinberger err = dbg_leb_change(c, lnum, buf, len); 13783cef708SArtem Bityutskiy if (err) { 138235c362bSSheng Yong ubifs_err(c, "changing %d bytes in LEB %d failed, error %d", 13983cef708SArtem Bityutskiy len, lnum, err); 14083cef708SArtem Bityutskiy ubifs_ro_mode(c, err); 1417c46d0aeSArtem Bityutskiy dump_stack(); 14283cef708SArtem Bityutskiy } 14383cef708SArtem Bityutskiy return err; 14483cef708SArtem Bityutskiy } 14583cef708SArtem Bityutskiy 14683cef708SArtem Bityutskiy int ubifs_leb_unmap(struct ubifs_info *c, int lnum) 14783cef708SArtem Bityutskiy { 14883cef708SArtem Bityutskiy int err; 14983cef708SArtem Bityutskiy 1506eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 15183cef708SArtem Bityutskiy if (c->ro_error) 15283cef708SArtem Bityutskiy return -EROFS; 15383cef708SArtem Bityutskiy if (!dbg_is_tst_rcvry(c)) 15483cef708SArtem Bityutskiy err = ubi_leb_unmap(c->ubi, lnum); 15583cef708SArtem Bityutskiy else 156f57cb188SArtem Bityutskiy err = dbg_leb_unmap(c, lnum); 15783cef708SArtem Bityutskiy if (err) { 158235c362bSSheng Yong ubifs_err(c, "unmap LEB %d failed, error %d", lnum, err); 15983cef708SArtem Bityutskiy ubifs_ro_mode(c, err); 1607c46d0aeSArtem Bityutskiy dump_stack(); 16183cef708SArtem Bityutskiy } 16283cef708SArtem Bityutskiy return err; 16383cef708SArtem Bityutskiy } 16483cef708SArtem Bityutskiy 165b36a261eSRichard Weinberger int ubifs_leb_map(struct ubifs_info *c, int lnum) 16683cef708SArtem Bityutskiy { 16783cef708SArtem Bityutskiy int err; 16883cef708SArtem Bityutskiy 1696eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 17083cef708SArtem Bityutskiy if (c->ro_error) 17183cef708SArtem Bityutskiy return -EROFS; 17283cef708SArtem Bityutskiy if (!dbg_is_tst_rcvry(c)) 173b36a261eSRichard Weinberger err = ubi_leb_map(c->ubi, lnum); 17483cef708SArtem Bityutskiy else 175b36a261eSRichard Weinberger err = dbg_leb_map(c, lnum); 17683cef708SArtem Bityutskiy if (err) { 177235c362bSSheng Yong ubifs_err(c, "mapping LEB %d failed, error %d", lnum, err); 17883cef708SArtem Bityutskiy ubifs_ro_mode(c, err); 1797c46d0aeSArtem Bityutskiy dump_stack(); 18083cef708SArtem Bityutskiy } 18183cef708SArtem Bityutskiy return err; 18283cef708SArtem Bityutskiy } 18383cef708SArtem Bityutskiy 18483cef708SArtem Bityutskiy int ubifs_is_mapped(const struct ubifs_info *c, int lnum) 18583cef708SArtem Bityutskiy { 18683cef708SArtem Bityutskiy int err; 18783cef708SArtem Bityutskiy 18883cef708SArtem Bityutskiy err = ubi_is_mapped(c->ubi, lnum); 18983cef708SArtem Bityutskiy if (err < 0) { 190235c362bSSheng Yong ubifs_err(c, "ubi_is_mapped failed for LEB %d, error %d", 19183cef708SArtem Bityutskiy lnum, err); 1927c46d0aeSArtem Bityutskiy dump_stack(); 19383cef708SArtem Bityutskiy } 19483cef708SArtem Bityutskiy return err; 19583cef708SArtem Bityutskiy } 19683cef708SArtem Bityutskiy 197ff46d7b3SAdrian Hunter /** 1981e51764aSArtem Bityutskiy * ubifs_check_node - check node. 1991e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 2001e51764aSArtem Bityutskiy * @buf: node to check 201*a33e30a0SZhihao Cheng * @len: node length 2021e51764aSArtem Bityutskiy * @lnum: logical eraseblock number 2031e51764aSArtem Bityutskiy * @offs: offset within the logical eraseblock 2041e51764aSArtem Bityutskiy * @quiet: print no messages 2056f7ab6d4SArtem Bityutskiy * @must_chk_crc: indicates whether to always check the CRC 2061e51764aSArtem Bityutskiy * 2071e51764aSArtem Bityutskiy * This function checks node magic number and CRC checksum. This function also 2081e51764aSArtem Bityutskiy * validates node length to prevent UBIFS from becoming crazy when an attacker 2091e51764aSArtem Bityutskiy * feeds it a file-system image with incorrect nodes. For example, too large 2101e51764aSArtem Bityutskiy * node length in the common header could cause UBIFS to read memory outside of 2111e51764aSArtem Bityutskiy * allocated buffer when checking the CRC checksum. 2121e51764aSArtem Bityutskiy * 2136f7ab6d4SArtem Bityutskiy * This function may skip data nodes CRC checking if @c->no_chk_data_crc is 2146f7ab6d4SArtem Bityutskiy * true, which is controlled by corresponding UBIFS mount option. However, if 2156f7ab6d4SArtem Bityutskiy * @must_chk_crc is true, then @c->no_chk_data_crc is ignored and CRC is 21618d1d7fbSArtem Bityutskiy * checked. Similarly, if @c->mounting or @c->remounting_rw is true (we are 21718d1d7fbSArtem Bityutskiy * mounting or re-mounting to R/W mode), @c->no_chk_data_crc is ignored and CRC 21818d1d7fbSArtem Bityutskiy * is checked. This is because during mounting or re-mounting from R/O mode to 21918d1d7fbSArtem Bityutskiy * R/W mode we may read journal nodes (when replying the journal or doing the 22018d1d7fbSArtem Bityutskiy * recovery) and the journal nodes may potentially be corrupted, so checking is 22118d1d7fbSArtem Bityutskiy * required. 2226f7ab6d4SArtem Bityutskiy * 2236f7ab6d4SArtem Bityutskiy * This function returns zero in case of success and %-EUCLEAN in case of bad 2246f7ab6d4SArtem Bityutskiy * CRC or magic. 2251e51764aSArtem Bityutskiy */ 226*a33e30a0SZhihao Cheng int ubifs_check_node(const struct ubifs_info *c, const void *buf, int len, 227*a33e30a0SZhihao Cheng int lnum, int offs, int quiet, int must_chk_crc) 2281e51764aSArtem Bityutskiy { 229c8be0975SZhihao Cheng int err = -EINVAL, type, node_len; 2301e51764aSArtem Bityutskiy uint32_t crc, node_crc, magic; 2311e51764aSArtem Bityutskiy const struct ubifs_ch *ch = buf; 2321e51764aSArtem Bityutskiy 2336eb61d58SRichard Weinberger ubifs_assert(c, lnum >= 0 && lnum < c->leb_cnt && offs >= 0); 2346eb61d58SRichard Weinberger ubifs_assert(c, !(offs & 7) && offs < c->leb_size); 2351e51764aSArtem Bityutskiy 2361e51764aSArtem Bityutskiy magic = le32_to_cpu(ch->magic); 2371e51764aSArtem Bityutskiy if (magic != UBIFS_NODE_MAGIC) { 2381e51764aSArtem Bityutskiy if (!quiet) 239235c362bSSheng Yong ubifs_err(c, "bad magic %#08x, expected %#08x", 2401e51764aSArtem Bityutskiy magic, UBIFS_NODE_MAGIC); 2411e51764aSArtem Bityutskiy err = -EUCLEAN; 2421e51764aSArtem Bityutskiy goto out; 2431e51764aSArtem Bityutskiy } 2441e51764aSArtem Bityutskiy 2451e51764aSArtem Bityutskiy type = ch->node_type; 2461e51764aSArtem Bityutskiy if (type < 0 || type >= UBIFS_NODE_TYPES_CNT) { 2471e51764aSArtem Bityutskiy if (!quiet) 248235c362bSSheng Yong ubifs_err(c, "bad node type %d", type); 2491e51764aSArtem Bityutskiy goto out; 2501e51764aSArtem Bityutskiy } 2511e51764aSArtem Bityutskiy 2521e51764aSArtem Bityutskiy node_len = le32_to_cpu(ch->len); 2531e51764aSArtem Bityutskiy if (node_len + offs > c->leb_size) 2541e51764aSArtem Bityutskiy goto out_len; 2551e51764aSArtem Bityutskiy 2561e51764aSArtem Bityutskiy if (c->ranges[type].max_len == 0) { 2571e51764aSArtem Bityutskiy if (node_len != c->ranges[type].len) 2581e51764aSArtem Bityutskiy goto out_len; 2591e51764aSArtem Bityutskiy } else if (node_len < c->ranges[type].min_len || 2601e51764aSArtem Bityutskiy node_len > c->ranges[type].max_len) 2611e51764aSArtem Bityutskiy goto out_len; 2621e51764aSArtem Bityutskiy 26318d1d7fbSArtem Bityutskiy if (!must_chk_crc && type == UBIFS_DATA_NODE && !c->mounting && 26418d1d7fbSArtem Bityutskiy !c->remounting_rw && c->no_chk_data_crc) 2652953e73fSAdrian Hunter return 0; 2662953e73fSAdrian Hunter 2671e51764aSArtem Bityutskiy crc = crc32(UBIFS_CRC32_INIT, buf + 8, node_len - 8); 2681e51764aSArtem Bityutskiy node_crc = le32_to_cpu(ch->crc); 2691e51764aSArtem Bityutskiy if (crc != node_crc) { 2701e51764aSArtem Bityutskiy if (!quiet) 271235c362bSSheng Yong ubifs_err(c, "bad CRC: calculated %#08x, read %#08x", 2721e51764aSArtem Bityutskiy crc, node_crc); 2731e51764aSArtem Bityutskiy err = -EUCLEAN; 2741e51764aSArtem Bityutskiy goto out; 2751e51764aSArtem Bityutskiy } 2761e51764aSArtem Bityutskiy 2771e51764aSArtem Bityutskiy return 0; 2781e51764aSArtem Bityutskiy 2791e51764aSArtem Bityutskiy out_len: 2801e51764aSArtem Bityutskiy if (!quiet) 281235c362bSSheng Yong ubifs_err(c, "bad node length %d", node_len); 2821e51764aSArtem Bityutskiy out: 2831e51764aSArtem Bityutskiy if (!quiet) { 284235c362bSSheng Yong ubifs_err(c, "bad node at LEB %d:%d", lnum, offs); 285*a33e30a0SZhihao Cheng ubifs_dump_node(c, buf, len); 2867c46d0aeSArtem Bityutskiy dump_stack(); 2871e51764aSArtem Bityutskiy } 2881e51764aSArtem Bityutskiy return err; 2891e51764aSArtem Bityutskiy } 2901e51764aSArtem Bityutskiy 2911e51764aSArtem Bityutskiy /** 2921e51764aSArtem Bityutskiy * ubifs_pad - pad flash space. 2931e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 2941e51764aSArtem Bityutskiy * @buf: buffer to put padding to 2951e51764aSArtem Bityutskiy * @pad: how many bytes to pad 2961e51764aSArtem Bityutskiy * 2971e51764aSArtem Bityutskiy * The flash media obliges us to write only in chunks of %c->min_io_size and 2981e51764aSArtem Bityutskiy * when we have to write less data we add padding node to the write-buffer and 2991e51764aSArtem Bityutskiy * pad it to the next minimal I/O unit's boundary. Padding nodes help when the 3001e51764aSArtem Bityutskiy * media is being scanned. If the amount of wasted space is not enough to fit a 3011e51764aSArtem Bityutskiy * padding node which takes %UBIFS_PAD_NODE_SZ bytes, we write padding bytes 3021e51764aSArtem Bityutskiy * pattern (%UBIFS_PADDING_BYTE). 3031e51764aSArtem Bityutskiy * 3041e51764aSArtem Bityutskiy * Padding nodes are also used to fill gaps when the "commit-in-gaps" method is 3051e51764aSArtem Bityutskiy * used. 3061e51764aSArtem Bityutskiy */ 3071e51764aSArtem Bityutskiy void ubifs_pad(const struct ubifs_info *c, void *buf, int pad) 3081e51764aSArtem Bityutskiy { 3091e51764aSArtem Bityutskiy uint32_t crc; 3101e51764aSArtem Bityutskiy 31120f14311SRichard Weinberger ubifs_assert(c, pad >= 0); 3121e51764aSArtem Bityutskiy 3131e51764aSArtem Bityutskiy if (pad >= UBIFS_PAD_NODE_SZ) { 3141e51764aSArtem Bityutskiy struct ubifs_ch *ch = buf; 3151e51764aSArtem Bityutskiy struct ubifs_pad_node *pad_node = buf; 3161e51764aSArtem Bityutskiy 3171e51764aSArtem Bityutskiy ch->magic = cpu_to_le32(UBIFS_NODE_MAGIC); 3181e51764aSArtem Bityutskiy ch->node_type = UBIFS_PAD_NODE; 3191e51764aSArtem Bityutskiy ch->group_type = UBIFS_NO_NODE_GROUP; 3201e51764aSArtem Bityutskiy ch->padding[0] = ch->padding[1] = 0; 3211e51764aSArtem Bityutskiy ch->sqnum = 0; 3221e51764aSArtem Bityutskiy ch->len = cpu_to_le32(UBIFS_PAD_NODE_SZ); 3231e51764aSArtem Bityutskiy pad -= UBIFS_PAD_NODE_SZ; 3241e51764aSArtem Bityutskiy pad_node->pad_len = cpu_to_le32(pad); 3251e51764aSArtem Bityutskiy crc = crc32(UBIFS_CRC32_INIT, buf + 8, UBIFS_PAD_NODE_SZ - 8); 3261e51764aSArtem Bityutskiy ch->crc = cpu_to_le32(crc); 3271e51764aSArtem Bityutskiy memset(buf + UBIFS_PAD_NODE_SZ, 0, pad); 3281e51764aSArtem Bityutskiy } else if (pad > 0) 3291e51764aSArtem Bityutskiy /* Too little space, padding node won't fit */ 3301e51764aSArtem Bityutskiy memset(buf, UBIFS_PADDING_BYTE, pad); 3311e51764aSArtem Bityutskiy } 3321e51764aSArtem Bityutskiy 3331e51764aSArtem Bityutskiy /** 3341e51764aSArtem Bityutskiy * next_sqnum - get next sequence number. 3351e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 3361e51764aSArtem Bityutskiy */ 3371e51764aSArtem Bityutskiy static unsigned long long next_sqnum(struct ubifs_info *c) 3381e51764aSArtem Bityutskiy { 3391e51764aSArtem Bityutskiy unsigned long long sqnum; 3401e51764aSArtem Bityutskiy 3411e51764aSArtem Bityutskiy spin_lock(&c->cnt_lock); 3421e51764aSArtem Bityutskiy sqnum = ++c->max_sqnum; 3431e51764aSArtem Bityutskiy spin_unlock(&c->cnt_lock); 3441e51764aSArtem Bityutskiy 3451e51764aSArtem Bityutskiy if (unlikely(sqnum >= SQNUM_WARN_WATERMARK)) { 3461e51764aSArtem Bityutskiy if (sqnum >= SQNUM_WATERMARK) { 347235c362bSSheng Yong ubifs_err(c, "sequence number overflow %llu, end of life", 3481e51764aSArtem Bityutskiy sqnum); 3491e51764aSArtem Bityutskiy ubifs_ro_mode(c, -EINVAL); 3501e51764aSArtem Bityutskiy } 351235c362bSSheng Yong ubifs_warn(c, "running out of sequence numbers, end of life soon"); 3521e51764aSArtem Bityutskiy } 3531e51764aSArtem Bityutskiy 3541e51764aSArtem Bityutskiy return sqnum; 3551e51764aSArtem Bityutskiy } 3561e51764aSArtem Bityutskiy 357dead9726SSascha Hauer void ubifs_init_node(struct ubifs_info *c, void *node, int len, int pad) 358dead9726SSascha Hauer { 359dead9726SSascha Hauer struct ubifs_ch *ch = node; 360dead9726SSascha Hauer unsigned long long sqnum = next_sqnum(c); 361dead9726SSascha Hauer 362dead9726SSascha Hauer ubifs_assert(c, len >= UBIFS_CH_SZ); 363dead9726SSascha Hauer 364dead9726SSascha Hauer ch->magic = cpu_to_le32(UBIFS_NODE_MAGIC); 365dead9726SSascha Hauer ch->len = cpu_to_le32(len); 366dead9726SSascha Hauer ch->group_type = UBIFS_NO_NODE_GROUP; 367dead9726SSascha Hauer ch->sqnum = cpu_to_le64(sqnum); 368dead9726SSascha Hauer ch->padding[0] = ch->padding[1] = 0; 369dead9726SSascha Hauer 370dead9726SSascha Hauer if (pad) { 371dead9726SSascha Hauer len = ALIGN(len, 8); 372dead9726SSascha Hauer pad = ALIGN(len, c->min_io_size) - len; 373dead9726SSascha Hauer ubifs_pad(c, node + len, pad); 374dead9726SSascha Hauer } 375dead9726SSascha Hauer } 376dead9726SSascha Hauer 377dead9726SSascha Hauer void ubifs_crc_node(struct ubifs_info *c, void *node, int len) 378dead9726SSascha Hauer { 379dead9726SSascha Hauer struct ubifs_ch *ch = node; 380dead9726SSascha Hauer uint32_t crc; 381dead9726SSascha Hauer 382dead9726SSascha Hauer crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); 383dead9726SSascha Hauer ch->crc = cpu_to_le32(crc); 384dead9726SSascha Hauer } 385dead9726SSascha Hauer 3861e51764aSArtem Bityutskiy /** 387a384b47eSSascha Hauer * ubifs_prepare_node_hmac - prepare node to be written to flash. 388a384b47eSSascha Hauer * @c: UBIFS file-system description object 389a384b47eSSascha Hauer * @node: the node to pad 390a384b47eSSascha Hauer * @len: node length 391a384b47eSSascha Hauer * @hmac_offs: offset of the HMAC in the node 392a384b47eSSascha Hauer * @pad: if the buffer has to be padded 393a384b47eSSascha Hauer * 394a384b47eSSascha Hauer * This function prepares node at @node to be written to the media - it 395a384b47eSSascha Hauer * calculates node CRC, fills the common header, and adds proper padding up to 396a384b47eSSascha Hauer * the next minimum I/O unit if @pad is not zero. if @hmac_offs is positive then 397a384b47eSSascha Hauer * a HMAC is inserted into the node at the given offset. 398a384b47eSSascha Hauer * 399a384b47eSSascha Hauer * This function returns 0 for success or a negative error code otherwise. 400a384b47eSSascha Hauer */ 401a384b47eSSascha Hauer int ubifs_prepare_node_hmac(struct ubifs_info *c, void *node, int len, 402a384b47eSSascha Hauer int hmac_offs, int pad) 403a384b47eSSascha Hauer { 404a384b47eSSascha Hauer int err; 405a384b47eSSascha Hauer 406a384b47eSSascha Hauer ubifs_init_node(c, node, len, pad); 407a384b47eSSascha Hauer 408a384b47eSSascha Hauer if (hmac_offs > 0) { 409a384b47eSSascha Hauer err = ubifs_node_insert_hmac(c, node, len, hmac_offs); 410a384b47eSSascha Hauer if (err) 411a384b47eSSascha Hauer return err; 412a384b47eSSascha Hauer } 413a384b47eSSascha Hauer 414a384b47eSSascha Hauer ubifs_crc_node(c, node, len); 415a384b47eSSascha Hauer 416a384b47eSSascha Hauer return 0; 417a384b47eSSascha Hauer } 418a384b47eSSascha Hauer 419a384b47eSSascha Hauer /** 4201e51764aSArtem Bityutskiy * ubifs_prepare_node - prepare node to be written to flash. 4211e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 4221e51764aSArtem Bityutskiy * @node: the node to pad 4231e51764aSArtem Bityutskiy * @len: node length 4241e51764aSArtem Bityutskiy * @pad: if the buffer has to be padded 4251e51764aSArtem Bityutskiy * 4261e51764aSArtem Bityutskiy * This function prepares node at @node to be written to the media - it 4271e51764aSArtem Bityutskiy * calculates node CRC, fills the common header, and adds proper padding up to 4281e51764aSArtem Bityutskiy * the next minimum I/O unit if @pad is not zero. 4291e51764aSArtem Bityutskiy */ 4301e51764aSArtem Bityutskiy void ubifs_prepare_node(struct ubifs_info *c, void *node, int len, int pad) 4311e51764aSArtem Bityutskiy { 432a384b47eSSascha Hauer /* 433a384b47eSSascha Hauer * Deliberately ignore return value since this function can only fail 434a384b47eSSascha Hauer * when a hmac offset is given. 435a384b47eSSascha Hauer */ 436a384b47eSSascha Hauer ubifs_prepare_node_hmac(c, node, len, 0, pad); 4371e51764aSArtem Bityutskiy } 4381e51764aSArtem Bityutskiy 4391e51764aSArtem Bityutskiy /** 4401e51764aSArtem Bityutskiy * ubifs_prep_grp_node - prepare node of a group to be written to flash. 4411e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 4421e51764aSArtem Bityutskiy * @node: the node to pad 4431e51764aSArtem Bityutskiy * @len: node length 4441e51764aSArtem Bityutskiy * @last: indicates the last node of the group 4451e51764aSArtem Bityutskiy * 4461e51764aSArtem Bityutskiy * This function prepares node at @node to be written to the media - it 4471e51764aSArtem Bityutskiy * calculates node CRC and fills the common header. 4481e51764aSArtem Bityutskiy */ 4491e51764aSArtem Bityutskiy void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last) 4501e51764aSArtem Bityutskiy { 4511e51764aSArtem Bityutskiy uint32_t crc; 4521e51764aSArtem Bityutskiy struct ubifs_ch *ch = node; 4531e51764aSArtem Bityutskiy unsigned long long sqnum = next_sqnum(c); 4541e51764aSArtem Bityutskiy 4556eb61d58SRichard Weinberger ubifs_assert(c, len >= UBIFS_CH_SZ); 4561e51764aSArtem Bityutskiy 4571e51764aSArtem Bityutskiy ch->magic = cpu_to_le32(UBIFS_NODE_MAGIC); 4581e51764aSArtem Bityutskiy ch->len = cpu_to_le32(len); 4591e51764aSArtem Bityutskiy if (last) 4601e51764aSArtem Bityutskiy ch->group_type = UBIFS_LAST_OF_NODE_GROUP; 4611e51764aSArtem Bityutskiy else 4621e51764aSArtem Bityutskiy ch->group_type = UBIFS_IN_NODE_GROUP; 4631e51764aSArtem Bityutskiy ch->sqnum = cpu_to_le64(sqnum); 4641e51764aSArtem Bityutskiy ch->padding[0] = ch->padding[1] = 0; 4651e51764aSArtem Bityutskiy crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); 4661e51764aSArtem Bityutskiy ch->crc = cpu_to_le32(crc); 4671e51764aSArtem Bityutskiy } 4681e51764aSArtem Bityutskiy 4691e51764aSArtem Bityutskiy /** 4701e51764aSArtem Bityutskiy * wbuf_timer_callback - write-buffer timer callback function. 47139274a1eSFabian Frederick * @timer: timer data (write-buffer descriptor) 4721e51764aSArtem Bityutskiy * 4731e51764aSArtem Bityutskiy * This function is called when the write-buffer timer expires. 4741e51764aSArtem Bityutskiy */ 475f2c5dbd7SArtem Bityutskiy static enum hrtimer_restart wbuf_timer_callback_nolock(struct hrtimer *timer) 4761e51764aSArtem Bityutskiy { 477f2c5dbd7SArtem Bityutskiy struct ubifs_wbuf *wbuf = container_of(timer, struct ubifs_wbuf, timer); 4781e51764aSArtem Bityutskiy 47977a7ae58SArtem Bityutskiy dbg_io("jhead %s", dbg_jhead(wbuf->jhead)); 4801e51764aSArtem Bityutskiy wbuf->need_sync = 1; 4811e51764aSArtem Bityutskiy wbuf->c->need_wbuf_sync = 1; 4821e51764aSArtem Bityutskiy ubifs_wake_up_bgt(wbuf->c); 483f2c5dbd7SArtem Bityutskiy return HRTIMER_NORESTART; 4841e51764aSArtem Bityutskiy } 4851e51764aSArtem Bityutskiy 4861e51764aSArtem Bityutskiy /** 4871e51764aSArtem Bityutskiy * new_wbuf_timer - start new write-buffer timer. 4886eb61d58SRichard Weinberger * @c: UBIFS file-system description object 4891e51764aSArtem Bityutskiy * @wbuf: write-buffer descriptor 4901e51764aSArtem Bityutskiy */ 4916eb61d58SRichard Weinberger static void new_wbuf_timer_nolock(struct ubifs_info *c, struct ubifs_wbuf *wbuf) 4921e51764aSArtem Bityutskiy { 4931b7fc2c0SRafał Miłecki ktime_t softlimit = ms_to_ktime(dirty_writeback_interval * 10); 4941b7fc2c0SRafał Miłecki unsigned long long delta = dirty_writeback_interval; 495854826c9SRafał Miłecki 4961b7fc2c0SRafał Miłecki /* centi to milli, milli to nano, then 10% */ 4971b7fc2c0SRafał Miłecki delta *= 10ULL * NSEC_PER_MSEC / 10ULL; 498854826c9SRafał Miłecki 4996eb61d58SRichard Weinberger ubifs_assert(c, !hrtimer_active(&wbuf->timer)); 5006eb61d58SRichard Weinberger ubifs_assert(c, delta <= ULONG_MAX); 5011e51764aSArtem Bityutskiy 5020b335b9dSArtem Bityutskiy if (wbuf->no_timer) 5031e51764aSArtem Bityutskiy return; 50477a7ae58SArtem Bityutskiy dbg_io("set timer for jhead %s, %llu-%llu millisecs", 50577a7ae58SArtem Bityutskiy dbg_jhead(wbuf->jhead), 506854826c9SRafał Miłecki div_u64(ktime_to_ns(softlimit), USEC_PER_SEC), 507854826c9SRafał Miłecki div_u64(ktime_to_ns(softlimit) + delta, USEC_PER_SEC)); 508854826c9SRafał Miłecki hrtimer_start_range_ns(&wbuf->timer, softlimit, delta, 509f2c5dbd7SArtem Bityutskiy HRTIMER_MODE_REL); 5101e51764aSArtem Bityutskiy } 5111e51764aSArtem Bityutskiy 5121e51764aSArtem Bityutskiy /** 5131e51764aSArtem Bityutskiy * cancel_wbuf_timer - cancel write-buffer timer. 5141e51764aSArtem Bityutskiy * @wbuf: write-buffer descriptor 5151e51764aSArtem Bityutskiy */ 5161e51764aSArtem Bityutskiy static void cancel_wbuf_timer_nolock(struct ubifs_wbuf *wbuf) 5171e51764aSArtem Bityutskiy { 5180b335b9dSArtem Bityutskiy if (wbuf->no_timer) 5190b335b9dSArtem Bityutskiy return; 5201e51764aSArtem Bityutskiy wbuf->need_sync = 0; 521f2c5dbd7SArtem Bityutskiy hrtimer_cancel(&wbuf->timer); 5221e51764aSArtem Bityutskiy } 5231e51764aSArtem Bityutskiy 5241e51764aSArtem Bityutskiy /** 5251e51764aSArtem Bityutskiy * ubifs_wbuf_sync_nolock - synchronize write-buffer. 5261e51764aSArtem Bityutskiy * @wbuf: write-buffer to synchronize 5271e51764aSArtem Bityutskiy * 5281e51764aSArtem Bityutskiy * This function synchronizes write-buffer @buf and returns zero in case of 5291e51764aSArtem Bityutskiy * success or a negative error code in case of failure. 5306c7f74f7SArtem Bityutskiy * 5316c7f74f7SArtem Bityutskiy * Note, although write-buffers are of @c->max_write_size, this function does 5326c7f74f7SArtem Bityutskiy * not necessarily writes all @c->max_write_size bytes to the flash. Instead, 5336c7f74f7SArtem Bityutskiy * if the write-buffer is only partially filled with data, only the used part 5346c7f74f7SArtem Bityutskiy * of the write-buffer (aligned on @c->min_io_size boundary) is synchronized. 5356c7f74f7SArtem Bityutskiy * This way we waste less space. 5361e51764aSArtem Bityutskiy */ 5371e51764aSArtem Bityutskiy int ubifs_wbuf_sync_nolock(struct ubifs_wbuf *wbuf) 5381e51764aSArtem Bityutskiy { 5391e51764aSArtem Bityutskiy struct ubifs_info *c = wbuf->c; 5406c7f74f7SArtem Bityutskiy int err, dirt, sync_len; 5411e51764aSArtem Bityutskiy 5421e51764aSArtem Bityutskiy cancel_wbuf_timer_nolock(wbuf); 5431e51764aSArtem Bityutskiy if (!wbuf->used || wbuf->lnum == -1) 5441e51764aSArtem Bityutskiy /* Write-buffer is empty or not seeked */ 5451e51764aSArtem Bityutskiy return 0; 5461e51764aSArtem Bityutskiy 54777a7ae58SArtem Bityutskiy dbg_io("LEB %d:%d, %d bytes, jhead %s", 54877a7ae58SArtem Bityutskiy wbuf->lnum, wbuf->offs, wbuf->used, dbg_jhead(wbuf->jhead)); 5496eb61d58SRichard Weinberger ubifs_assert(c, !(wbuf->avail & 7)); 5506eb61d58SRichard Weinberger ubifs_assert(c, wbuf->offs + wbuf->size <= c->leb_size); 5516eb61d58SRichard Weinberger ubifs_assert(c, wbuf->size >= c->min_io_size); 5526eb61d58SRichard Weinberger ubifs_assert(c, wbuf->size <= c->max_write_size); 5536eb61d58SRichard Weinberger ubifs_assert(c, wbuf->size % c->min_io_size == 0); 5546eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 5556c7f74f7SArtem Bityutskiy if (c->leb_size - wbuf->offs >= c->max_write_size) 5566eb61d58SRichard Weinberger ubifs_assert(c, !((wbuf->offs + wbuf->size) % c->max_write_size)); 5571e51764aSArtem Bityutskiy 5582680d722SArtem Bityutskiy if (c->ro_error) 5591e51764aSArtem Bityutskiy return -EROFS; 5601e51764aSArtem Bityutskiy 5616c7f74f7SArtem Bityutskiy /* 5626c7f74f7SArtem Bityutskiy * Do not write whole write buffer but write only the minimum necessary 5636c7f74f7SArtem Bityutskiy * amount of min. I/O units. 5646c7f74f7SArtem Bityutskiy */ 5656c7f74f7SArtem Bityutskiy sync_len = ALIGN(wbuf->used, c->min_io_size); 5666c7f74f7SArtem Bityutskiy dirt = sync_len - wbuf->used; 5676c7f74f7SArtem Bityutskiy if (dirt) 5686c7f74f7SArtem Bityutskiy ubifs_pad(c, wbuf->buf + wbuf->used, dirt); 569b36a261eSRichard Weinberger err = ubifs_leb_write(c, wbuf->lnum, wbuf->buf, wbuf->offs, sync_len); 570987226a5SArtem Bityutskiy if (err) 5711e51764aSArtem Bityutskiy return err; 5721e51764aSArtem Bityutskiy 5731e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 5746c7f74f7SArtem Bityutskiy wbuf->offs += sync_len; 5756c7f74f7SArtem Bityutskiy /* 5766c7f74f7SArtem Bityutskiy * Now @wbuf->offs is not necessarily aligned to @c->max_write_size. 5776c7f74f7SArtem Bityutskiy * But our goal is to optimize writes and make sure we write in 5786c7f74f7SArtem Bityutskiy * @c->max_write_size chunks and to @c->max_write_size-aligned offset. 5796c7f74f7SArtem Bityutskiy * Thus, if @wbuf->offs is not aligned to @c->max_write_size now, make 5806c7f74f7SArtem Bityutskiy * sure that @wbuf->offs + @wbuf->size is aligned to 5816c7f74f7SArtem Bityutskiy * @c->max_write_size. This way we make sure that after next 5826c7f74f7SArtem Bityutskiy * write-buffer flush we are again at the optimal offset (aligned to 5836c7f74f7SArtem Bityutskiy * @c->max_write_size). 5846c7f74f7SArtem Bityutskiy */ 5856c7f74f7SArtem Bityutskiy if (c->leb_size - wbuf->offs < c->max_write_size) 5866c7f74f7SArtem Bityutskiy wbuf->size = c->leb_size - wbuf->offs; 5876c7f74f7SArtem Bityutskiy else if (wbuf->offs & (c->max_write_size - 1)) 5886c7f74f7SArtem Bityutskiy wbuf->size = ALIGN(wbuf->offs, c->max_write_size) - wbuf->offs; 5896c7f74f7SArtem Bityutskiy else 5906c7f74f7SArtem Bityutskiy wbuf->size = c->max_write_size; 5916c7f74f7SArtem Bityutskiy wbuf->avail = wbuf->size; 5921e51764aSArtem Bityutskiy wbuf->used = 0; 5931e51764aSArtem Bityutskiy wbuf->next_ino = 0; 5941e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 5951e51764aSArtem Bityutskiy 5961e51764aSArtem Bityutskiy if (wbuf->sync_callback) 5971e51764aSArtem Bityutskiy err = wbuf->sync_callback(c, wbuf->lnum, 5981e51764aSArtem Bityutskiy c->leb_size - wbuf->offs, dirt); 5991e51764aSArtem Bityutskiy return err; 6001e51764aSArtem Bityutskiy } 6011e51764aSArtem Bityutskiy 6021e51764aSArtem Bityutskiy /** 6031e51764aSArtem Bityutskiy * ubifs_wbuf_seek_nolock - seek write-buffer. 6041e51764aSArtem Bityutskiy * @wbuf: write-buffer 6051e51764aSArtem Bityutskiy * @lnum: logical eraseblock number to seek to 6061e51764aSArtem Bityutskiy * @offs: logical eraseblock offset to seek to 6071e51764aSArtem Bityutskiy * 608cb54ef8bSArtem Bityutskiy * This function targets the write-buffer to logical eraseblock @lnum:@offs. 609cb14a184SArtem Bityutskiy * The write-buffer has to be empty. Returns zero in case of success and a 610cb14a184SArtem Bityutskiy * negative error code in case of failure. 6111e51764aSArtem Bityutskiy */ 612b36a261eSRichard Weinberger int ubifs_wbuf_seek_nolock(struct ubifs_wbuf *wbuf, int lnum, int offs) 6131e51764aSArtem Bityutskiy { 6141e51764aSArtem Bityutskiy const struct ubifs_info *c = wbuf->c; 6151e51764aSArtem Bityutskiy 61677a7ae58SArtem Bityutskiy dbg_io("LEB %d:%d, jhead %s", lnum, offs, dbg_jhead(wbuf->jhead)); 6176eb61d58SRichard Weinberger ubifs_assert(c, lnum >= 0 && lnum < c->leb_cnt); 6186eb61d58SRichard Weinberger ubifs_assert(c, offs >= 0 && offs <= c->leb_size); 6196eb61d58SRichard Weinberger ubifs_assert(c, offs % c->min_io_size == 0 && !(offs & 7)); 6206eb61d58SRichard Weinberger ubifs_assert(c, lnum != wbuf->lnum); 6216eb61d58SRichard Weinberger ubifs_assert(c, wbuf->used == 0); 6221e51764aSArtem Bityutskiy 6231e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 6241e51764aSArtem Bityutskiy wbuf->lnum = lnum; 6251e51764aSArtem Bityutskiy wbuf->offs = offs; 6266c7f74f7SArtem Bityutskiy if (c->leb_size - wbuf->offs < c->max_write_size) 6276c7f74f7SArtem Bityutskiy wbuf->size = c->leb_size - wbuf->offs; 6286c7f74f7SArtem Bityutskiy else if (wbuf->offs & (c->max_write_size - 1)) 6296c7f74f7SArtem Bityutskiy wbuf->size = ALIGN(wbuf->offs, c->max_write_size) - wbuf->offs; 6306c7f74f7SArtem Bityutskiy else 6316c7f74f7SArtem Bityutskiy wbuf->size = c->max_write_size; 6326c7f74f7SArtem Bityutskiy wbuf->avail = wbuf->size; 6331e51764aSArtem Bityutskiy wbuf->used = 0; 6341e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 6351e51764aSArtem Bityutskiy 6361e51764aSArtem Bityutskiy return 0; 6371e51764aSArtem Bityutskiy } 6381e51764aSArtem Bityutskiy 6391e51764aSArtem Bityutskiy /** 6401e51764aSArtem Bityutskiy * ubifs_bg_wbufs_sync - synchronize write-buffers. 6411e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 6421e51764aSArtem Bityutskiy * 6431e51764aSArtem Bityutskiy * This function is called by background thread to synchronize write-buffers. 6441e51764aSArtem Bityutskiy * Returns zero in case of success and a negative error code in case of 6451e51764aSArtem Bityutskiy * failure. 6461e51764aSArtem Bityutskiy */ 6471e51764aSArtem Bityutskiy int ubifs_bg_wbufs_sync(struct ubifs_info *c) 6481e51764aSArtem Bityutskiy { 6491e51764aSArtem Bityutskiy int err, i; 6501e51764aSArtem Bityutskiy 6516eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 6521e51764aSArtem Bityutskiy if (!c->need_wbuf_sync) 6531e51764aSArtem Bityutskiy return 0; 6541e51764aSArtem Bityutskiy c->need_wbuf_sync = 0; 6551e51764aSArtem Bityutskiy 6562680d722SArtem Bityutskiy if (c->ro_error) { 6571e51764aSArtem Bityutskiy err = -EROFS; 6581e51764aSArtem Bityutskiy goto out_timers; 6591e51764aSArtem Bityutskiy } 6601e51764aSArtem Bityutskiy 6611e51764aSArtem Bityutskiy dbg_io("synchronize"); 6621e51764aSArtem Bityutskiy for (i = 0; i < c->jhead_cnt; i++) { 6631e51764aSArtem Bityutskiy struct ubifs_wbuf *wbuf = &c->jheads[i].wbuf; 6641e51764aSArtem Bityutskiy 6651e51764aSArtem Bityutskiy cond_resched(); 6661e51764aSArtem Bityutskiy 6671e51764aSArtem Bityutskiy /* 6681e51764aSArtem Bityutskiy * If the mutex is locked then wbuf is being changed, so 6691e51764aSArtem Bityutskiy * synchronization is not necessary. 6701e51764aSArtem Bityutskiy */ 6711e51764aSArtem Bityutskiy if (mutex_is_locked(&wbuf->io_mutex)) 6721e51764aSArtem Bityutskiy continue; 6731e51764aSArtem Bityutskiy 6741e51764aSArtem Bityutskiy mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); 6751e51764aSArtem Bityutskiy if (!wbuf->need_sync) { 6761e51764aSArtem Bityutskiy mutex_unlock(&wbuf->io_mutex); 6771e51764aSArtem Bityutskiy continue; 6781e51764aSArtem Bityutskiy } 6791e51764aSArtem Bityutskiy 6801e51764aSArtem Bityutskiy err = ubifs_wbuf_sync_nolock(wbuf); 6811e51764aSArtem Bityutskiy mutex_unlock(&wbuf->io_mutex); 6821e51764aSArtem Bityutskiy if (err) { 683235c362bSSheng Yong ubifs_err(c, "cannot sync write-buffer, error %d", err); 6841e51764aSArtem Bityutskiy ubifs_ro_mode(c, err); 6851e51764aSArtem Bityutskiy goto out_timers; 6861e51764aSArtem Bityutskiy } 6871e51764aSArtem Bityutskiy } 6881e51764aSArtem Bityutskiy 6891e51764aSArtem Bityutskiy return 0; 6901e51764aSArtem Bityutskiy 6911e51764aSArtem Bityutskiy out_timers: 6921e51764aSArtem Bityutskiy /* Cancel all timers to prevent repeated errors */ 6931e51764aSArtem Bityutskiy for (i = 0; i < c->jhead_cnt; i++) { 6941e51764aSArtem Bityutskiy struct ubifs_wbuf *wbuf = &c->jheads[i].wbuf; 6951e51764aSArtem Bityutskiy 6961e51764aSArtem Bityutskiy mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); 6971e51764aSArtem Bityutskiy cancel_wbuf_timer_nolock(wbuf); 6981e51764aSArtem Bityutskiy mutex_unlock(&wbuf->io_mutex); 6991e51764aSArtem Bityutskiy } 7001e51764aSArtem Bityutskiy return err; 7011e51764aSArtem Bityutskiy } 7021e51764aSArtem Bityutskiy 7031e51764aSArtem Bityutskiy /** 7041e51764aSArtem Bityutskiy * ubifs_wbuf_write_nolock - write data to flash via write-buffer. 7051e51764aSArtem Bityutskiy * @wbuf: write-buffer 7061e51764aSArtem Bityutskiy * @buf: node to write 7071e51764aSArtem Bityutskiy * @len: node length 7081e51764aSArtem Bityutskiy * 7091e51764aSArtem Bityutskiy * This function writes data to flash via write-buffer @wbuf. This means that 7101e51764aSArtem Bityutskiy * the last piece of the node won't reach the flash media immediately if it 7116c7f74f7SArtem Bityutskiy * does not take whole max. write unit (@c->max_write_size). Instead, the node 7126c7f74f7SArtem Bityutskiy * will sit in RAM until the write-buffer is synchronized (e.g., by timer, or 7136c7f74f7SArtem Bityutskiy * because more data are appended to the write-buffer). 7141e51764aSArtem Bityutskiy * 7151e51764aSArtem Bityutskiy * This function returns zero in case of success and a negative error code in 7161e51764aSArtem Bityutskiy * case of failure. If the node cannot be written because there is no more 7171e51764aSArtem Bityutskiy * space in this logical eraseblock, %-ENOSPC is returned. 7181e51764aSArtem Bityutskiy */ 7191e51764aSArtem Bityutskiy int ubifs_wbuf_write_nolock(struct ubifs_wbuf *wbuf, void *buf, int len) 7201e51764aSArtem Bityutskiy { 7211e51764aSArtem Bityutskiy struct ubifs_info *c = wbuf->c; 722*a33e30a0SZhihao Cheng int err, n, written = 0, aligned_len = ALIGN(len, 8); 7231e51764aSArtem Bityutskiy 72477a7ae58SArtem Bityutskiy dbg_io("%d bytes (%s) to jhead %s wbuf at LEB %d:%d", len, 72577a7ae58SArtem Bityutskiy dbg_ntype(((struct ubifs_ch *)buf)->node_type), 72677a7ae58SArtem Bityutskiy dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs + wbuf->used); 7276eb61d58SRichard Weinberger ubifs_assert(c, len > 0 && wbuf->lnum >= 0 && wbuf->lnum < c->leb_cnt); 7286eb61d58SRichard Weinberger ubifs_assert(c, wbuf->offs >= 0 && wbuf->offs % c->min_io_size == 0); 7296eb61d58SRichard Weinberger ubifs_assert(c, !(wbuf->offs & 7) && wbuf->offs <= c->leb_size); 7306eb61d58SRichard Weinberger ubifs_assert(c, wbuf->avail > 0 && wbuf->avail <= wbuf->size); 7316eb61d58SRichard Weinberger ubifs_assert(c, wbuf->size >= c->min_io_size); 7326eb61d58SRichard Weinberger ubifs_assert(c, wbuf->size <= c->max_write_size); 7336eb61d58SRichard Weinberger ubifs_assert(c, wbuf->size % c->min_io_size == 0); 7346eb61d58SRichard Weinberger ubifs_assert(c, mutex_is_locked(&wbuf->io_mutex)); 7356eb61d58SRichard Weinberger ubifs_assert(c, !c->ro_media && !c->ro_mount); 7366eb61d58SRichard Weinberger ubifs_assert(c, !c->space_fixup); 7376c7f74f7SArtem Bityutskiy if (c->leb_size - wbuf->offs >= c->max_write_size) 7386eb61d58SRichard Weinberger ubifs_assert(c, !((wbuf->offs + wbuf->size) % c->max_write_size)); 7391e51764aSArtem Bityutskiy 7401e51764aSArtem Bityutskiy if (c->leb_size - wbuf->offs - wbuf->used < aligned_len) { 7411e51764aSArtem Bityutskiy err = -ENOSPC; 7421e51764aSArtem Bityutskiy goto out; 7431e51764aSArtem Bityutskiy } 7441e51764aSArtem Bityutskiy 7451e51764aSArtem Bityutskiy cancel_wbuf_timer_nolock(wbuf); 7461e51764aSArtem Bityutskiy 7472680d722SArtem Bityutskiy if (c->ro_error) 7481e51764aSArtem Bityutskiy return -EROFS; 7491e51764aSArtem Bityutskiy 7501e51764aSArtem Bityutskiy if (aligned_len <= wbuf->avail) { 7511e51764aSArtem Bityutskiy /* 7521e51764aSArtem Bityutskiy * The node is not very large and fits entirely within 7531e51764aSArtem Bityutskiy * write-buffer. 7541e51764aSArtem Bityutskiy */ 7551e51764aSArtem Bityutskiy memcpy(wbuf->buf + wbuf->used, buf, len); 75620f14311SRichard Weinberger if (aligned_len > len) { 75720f14311SRichard Weinberger ubifs_assert(c, aligned_len - len < 8); 75820f14311SRichard Weinberger ubifs_pad(c, wbuf->buf + wbuf->used + len, aligned_len - len); 75920f14311SRichard Weinberger } 7601e51764aSArtem Bityutskiy 7611e51764aSArtem Bityutskiy if (aligned_len == wbuf->avail) { 76277a7ae58SArtem Bityutskiy dbg_io("flush jhead %s wbuf to LEB %d:%d", 76377a7ae58SArtem Bityutskiy dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs); 764987226a5SArtem Bityutskiy err = ubifs_leb_write(c, wbuf->lnum, wbuf->buf, 765b36a261eSRichard Weinberger wbuf->offs, wbuf->size); 7661e51764aSArtem Bityutskiy if (err) 7671e51764aSArtem Bityutskiy goto out; 7681e51764aSArtem Bityutskiy 7691e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 7706c7f74f7SArtem Bityutskiy wbuf->offs += wbuf->size; 7716c7f74f7SArtem Bityutskiy if (c->leb_size - wbuf->offs >= c->max_write_size) 7726c7f74f7SArtem Bityutskiy wbuf->size = c->max_write_size; 7736c7f74f7SArtem Bityutskiy else 7746c7f74f7SArtem Bityutskiy wbuf->size = c->leb_size - wbuf->offs; 7756c7f74f7SArtem Bityutskiy wbuf->avail = wbuf->size; 7761e51764aSArtem Bityutskiy wbuf->used = 0; 7771e51764aSArtem Bityutskiy wbuf->next_ino = 0; 7781e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 7791e51764aSArtem Bityutskiy } else { 7801e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 7811e51764aSArtem Bityutskiy wbuf->avail -= aligned_len; 7821e51764aSArtem Bityutskiy wbuf->used += aligned_len; 7831e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 7841e51764aSArtem Bityutskiy } 7851e51764aSArtem Bityutskiy 7861e51764aSArtem Bityutskiy goto exit; 7871e51764aSArtem Bityutskiy } 7881e51764aSArtem Bityutskiy 7896c7f74f7SArtem Bityutskiy if (wbuf->used) { 7901e51764aSArtem Bityutskiy /* 7916c7f74f7SArtem Bityutskiy * The node is large enough and does not fit entirely within 7926c7f74f7SArtem Bityutskiy * current available space. We have to fill and flush 7936c7f74f7SArtem Bityutskiy * write-buffer and switch to the next max. write unit. 7941e51764aSArtem Bityutskiy */ 79577a7ae58SArtem Bityutskiy dbg_io("flush jhead %s wbuf to LEB %d:%d", 79677a7ae58SArtem Bityutskiy dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs); 7971e51764aSArtem Bityutskiy memcpy(wbuf->buf + wbuf->used, buf, wbuf->avail); 798987226a5SArtem Bityutskiy err = ubifs_leb_write(c, wbuf->lnum, wbuf->buf, wbuf->offs, 799b36a261eSRichard Weinberger wbuf->size); 8001e51764aSArtem Bityutskiy if (err) 8011e51764aSArtem Bityutskiy goto out; 8021e51764aSArtem Bityutskiy 80312f33891SArtem Bityutskiy wbuf->offs += wbuf->size; 8041e51764aSArtem Bityutskiy len -= wbuf->avail; 8051e51764aSArtem Bityutskiy aligned_len -= wbuf->avail; 8066c7f74f7SArtem Bityutskiy written += wbuf->avail; 8076c7f74f7SArtem Bityutskiy } else if (wbuf->offs & (c->max_write_size - 1)) { 8086c7f74f7SArtem Bityutskiy /* 8096c7f74f7SArtem Bityutskiy * The write-buffer offset is not aligned to 8106c7f74f7SArtem Bityutskiy * @c->max_write_size and @wbuf->size is less than 8116c7f74f7SArtem Bityutskiy * @c->max_write_size. Write @wbuf->size bytes to make sure the 8126c7f74f7SArtem Bityutskiy * following writes are done in optimal @c->max_write_size 8136c7f74f7SArtem Bityutskiy * chunks. 8146c7f74f7SArtem Bityutskiy */ 8156c7f74f7SArtem Bityutskiy dbg_io("write %d bytes to LEB %d:%d", 8166c7f74f7SArtem Bityutskiy wbuf->size, wbuf->lnum, wbuf->offs); 817987226a5SArtem Bityutskiy err = ubifs_leb_write(c, wbuf->lnum, buf, wbuf->offs, 818b36a261eSRichard Weinberger wbuf->size); 8196c7f74f7SArtem Bityutskiy if (err) 8206c7f74f7SArtem Bityutskiy goto out; 8216c7f74f7SArtem Bityutskiy 82212f33891SArtem Bityutskiy wbuf->offs += wbuf->size; 8236c7f74f7SArtem Bityutskiy len -= wbuf->size; 8246c7f74f7SArtem Bityutskiy aligned_len -= wbuf->size; 8256c7f74f7SArtem Bityutskiy written += wbuf->size; 8266c7f74f7SArtem Bityutskiy } 8271e51764aSArtem Bityutskiy 8281e51764aSArtem Bityutskiy /* 8296c7f74f7SArtem Bityutskiy * The remaining data may take more whole max. write units, so write the 8306c7f74f7SArtem Bityutskiy * remains multiple to max. write unit size directly to the flash media. 8311e51764aSArtem Bityutskiy * We align node length to 8-byte boundary because we anyway flash wbuf 8321e51764aSArtem Bityutskiy * if the remaining space is less than 8 bytes. 8331e51764aSArtem Bityutskiy */ 8346c7f74f7SArtem Bityutskiy n = aligned_len >> c->max_write_shift; 8351e51764aSArtem Bityutskiy if (n) { 8366c7f74f7SArtem Bityutskiy n <<= c->max_write_shift; 83712f33891SArtem Bityutskiy dbg_io("write %d bytes to LEB %d:%d", n, wbuf->lnum, 83812f33891SArtem Bityutskiy wbuf->offs); 839987226a5SArtem Bityutskiy err = ubifs_leb_write(c, wbuf->lnum, buf + written, 840b36a261eSRichard Weinberger wbuf->offs, n); 8411e51764aSArtem Bityutskiy if (err) 8421e51764aSArtem Bityutskiy goto out; 84312f33891SArtem Bityutskiy wbuf->offs += n; 8441e51764aSArtem Bityutskiy aligned_len -= n; 8451e51764aSArtem Bityutskiy len -= n; 8461e51764aSArtem Bityutskiy written += n; 8471e51764aSArtem Bityutskiy } 8481e51764aSArtem Bityutskiy 8491e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 85020f14311SRichard Weinberger if (aligned_len) { 8511e51764aSArtem Bityutskiy /* 8521e51764aSArtem Bityutskiy * And now we have what's left and what does not take whole 8536c7f74f7SArtem Bityutskiy * max. write unit, so write it to the write-buffer and we are 8541e51764aSArtem Bityutskiy * done. 8551e51764aSArtem Bityutskiy */ 8561e51764aSArtem Bityutskiy memcpy(wbuf->buf, buf + written, len); 85720f14311SRichard Weinberger if (aligned_len > len) { 85820f14311SRichard Weinberger ubifs_assert(c, aligned_len - len < 8); 85920f14311SRichard Weinberger ubifs_pad(c, wbuf->buf + len, aligned_len - len); 86020f14311SRichard Weinberger } 86120f14311SRichard Weinberger } 8621e51764aSArtem Bityutskiy 8636c7f74f7SArtem Bityutskiy if (c->leb_size - wbuf->offs >= c->max_write_size) 8646c7f74f7SArtem Bityutskiy wbuf->size = c->max_write_size; 8656c7f74f7SArtem Bityutskiy else 8666c7f74f7SArtem Bityutskiy wbuf->size = c->leb_size - wbuf->offs; 8676c7f74f7SArtem Bityutskiy wbuf->avail = wbuf->size - aligned_len; 8681e51764aSArtem Bityutskiy wbuf->used = aligned_len; 8691e51764aSArtem Bityutskiy wbuf->next_ino = 0; 8701e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 8711e51764aSArtem Bityutskiy 8721e51764aSArtem Bityutskiy exit: 8731e51764aSArtem Bityutskiy if (wbuf->sync_callback) { 8741e51764aSArtem Bityutskiy int free = c->leb_size - wbuf->offs - wbuf->used; 8751e51764aSArtem Bityutskiy 8761e51764aSArtem Bityutskiy err = wbuf->sync_callback(c, wbuf->lnum, free, 0); 8771e51764aSArtem Bityutskiy if (err) 8781e51764aSArtem Bityutskiy goto out; 8791e51764aSArtem Bityutskiy } 8801e51764aSArtem Bityutskiy 8811e51764aSArtem Bityutskiy if (wbuf->used) 8826eb61d58SRichard Weinberger new_wbuf_timer_nolock(c, wbuf); 8831e51764aSArtem Bityutskiy 8841e51764aSArtem Bityutskiy return 0; 8851e51764aSArtem Bityutskiy 8861e51764aSArtem Bityutskiy out: 887235c362bSSheng Yong ubifs_err(c, "cannot write %d bytes to LEB %d:%d, error %d", 8881e51764aSArtem Bityutskiy len, wbuf->lnum, wbuf->offs, err); 889*a33e30a0SZhihao Cheng ubifs_dump_node(c, buf, written + len); 8907c46d0aeSArtem Bityutskiy dump_stack(); 891edf6be24SArtem Bityutskiy ubifs_dump_leb(c, wbuf->lnum); 8921e51764aSArtem Bityutskiy return err; 8931e51764aSArtem Bityutskiy } 8941e51764aSArtem Bityutskiy 8951e51764aSArtem Bityutskiy /** 896a384b47eSSascha Hauer * ubifs_write_node_hmac - write node to the media. 897a384b47eSSascha Hauer * @c: UBIFS file-system description object 898a384b47eSSascha Hauer * @buf: the node to write 899a384b47eSSascha Hauer * @len: node length 900a384b47eSSascha Hauer * @lnum: logical eraseblock number 901a384b47eSSascha Hauer * @offs: offset within the logical eraseblock 902a384b47eSSascha Hauer * @hmac_offs: offset of the HMAC within the node 903a384b47eSSascha Hauer * 904a384b47eSSascha Hauer * This function automatically fills node magic number, assigns sequence 905a384b47eSSascha Hauer * number, and calculates node CRC checksum. The length of the @buf buffer has 906a384b47eSSascha Hauer * to be aligned to the minimal I/O unit size. This function automatically 907a384b47eSSascha Hauer * appends padding node and padding bytes if needed. Returns zero in case of 908a384b47eSSascha Hauer * success and a negative error code in case of failure. 909a384b47eSSascha Hauer */ 910a384b47eSSascha Hauer int ubifs_write_node_hmac(struct ubifs_info *c, void *buf, int len, int lnum, 911a384b47eSSascha Hauer int offs, int hmac_offs) 912a384b47eSSascha Hauer { 913a384b47eSSascha Hauer int err, buf_len = ALIGN(len, c->min_io_size); 914a384b47eSSascha Hauer 915a384b47eSSascha Hauer dbg_io("LEB %d:%d, %s, length %d (aligned %d)", 916a384b47eSSascha Hauer lnum, offs, dbg_ntype(((struct ubifs_ch *)buf)->node_type), len, 917a384b47eSSascha Hauer buf_len); 918a384b47eSSascha Hauer ubifs_assert(c, lnum >= 0 && lnum < c->leb_cnt && offs >= 0); 919a384b47eSSascha Hauer ubifs_assert(c, offs % c->min_io_size == 0 && offs < c->leb_size); 920a384b47eSSascha Hauer ubifs_assert(c, !c->ro_media && !c->ro_mount); 921a384b47eSSascha Hauer ubifs_assert(c, !c->space_fixup); 922a384b47eSSascha Hauer 923a384b47eSSascha Hauer if (c->ro_error) 924a384b47eSSascha Hauer return -EROFS; 925a384b47eSSascha Hauer 926a384b47eSSascha Hauer err = ubifs_prepare_node_hmac(c, buf, len, hmac_offs, 1); 927a384b47eSSascha Hauer if (err) 928a384b47eSSascha Hauer return err; 929a384b47eSSascha Hauer 930a384b47eSSascha Hauer err = ubifs_leb_write(c, lnum, buf, offs, buf_len); 931a384b47eSSascha Hauer if (err) 932*a33e30a0SZhihao Cheng ubifs_dump_node(c, buf, len); 933a384b47eSSascha Hauer 934a384b47eSSascha Hauer return err; 935a384b47eSSascha Hauer } 936a384b47eSSascha Hauer 937a384b47eSSascha Hauer /** 9381e51764aSArtem Bityutskiy * ubifs_write_node - write node to the media. 9391e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 9401e51764aSArtem Bityutskiy * @buf: the node to write 9411e51764aSArtem Bityutskiy * @len: node length 9421e51764aSArtem Bityutskiy * @lnum: logical eraseblock number 9431e51764aSArtem Bityutskiy * @offs: offset within the logical eraseblock 9441e51764aSArtem Bityutskiy * 9451e51764aSArtem Bityutskiy * This function automatically fills node magic number, assigns sequence 9461e51764aSArtem Bityutskiy * number, and calculates node CRC checksum. The length of the @buf buffer has 9471e51764aSArtem Bityutskiy * to be aligned to the minimal I/O unit size. This function automatically 9481e51764aSArtem Bityutskiy * appends padding node and padding bytes if needed. Returns zero in case of 9491e51764aSArtem Bityutskiy * success and a negative error code in case of failure. 9501e51764aSArtem Bityutskiy */ 9511e51764aSArtem Bityutskiy int ubifs_write_node(struct ubifs_info *c, void *buf, int len, int lnum, 952b36a261eSRichard Weinberger int offs) 9531e51764aSArtem Bityutskiy { 954a384b47eSSascha Hauer return ubifs_write_node_hmac(c, buf, len, lnum, offs, -1); 9551e51764aSArtem Bityutskiy } 9561e51764aSArtem Bityutskiy 9571e51764aSArtem Bityutskiy /** 9581e51764aSArtem Bityutskiy * ubifs_read_node_wbuf - read node from the media or write-buffer. 9591e51764aSArtem Bityutskiy * @wbuf: wbuf to check for un-written data 9601e51764aSArtem Bityutskiy * @buf: buffer to read to 9611e51764aSArtem Bityutskiy * @type: node type 9621e51764aSArtem Bityutskiy * @len: node length 9631e51764aSArtem Bityutskiy * @lnum: logical eraseblock number 9641e51764aSArtem Bityutskiy * @offs: offset within the logical eraseblock 9651e51764aSArtem Bityutskiy * 9661e51764aSArtem Bityutskiy * This function reads a node of known type and length, checks it and stores 9671e51764aSArtem Bityutskiy * in @buf. If the node partially or fully sits in the write-buffer, this 9681e51764aSArtem Bityutskiy * function takes data from the buffer, otherwise it reads the flash media. 9691e51764aSArtem Bityutskiy * Returns zero in case of success, %-EUCLEAN if CRC mismatched and a negative 9701e51764aSArtem Bityutskiy * error code in case of failure. 9711e51764aSArtem Bityutskiy */ 9721e51764aSArtem Bityutskiy int ubifs_read_node_wbuf(struct ubifs_wbuf *wbuf, void *buf, int type, int len, 9731e51764aSArtem Bityutskiy int lnum, int offs) 9741e51764aSArtem Bityutskiy { 9751e51764aSArtem Bityutskiy const struct ubifs_info *c = wbuf->c; 9761e51764aSArtem Bityutskiy int err, rlen, overlap; 9771e51764aSArtem Bityutskiy struct ubifs_ch *ch = buf; 9781e51764aSArtem Bityutskiy 97977a7ae58SArtem Bityutskiy dbg_io("LEB %d:%d, %s, length %d, jhead %s", lnum, offs, 98077a7ae58SArtem Bityutskiy dbg_ntype(type), len, dbg_jhead(wbuf->jhead)); 9816eb61d58SRichard Weinberger ubifs_assert(c, wbuf && lnum >= 0 && lnum < c->leb_cnt && offs >= 0); 9826eb61d58SRichard Weinberger ubifs_assert(c, !(offs & 7) && offs < c->leb_size); 9836eb61d58SRichard Weinberger ubifs_assert(c, type >= 0 && type < UBIFS_NODE_TYPES_CNT); 9841e51764aSArtem Bityutskiy 9851e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 9861e51764aSArtem Bityutskiy overlap = (lnum == wbuf->lnum && offs + len > wbuf->offs); 9871e51764aSArtem Bityutskiy if (!overlap) { 9881e51764aSArtem Bityutskiy /* We may safely unlock the write-buffer and read the data */ 9891e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 9901e51764aSArtem Bityutskiy return ubifs_read_node(c, buf, type, len, lnum, offs); 9911e51764aSArtem Bityutskiy } 9921e51764aSArtem Bityutskiy 9931e51764aSArtem Bityutskiy /* Don't read under wbuf */ 9941e51764aSArtem Bityutskiy rlen = wbuf->offs - offs; 9951e51764aSArtem Bityutskiy if (rlen < 0) 9961e51764aSArtem Bityutskiy rlen = 0; 9971e51764aSArtem Bityutskiy 9981e51764aSArtem Bityutskiy /* Copy the rest from the write-buffer */ 9991e51764aSArtem Bityutskiy memcpy(buf + rlen, wbuf->buf + offs + rlen - wbuf->offs, len - rlen); 10001e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 10011e51764aSArtem Bityutskiy 10021e51764aSArtem Bityutskiy if (rlen > 0) { 10031e51764aSArtem Bityutskiy /* Read everything that goes before write-buffer */ 1004d304820aSArtem Bityutskiy err = ubifs_leb_read(c, lnum, buf, offs, rlen, 0); 1005d304820aSArtem Bityutskiy if (err && err != -EBADMSG) 10061e51764aSArtem Bityutskiy return err; 10071e51764aSArtem Bityutskiy } 10081e51764aSArtem Bityutskiy 10091e51764aSArtem Bityutskiy if (type != ch->node_type) { 1010235c362bSSheng Yong ubifs_err(c, "bad node type (%d but expected %d)", 10111e51764aSArtem Bityutskiy ch->node_type, type); 10121e51764aSArtem Bityutskiy goto out; 10131e51764aSArtem Bityutskiy } 10141e51764aSArtem Bityutskiy 1015*a33e30a0SZhihao Cheng err = ubifs_check_node(c, buf, len, lnum, offs, 0, 0); 10161e51764aSArtem Bityutskiy if (err) { 1017235c362bSSheng Yong ubifs_err(c, "expected node type %d", type); 10181e51764aSArtem Bityutskiy return err; 10191e51764aSArtem Bityutskiy } 10201e51764aSArtem Bityutskiy 10211e51764aSArtem Bityutskiy rlen = le32_to_cpu(ch->len); 10221e51764aSArtem Bityutskiy if (rlen != len) { 1023235c362bSSheng Yong ubifs_err(c, "bad node length %d, expected %d", rlen, len); 10241e51764aSArtem Bityutskiy goto out; 10251e51764aSArtem Bityutskiy } 10261e51764aSArtem Bityutskiy 10271e51764aSArtem Bityutskiy return 0; 10281e51764aSArtem Bityutskiy 10291e51764aSArtem Bityutskiy out: 1030235c362bSSheng Yong ubifs_err(c, "bad node at LEB %d:%d", lnum, offs); 1031*a33e30a0SZhihao Cheng ubifs_dump_node(c, buf, len); 10327c46d0aeSArtem Bityutskiy dump_stack(); 10331e51764aSArtem Bityutskiy return -EINVAL; 10341e51764aSArtem Bityutskiy } 10351e51764aSArtem Bityutskiy 10361e51764aSArtem Bityutskiy /** 10371e51764aSArtem Bityutskiy * ubifs_read_node - read node. 10381e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 10391e51764aSArtem Bityutskiy * @buf: buffer to read to 10401e51764aSArtem Bityutskiy * @type: node type 10411e51764aSArtem Bityutskiy * @len: node length (not aligned) 10421e51764aSArtem Bityutskiy * @lnum: logical eraseblock number 10431e51764aSArtem Bityutskiy * @offs: offset within the logical eraseblock 10441e51764aSArtem Bityutskiy * 1045b8f1da98SRandy Dunlap * This function reads a node of known type and length, checks it and 10461e51764aSArtem Bityutskiy * stores in @buf. Returns zero in case of success, %-EUCLEAN if CRC mismatched 10471e51764aSArtem Bityutskiy * and a negative error code in case of failure. 10481e51764aSArtem Bityutskiy */ 10491e51764aSArtem Bityutskiy int ubifs_read_node(const struct ubifs_info *c, void *buf, int type, int len, 10501e51764aSArtem Bityutskiy int lnum, int offs) 10511e51764aSArtem Bityutskiy { 10521e51764aSArtem Bityutskiy int err, l; 10531e51764aSArtem Bityutskiy struct ubifs_ch *ch = buf; 10541e51764aSArtem Bityutskiy 10551e51764aSArtem Bityutskiy dbg_io("LEB %d:%d, %s, length %d", lnum, offs, dbg_ntype(type), len); 10566eb61d58SRichard Weinberger ubifs_assert(c, lnum >= 0 && lnum < c->leb_cnt && offs >= 0); 10576eb61d58SRichard Weinberger ubifs_assert(c, len >= UBIFS_CH_SZ && offs + len <= c->leb_size); 10586eb61d58SRichard Weinberger ubifs_assert(c, !(offs & 7) && offs < c->leb_size); 10596eb61d58SRichard Weinberger ubifs_assert(c, type >= 0 && type < UBIFS_NODE_TYPES_CNT); 10601e51764aSArtem Bityutskiy 1061d304820aSArtem Bityutskiy err = ubifs_leb_read(c, lnum, buf, offs, len, 0); 1062d304820aSArtem Bityutskiy if (err && err != -EBADMSG) 10631e51764aSArtem Bityutskiy return err; 10641e51764aSArtem Bityutskiy 10651e51764aSArtem Bityutskiy if (type != ch->node_type) { 106690bea5a3SDaniel Golle ubifs_errc(c, "bad node type (%d but expected %d)", 10671e51764aSArtem Bityutskiy ch->node_type, type); 10681e51764aSArtem Bityutskiy goto out; 10691e51764aSArtem Bityutskiy } 10701e51764aSArtem Bityutskiy 1071*a33e30a0SZhihao Cheng err = ubifs_check_node(c, buf, len, lnum, offs, 0, 0); 10721e51764aSArtem Bityutskiy if (err) { 107390bea5a3SDaniel Golle ubifs_errc(c, "expected node type %d", type); 10741e51764aSArtem Bityutskiy return err; 10751e51764aSArtem Bityutskiy } 10761e51764aSArtem Bityutskiy 10771e51764aSArtem Bityutskiy l = le32_to_cpu(ch->len); 10781e51764aSArtem Bityutskiy if (l != len) { 107990bea5a3SDaniel Golle ubifs_errc(c, "bad node length %d, expected %d", l, len); 10801e51764aSArtem Bityutskiy goto out; 10811e51764aSArtem Bityutskiy } 10821e51764aSArtem Bityutskiy 10831e51764aSArtem Bityutskiy return 0; 10841e51764aSArtem Bityutskiy 10851e51764aSArtem Bityutskiy out: 108690bea5a3SDaniel Golle ubifs_errc(c, "bad node at LEB %d:%d, LEB mapping status %d", lnum, 108790bea5a3SDaniel Golle offs, ubi_is_mapped(c->ubi, lnum)); 108890bea5a3SDaniel Golle if (!c->probing) { 1089*a33e30a0SZhihao Cheng ubifs_dump_node(c, buf, len); 10907c46d0aeSArtem Bityutskiy dump_stack(); 109190bea5a3SDaniel Golle } 10921e51764aSArtem Bityutskiy return -EINVAL; 10931e51764aSArtem Bityutskiy } 10941e51764aSArtem Bityutskiy 10951e51764aSArtem Bityutskiy /** 10961e51764aSArtem Bityutskiy * ubifs_wbuf_init - initialize write-buffer. 10971e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 10981e51764aSArtem Bityutskiy * @wbuf: write-buffer to initialize 10991e51764aSArtem Bityutskiy * 1100cb54ef8bSArtem Bityutskiy * This function initializes write-buffer. Returns zero in case of success 11011e51764aSArtem Bityutskiy * %-ENOMEM in case of failure. 11021e51764aSArtem Bityutskiy */ 11031e51764aSArtem Bityutskiy int ubifs_wbuf_init(struct ubifs_info *c, struct ubifs_wbuf *wbuf) 11041e51764aSArtem Bityutskiy { 11051e51764aSArtem Bityutskiy size_t size; 11061e51764aSArtem Bityutskiy 11076c7f74f7SArtem Bityutskiy wbuf->buf = kmalloc(c->max_write_size, GFP_KERNEL); 11081e51764aSArtem Bityutskiy if (!wbuf->buf) 11091e51764aSArtem Bityutskiy return -ENOMEM; 11101e51764aSArtem Bityutskiy 11116c7f74f7SArtem Bityutskiy size = (c->max_write_size / UBIFS_CH_SZ + 1) * sizeof(ino_t); 11121e51764aSArtem Bityutskiy wbuf->inodes = kmalloc(size, GFP_KERNEL); 11131e51764aSArtem Bityutskiy if (!wbuf->inodes) { 11141e51764aSArtem Bityutskiy kfree(wbuf->buf); 11151e51764aSArtem Bityutskiy wbuf->buf = NULL; 11161e51764aSArtem Bityutskiy return -ENOMEM; 11171e51764aSArtem Bityutskiy } 11181e51764aSArtem Bityutskiy 11191e51764aSArtem Bityutskiy wbuf->used = 0; 11201e51764aSArtem Bityutskiy wbuf->lnum = wbuf->offs = -1; 11216c7f74f7SArtem Bityutskiy /* 11226c7f74f7SArtem Bityutskiy * If the LEB starts at the max. write size aligned address, then 11236c7f74f7SArtem Bityutskiy * write-buffer size has to be set to @c->max_write_size. Otherwise, 11246c7f74f7SArtem Bityutskiy * set it to something smaller so that it ends at the closest max. 11256c7f74f7SArtem Bityutskiy * write size boundary. 11266c7f74f7SArtem Bityutskiy */ 11276c7f74f7SArtem Bityutskiy size = c->max_write_size - (c->leb_start % c->max_write_size); 11286c7f74f7SArtem Bityutskiy wbuf->avail = wbuf->size = size; 11291e51764aSArtem Bityutskiy wbuf->sync_callback = NULL; 11301e51764aSArtem Bityutskiy mutex_init(&wbuf->io_mutex); 11311e51764aSArtem Bityutskiy spin_lock_init(&wbuf->lock); 11321e51764aSArtem Bityutskiy wbuf->c = c; 11331e51764aSArtem Bityutskiy wbuf->next_ino = 0; 11341e51764aSArtem Bityutskiy 1135f2c5dbd7SArtem Bityutskiy hrtimer_init(&wbuf->timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL); 1136f2c5dbd7SArtem Bityutskiy wbuf->timer.function = wbuf_timer_callback_nolock; 11371e51764aSArtem Bityutskiy return 0; 11381e51764aSArtem Bityutskiy } 11391e51764aSArtem Bityutskiy 11401e51764aSArtem Bityutskiy /** 11411e51764aSArtem Bityutskiy * ubifs_wbuf_add_ino_nolock - add an inode number into the wbuf inode array. 11421e51764aSArtem Bityutskiy * @wbuf: the write-buffer where to add 11431e51764aSArtem Bityutskiy * @inum: the inode number 11441e51764aSArtem Bityutskiy * 11451e51764aSArtem Bityutskiy * This function adds an inode number to the inode array of the write-buffer. 11461e51764aSArtem Bityutskiy */ 11471e51764aSArtem Bityutskiy void ubifs_wbuf_add_ino_nolock(struct ubifs_wbuf *wbuf, ino_t inum) 11481e51764aSArtem Bityutskiy { 11491e51764aSArtem Bityutskiy if (!wbuf->buf) 11501e51764aSArtem Bityutskiy /* NOR flash or something similar */ 11511e51764aSArtem Bityutskiy return; 11521e51764aSArtem Bityutskiy 11531e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 11541e51764aSArtem Bityutskiy if (wbuf->used) 11551e51764aSArtem Bityutskiy wbuf->inodes[wbuf->next_ino++] = inum; 11561e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 11571e51764aSArtem Bityutskiy } 11581e51764aSArtem Bityutskiy 11591e51764aSArtem Bityutskiy /** 11601e51764aSArtem Bityutskiy * wbuf_has_ino - returns if the wbuf contains data from the inode. 11611e51764aSArtem Bityutskiy * @wbuf: the write-buffer 11621e51764aSArtem Bityutskiy * @inum: the inode number 11631e51764aSArtem Bityutskiy * 11641e51764aSArtem Bityutskiy * This function returns with %1 if the write-buffer contains some data from the 11651e51764aSArtem Bityutskiy * given inode otherwise it returns with %0. 11661e51764aSArtem Bityutskiy */ 11671e51764aSArtem Bityutskiy static int wbuf_has_ino(struct ubifs_wbuf *wbuf, ino_t inum) 11681e51764aSArtem Bityutskiy { 11691e51764aSArtem Bityutskiy int i, ret = 0; 11701e51764aSArtem Bityutskiy 11711e51764aSArtem Bityutskiy spin_lock(&wbuf->lock); 11721e51764aSArtem Bityutskiy for (i = 0; i < wbuf->next_ino; i++) 11731e51764aSArtem Bityutskiy if (inum == wbuf->inodes[i]) { 11741e51764aSArtem Bityutskiy ret = 1; 11751e51764aSArtem Bityutskiy break; 11761e51764aSArtem Bityutskiy } 11771e51764aSArtem Bityutskiy spin_unlock(&wbuf->lock); 11781e51764aSArtem Bityutskiy 11791e51764aSArtem Bityutskiy return ret; 11801e51764aSArtem Bityutskiy } 11811e51764aSArtem Bityutskiy 11821e51764aSArtem Bityutskiy /** 11831e51764aSArtem Bityutskiy * ubifs_sync_wbufs_by_inode - synchronize write-buffers for an inode. 11841e51764aSArtem Bityutskiy * @c: UBIFS file-system description object 11851e51764aSArtem Bityutskiy * @inode: inode to synchronize 11861e51764aSArtem Bityutskiy * 11871e51764aSArtem Bityutskiy * This function synchronizes write-buffers which contain nodes belonging to 11881e51764aSArtem Bityutskiy * @inode. Returns zero in case of success and a negative error code in case of 11891e51764aSArtem Bityutskiy * failure. 11901e51764aSArtem Bityutskiy */ 11911e51764aSArtem Bityutskiy int ubifs_sync_wbufs_by_inode(struct ubifs_info *c, struct inode *inode) 11921e51764aSArtem Bityutskiy { 11931e51764aSArtem Bityutskiy int i, err = 0; 11941e51764aSArtem Bityutskiy 11951e51764aSArtem Bityutskiy for (i = 0; i < c->jhead_cnt; i++) { 11961e51764aSArtem Bityutskiy struct ubifs_wbuf *wbuf = &c->jheads[i].wbuf; 11971e51764aSArtem Bityutskiy 11981e51764aSArtem Bityutskiy if (i == GCHD) 11991e51764aSArtem Bityutskiy /* 12001e51764aSArtem Bityutskiy * GC head is special, do not look at it. Even if the 12011e51764aSArtem Bityutskiy * head contains something related to this inode, it is 12021e51764aSArtem Bityutskiy * a _copy_ of corresponding on-flash node which sits 12031e51764aSArtem Bityutskiy * somewhere else. 12041e51764aSArtem Bityutskiy */ 12051e51764aSArtem Bityutskiy continue; 12061e51764aSArtem Bityutskiy 12071e51764aSArtem Bityutskiy if (!wbuf_has_ino(wbuf, inode->i_ino)) 12081e51764aSArtem Bityutskiy continue; 12091e51764aSArtem Bityutskiy 12101e51764aSArtem Bityutskiy mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); 12111e51764aSArtem Bityutskiy if (wbuf_has_ino(wbuf, inode->i_ino)) 12121e51764aSArtem Bityutskiy err = ubifs_wbuf_sync_nolock(wbuf); 12131e51764aSArtem Bityutskiy mutex_unlock(&wbuf->io_mutex); 12141e51764aSArtem Bityutskiy 12151e51764aSArtem Bityutskiy if (err) { 12161e51764aSArtem Bityutskiy ubifs_ro_mode(c, err); 12171e51764aSArtem Bityutskiy return err; 12181e51764aSArtem Bityutskiy } 12191e51764aSArtem Bityutskiy } 12201e51764aSArtem Bityutskiy return 0; 12211e51764aSArtem Bityutskiy } 1222