1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * linux/fs/hpfs/file.c 4 * 5 * Mikulas Patocka (mikulas@artax.karlin.mff.cuni.cz), 1998-1999 6 * 7 * file VFS functions 8 */ 9 10 #include "hpfs_fn.h" 11 #include <linux/mpage.h> 12 #include <linux/iomap.h> 13 #include <linux/fiemap.h> 14 15 #define BLOCKS(size) (((size) + 511) >> 9) 16 17 static int hpfs_file_release(struct inode *inode, struct file *file) 18 { 19 hpfs_lock(inode->i_sb); 20 hpfs_write_if_changed(inode); 21 hpfs_unlock(inode->i_sb); 22 return 0; 23 } 24 25 int hpfs_file_fsync(struct file *file, loff_t start, loff_t end, int datasync) 26 { 27 struct inode *inode = file->f_mapping->host; 28 int ret; 29 30 ret = file_write_and_wait_range(file, start, end); 31 if (ret) 32 return ret; 33 return sync_blockdev(inode->i_sb->s_bdev); 34 } 35 36 /* 37 * generic_file_read often calls bmap with non-existing sector, 38 * so we must ignore such errors. 39 */ 40 41 static secno hpfs_bmap(struct inode *inode, unsigned file_secno, unsigned *n_secs) 42 { 43 struct hpfs_inode_info *hpfs_inode = hpfs_i(inode); 44 unsigned n, disk_secno; 45 struct fnode *fnode; 46 struct buffer_head *bh; 47 if (BLOCKS(hpfs_i(inode)->mmu_private) <= file_secno) return 0; 48 n = file_secno - hpfs_inode->i_file_sec; 49 if (n < hpfs_inode->i_n_secs) { 50 *n_secs = hpfs_inode->i_n_secs - n; 51 return hpfs_inode->i_disk_sec + n; 52 } 53 if (!(fnode = hpfs_map_fnode(inode->i_sb, inode->i_ino, &bh))) return 0; 54 disk_secno = hpfs_bplus_lookup(inode->i_sb, inode, &fnode->btree, file_secno, bh); 55 if (disk_secno == -1) return 0; 56 if (hpfs_chk_sectors(inode->i_sb, disk_secno, 1, "bmap")) return 0; 57 n = file_secno - hpfs_inode->i_file_sec; 58 if (n < hpfs_inode->i_n_secs) { 59 *n_secs = hpfs_inode->i_n_secs - n; 60 return hpfs_inode->i_disk_sec + n; 61 } 62 *n_secs = 1; 63 return disk_secno; 64 } 65 66 void hpfs_truncate(struct inode *i) 67 { 68 if (IS_IMMUTABLE(i)) return /*-EPERM*/; 69 hpfs_lock_assert(i->i_sb); 70 71 hpfs_i(i)->i_n_secs = 0; 72 i->i_blocks = 1 + ((i->i_size + 511) >> 9); 73 hpfs_i(i)->mmu_private = i->i_size; 74 hpfs_truncate_btree(i->i_sb, i->i_ino, 1, ((i->i_size + 511) >> 9)); 75 hpfs_write_inode(i); 76 hpfs_i(i)->i_n_secs = 0; 77 } 78 79 static int hpfs_get_block(struct inode *inode, sector_t iblock, struct buffer_head *bh_result, int create) 80 { 81 int r; 82 secno s; 83 unsigned n_secs; 84 hpfs_lock(inode->i_sb); 85 s = hpfs_bmap(inode, iblock, &n_secs); 86 if (s) { 87 if (bh_result->b_size >> 9 < n_secs) 88 n_secs = bh_result->b_size >> 9; 89 n_secs = hpfs_search_hotfix_map_for_range(inode->i_sb, s, n_secs); 90 if (unlikely(!n_secs)) { 91 s = hpfs_search_hotfix_map(inode->i_sb, s); 92 n_secs = 1; 93 } 94 map_bh(bh_result, inode->i_sb, s); 95 bh_result->b_size = n_secs << 9; 96 goto ret_0; 97 } 98 if (!create) goto ret_0; 99 if (iblock<<9 != hpfs_i(inode)->mmu_private) { 100 BUG(); 101 r = -EIO; 102 goto ret_r; 103 } 104 if ((s = hpfs_add_sector_to_btree(inode->i_sb, inode->i_ino, 1, inode->i_blocks - 1)) == -1) { 105 hpfs_truncate_btree(inode->i_sb, inode->i_ino, 1, inode->i_blocks - 1); 106 r = -ENOSPC; 107 goto ret_r; 108 } 109 inode->i_blocks++; 110 hpfs_i(inode)->mmu_private += 512; 111 set_buffer_new(bh_result); 112 map_bh(bh_result, inode->i_sb, hpfs_search_hotfix_map(inode->i_sb, s)); 113 ret_0: 114 r = 0; 115 ret_r: 116 hpfs_unlock(inode->i_sb); 117 return r; 118 } 119 120 static int hpfs_iomap_begin(struct inode *inode, loff_t offset, loff_t length, 121 unsigned flags, struct iomap *iomap, struct iomap *srcmap) 122 { 123 struct super_block *sb = inode->i_sb; 124 unsigned int blkbits = inode->i_blkbits; 125 unsigned int n_secs; 126 secno s; 127 128 if (WARN_ON_ONCE(flags & (IOMAP_WRITE | IOMAP_ZERO))) 129 return -EINVAL; 130 131 iomap->bdev = inode->i_sb->s_bdev; 132 iomap->offset = offset; 133 134 hpfs_lock(sb); 135 s = hpfs_bmap(inode, offset >> blkbits, &n_secs); 136 if (s) { 137 n_secs = hpfs_search_hotfix_map_for_range(sb, s, 138 min_t(loff_t, n_secs, length)); 139 if (unlikely(!n_secs)) { 140 s = hpfs_search_hotfix_map(sb, s); 141 n_secs = 1; 142 } 143 iomap->type = IOMAP_MAPPED; 144 iomap->flags = IOMAP_F_MERGED; 145 iomap->addr = (u64)s << blkbits; 146 iomap->length = (u64)n_secs << blkbits; 147 } else { 148 iomap->type = IOMAP_HOLE; 149 iomap->addr = IOMAP_NULL_ADDR; 150 iomap->length = 1 << blkbits; 151 } 152 153 hpfs_unlock(sb); 154 return 0; 155 } 156 157 static const struct iomap_ops hpfs_iomap_ops = { 158 .iomap_begin = hpfs_iomap_begin, 159 }; 160 161 static int hpfs_read_folio(struct file *file, struct folio *folio) 162 { 163 return mpage_read_folio(folio, hpfs_get_block); 164 } 165 166 static int hpfs_writepage(struct page *page, struct writeback_control *wbc) 167 { 168 return block_write_full_page(page, hpfs_get_block, wbc); 169 } 170 171 static void hpfs_readahead(struct readahead_control *rac) 172 { 173 mpage_readahead(rac, hpfs_get_block); 174 } 175 176 static int hpfs_writepages(struct address_space *mapping, 177 struct writeback_control *wbc) 178 { 179 return mpage_writepages(mapping, wbc, hpfs_get_block); 180 } 181 182 static void hpfs_write_failed(struct address_space *mapping, loff_t to) 183 { 184 struct inode *inode = mapping->host; 185 186 hpfs_lock(inode->i_sb); 187 188 if (to > inode->i_size) { 189 truncate_pagecache(inode, inode->i_size); 190 hpfs_truncate(inode); 191 } 192 193 hpfs_unlock(inode->i_sb); 194 } 195 196 static int hpfs_write_begin(struct file *file, struct address_space *mapping, 197 loff_t pos, unsigned len, 198 struct page **pagep, void **fsdata) 199 { 200 int ret; 201 202 *pagep = NULL; 203 ret = cont_write_begin(file, mapping, pos, len, pagep, fsdata, 204 hpfs_get_block, 205 &hpfs_i(mapping->host)->mmu_private); 206 if (unlikely(ret)) 207 hpfs_write_failed(mapping, pos + len); 208 209 return ret; 210 } 211 212 static int hpfs_write_end(struct file *file, struct address_space *mapping, 213 loff_t pos, unsigned len, unsigned copied, 214 struct page *pagep, void *fsdata) 215 { 216 struct inode *inode = mapping->host; 217 int err; 218 err = generic_write_end(file, mapping, pos, len, copied, pagep, fsdata); 219 if (err < len) 220 hpfs_write_failed(mapping, pos + len); 221 if (!(err < 0)) { 222 /* make sure we write it on close, if not earlier */ 223 hpfs_lock(inode->i_sb); 224 hpfs_i(inode)->i_dirty = 1; 225 hpfs_unlock(inode->i_sb); 226 } 227 return err; 228 } 229 230 static sector_t _hpfs_bmap(struct address_space *mapping, sector_t block) 231 { 232 return generic_block_bmap(mapping, block, hpfs_get_block); 233 } 234 235 static int hpfs_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, u64 start, u64 len) 236 { 237 int ret; 238 239 inode_lock(inode); 240 len = min_t(u64, len, i_size_read(inode)); 241 ret = iomap_fiemap(inode, fieinfo, start, len, &hpfs_iomap_ops); 242 inode_unlock(inode); 243 244 return ret; 245 } 246 247 const struct address_space_operations hpfs_aops = { 248 .dirty_folio = block_dirty_folio, 249 .invalidate_folio = block_invalidate_folio, 250 .read_folio = hpfs_read_folio, 251 .writepage = hpfs_writepage, 252 .readahead = hpfs_readahead, 253 .writepages = hpfs_writepages, 254 .write_begin = hpfs_write_begin, 255 .write_end = hpfs_write_end, 256 .bmap = _hpfs_bmap 257 }; 258 259 const struct file_operations hpfs_file_ops = 260 { 261 .llseek = generic_file_llseek, 262 .read_iter = generic_file_read_iter, 263 .write_iter = generic_file_write_iter, 264 .mmap = generic_file_mmap, 265 .release = hpfs_file_release, 266 .fsync = hpfs_file_fsync, 267 .splice_read = generic_file_splice_read, 268 .unlocked_ioctl = hpfs_ioctl, 269 .compat_ioctl = compat_ptr_ioctl, 270 }; 271 272 const struct inode_operations hpfs_file_iops = 273 { 274 .setattr = hpfs_setattr, 275 .fiemap = hpfs_fiemap, 276 }; 277