1 /* 2 * cpfile.c - NILFS checkpoint file. 3 * 4 * Copyright (C) 2006-2008 Nippon Telegraph and Telephone Corporation. 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation; either version 2 of the License, or 9 * (at your option) any later version. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 * You should have received a copy of the GNU General Public License 17 * along with this program; if not, write to the Free Software 18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA 19 * 20 * Written by Koji Sato <koji@osrg.net>. 21 */ 22 23 #include <linux/kernel.h> 24 #include <linux/fs.h> 25 #include <linux/string.h> 26 #include <linux/buffer_head.h> 27 #include <linux/errno.h> 28 #include <linux/nilfs2_fs.h> 29 #include "mdt.h" 30 #include "cpfile.h" 31 32 33 static inline unsigned long 34 nilfs_cpfile_checkpoints_per_block(const struct inode *cpfile) 35 { 36 return NILFS_MDT(cpfile)->mi_entries_per_block; 37 } 38 39 /* block number from the beginning of the file */ 40 static unsigned long 41 nilfs_cpfile_get_blkoff(const struct inode *cpfile, __u64 cno) 42 { 43 __u64 tcno = cno + NILFS_MDT(cpfile)->mi_first_entry_offset - 1; 44 do_div(tcno, nilfs_cpfile_checkpoints_per_block(cpfile)); 45 return (unsigned long)tcno; 46 } 47 48 /* offset in block */ 49 static unsigned long 50 nilfs_cpfile_get_offset(const struct inode *cpfile, __u64 cno) 51 { 52 __u64 tcno = cno + NILFS_MDT(cpfile)->mi_first_entry_offset - 1; 53 return do_div(tcno, nilfs_cpfile_checkpoints_per_block(cpfile)); 54 } 55 56 static unsigned long 57 nilfs_cpfile_checkpoints_in_block(const struct inode *cpfile, 58 __u64 curr, 59 __u64 max) 60 { 61 return min_t(__u64, 62 nilfs_cpfile_checkpoints_per_block(cpfile) - 63 nilfs_cpfile_get_offset(cpfile, curr), 64 max - curr); 65 } 66 67 static inline int nilfs_cpfile_is_in_first(const struct inode *cpfile, 68 __u64 cno) 69 { 70 return nilfs_cpfile_get_blkoff(cpfile, cno) == 0; 71 } 72 73 static unsigned int 74 nilfs_cpfile_block_add_valid_checkpoints(const struct inode *cpfile, 75 struct buffer_head *bh, 76 void *kaddr, 77 unsigned int n) 78 { 79 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); 80 unsigned int count; 81 82 count = le32_to_cpu(cp->cp_checkpoints_count) + n; 83 cp->cp_checkpoints_count = cpu_to_le32(count); 84 return count; 85 } 86 87 static unsigned int 88 nilfs_cpfile_block_sub_valid_checkpoints(const struct inode *cpfile, 89 struct buffer_head *bh, 90 void *kaddr, 91 unsigned int n) 92 { 93 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); 94 unsigned int count; 95 96 WARN_ON(le32_to_cpu(cp->cp_checkpoints_count) < n); 97 count = le32_to_cpu(cp->cp_checkpoints_count) - n; 98 cp->cp_checkpoints_count = cpu_to_le32(count); 99 return count; 100 } 101 102 static inline struct nilfs_cpfile_header * 103 nilfs_cpfile_block_get_header(const struct inode *cpfile, 104 struct buffer_head *bh, 105 void *kaddr) 106 { 107 return kaddr + bh_offset(bh); 108 } 109 110 static struct nilfs_checkpoint * 111 nilfs_cpfile_block_get_checkpoint(const struct inode *cpfile, __u64 cno, 112 struct buffer_head *bh, 113 void *kaddr) 114 { 115 return kaddr + bh_offset(bh) + nilfs_cpfile_get_offset(cpfile, cno) * 116 NILFS_MDT(cpfile)->mi_entry_size; 117 } 118 119 static void nilfs_cpfile_block_init(struct inode *cpfile, 120 struct buffer_head *bh, 121 void *kaddr) 122 { 123 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); 124 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size; 125 int n = nilfs_cpfile_checkpoints_per_block(cpfile); 126 127 while (n-- > 0) { 128 nilfs_checkpoint_set_invalid(cp); 129 cp = (void *)cp + cpsz; 130 } 131 } 132 133 static inline int nilfs_cpfile_get_header_block(struct inode *cpfile, 134 struct buffer_head **bhp) 135 { 136 return nilfs_mdt_get_block(cpfile, 0, 0, NULL, bhp); 137 } 138 139 static inline int nilfs_cpfile_get_checkpoint_block(struct inode *cpfile, 140 __u64 cno, 141 int create, 142 struct buffer_head **bhp) 143 { 144 return nilfs_mdt_get_block(cpfile, 145 nilfs_cpfile_get_blkoff(cpfile, cno), 146 create, nilfs_cpfile_block_init, bhp); 147 } 148 149 static inline int nilfs_cpfile_delete_checkpoint_block(struct inode *cpfile, 150 __u64 cno) 151 { 152 return nilfs_mdt_delete_block(cpfile, 153 nilfs_cpfile_get_blkoff(cpfile, cno)); 154 } 155 156 /** 157 * nilfs_cpfile_get_checkpoint - get a checkpoint 158 * @cpfile: inode of checkpoint file 159 * @cno: checkpoint number 160 * @create: create flag 161 * @cpp: pointer to a checkpoint 162 * @bhp: pointer to a buffer head 163 * 164 * Description: nilfs_cpfile_get_checkpoint() acquires the checkpoint 165 * specified by @cno. A new checkpoint will be created if @cno is the current 166 * checkpoint number and @create is nonzero. 167 * 168 * Return Value: On success, 0 is returned, and the checkpoint and the 169 * buffer head of the buffer on which the checkpoint is located are stored in 170 * the place pointed by @cpp and @bhp, respectively. On error, one of the 171 * following negative error codes is returned. 172 * 173 * %-EIO - I/O error. 174 * 175 * %-ENOMEM - Insufficient amount of memory available. 176 * 177 * %-ENOENT - No such checkpoint. 178 * 179 * %-EINVAL - invalid checkpoint. 180 */ 181 int nilfs_cpfile_get_checkpoint(struct inode *cpfile, 182 __u64 cno, 183 int create, 184 struct nilfs_checkpoint **cpp, 185 struct buffer_head **bhp) 186 { 187 struct buffer_head *header_bh, *cp_bh; 188 struct nilfs_cpfile_header *header; 189 struct nilfs_checkpoint *cp; 190 void *kaddr; 191 int ret; 192 193 if (unlikely(cno < 1 || cno > nilfs_mdt_cno(cpfile) || 194 (cno < nilfs_mdt_cno(cpfile) && create))) 195 return -EINVAL; 196 197 down_write(&NILFS_MDT(cpfile)->mi_sem); 198 199 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh); 200 if (ret < 0) 201 goto out_sem; 202 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, create, &cp_bh); 203 if (ret < 0) 204 goto out_header; 205 kaddr = kmap(cp_bh->b_page); 206 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr); 207 if (nilfs_checkpoint_invalid(cp)) { 208 if (!create) { 209 kunmap(cp_bh->b_page); 210 brelse(cp_bh); 211 ret = -ENOENT; 212 goto out_header; 213 } 214 /* a newly-created checkpoint */ 215 nilfs_checkpoint_clear_invalid(cp); 216 if (!nilfs_cpfile_is_in_first(cpfile, cno)) 217 nilfs_cpfile_block_add_valid_checkpoints(cpfile, cp_bh, 218 kaddr, 1); 219 nilfs_mdt_mark_buffer_dirty(cp_bh); 220 221 kaddr = kmap_atomic(header_bh->b_page, KM_USER0); 222 header = nilfs_cpfile_block_get_header(cpfile, header_bh, 223 kaddr); 224 le64_add_cpu(&header->ch_ncheckpoints, 1); 225 kunmap_atomic(kaddr, KM_USER0); 226 nilfs_mdt_mark_buffer_dirty(header_bh); 227 nilfs_mdt_mark_dirty(cpfile); 228 } 229 230 if (cpp != NULL) 231 *cpp = cp; 232 *bhp = cp_bh; 233 234 out_header: 235 brelse(header_bh); 236 237 out_sem: 238 up_write(&NILFS_MDT(cpfile)->mi_sem); 239 return ret; 240 } 241 242 /** 243 * nilfs_cpfile_put_checkpoint - put a checkpoint 244 * @cpfile: inode of checkpoint file 245 * @cno: checkpoint number 246 * @bh: buffer head 247 * 248 * Description: nilfs_cpfile_put_checkpoint() releases the checkpoint 249 * specified by @cno. @bh must be the buffer head which has been returned by 250 * a previous call to nilfs_cpfile_get_checkpoint() with @cno. 251 */ 252 void nilfs_cpfile_put_checkpoint(struct inode *cpfile, __u64 cno, 253 struct buffer_head *bh) 254 { 255 kunmap(bh->b_page); 256 brelse(bh); 257 } 258 259 /** 260 * nilfs_cpfile_delete_checkpoints - delete checkpoints 261 * @cpfile: inode of checkpoint file 262 * @start: start checkpoint number 263 * @end: end checkpoint numer 264 * 265 * Description: nilfs_cpfile_delete_checkpoints() deletes the checkpoints in 266 * the period from @start to @end, excluding @end itself. The checkpoints 267 * which have been already deleted are ignored. 268 * 269 * Return Value: On success, 0 is returned. On error, one of the following 270 * negative error codes is returned. 271 * 272 * %-EIO - I/O error. 273 * 274 * %-ENOMEM - Insufficient amount of memory available. 275 * 276 * %-EINVAL - invalid checkpoints. 277 */ 278 int nilfs_cpfile_delete_checkpoints(struct inode *cpfile, 279 __u64 start, 280 __u64 end) 281 { 282 struct buffer_head *header_bh, *cp_bh; 283 struct nilfs_cpfile_header *header; 284 struct nilfs_checkpoint *cp; 285 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size; 286 __u64 cno; 287 void *kaddr; 288 unsigned long tnicps; 289 int ret, ncps, nicps, count, i; 290 291 if (unlikely(start == 0 || start > end)) { 292 printk(KERN_ERR "%s: invalid range of checkpoint numbers: " 293 "[%llu, %llu)\n", __func__, 294 (unsigned long long)start, (unsigned long long)end); 295 return -EINVAL; 296 } 297 298 /* cannot delete the latest checkpoint */ 299 if (start == nilfs_mdt_cno(cpfile) - 1) 300 return -EPERM; 301 302 down_write(&NILFS_MDT(cpfile)->mi_sem); 303 304 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh); 305 if (ret < 0) 306 goto out_sem; 307 tnicps = 0; 308 309 for (cno = start; cno < end; cno += ncps) { 310 ncps = nilfs_cpfile_checkpoints_in_block(cpfile, cno, end); 311 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh); 312 if (ret < 0) { 313 if (ret != -ENOENT) 314 goto out_header; 315 /* skip hole */ 316 ret = 0; 317 continue; 318 } 319 320 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0); 321 cp = nilfs_cpfile_block_get_checkpoint( 322 cpfile, cno, cp_bh, kaddr); 323 nicps = 0; 324 for (i = 0; i < ncps; i++, cp = (void *)cp + cpsz) { 325 WARN_ON(nilfs_checkpoint_snapshot(cp)); 326 if (!nilfs_checkpoint_invalid(cp)) { 327 nilfs_checkpoint_set_invalid(cp); 328 nicps++; 329 } 330 } 331 if (nicps > 0) { 332 tnicps += nicps; 333 nilfs_mdt_mark_buffer_dirty(cp_bh); 334 nilfs_mdt_mark_dirty(cpfile); 335 if (!nilfs_cpfile_is_in_first(cpfile, cno) && 336 (count = nilfs_cpfile_block_sub_valid_checkpoints( 337 cpfile, cp_bh, kaddr, nicps)) == 0) { 338 /* make hole */ 339 kunmap_atomic(kaddr, KM_USER0); 340 brelse(cp_bh); 341 ret = nilfs_cpfile_delete_checkpoint_block( 342 cpfile, cno); 343 if (ret == 0) 344 continue; 345 printk(KERN_ERR "%s: cannot delete block\n", 346 __func__); 347 goto out_header; 348 } 349 } 350 351 kunmap_atomic(kaddr, KM_USER0); 352 brelse(cp_bh); 353 } 354 355 if (tnicps > 0) { 356 kaddr = kmap_atomic(header_bh->b_page, KM_USER0); 357 header = nilfs_cpfile_block_get_header(cpfile, header_bh, 358 kaddr); 359 le64_add_cpu(&header->ch_ncheckpoints, -(u64)tnicps); 360 nilfs_mdt_mark_buffer_dirty(header_bh); 361 nilfs_mdt_mark_dirty(cpfile); 362 kunmap_atomic(kaddr, KM_USER0); 363 } 364 365 out_header: 366 brelse(header_bh); 367 368 out_sem: 369 up_write(&NILFS_MDT(cpfile)->mi_sem); 370 return ret; 371 } 372 373 static void nilfs_cpfile_checkpoint_to_cpinfo(struct inode *cpfile, 374 struct nilfs_checkpoint *cp, 375 struct nilfs_cpinfo *ci) 376 { 377 ci->ci_flags = le32_to_cpu(cp->cp_flags); 378 ci->ci_cno = le64_to_cpu(cp->cp_cno); 379 ci->ci_create = le64_to_cpu(cp->cp_create); 380 ci->ci_nblk_inc = le64_to_cpu(cp->cp_nblk_inc); 381 ci->ci_inodes_count = le64_to_cpu(cp->cp_inodes_count); 382 ci->ci_blocks_count = le64_to_cpu(cp->cp_blocks_count); 383 ci->ci_next = le64_to_cpu(cp->cp_snapshot_list.ssl_next); 384 } 385 386 static ssize_t nilfs_cpfile_do_get_cpinfo(struct inode *cpfile, __u64 *cnop, 387 struct nilfs_cpinfo *ci, size_t nci) 388 { 389 struct nilfs_checkpoint *cp; 390 struct buffer_head *bh; 391 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size; 392 __u64 cur_cno = nilfs_mdt_cno(cpfile), cno = *cnop; 393 void *kaddr; 394 int n, ret; 395 int ncps, i; 396 397 if (cno == 0) 398 return -ENOENT; /* checkpoint number 0 is invalid */ 399 down_read(&NILFS_MDT(cpfile)->mi_sem); 400 401 for (n = 0; cno < cur_cno && n < nci; cno += ncps) { 402 ncps = nilfs_cpfile_checkpoints_in_block(cpfile, cno, cur_cno); 403 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh); 404 if (ret < 0) { 405 if (ret != -ENOENT) 406 goto out; 407 continue; /* skip hole */ 408 } 409 410 kaddr = kmap_atomic(bh->b_page, KM_USER0); 411 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr); 412 for (i = 0; i < ncps && n < nci; i++, cp = (void *)cp + cpsz) { 413 if (!nilfs_checkpoint_invalid(cp)) 414 nilfs_cpfile_checkpoint_to_cpinfo( 415 cpfile, cp, &ci[n++]); 416 } 417 kunmap_atomic(kaddr, KM_USER0); 418 brelse(bh); 419 } 420 421 ret = n; 422 if (n > 0) 423 *cnop = ci[n - 1].ci_cno + 1; 424 425 out: 426 up_read(&NILFS_MDT(cpfile)->mi_sem); 427 return ret; 428 } 429 430 static ssize_t nilfs_cpfile_do_get_ssinfo(struct inode *cpfile, __u64 *cnop, 431 struct nilfs_cpinfo *ci, size_t nci) 432 { 433 struct buffer_head *bh; 434 struct nilfs_cpfile_header *header; 435 struct nilfs_checkpoint *cp; 436 __u64 curr = *cnop, next; 437 unsigned long curr_blkoff, next_blkoff; 438 void *kaddr; 439 int n = 0, ret; 440 441 down_read(&NILFS_MDT(cpfile)->mi_sem); 442 443 if (curr == 0) { 444 ret = nilfs_cpfile_get_header_block(cpfile, &bh); 445 if (ret < 0) 446 goto out; 447 kaddr = kmap_atomic(bh->b_page, KM_USER0); 448 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr); 449 curr = le64_to_cpu(header->ch_snapshot_list.ssl_next); 450 kunmap_atomic(kaddr, KM_USER0); 451 brelse(bh); 452 if (curr == 0) { 453 ret = 0; 454 goto out; 455 } 456 } else if (unlikely(curr == ~(__u64)0)) { 457 ret = 0; 458 goto out; 459 } 460 461 curr_blkoff = nilfs_cpfile_get_blkoff(cpfile, curr); 462 ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr, 0, &bh); 463 if (unlikely(ret < 0)) { 464 if (ret == -ENOENT) 465 ret = 0; /* No snapshots (started from a hole block) */ 466 goto out; 467 } 468 kaddr = kmap_atomic(bh->b_page, KM_USER0); 469 while (n < nci) { 470 cp = nilfs_cpfile_block_get_checkpoint(cpfile, curr, bh, kaddr); 471 curr = ~(__u64)0; /* Terminator */ 472 if (unlikely(nilfs_checkpoint_invalid(cp) || 473 !nilfs_checkpoint_snapshot(cp))) 474 break; 475 nilfs_cpfile_checkpoint_to_cpinfo(cpfile, cp, &ci[n++]); 476 next = le64_to_cpu(cp->cp_snapshot_list.ssl_next); 477 if (next == 0) 478 break; /* reach end of the snapshot list */ 479 480 next_blkoff = nilfs_cpfile_get_blkoff(cpfile, next); 481 if (curr_blkoff != next_blkoff) { 482 kunmap_atomic(kaddr, KM_USER0); 483 brelse(bh); 484 ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 485 0, &bh); 486 if (unlikely(ret < 0)) { 487 WARN_ON(ret == -ENOENT); 488 goto out; 489 } 490 kaddr = kmap_atomic(bh->b_page, KM_USER0); 491 } 492 curr = next; 493 curr_blkoff = next_blkoff; 494 } 495 kunmap_atomic(kaddr, KM_USER0); 496 brelse(bh); 497 *cnop = curr; 498 ret = n; 499 500 out: 501 up_read(&NILFS_MDT(cpfile)->mi_sem); 502 return ret; 503 } 504 505 /** 506 * nilfs_cpfile_get_cpinfo - 507 * @cpfile: 508 * @cno: 509 * @ci: 510 * @nci: 511 */ 512 513 ssize_t nilfs_cpfile_get_cpinfo(struct inode *cpfile, __u64 *cnop, int mode, 514 struct nilfs_cpinfo *ci, size_t nci) 515 { 516 switch (mode) { 517 case NILFS_CHECKPOINT: 518 return nilfs_cpfile_do_get_cpinfo(cpfile, cnop, ci, nci); 519 case NILFS_SNAPSHOT: 520 return nilfs_cpfile_do_get_ssinfo(cpfile, cnop, ci, nci); 521 default: 522 return -EINVAL; 523 } 524 } 525 526 /** 527 * nilfs_cpfile_delete_checkpoint - 528 * @cpfile: 529 * @cno: 530 */ 531 int nilfs_cpfile_delete_checkpoint(struct inode *cpfile, __u64 cno) 532 { 533 struct nilfs_cpinfo ci; 534 __u64 tcno = cno; 535 ssize_t nci; 536 int ret; 537 538 nci = nilfs_cpfile_do_get_cpinfo(cpfile, &tcno, &ci, 1); 539 if (nci < 0) 540 return nci; 541 else if (nci == 0 || ci.ci_cno != cno) 542 return -ENOENT; 543 544 /* cannot delete the latest checkpoint nor snapshots */ 545 ret = nilfs_cpinfo_snapshot(&ci); 546 if (ret < 0) 547 return ret; 548 else if (ret > 0 || cno == nilfs_mdt_cno(cpfile) - 1) 549 return -EPERM; 550 551 return nilfs_cpfile_delete_checkpoints(cpfile, cno, cno + 1); 552 } 553 554 static struct nilfs_snapshot_list * 555 nilfs_cpfile_block_get_snapshot_list(const struct inode *cpfile, 556 __u64 cno, 557 struct buffer_head *bh, 558 void *kaddr) 559 { 560 struct nilfs_cpfile_header *header; 561 struct nilfs_checkpoint *cp; 562 struct nilfs_snapshot_list *list; 563 564 if (cno != 0) { 565 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr); 566 list = &cp->cp_snapshot_list; 567 } else { 568 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr); 569 list = &header->ch_snapshot_list; 570 } 571 return list; 572 } 573 574 static int nilfs_cpfile_set_snapshot(struct inode *cpfile, __u64 cno) 575 { 576 struct buffer_head *header_bh, *curr_bh, *prev_bh, *cp_bh; 577 struct nilfs_cpfile_header *header; 578 struct nilfs_checkpoint *cp; 579 struct nilfs_snapshot_list *list; 580 __u64 curr, prev; 581 unsigned long curr_blkoff, prev_blkoff; 582 void *kaddr; 583 int ret; 584 585 if (cno == 0) 586 return -ENOENT; /* checkpoint number 0 is invalid */ 587 down_write(&NILFS_MDT(cpfile)->mi_sem); 588 589 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh); 590 if (ret < 0) 591 goto out_sem; 592 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0); 593 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr); 594 if (nilfs_checkpoint_invalid(cp)) { 595 ret = -ENOENT; 596 kunmap_atomic(kaddr, KM_USER0); 597 goto out_cp; 598 } 599 if (nilfs_checkpoint_snapshot(cp)) { 600 ret = 0; 601 kunmap_atomic(kaddr, KM_USER0); 602 goto out_cp; 603 } 604 kunmap_atomic(kaddr, KM_USER0); 605 606 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh); 607 if (ret < 0) 608 goto out_cp; 609 kaddr = kmap_atomic(header_bh->b_page, KM_USER0); 610 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr); 611 list = &header->ch_snapshot_list; 612 curr_bh = header_bh; 613 get_bh(curr_bh); 614 curr = 0; 615 curr_blkoff = 0; 616 prev = le64_to_cpu(list->ssl_prev); 617 while (prev > cno) { 618 prev_blkoff = nilfs_cpfile_get_blkoff(cpfile, prev); 619 curr = prev; 620 if (curr_blkoff != prev_blkoff) { 621 kunmap_atomic(kaddr, KM_USER0); 622 brelse(curr_bh); 623 ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr, 624 0, &curr_bh); 625 if (ret < 0) 626 goto out_header; 627 kaddr = kmap_atomic(curr_bh->b_page, KM_USER0); 628 } 629 curr_blkoff = prev_blkoff; 630 cp = nilfs_cpfile_block_get_checkpoint( 631 cpfile, curr, curr_bh, kaddr); 632 list = &cp->cp_snapshot_list; 633 prev = le64_to_cpu(list->ssl_prev); 634 } 635 kunmap_atomic(kaddr, KM_USER0); 636 637 if (prev != 0) { 638 ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0, 639 &prev_bh); 640 if (ret < 0) 641 goto out_curr; 642 } else { 643 prev_bh = header_bh; 644 get_bh(prev_bh); 645 } 646 647 kaddr = kmap_atomic(curr_bh->b_page, KM_USER0); 648 list = nilfs_cpfile_block_get_snapshot_list( 649 cpfile, curr, curr_bh, kaddr); 650 list->ssl_prev = cpu_to_le64(cno); 651 kunmap_atomic(kaddr, KM_USER0); 652 653 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0); 654 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr); 655 cp->cp_snapshot_list.ssl_next = cpu_to_le64(curr); 656 cp->cp_snapshot_list.ssl_prev = cpu_to_le64(prev); 657 nilfs_checkpoint_set_snapshot(cp); 658 kunmap_atomic(kaddr, KM_USER0); 659 660 kaddr = kmap_atomic(prev_bh->b_page, KM_USER0); 661 list = nilfs_cpfile_block_get_snapshot_list( 662 cpfile, prev, prev_bh, kaddr); 663 list->ssl_next = cpu_to_le64(cno); 664 kunmap_atomic(kaddr, KM_USER0); 665 666 kaddr = kmap_atomic(header_bh->b_page, KM_USER0); 667 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr); 668 le64_add_cpu(&header->ch_nsnapshots, 1); 669 kunmap_atomic(kaddr, KM_USER0); 670 671 nilfs_mdt_mark_buffer_dirty(prev_bh); 672 nilfs_mdt_mark_buffer_dirty(curr_bh); 673 nilfs_mdt_mark_buffer_dirty(cp_bh); 674 nilfs_mdt_mark_buffer_dirty(header_bh); 675 nilfs_mdt_mark_dirty(cpfile); 676 677 brelse(prev_bh); 678 679 out_curr: 680 brelse(curr_bh); 681 682 out_header: 683 brelse(header_bh); 684 685 out_cp: 686 brelse(cp_bh); 687 688 out_sem: 689 up_write(&NILFS_MDT(cpfile)->mi_sem); 690 return ret; 691 } 692 693 static int nilfs_cpfile_clear_snapshot(struct inode *cpfile, __u64 cno) 694 { 695 struct buffer_head *header_bh, *next_bh, *prev_bh, *cp_bh; 696 struct nilfs_cpfile_header *header; 697 struct nilfs_checkpoint *cp; 698 struct nilfs_snapshot_list *list; 699 __u64 next, prev; 700 void *kaddr; 701 int ret; 702 703 if (cno == 0) 704 return -ENOENT; /* checkpoint number 0 is invalid */ 705 down_write(&NILFS_MDT(cpfile)->mi_sem); 706 707 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh); 708 if (ret < 0) 709 goto out_sem; 710 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0); 711 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr); 712 if (nilfs_checkpoint_invalid(cp)) { 713 ret = -ENOENT; 714 kunmap_atomic(kaddr, KM_USER0); 715 goto out_cp; 716 } 717 if (!nilfs_checkpoint_snapshot(cp)) { 718 ret = 0; 719 kunmap_atomic(kaddr, KM_USER0); 720 goto out_cp; 721 } 722 723 list = &cp->cp_snapshot_list; 724 next = le64_to_cpu(list->ssl_next); 725 prev = le64_to_cpu(list->ssl_prev); 726 kunmap_atomic(kaddr, KM_USER0); 727 728 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh); 729 if (ret < 0) 730 goto out_cp; 731 if (next != 0) { 732 ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 0, 733 &next_bh); 734 if (ret < 0) 735 goto out_header; 736 } else { 737 next_bh = header_bh; 738 get_bh(next_bh); 739 } 740 if (prev != 0) { 741 ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0, 742 &prev_bh); 743 if (ret < 0) 744 goto out_next; 745 } else { 746 prev_bh = header_bh; 747 get_bh(prev_bh); 748 } 749 750 kaddr = kmap_atomic(next_bh->b_page, KM_USER0); 751 list = nilfs_cpfile_block_get_snapshot_list( 752 cpfile, next, next_bh, kaddr); 753 list->ssl_prev = cpu_to_le64(prev); 754 kunmap_atomic(kaddr, KM_USER0); 755 756 kaddr = kmap_atomic(prev_bh->b_page, KM_USER0); 757 list = nilfs_cpfile_block_get_snapshot_list( 758 cpfile, prev, prev_bh, kaddr); 759 list->ssl_next = cpu_to_le64(next); 760 kunmap_atomic(kaddr, KM_USER0); 761 762 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0); 763 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr); 764 cp->cp_snapshot_list.ssl_next = cpu_to_le64(0); 765 cp->cp_snapshot_list.ssl_prev = cpu_to_le64(0); 766 nilfs_checkpoint_clear_snapshot(cp); 767 kunmap_atomic(kaddr, KM_USER0); 768 769 kaddr = kmap_atomic(header_bh->b_page, KM_USER0); 770 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr); 771 le64_add_cpu(&header->ch_nsnapshots, -1); 772 kunmap_atomic(kaddr, KM_USER0); 773 774 nilfs_mdt_mark_buffer_dirty(next_bh); 775 nilfs_mdt_mark_buffer_dirty(prev_bh); 776 nilfs_mdt_mark_buffer_dirty(cp_bh); 777 nilfs_mdt_mark_buffer_dirty(header_bh); 778 nilfs_mdt_mark_dirty(cpfile); 779 780 brelse(prev_bh); 781 782 out_next: 783 brelse(next_bh); 784 785 out_header: 786 brelse(header_bh); 787 788 out_cp: 789 brelse(cp_bh); 790 791 out_sem: 792 up_write(&NILFS_MDT(cpfile)->mi_sem); 793 return ret; 794 } 795 796 /** 797 * nilfs_cpfile_is_snapshot - 798 * @cpfile: inode of checkpoint file 799 * @cno: checkpoint number 800 * 801 * Description: 802 * 803 * Return Value: On success, 1 is returned if the checkpoint specified by 804 * @cno is a snapshot, or 0 if not. On error, one of the following negative 805 * error codes is returned. 806 * 807 * %-EIO - I/O error. 808 * 809 * %-ENOMEM - Insufficient amount of memory available. 810 * 811 * %-ENOENT - No such checkpoint. 812 */ 813 int nilfs_cpfile_is_snapshot(struct inode *cpfile, __u64 cno) 814 { 815 struct buffer_head *bh; 816 struct nilfs_checkpoint *cp; 817 void *kaddr; 818 int ret; 819 820 if (cno == 0) 821 return -ENOENT; /* checkpoint number 0 is invalid */ 822 down_read(&NILFS_MDT(cpfile)->mi_sem); 823 824 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh); 825 if (ret < 0) 826 goto out; 827 kaddr = kmap_atomic(bh->b_page, KM_USER0); 828 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr); 829 ret = nilfs_checkpoint_snapshot(cp); 830 kunmap_atomic(kaddr, KM_USER0); 831 brelse(bh); 832 833 out: 834 up_read(&NILFS_MDT(cpfile)->mi_sem); 835 return ret; 836 } 837 838 /** 839 * nilfs_cpfile_change_cpmode - change checkpoint mode 840 * @cpfile: inode of checkpoint file 841 * @cno: checkpoint number 842 * @status: mode of checkpoint 843 * 844 * Description: nilfs_change_cpmode() changes the mode of the checkpoint 845 * specified by @cno. The mode @mode is NILFS_CHECKPOINT or NILFS_SNAPSHOT. 846 * 847 * Return Value: On success, 0 is returned. On error, one of the following 848 * negative error codes is returned. 849 * 850 * %-EIO - I/O error. 851 * 852 * %-ENOMEM - Insufficient amount of memory available. 853 * 854 * %-ENOENT - No such checkpoint. 855 */ 856 int nilfs_cpfile_change_cpmode(struct inode *cpfile, __u64 cno, int mode) 857 { 858 struct the_nilfs *nilfs; 859 int ret; 860 861 nilfs = NILFS_MDT(cpfile)->mi_nilfs; 862 863 switch (mode) { 864 case NILFS_CHECKPOINT: 865 /* 866 * Check for protecting existing snapshot mounts: 867 * bd_mount_sem is used to make this operation atomic and 868 * exclusive with a new mount job. Though it doesn't cover 869 * umount, it's enough for the purpose. 870 */ 871 down(&nilfs->ns_bdev->bd_mount_sem); 872 if (nilfs_checkpoint_is_mounted(nilfs, cno, 1)) { 873 /* Current implementation does not have to protect 874 plain read-only mounts since they are exclusive 875 with a read/write mount and are protected from the 876 cleaner. */ 877 ret = -EBUSY; 878 } else 879 ret = nilfs_cpfile_clear_snapshot(cpfile, cno); 880 up(&nilfs->ns_bdev->bd_mount_sem); 881 return ret; 882 case NILFS_SNAPSHOT: 883 return nilfs_cpfile_set_snapshot(cpfile, cno); 884 default: 885 return -EINVAL; 886 } 887 } 888 889 /** 890 * nilfs_cpfile_get_stat - get checkpoint statistics 891 * @cpfile: inode of checkpoint file 892 * @stat: pointer to a structure of checkpoint statistics 893 * 894 * Description: nilfs_cpfile_get_stat() returns information about checkpoints. 895 * 896 * Return Value: On success, 0 is returned, and checkpoints information is 897 * stored in the place pointed by @stat. On error, one of the following 898 * negative error codes is returned. 899 * 900 * %-EIO - I/O error. 901 * 902 * %-ENOMEM - Insufficient amount of memory available. 903 */ 904 int nilfs_cpfile_get_stat(struct inode *cpfile, struct nilfs_cpstat *cpstat) 905 { 906 struct buffer_head *bh; 907 struct nilfs_cpfile_header *header; 908 void *kaddr; 909 int ret; 910 911 down_read(&NILFS_MDT(cpfile)->mi_sem); 912 913 ret = nilfs_cpfile_get_header_block(cpfile, &bh); 914 if (ret < 0) 915 goto out_sem; 916 kaddr = kmap_atomic(bh->b_page, KM_USER0); 917 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr); 918 cpstat->cs_cno = nilfs_mdt_cno(cpfile); 919 cpstat->cs_ncps = le64_to_cpu(header->ch_ncheckpoints); 920 cpstat->cs_nsss = le64_to_cpu(header->ch_nsnapshots); 921 kunmap_atomic(kaddr, KM_USER0); 922 brelse(bh); 923 924 out_sem: 925 up_read(&NILFS_MDT(cpfile)->mi_sem); 926 return ret; 927 } 928