1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * linux/fs/stat.c 4 * 5 * Copyright (C) 1991, 1992 Linus Torvalds 6 */ 7 8 #include <linux/export.h> 9 #include <linux/mm.h> 10 #include <linux/errno.h> 11 #include <linux/file.h> 12 #include <linux/highuid.h> 13 #include <linux/fs.h> 14 #include <linux/namei.h> 15 #include <linux/security.h> 16 #include <linux/cred.h> 17 #include <linux/syscalls.h> 18 #include <linux/pagemap.h> 19 #include <linux/compat.h> 20 21 #include <linux/uaccess.h> 22 #include <asm/unistd.h> 23 24 /** 25 * generic_fillattr - Fill in the basic attributes from the inode struct 26 * @inode: Inode to use as the source 27 * @stat: Where to fill in the attributes 28 * 29 * Fill in the basic attributes in the kstat structure from data that's to be 30 * found on the VFS inode structure. This is the default if no getattr inode 31 * operation is supplied. 32 */ 33 void generic_fillattr(struct inode *inode, struct kstat *stat) 34 { 35 stat->dev = inode->i_sb->s_dev; 36 stat->ino = inode->i_ino; 37 stat->mode = inode->i_mode; 38 stat->nlink = inode->i_nlink; 39 stat->uid = inode->i_uid; 40 stat->gid = inode->i_gid; 41 stat->rdev = inode->i_rdev; 42 stat->size = i_size_read(inode); 43 stat->atime = inode->i_atime; 44 stat->mtime = inode->i_mtime; 45 stat->ctime = inode->i_ctime; 46 stat->blksize = i_blocksize(inode); 47 stat->blocks = inode->i_blocks; 48 49 if (IS_NOATIME(inode)) 50 stat->result_mask &= ~STATX_ATIME; 51 if (IS_AUTOMOUNT(inode)) 52 stat->attributes |= STATX_ATTR_AUTOMOUNT; 53 } 54 EXPORT_SYMBOL(generic_fillattr); 55 56 /** 57 * vfs_getattr_nosec - getattr without security checks 58 * @path: file to get attributes from 59 * @stat: structure to return attributes in 60 * @request_mask: STATX_xxx flags indicating what the caller wants 61 * @query_flags: Query mode (KSTAT_QUERY_FLAGS) 62 * 63 * Get attributes without calling security_inode_getattr. 64 * 65 * Currently the only caller other than vfs_getattr is internal to the 66 * filehandle lookup code, which uses only the inode number and returns no 67 * attributes to any user. Any other code probably wants vfs_getattr. 68 */ 69 int vfs_getattr_nosec(const struct path *path, struct kstat *stat, 70 u32 request_mask, unsigned int query_flags) 71 { 72 struct inode *inode = d_backing_inode(path->dentry); 73 74 memset(stat, 0, sizeof(*stat)); 75 stat->result_mask |= STATX_BASIC_STATS; 76 request_mask &= STATX_ALL; 77 query_flags &= KSTAT_QUERY_FLAGS; 78 if (inode->i_op->getattr) 79 return inode->i_op->getattr(path, stat, request_mask, 80 query_flags); 81 82 generic_fillattr(inode, stat); 83 return 0; 84 } 85 EXPORT_SYMBOL(vfs_getattr_nosec); 86 87 /* 88 * vfs_getattr - Get the enhanced basic attributes of a file 89 * @path: The file of interest 90 * @stat: Where to return the statistics 91 * @request_mask: STATX_xxx flags indicating what the caller wants 92 * @query_flags: Query mode (KSTAT_QUERY_FLAGS) 93 * 94 * Ask the filesystem for a file's attributes. The caller must indicate in 95 * request_mask and query_flags to indicate what they want. 96 * 97 * If the file is remote, the filesystem can be forced to update the attributes 98 * from the backing store by passing AT_STATX_FORCE_SYNC in query_flags or can 99 * suppress the update by passing AT_STATX_DONT_SYNC. 100 * 101 * Bits must have been set in request_mask to indicate which attributes the 102 * caller wants retrieving. Any such attribute not requested may be returned 103 * anyway, but the value may be approximate, and, if remote, may not have been 104 * synchronised with the server. 105 * 106 * 0 will be returned on success, and a -ve error code if unsuccessful. 107 */ 108 int vfs_getattr(const struct path *path, struct kstat *stat, 109 u32 request_mask, unsigned int query_flags) 110 { 111 int retval; 112 113 retval = security_inode_getattr(path); 114 if (retval) 115 return retval; 116 return vfs_getattr_nosec(path, stat, request_mask, query_flags); 117 } 118 EXPORT_SYMBOL(vfs_getattr); 119 120 /** 121 * vfs_statx_fd - Get the enhanced basic attributes by file descriptor 122 * @fd: The file descriptor referring to the file of interest 123 * @stat: The result structure to fill in. 124 * @request_mask: STATX_xxx flags indicating what the caller wants 125 * @query_flags: Query mode (KSTAT_QUERY_FLAGS) 126 * 127 * This function is a wrapper around vfs_getattr(). The main difference is 128 * that it uses a file descriptor to determine the file location. 129 * 130 * 0 will be returned on success, and a -ve error code if unsuccessful. 131 */ 132 int vfs_statx_fd(unsigned int fd, struct kstat *stat, 133 u32 request_mask, unsigned int query_flags) 134 { 135 struct fd f; 136 int error = -EBADF; 137 138 if (query_flags & ~KSTAT_QUERY_FLAGS) 139 return -EINVAL; 140 141 f = fdget_raw(fd); 142 if (f.file) { 143 error = vfs_getattr(&f.file->f_path, stat, 144 request_mask, query_flags); 145 fdput(f); 146 } 147 return error; 148 } 149 EXPORT_SYMBOL(vfs_statx_fd); 150 151 /** 152 * vfs_statx - Get basic and extra attributes by filename 153 * @dfd: A file descriptor representing the base dir for a relative filename 154 * @filename: The name of the file of interest 155 * @flags: Flags to control the query 156 * @stat: The result structure to fill in. 157 * @request_mask: STATX_xxx flags indicating what the caller wants 158 * 159 * This function is a wrapper around vfs_getattr(). The main difference is 160 * that it uses a filename and base directory to determine the file location. 161 * Additionally, the use of AT_SYMLINK_NOFOLLOW in flags will prevent a symlink 162 * at the given name from being referenced. 163 * 164 * 0 will be returned on success, and a -ve error code if unsuccessful. 165 */ 166 int vfs_statx(int dfd, const char __user *filename, int flags, 167 struct kstat *stat, u32 request_mask) 168 { 169 struct path path; 170 int error = -EINVAL; 171 unsigned int lookup_flags = LOOKUP_FOLLOW | LOOKUP_AUTOMOUNT; 172 173 if ((flags & ~(AT_SYMLINK_NOFOLLOW | AT_NO_AUTOMOUNT | 174 AT_EMPTY_PATH | KSTAT_QUERY_FLAGS)) != 0) 175 return -EINVAL; 176 177 if (flags & AT_SYMLINK_NOFOLLOW) 178 lookup_flags &= ~LOOKUP_FOLLOW; 179 if (flags & AT_NO_AUTOMOUNT) 180 lookup_flags &= ~LOOKUP_AUTOMOUNT; 181 if (flags & AT_EMPTY_PATH) 182 lookup_flags |= LOOKUP_EMPTY; 183 184 retry: 185 error = user_path_at(dfd, filename, lookup_flags, &path); 186 if (error) 187 goto out; 188 189 error = vfs_getattr(&path, stat, request_mask, flags); 190 path_put(&path); 191 if (retry_estale(error, lookup_flags)) { 192 lookup_flags |= LOOKUP_REVAL; 193 goto retry; 194 } 195 out: 196 return error; 197 } 198 EXPORT_SYMBOL(vfs_statx); 199 200 201 #ifdef __ARCH_WANT_OLD_STAT 202 203 /* 204 * For backward compatibility? Maybe this should be moved 205 * into arch/i386 instead? 206 */ 207 static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * statbuf) 208 { 209 static int warncount = 5; 210 struct __old_kernel_stat tmp; 211 212 if (warncount > 0) { 213 warncount--; 214 printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n", 215 current->comm); 216 } else if (warncount < 0) { 217 /* it's laughable, but... */ 218 warncount = 0; 219 } 220 221 memset(&tmp, 0, sizeof(struct __old_kernel_stat)); 222 tmp.st_dev = old_encode_dev(stat->dev); 223 tmp.st_ino = stat->ino; 224 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 225 return -EOVERFLOW; 226 tmp.st_mode = stat->mode; 227 tmp.st_nlink = stat->nlink; 228 if (tmp.st_nlink != stat->nlink) 229 return -EOVERFLOW; 230 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 231 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 232 tmp.st_rdev = old_encode_dev(stat->rdev); 233 #if BITS_PER_LONG == 32 234 if (stat->size > MAX_NON_LFS) 235 return -EOVERFLOW; 236 #endif 237 tmp.st_size = stat->size; 238 tmp.st_atime = stat->atime.tv_sec; 239 tmp.st_mtime = stat->mtime.tv_sec; 240 tmp.st_ctime = stat->ctime.tv_sec; 241 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 242 } 243 244 SYSCALL_DEFINE2(stat, const char __user *, filename, 245 struct __old_kernel_stat __user *, statbuf) 246 { 247 struct kstat stat; 248 int error; 249 250 error = vfs_stat(filename, &stat); 251 if (error) 252 return error; 253 254 return cp_old_stat(&stat, statbuf); 255 } 256 257 SYSCALL_DEFINE2(lstat, const char __user *, filename, 258 struct __old_kernel_stat __user *, statbuf) 259 { 260 struct kstat stat; 261 int error; 262 263 error = vfs_lstat(filename, &stat); 264 if (error) 265 return error; 266 267 return cp_old_stat(&stat, statbuf); 268 } 269 270 SYSCALL_DEFINE2(fstat, unsigned int, fd, struct __old_kernel_stat __user *, statbuf) 271 { 272 struct kstat stat; 273 int error = vfs_fstat(fd, &stat); 274 275 if (!error) 276 error = cp_old_stat(&stat, statbuf); 277 278 return error; 279 } 280 281 #endif /* __ARCH_WANT_OLD_STAT */ 282 283 #ifdef __ARCH_WANT_NEW_STAT 284 285 #if BITS_PER_LONG == 32 286 # define choose_32_64(a,b) a 287 #else 288 # define choose_32_64(a,b) b 289 #endif 290 291 #define valid_dev(x) choose_32_64(old_valid_dev(x),true) 292 #define encode_dev(x) choose_32_64(old_encode_dev,new_encode_dev)(x) 293 294 #ifndef INIT_STRUCT_STAT_PADDING 295 # define INIT_STRUCT_STAT_PADDING(st) memset(&st, 0, sizeof(st)) 296 #endif 297 298 static int cp_new_stat(struct kstat *stat, struct stat __user *statbuf) 299 { 300 struct stat tmp; 301 302 if (!valid_dev(stat->dev) || !valid_dev(stat->rdev)) 303 return -EOVERFLOW; 304 #if BITS_PER_LONG == 32 305 if (stat->size > MAX_NON_LFS) 306 return -EOVERFLOW; 307 #endif 308 309 INIT_STRUCT_STAT_PADDING(tmp); 310 tmp.st_dev = encode_dev(stat->dev); 311 tmp.st_ino = stat->ino; 312 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 313 return -EOVERFLOW; 314 tmp.st_mode = stat->mode; 315 tmp.st_nlink = stat->nlink; 316 if (tmp.st_nlink != stat->nlink) 317 return -EOVERFLOW; 318 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 319 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 320 tmp.st_rdev = encode_dev(stat->rdev); 321 tmp.st_size = stat->size; 322 tmp.st_atime = stat->atime.tv_sec; 323 tmp.st_mtime = stat->mtime.tv_sec; 324 tmp.st_ctime = stat->ctime.tv_sec; 325 #ifdef STAT_HAVE_NSEC 326 tmp.st_atime_nsec = stat->atime.tv_nsec; 327 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 328 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 329 #endif 330 tmp.st_blocks = stat->blocks; 331 tmp.st_blksize = stat->blksize; 332 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 333 } 334 335 SYSCALL_DEFINE2(newstat, const char __user *, filename, 336 struct stat __user *, statbuf) 337 { 338 struct kstat stat; 339 int error = vfs_stat(filename, &stat); 340 341 if (error) 342 return error; 343 return cp_new_stat(&stat, statbuf); 344 } 345 346 SYSCALL_DEFINE2(newlstat, const char __user *, filename, 347 struct stat __user *, statbuf) 348 { 349 struct kstat stat; 350 int error; 351 352 error = vfs_lstat(filename, &stat); 353 if (error) 354 return error; 355 356 return cp_new_stat(&stat, statbuf); 357 } 358 359 #if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT) 360 SYSCALL_DEFINE4(newfstatat, int, dfd, const char __user *, filename, 361 struct stat __user *, statbuf, int, flag) 362 { 363 struct kstat stat; 364 int error; 365 366 error = vfs_fstatat(dfd, filename, &stat, flag); 367 if (error) 368 return error; 369 return cp_new_stat(&stat, statbuf); 370 } 371 #endif 372 373 SYSCALL_DEFINE2(newfstat, unsigned int, fd, struct stat __user *, statbuf) 374 { 375 struct kstat stat; 376 int error = vfs_fstat(fd, &stat); 377 378 if (!error) 379 error = cp_new_stat(&stat, statbuf); 380 381 return error; 382 } 383 #endif 384 385 static int do_readlinkat(int dfd, const char __user *pathname, 386 char __user *buf, int bufsiz) 387 { 388 struct path path; 389 int error; 390 int empty = 0; 391 unsigned int lookup_flags = LOOKUP_EMPTY; 392 393 if (bufsiz <= 0) 394 return -EINVAL; 395 396 retry: 397 error = user_path_at_empty(dfd, pathname, lookup_flags, &path, &empty); 398 if (!error) { 399 struct inode *inode = d_backing_inode(path.dentry); 400 401 error = empty ? -ENOENT : -EINVAL; 402 /* 403 * AFS mountpoints allow readlink(2) but are not symlinks 404 */ 405 if (d_is_symlink(path.dentry) || inode->i_op->readlink) { 406 error = security_inode_readlink(path.dentry); 407 if (!error) { 408 touch_atime(&path); 409 error = vfs_readlink(path.dentry, buf, bufsiz); 410 } 411 } 412 path_put(&path); 413 if (retry_estale(error, lookup_flags)) { 414 lookup_flags |= LOOKUP_REVAL; 415 goto retry; 416 } 417 } 418 return error; 419 } 420 421 SYSCALL_DEFINE4(readlinkat, int, dfd, const char __user *, pathname, 422 char __user *, buf, int, bufsiz) 423 { 424 return do_readlinkat(dfd, pathname, buf, bufsiz); 425 } 426 427 SYSCALL_DEFINE3(readlink, const char __user *, path, char __user *, buf, 428 int, bufsiz) 429 { 430 return do_readlinkat(AT_FDCWD, path, buf, bufsiz); 431 } 432 433 434 /* ---------- LFS-64 ----------- */ 435 #if defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_COMPAT_STAT64) 436 437 #ifndef INIT_STRUCT_STAT64_PADDING 438 # define INIT_STRUCT_STAT64_PADDING(st) memset(&st, 0, sizeof(st)) 439 #endif 440 441 static long cp_new_stat64(struct kstat *stat, struct stat64 __user *statbuf) 442 { 443 struct stat64 tmp; 444 445 INIT_STRUCT_STAT64_PADDING(tmp); 446 #ifdef CONFIG_MIPS 447 /* mips has weird padding, so we don't get 64 bits there */ 448 tmp.st_dev = new_encode_dev(stat->dev); 449 tmp.st_rdev = new_encode_dev(stat->rdev); 450 #else 451 tmp.st_dev = huge_encode_dev(stat->dev); 452 tmp.st_rdev = huge_encode_dev(stat->rdev); 453 #endif 454 tmp.st_ino = stat->ino; 455 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 456 return -EOVERFLOW; 457 #ifdef STAT64_HAS_BROKEN_ST_INO 458 tmp.__st_ino = stat->ino; 459 #endif 460 tmp.st_mode = stat->mode; 461 tmp.st_nlink = stat->nlink; 462 tmp.st_uid = from_kuid_munged(current_user_ns(), stat->uid); 463 tmp.st_gid = from_kgid_munged(current_user_ns(), stat->gid); 464 tmp.st_atime = stat->atime.tv_sec; 465 tmp.st_atime_nsec = stat->atime.tv_nsec; 466 tmp.st_mtime = stat->mtime.tv_sec; 467 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 468 tmp.st_ctime = stat->ctime.tv_sec; 469 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 470 tmp.st_size = stat->size; 471 tmp.st_blocks = stat->blocks; 472 tmp.st_blksize = stat->blksize; 473 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 474 } 475 476 SYSCALL_DEFINE2(stat64, const char __user *, filename, 477 struct stat64 __user *, statbuf) 478 { 479 struct kstat stat; 480 int error = vfs_stat(filename, &stat); 481 482 if (!error) 483 error = cp_new_stat64(&stat, statbuf); 484 485 return error; 486 } 487 488 SYSCALL_DEFINE2(lstat64, const char __user *, filename, 489 struct stat64 __user *, statbuf) 490 { 491 struct kstat stat; 492 int error = vfs_lstat(filename, &stat); 493 494 if (!error) 495 error = cp_new_stat64(&stat, statbuf); 496 497 return error; 498 } 499 500 SYSCALL_DEFINE2(fstat64, unsigned long, fd, struct stat64 __user *, statbuf) 501 { 502 struct kstat stat; 503 int error = vfs_fstat(fd, &stat); 504 505 if (!error) 506 error = cp_new_stat64(&stat, statbuf); 507 508 return error; 509 } 510 511 SYSCALL_DEFINE4(fstatat64, int, dfd, const char __user *, filename, 512 struct stat64 __user *, statbuf, int, flag) 513 { 514 struct kstat stat; 515 int error; 516 517 error = vfs_fstatat(dfd, filename, &stat, flag); 518 if (error) 519 return error; 520 return cp_new_stat64(&stat, statbuf); 521 } 522 #endif /* __ARCH_WANT_STAT64 || __ARCH_WANT_COMPAT_STAT64 */ 523 524 static noinline_for_stack int 525 cp_statx(const struct kstat *stat, struct statx __user *buffer) 526 { 527 struct statx tmp; 528 529 memset(&tmp, 0, sizeof(tmp)); 530 531 tmp.stx_mask = stat->result_mask; 532 tmp.stx_blksize = stat->blksize; 533 tmp.stx_attributes = stat->attributes; 534 tmp.stx_nlink = stat->nlink; 535 tmp.stx_uid = from_kuid_munged(current_user_ns(), stat->uid); 536 tmp.stx_gid = from_kgid_munged(current_user_ns(), stat->gid); 537 tmp.stx_mode = stat->mode; 538 tmp.stx_ino = stat->ino; 539 tmp.stx_size = stat->size; 540 tmp.stx_blocks = stat->blocks; 541 tmp.stx_attributes_mask = stat->attributes_mask; 542 tmp.stx_atime.tv_sec = stat->atime.tv_sec; 543 tmp.stx_atime.tv_nsec = stat->atime.tv_nsec; 544 tmp.stx_btime.tv_sec = stat->btime.tv_sec; 545 tmp.stx_btime.tv_nsec = stat->btime.tv_nsec; 546 tmp.stx_ctime.tv_sec = stat->ctime.tv_sec; 547 tmp.stx_ctime.tv_nsec = stat->ctime.tv_nsec; 548 tmp.stx_mtime.tv_sec = stat->mtime.tv_sec; 549 tmp.stx_mtime.tv_nsec = stat->mtime.tv_nsec; 550 tmp.stx_rdev_major = MAJOR(stat->rdev); 551 tmp.stx_rdev_minor = MINOR(stat->rdev); 552 tmp.stx_dev_major = MAJOR(stat->dev); 553 tmp.stx_dev_minor = MINOR(stat->dev); 554 555 return copy_to_user(buffer, &tmp, sizeof(tmp)) ? -EFAULT : 0; 556 } 557 558 /** 559 * sys_statx - System call to get enhanced stats 560 * @dfd: Base directory to pathwalk from *or* fd to stat. 561 * @filename: File to stat or "" with AT_EMPTY_PATH 562 * @flags: AT_* flags to control pathwalk. 563 * @mask: Parts of statx struct actually required. 564 * @buffer: Result buffer. 565 * 566 * Note that fstat() can be emulated by setting dfd to the fd of interest, 567 * supplying "" as the filename and setting AT_EMPTY_PATH in the flags. 568 */ 569 SYSCALL_DEFINE5(statx, 570 int, dfd, const char __user *, filename, unsigned, flags, 571 unsigned int, mask, 572 struct statx __user *, buffer) 573 { 574 struct kstat stat; 575 int error; 576 577 if (mask & STATX__RESERVED) 578 return -EINVAL; 579 if ((flags & AT_STATX_SYNC_TYPE) == AT_STATX_SYNC_TYPE) 580 return -EINVAL; 581 582 error = vfs_statx(dfd, filename, flags, &stat, mask); 583 if (error) 584 return error; 585 586 return cp_statx(&stat, buffer); 587 } 588 589 #ifdef CONFIG_COMPAT 590 static int cp_compat_stat(struct kstat *stat, struct compat_stat __user *ubuf) 591 { 592 struct compat_stat tmp; 593 594 if (!old_valid_dev(stat->dev) || !old_valid_dev(stat->rdev)) 595 return -EOVERFLOW; 596 597 memset(&tmp, 0, sizeof(tmp)); 598 tmp.st_dev = old_encode_dev(stat->dev); 599 tmp.st_ino = stat->ino; 600 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 601 return -EOVERFLOW; 602 tmp.st_mode = stat->mode; 603 tmp.st_nlink = stat->nlink; 604 if (tmp.st_nlink != stat->nlink) 605 return -EOVERFLOW; 606 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 607 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 608 tmp.st_rdev = old_encode_dev(stat->rdev); 609 if ((u64) stat->size > MAX_NON_LFS) 610 return -EOVERFLOW; 611 tmp.st_size = stat->size; 612 tmp.st_atime = stat->atime.tv_sec; 613 tmp.st_atime_nsec = stat->atime.tv_nsec; 614 tmp.st_mtime = stat->mtime.tv_sec; 615 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 616 tmp.st_ctime = stat->ctime.tv_sec; 617 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 618 tmp.st_blocks = stat->blocks; 619 tmp.st_blksize = stat->blksize; 620 return copy_to_user(ubuf, &tmp, sizeof(tmp)) ? -EFAULT : 0; 621 } 622 623 COMPAT_SYSCALL_DEFINE2(newstat, const char __user *, filename, 624 struct compat_stat __user *, statbuf) 625 { 626 struct kstat stat; 627 int error; 628 629 error = vfs_stat(filename, &stat); 630 if (error) 631 return error; 632 return cp_compat_stat(&stat, statbuf); 633 } 634 635 COMPAT_SYSCALL_DEFINE2(newlstat, const char __user *, filename, 636 struct compat_stat __user *, statbuf) 637 { 638 struct kstat stat; 639 int error; 640 641 error = vfs_lstat(filename, &stat); 642 if (error) 643 return error; 644 return cp_compat_stat(&stat, statbuf); 645 } 646 647 #ifndef __ARCH_WANT_STAT64 648 COMPAT_SYSCALL_DEFINE4(newfstatat, unsigned int, dfd, 649 const char __user *, filename, 650 struct compat_stat __user *, statbuf, int, flag) 651 { 652 struct kstat stat; 653 int error; 654 655 error = vfs_fstatat(dfd, filename, &stat, flag); 656 if (error) 657 return error; 658 return cp_compat_stat(&stat, statbuf); 659 } 660 #endif 661 662 COMPAT_SYSCALL_DEFINE2(newfstat, unsigned int, fd, 663 struct compat_stat __user *, statbuf) 664 { 665 struct kstat stat; 666 int error = vfs_fstat(fd, &stat); 667 668 if (!error) 669 error = cp_compat_stat(&stat, statbuf); 670 return error; 671 } 672 #endif 673 674 /* Caller is here responsible for sufficient locking (ie. inode->i_lock) */ 675 void __inode_add_bytes(struct inode *inode, loff_t bytes) 676 { 677 inode->i_blocks += bytes >> 9; 678 bytes &= 511; 679 inode->i_bytes += bytes; 680 if (inode->i_bytes >= 512) { 681 inode->i_blocks++; 682 inode->i_bytes -= 512; 683 } 684 } 685 EXPORT_SYMBOL(__inode_add_bytes); 686 687 void inode_add_bytes(struct inode *inode, loff_t bytes) 688 { 689 spin_lock(&inode->i_lock); 690 __inode_add_bytes(inode, bytes); 691 spin_unlock(&inode->i_lock); 692 } 693 694 EXPORT_SYMBOL(inode_add_bytes); 695 696 void __inode_sub_bytes(struct inode *inode, loff_t bytes) 697 { 698 inode->i_blocks -= bytes >> 9; 699 bytes &= 511; 700 if (inode->i_bytes < bytes) { 701 inode->i_blocks--; 702 inode->i_bytes += 512; 703 } 704 inode->i_bytes -= bytes; 705 } 706 707 EXPORT_SYMBOL(__inode_sub_bytes); 708 709 void inode_sub_bytes(struct inode *inode, loff_t bytes) 710 { 711 spin_lock(&inode->i_lock); 712 __inode_sub_bytes(inode, bytes); 713 spin_unlock(&inode->i_lock); 714 } 715 716 EXPORT_SYMBOL(inode_sub_bytes); 717 718 loff_t inode_get_bytes(struct inode *inode) 719 { 720 loff_t ret; 721 722 spin_lock(&inode->i_lock); 723 ret = __inode_get_bytes(inode); 724 spin_unlock(&inode->i_lock); 725 return ret; 726 } 727 728 EXPORT_SYMBOL(inode_get_bytes); 729 730 void inode_set_bytes(struct inode *inode, loff_t bytes) 731 { 732 /* Caller is here responsible for sufficient locking 733 * (ie. inode->i_lock) */ 734 inode->i_blocks = bytes >> 9; 735 inode->i_bytes = bytes & 511; 736 } 737 738 EXPORT_SYMBOL(inode_set_bytes); 739