1 /* 2 * linux/fs/stat.c 3 * 4 * Copyright (C) 1991, 1992 Linus Torvalds 5 */ 6 7 #include <linux/export.h> 8 #include <linux/mm.h> 9 #include <linux/errno.h> 10 #include <linux/file.h> 11 #include <linux/highuid.h> 12 #include <linux/fs.h> 13 #include <linux/namei.h> 14 #include <linux/security.h> 15 #include <linux/syscalls.h> 16 #include <linux/pagemap.h> 17 18 #include <asm/uaccess.h> 19 #include <asm/unistd.h> 20 21 void generic_fillattr(struct inode *inode, struct kstat *stat) 22 { 23 stat->dev = inode->i_sb->s_dev; 24 stat->ino = inode->i_ino; 25 stat->mode = inode->i_mode; 26 stat->nlink = inode->i_nlink; 27 stat->uid = inode->i_uid; 28 stat->gid = inode->i_gid; 29 stat->rdev = inode->i_rdev; 30 stat->size = i_size_read(inode); 31 stat->atime = inode->i_atime; 32 stat->mtime = inode->i_mtime; 33 stat->ctime = inode->i_ctime; 34 stat->blksize = (1 << inode->i_blkbits); 35 stat->blocks = inode->i_blocks; 36 } 37 38 EXPORT_SYMBOL(generic_fillattr); 39 40 int vfs_getattr(struct path *path, struct kstat *stat) 41 { 42 struct inode *inode = path->dentry->d_inode; 43 int retval; 44 45 retval = security_inode_getattr(path->mnt, path->dentry); 46 if (retval) 47 return retval; 48 49 if (inode->i_op->getattr) 50 return inode->i_op->getattr(path->mnt, path->dentry, stat); 51 52 generic_fillattr(inode, stat); 53 return 0; 54 } 55 56 EXPORT_SYMBOL(vfs_getattr); 57 58 int vfs_fstat(unsigned int fd, struct kstat *stat) 59 { 60 struct fd f = fdget_raw(fd); 61 int error = -EBADF; 62 63 if (f.file) { 64 error = vfs_getattr(&f.file->f_path, stat); 65 fdput(f); 66 } 67 return error; 68 } 69 EXPORT_SYMBOL(vfs_fstat); 70 71 int vfs_fstatat(int dfd, const char __user *filename, struct kstat *stat, 72 int flag) 73 { 74 struct path path; 75 int error = -EINVAL; 76 unsigned int lookup_flags = 0; 77 78 if ((flag & ~(AT_SYMLINK_NOFOLLOW | AT_NO_AUTOMOUNT | 79 AT_EMPTY_PATH)) != 0) 80 goto out; 81 82 if (!(flag & AT_SYMLINK_NOFOLLOW)) 83 lookup_flags |= LOOKUP_FOLLOW; 84 if (flag & AT_EMPTY_PATH) 85 lookup_flags |= LOOKUP_EMPTY; 86 retry: 87 error = user_path_at(dfd, filename, lookup_flags, &path); 88 if (error) 89 goto out; 90 91 error = vfs_getattr(&path, stat); 92 path_put(&path); 93 if (retry_estale(error, lookup_flags)) { 94 lookup_flags |= LOOKUP_REVAL; 95 goto retry; 96 } 97 out: 98 return error; 99 } 100 EXPORT_SYMBOL(vfs_fstatat); 101 102 int vfs_stat(const char __user *name, struct kstat *stat) 103 { 104 return vfs_fstatat(AT_FDCWD, name, stat, 0); 105 } 106 EXPORT_SYMBOL(vfs_stat); 107 108 int vfs_lstat(const char __user *name, struct kstat *stat) 109 { 110 return vfs_fstatat(AT_FDCWD, name, stat, AT_SYMLINK_NOFOLLOW); 111 } 112 EXPORT_SYMBOL(vfs_lstat); 113 114 115 #ifdef __ARCH_WANT_OLD_STAT 116 117 /* 118 * For backward compatibility? Maybe this should be moved 119 * into arch/i386 instead? 120 */ 121 static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * statbuf) 122 { 123 static int warncount = 5; 124 struct __old_kernel_stat tmp; 125 126 if (warncount > 0) { 127 warncount--; 128 printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n", 129 current->comm); 130 } else if (warncount < 0) { 131 /* it's laughable, but... */ 132 warncount = 0; 133 } 134 135 memset(&tmp, 0, sizeof(struct __old_kernel_stat)); 136 tmp.st_dev = old_encode_dev(stat->dev); 137 tmp.st_ino = stat->ino; 138 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 139 return -EOVERFLOW; 140 tmp.st_mode = stat->mode; 141 tmp.st_nlink = stat->nlink; 142 if (tmp.st_nlink != stat->nlink) 143 return -EOVERFLOW; 144 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 145 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 146 tmp.st_rdev = old_encode_dev(stat->rdev); 147 #if BITS_PER_LONG == 32 148 if (stat->size > MAX_NON_LFS) 149 return -EOVERFLOW; 150 #endif 151 tmp.st_size = stat->size; 152 tmp.st_atime = stat->atime.tv_sec; 153 tmp.st_mtime = stat->mtime.tv_sec; 154 tmp.st_ctime = stat->ctime.tv_sec; 155 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 156 } 157 158 SYSCALL_DEFINE2(stat, const char __user *, filename, 159 struct __old_kernel_stat __user *, statbuf) 160 { 161 struct kstat stat; 162 int error; 163 164 error = vfs_stat(filename, &stat); 165 if (error) 166 return error; 167 168 return cp_old_stat(&stat, statbuf); 169 } 170 171 SYSCALL_DEFINE2(lstat, const char __user *, filename, 172 struct __old_kernel_stat __user *, statbuf) 173 { 174 struct kstat stat; 175 int error; 176 177 error = vfs_lstat(filename, &stat); 178 if (error) 179 return error; 180 181 return cp_old_stat(&stat, statbuf); 182 } 183 184 SYSCALL_DEFINE2(fstat, unsigned int, fd, struct __old_kernel_stat __user *, statbuf) 185 { 186 struct kstat stat; 187 int error = vfs_fstat(fd, &stat); 188 189 if (!error) 190 error = cp_old_stat(&stat, statbuf); 191 192 return error; 193 } 194 195 #endif /* __ARCH_WANT_OLD_STAT */ 196 197 #if BITS_PER_LONG == 32 198 # define choose_32_64(a,b) a 199 #else 200 # define choose_32_64(a,b) b 201 #endif 202 203 #define valid_dev(x) choose_32_64(old_valid_dev,new_valid_dev)(x) 204 #define encode_dev(x) choose_32_64(old_encode_dev,new_encode_dev)(x) 205 206 #ifndef INIT_STRUCT_STAT_PADDING 207 # define INIT_STRUCT_STAT_PADDING(st) memset(&st, 0, sizeof(st)) 208 #endif 209 210 static int cp_new_stat(struct kstat *stat, struct stat __user *statbuf) 211 { 212 struct stat tmp; 213 214 if (!valid_dev(stat->dev) || !valid_dev(stat->rdev)) 215 return -EOVERFLOW; 216 #if BITS_PER_LONG == 32 217 if (stat->size > MAX_NON_LFS) 218 return -EOVERFLOW; 219 #endif 220 221 INIT_STRUCT_STAT_PADDING(tmp); 222 tmp.st_dev = encode_dev(stat->dev); 223 tmp.st_ino = stat->ino; 224 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 225 return -EOVERFLOW; 226 tmp.st_mode = stat->mode; 227 tmp.st_nlink = stat->nlink; 228 if (tmp.st_nlink != stat->nlink) 229 return -EOVERFLOW; 230 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 231 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 232 tmp.st_rdev = encode_dev(stat->rdev); 233 tmp.st_size = stat->size; 234 tmp.st_atime = stat->atime.tv_sec; 235 tmp.st_mtime = stat->mtime.tv_sec; 236 tmp.st_ctime = stat->ctime.tv_sec; 237 #ifdef STAT_HAVE_NSEC 238 tmp.st_atime_nsec = stat->atime.tv_nsec; 239 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 240 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 241 #endif 242 tmp.st_blocks = stat->blocks; 243 tmp.st_blksize = stat->blksize; 244 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 245 } 246 247 SYSCALL_DEFINE2(newstat, const char __user *, filename, 248 struct stat __user *, statbuf) 249 { 250 struct kstat stat; 251 int error = vfs_stat(filename, &stat); 252 253 if (error) 254 return error; 255 return cp_new_stat(&stat, statbuf); 256 } 257 258 SYSCALL_DEFINE2(newlstat, const char __user *, filename, 259 struct stat __user *, statbuf) 260 { 261 struct kstat stat; 262 int error; 263 264 error = vfs_lstat(filename, &stat); 265 if (error) 266 return error; 267 268 return cp_new_stat(&stat, statbuf); 269 } 270 271 #if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT) 272 SYSCALL_DEFINE4(newfstatat, int, dfd, const char __user *, filename, 273 struct stat __user *, statbuf, int, flag) 274 { 275 struct kstat stat; 276 int error; 277 278 error = vfs_fstatat(dfd, filename, &stat, flag); 279 if (error) 280 return error; 281 return cp_new_stat(&stat, statbuf); 282 } 283 #endif 284 285 SYSCALL_DEFINE2(newfstat, unsigned int, fd, struct stat __user *, statbuf) 286 { 287 struct kstat stat; 288 int error = vfs_fstat(fd, &stat); 289 290 if (!error) 291 error = cp_new_stat(&stat, statbuf); 292 293 return error; 294 } 295 296 SYSCALL_DEFINE4(readlinkat, int, dfd, const char __user *, pathname, 297 char __user *, buf, int, bufsiz) 298 { 299 struct path path; 300 int error; 301 int empty = 0; 302 unsigned int lookup_flags = LOOKUP_EMPTY; 303 304 if (bufsiz <= 0) 305 return -EINVAL; 306 307 retry: 308 error = user_path_at_empty(dfd, pathname, lookup_flags, &path, &empty); 309 if (!error) { 310 struct inode *inode = path.dentry->d_inode; 311 312 error = empty ? -ENOENT : -EINVAL; 313 if (inode->i_op->readlink) { 314 error = security_inode_readlink(path.dentry); 315 if (!error) { 316 touch_atime(&path); 317 error = inode->i_op->readlink(path.dentry, 318 buf, bufsiz); 319 } 320 } 321 path_put(&path); 322 if (retry_estale(error, lookup_flags)) { 323 lookup_flags |= LOOKUP_REVAL; 324 goto retry; 325 } 326 } 327 return error; 328 } 329 330 SYSCALL_DEFINE3(readlink, const char __user *, path, char __user *, buf, 331 int, bufsiz) 332 { 333 return sys_readlinkat(AT_FDCWD, path, buf, bufsiz); 334 } 335 336 337 /* ---------- LFS-64 ----------- */ 338 #if defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_COMPAT_STAT64) 339 340 #ifndef INIT_STRUCT_STAT64_PADDING 341 # define INIT_STRUCT_STAT64_PADDING(st) memset(&st, 0, sizeof(st)) 342 #endif 343 344 static long cp_new_stat64(struct kstat *stat, struct stat64 __user *statbuf) 345 { 346 struct stat64 tmp; 347 348 INIT_STRUCT_STAT64_PADDING(tmp); 349 #ifdef CONFIG_MIPS 350 /* mips has weird padding, so we don't get 64 bits there */ 351 if (!new_valid_dev(stat->dev) || !new_valid_dev(stat->rdev)) 352 return -EOVERFLOW; 353 tmp.st_dev = new_encode_dev(stat->dev); 354 tmp.st_rdev = new_encode_dev(stat->rdev); 355 #else 356 tmp.st_dev = huge_encode_dev(stat->dev); 357 tmp.st_rdev = huge_encode_dev(stat->rdev); 358 #endif 359 tmp.st_ino = stat->ino; 360 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 361 return -EOVERFLOW; 362 #ifdef STAT64_HAS_BROKEN_ST_INO 363 tmp.__st_ino = stat->ino; 364 #endif 365 tmp.st_mode = stat->mode; 366 tmp.st_nlink = stat->nlink; 367 tmp.st_uid = from_kuid_munged(current_user_ns(), stat->uid); 368 tmp.st_gid = from_kgid_munged(current_user_ns(), stat->gid); 369 tmp.st_atime = stat->atime.tv_sec; 370 tmp.st_atime_nsec = stat->atime.tv_nsec; 371 tmp.st_mtime = stat->mtime.tv_sec; 372 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 373 tmp.st_ctime = stat->ctime.tv_sec; 374 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 375 tmp.st_size = stat->size; 376 tmp.st_blocks = stat->blocks; 377 tmp.st_blksize = stat->blksize; 378 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 379 } 380 381 SYSCALL_DEFINE2(stat64, const char __user *, filename, 382 struct stat64 __user *, statbuf) 383 { 384 struct kstat stat; 385 int error = vfs_stat(filename, &stat); 386 387 if (!error) 388 error = cp_new_stat64(&stat, statbuf); 389 390 return error; 391 } 392 393 SYSCALL_DEFINE2(lstat64, const char __user *, filename, 394 struct stat64 __user *, statbuf) 395 { 396 struct kstat stat; 397 int error = vfs_lstat(filename, &stat); 398 399 if (!error) 400 error = cp_new_stat64(&stat, statbuf); 401 402 return error; 403 } 404 405 SYSCALL_DEFINE2(fstat64, unsigned long, fd, struct stat64 __user *, statbuf) 406 { 407 struct kstat stat; 408 int error = vfs_fstat(fd, &stat); 409 410 if (!error) 411 error = cp_new_stat64(&stat, statbuf); 412 413 return error; 414 } 415 416 SYSCALL_DEFINE4(fstatat64, int, dfd, const char __user *, filename, 417 struct stat64 __user *, statbuf, int, flag) 418 { 419 struct kstat stat; 420 int error; 421 422 error = vfs_fstatat(dfd, filename, &stat, flag); 423 if (error) 424 return error; 425 return cp_new_stat64(&stat, statbuf); 426 } 427 #endif /* __ARCH_WANT_STAT64 || __ARCH_WANT_COMPAT_STAT64 */ 428 429 /* Caller is here responsible for sufficient locking (ie. inode->i_lock) */ 430 void __inode_add_bytes(struct inode *inode, loff_t bytes) 431 { 432 inode->i_blocks += bytes >> 9; 433 bytes &= 511; 434 inode->i_bytes += bytes; 435 if (inode->i_bytes >= 512) { 436 inode->i_blocks++; 437 inode->i_bytes -= 512; 438 } 439 } 440 441 void inode_add_bytes(struct inode *inode, loff_t bytes) 442 { 443 spin_lock(&inode->i_lock); 444 __inode_add_bytes(inode, bytes); 445 spin_unlock(&inode->i_lock); 446 } 447 448 EXPORT_SYMBOL(inode_add_bytes); 449 450 void __inode_sub_bytes(struct inode *inode, loff_t bytes) 451 { 452 inode->i_blocks -= bytes >> 9; 453 bytes &= 511; 454 if (inode->i_bytes < bytes) { 455 inode->i_blocks--; 456 inode->i_bytes += 512; 457 } 458 inode->i_bytes -= bytes; 459 } 460 461 EXPORT_SYMBOL(__inode_sub_bytes); 462 463 void inode_sub_bytes(struct inode *inode, loff_t bytes) 464 { 465 spin_lock(&inode->i_lock); 466 __inode_sub_bytes(inode, bytes); 467 spin_unlock(&inode->i_lock); 468 } 469 470 EXPORT_SYMBOL(inode_sub_bytes); 471 472 loff_t inode_get_bytes(struct inode *inode) 473 { 474 loff_t ret; 475 476 spin_lock(&inode->i_lock); 477 ret = (((loff_t)inode->i_blocks) << 9) + inode->i_bytes; 478 spin_unlock(&inode->i_lock); 479 return ret; 480 } 481 482 EXPORT_SYMBOL(inode_get_bytes); 483 484 void inode_set_bytes(struct inode *inode, loff_t bytes) 485 { 486 /* Caller is here responsible for sufficient locking 487 * (ie. inode->i_lock) */ 488 inode->i_blocks = bytes >> 9; 489 inode->i_bytes = bytes & 511; 490 } 491 492 EXPORT_SYMBOL(inode_set_bytes); 493