1 /* 2 * linux/fs/stat.c 3 * 4 * Copyright (C) 1991, 1992 Linus Torvalds 5 */ 6 7 #include <linux/export.h> 8 #include <linux/mm.h> 9 #include <linux/errno.h> 10 #include <linux/file.h> 11 #include <linux/highuid.h> 12 #include <linux/fs.h> 13 #include <linux/namei.h> 14 #include <linux/security.h> 15 #include <linux/syscalls.h> 16 #include <linux/pagemap.h> 17 18 #include <asm/uaccess.h> 19 #include <asm/unistd.h> 20 21 void generic_fillattr(struct inode *inode, struct kstat *stat) 22 { 23 stat->dev = inode->i_sb->s_dev; 24 stat->ino = inode->i_ino; 25 stat->mode = inode->i_mode; 26 stat->nlink = inode->i_nlink; 27 stat->uid = inode->i_uid; 28 stat->gid = inode->i_gid; 29 stat->rdev = inode->i_rdev; 30 stat->size = i_size_read(inode); 31 stat->atime = inode->i_atime; 32 stat->mtime = inode->i_mtime; 33 stat->ctime = inode->i_ctime; 34 stat->blksize = (1 << inode->i_blkbits); 35 stat->blocks = inode->i_blocks; 36 } 37 38 EXPORT_SYMBOL(generic_fillattr); 39 40 int vfs_getattr(struct vfsmount *mnt, struct dentry *dentry, struct kstat *stat) 41 { 42 struct inode *inode = dentry->d_inode; 43 int retval; 44 45 retval = security_inode_getattr(mnt, dentry); 46 if (retval) 47 return retval; 48 49 if (inode->i_op->getattr) 50 return inode->i_op->getattr(mnt, dentry, stat); 51 52 generic_fillattr(inode, stat); 53 return 0; 54 } 55 56 EXPORT_SYMBOL(vfs_getattr); 57 58 int vfs_fstat(unsigned int fd, struct kstat *stat) 59 { 60 struct fd f = fdget_raw(fd); 61 int error = -EBADF; 62 63 if (f.file) { 64 error = vfs_getattr(f.file->f_path.mnt, f.file->f_path.dentry, 65 stat); 66 fdput(f); 67 } 68 return error; 69 } 70 EXPORT_SYMBOL(vfs_fstat); 71 72 int vfs_fstatat(int dfd, const char __user *filename, struct kstat *stat, 73 int flag) 74 { 75 struct path path; 76 int error = -EINVAL; 77 unsigned int lookup_flags = 0; 78 79 if ((flag & ~(AT_SYMLINK_NOFOLLOW | AT_NO_AUTOMOUNT | 80 AT_EMPTY_PATH)) != 0) 81 goto out; 82 83 if (!(flag & AT_SYMLINK_NOFOLLOW)) 84 lookup_flags |= LOOKUP_FOLLOW; 85 if (flag & AT_EMPTY_PATH) 86 lookup_flags |= LOOKUP_EMPTY; 87 retry: 88 error = user_path_at(dfd, filename, lookup_flags, &path); 89 if (error) 90 goto out; 91 92 error = vfs_getattr(path.mnt, path.dentry, stat); 93 path_put(&path); 94 if (retry_estale(error, lookup_flags)) { 95 lookup_flags |= LOOKUP_REVAL; 96 goto retry; 97 } 98 out: 99 return error; 100 } 101 EXPORT_SYMBOL(vfs_fstatat); 102 103 int vfs_stat(const char __user *name, struct kstat *stat) 104 { 105 return vfs_fstatat(AT_FDCWD, name, stat, 0); 106 } 107 EXPORT_SYMBOL(vfs_stat); 108 109 int vfs_lstat(const char __user *name, struct kstat *stat) 110 { 111 return vfs_fstatat(AT_FDCWD, name, stat, AT_SYMLINK_NOFOLLOW); 112 } 113 EXPORT_SYMBOL(vfs_lstat); 114 115 116 #ifdef __ARCH_WANT_OLD_STAT 117 118 /* 119 * For backward compatibility? Maybe this should be moved 120 * into arch/i386 instead? 121 */ 122 static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * statbuf) 123 { 124 static int warncount = 5; 125 struct __old_kernel_stat tmp; 126 127 if (warncount > 0) { 128 warncount--; 129 printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n", 130 current->comm); 131 } else if (warncount < 0) { 132 /* it's laughable, but... */ 133 warncount = 0; 134 } 135 136 memset(&tmp, 0, sizeof(struct __old_kernel_stat)); 137 tmp.st_dev = old_encode_dev(stat->dev); 138 tmp.st_ino = stat->ino; 139 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 140 return -EOVERFLOW; 141 tmp.st_mode = stat->mode; 142 tmp.st_nlink = stat->nlink; 143 if (tmp.st_nlink != stat->nlink) 144 return -EOVERFLOW; 145 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 146 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 147 tmp.st_rdev = old_encode_dev(stat->rdev); 148 #if BITS_PER_LONG == 32 149 if (stat->size > MAX_NON_LFS) 150 return -EOVERFLOW; 151 #endif 152 tmp.st_size = stat->size; 153 tmp.st_atime = stat->atime.tv_sec; 154 tmp.st_mtime = stat->mtime.tv_sec; 155 tmp.st_ctime = stat->ctime.tv_sec; 156 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 157 } 158 159 SYSCALL_DEFINE2(stat, const char __user *, filename, 160 struct __old_kernel_stat __user *, statbuf) 161 { 162 struct kstat stat; 163 int error; 164 165 error = vfs_stat(filename, &stat); 166 if (error) 167 return error; 168 169 return cp_old_stat(&stat, statbuf); 170 } 171 172 SYSCALL_DEFINE2(lstat, const char __user *, filename, 173 struct __old_kernel_stat __user *, statbuf) 174 { 175 struct kstat stat; 176 int error; 177 178 error = vfs_lstat(filename, &stat); 179 if (error) 180 return error; 181 182 return cp_old_stat(&stat, statbuf); 183 } 184 185 SYSCALL_DEFINE2(fstat, unsigned int, fd, struct __old_kernel_stat __user *, statbuf) 186 { 187 struct kstat stat; 188 int error = vfs_fstat(fd, &stat); 189 190 if (!error) 191 error = cp_old_stat(&stat, statbuf); 192 193 return error; 194 } 195 196 #endif /* __ARCH_WANT_OLD_STAT */ 197 198 #if BITS_PER_LONG == 32 199 # define choose_32_64(a,b) a 200 #else 201 # define choose_32_64(a,b) b 202 #endif 203 204 #define valid_dev(x) choose_32_64(old_valid_dev,new_valid_dev)(x) 205 #define encode_dev(x) choose_32_64(old_encode_dev,new_encode_dev)(x) 206 207 #ifndef INIT_STRUCT_STAT_PADDING 208 # define INIT_STRUCT_STAT_PADDING(st) memset(&st, 0, sizeof(st)) 209 #endif 210 211 static int cp_new_stat(struct kstat *stat, struct stat __user *statbuf) 212 { 213 struct stat tmp; 214 215 if (!valid_dev(stat->dev) || !valid_dev(stat->rdev)) 216 return -EOVERFLOW; 217 #if BITS_PER_LONG == 32 218 if (stat->size > MAX_NON_LFS) 219 return -EOVERFLOW; 220 #endif 221 222 INIT_STRUCT_STAT_PADDING(tmp); 223 tmp.st_dev = encode_dev(stat->dev); 224 tmp.st_ino = stat->ino; 225 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 226 return -EOVERFLOW; 227 tmp.st_mode = stat->mode; 228 tmp.st_nlink = stat->nlink; 229 if (tmp.st_nlink != stat->nlink) 230 return -EOVERFLOW; 231 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 232 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 233 tmp.st_rdev = encode_dev(stat->rdev); 234 tmp.st_size = stat->size; 235 tmp.st_atime = stat->atime.tv_sec; 236 tmp.st_mtime = stat->mtime.tv_sec; 237 tmp.st_ctime = stat->ctime.tv_sec; 238 #ifdef STAT_HAVE_NSEC 239 tmp.st_atime_nsec = stat->atime.tv_nsec; 240 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 241 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 242 #endif 243 tmp.st_blocks = stat->blocks; 244 tmp.st_blksize = stat->blksize; 245 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 246 } 247 248 SYSCALL_DEFINE2(newstat, const char __user *, filename, 249 struct stat __user *, statbuf) 250 { 251 struct kstat stat; 252 int error = vfs_stat(filename, &stat); 253 254 if (error) 255 return error; 256 return cp_new_stat(&stat, statbuf); 257 } 258 259 SYSCALL_DEFINE2(newlstat, const char __user *, filename, 260 struct stat __user *, statbuf) 261 { 262 struct kstat stat; 263 int error; 264 265 error = vfs_lstat(filename, &stat); 266 if (error) 267 return error; 268 269 return cp_new_stat(&stat, statbuf); 270 } 271 272 #if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT) 273 SYSCALL_DEFINE4(newfstatat, int, dfd, const char __user *, filename, 274 struct stat __user *, statbuf, int, flag) 275 { 276 struct kstat stat; 277 int error; 278 279 error = vfs_fstatat(dfd, filename, &stat, flag); 280 if (error) 281 return error; 282 return cp_new_stat(&stat, statbuf); 283 } 284 #endif 285 286 SYSCALL_DEFINE2(newfstat, unsigned int, fd, struct stat __user *, statbuf) 287 { 288 struct kstat stat; 289 int error = vfs_fstat(fd, &stat); 290 291 if (!error) 292 error = cp_new_stat(&stat, statbuf); 293 294 return error; 295 } 296 297 SYSCALL_DEFINE4(readlinkat, int, dfd, const char __user *, pathname, 298 char __user *, buf, int, bufsiz) 299 { 300 struct path path; 301 int error; 302 int empty = 0; 303 unsigned int lookup_flags = LOOKUP_EMPTY; 304 305 if (bufsiz <= 0) 306 return -EINVAL; 307 308 retry: 309 error = user_path_at_empty(dfd, pathname, lookup_flags, &path, &empty); 310 if (!error) { 311 struct inode *inode = path.dentry->d_inode; 312 313 error = empty ? -ENOENT : -EINVAL; 314 if (inode->i_op->readlink) { 315 error = security_inode_readlink(path.dentry); 316 if (!error) { 317 touch_atime(&path); 318 error = inode->i_op->readlink(path.dentry, 319 buf, bufsiz); 320 } 321 } 322 path_put(&path); 323 if (retry_estale(error, lookup_flags)) { 324 lookup_flags |= LOOKUP_REVAL; 325 goto retry; 326 } 327 } 328 return error; 329 } 330 331 SYSCALL_DEFINE3(readlink, const char __user *, path, char __user *, buf, 332 int, bufsiz) 333 { 334 return sys_readlinkat(AT_FDCWD, path, buf, bufsiz); 335 } 336 337 338 /* ---------- LFS-64 ----------- */ 339 #if defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_COMPAT_STAT64) 340 341 #ifndef INIT_STRUCT_STAT64_PADDING 342 # define INIT_STRUCT_STAT64_PADDING(st) memset(&st, 0, sizeof(st)) 343 #endif 344 345 static long cp_new_stat64(struct kstat *stat, struct stat64 __user *statbuf) 346 { 347 struct stat64 tmp; 348 349 INIT_STRUCT_STAT64_PADDING(tmp); 350 #ifdef CONFIG_MIPS 351 /* mips has weird padding, so we don't get 64 bits there */ 352 if (!new_valid_dev(stat->dev) || !new_valid_dev(stat->rdev)) 353 return -EOVERFLOW; 354 tmp.st_dev = new_encode_dev(stat->dev); 355 tmp.st_rdev = new_encode_dev(stat->rdev); 356 #else 357 tmp.st_dev = huge_encode_dev(stat->dev); 358 tmp.st_rdev = huge_encode_dev(stat->rdev); 359 #endif 360 tmp.st_ino = stat->ino; 361 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 362 return -EOVERFLOW; 363 #ifdef STAT64_HAS_BROKEN_ST_INO 364 tmp.__st_ino = stat->ino; 365 #endif 366 tmp.st_mode = stat->mode; 367 tmp.st_nlink = stat->nlink; 368 tmp.st_uid = from_kuid_munged(current_user_ns(), stat->uid); 369 tmp.st_gid = from_kgid_munged(current_user_ns(), stat->gid); 370 tmp.st_atime = stat->atime.tv_sec; 371 tmp.st_atime_nsec = stat->atime.tv_nsec; 372 tmp.st_mtime = stat->mtime.tv_sec; 373 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 374 tmp.st_ctime = stat->ctime.tv_sec; 375 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 376 tmp.st_size = stat->size; 377 tmp.st_blocks = stat->blocks; 378 tmp.st_blksize = stat->blksize; 379 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 380 } 381 382 SYSCALL_DEFINE2(stat64, const char __user *, filename, 383 struct stat64 __user *, statbuf) 384 { 385 struct kstat stat; 386 int error = vfs_stat(filename, &stat); 387 388 if (!error) 389 error = cp_new_stat64(&stat, statbuf); 390 391 return error; 392 } 393 394 SYSCALL_DEFINE2(lstat64, const char __user *, filename, 395 struct stat64 __user *, statbuf) 396 { 397 struct kstat stat; 398 int error = vfs_lstat(filename, &stat); 399 400 if (!error) 401 error = cp_new_stat64(&stat, statbuf); 402 403 return error; 404 } 405 406 SYSCALL_DEFINE2(fstat64, unsigned long, fd, struct stat64 __user *, statbuf) 407 { 408 struct kstat stat; 409 int error = vfs_fstat(fd, &stat); 410 411 if (!error) 412 error = cp_new_stat64(&stat, statbuf); 413 414 return error; 415 } 416 417 SYSCALL_DEFINE4(fstatat64, int, dfd, const char __user *, filename, 418 struct stat64 __user *, statbuf, int, flag) 419 { 420 struct kstat stat; 421 int error; 422 423 error = vfs_fstatat(dfd, filename, &stat, flag); 424 if (error) 425 return error; 426 return cp_new_stat64(&stat, statbuf); 427 } 428 #endif /* __ARCH_WANT_STAT64 || __ARCH_WANT_COMPAT_STAT64 */ 429 430 /* Caller is here responsible for sufficient locking (ie. inode->i_lock) */ 431 void __inode_add_bytes(struct inode *inode, loff_t bytes) 432 { 433 inode->i_blocks += bytes >> 9; 434 bytes &= 511; 435 inode->i_bytes += bytes; 436 if (inode->i_bytes >= 512) { 437 inode->i_blocks++; 438 inode->i_bytes -= 512; 439 } 440 } 441 442 void inode_add_bytes(struct inode *inode, loff_t bytes) 443 { 444 spin_lock(&inode->i_lock); 445 __inode_add_bytes(inode, bytes); 446 spin_unlock(&inode->i_lock); 447 } 448 449 EXPORT_SYMBOL(inode_add_bytes); 450 451 void inode_sub_bytes(struct inode *inode, loff_t bytes) 452 { 453 spin_lock(&inode->i_lock); 454 inode->i_blocks -= bytes >> 9; 455 bytes &= 511; 456 if (inode->i_bytes < bytes) { 457 inode->i_blocks--; 458 inode->i_bytes += 512; 459 } 460 inode->i_bytes -= bytes; 461 spin_unlock(&inode->i_lock); 462 } 463 464 EXPORT_SYMBOL(inode_sub_bytes); 465 466 loff_t inode_get_bytes(struct inode *inode) 467 { 468 loff_t ret; 469 470 spin_lock(&inode->i_lock); 471 ret = (((loff_t)inode->i_blocks) << 9) + inode->i_bytes; 472 spin_unlock(&inode->i_lock); 473 return ret; 474 } 475 476 EXPORT_SYMBOL(inode_get_bytes); 477 478 void inode_set_bytes(struct inode *inode, loff_t bytes) 479 { 480 /* Caller is here responsible for sufficient locking 481 * (ie. inode->i_lock) */ 482 inode->i_blocks = bytes >> 9; 483 inode->i_bytes = bytes & 511; 484 } 485 486 EXPORT_SYMBOL(inode_set_bytes); 487