xref: /openbmc/linux/fs/proc/generic.c (revision 75f25bd3)
1 /*
2  * proc/fs/generic.c --- generic routines for the proc-fs
3  *
4  * This file contains generic proc-fs routines for handling
5  * directories and files.
6  *
7  * Copyright (C) 1991, 1992 Linus Torvalds.
8  * Copyright (C) 1997 Theodore Ts'o
9  */
10 
11 #include <linux/errno.h>
12 #include <linux/time.h>
13 #include <linux/proc_fs.h>
14 #include <linux/stat.h>
15 #include <linux/mm.h>
16 #include <linux/module.h>
17 #include <linux/slab.h>
18 #include <linux/mount.h>
19 #include <linux/init.h>
20 #include <linux/idr.h>
21 #include <linux/namei.h>
22 #include <linux/bitops.h>
23 #include <linux/spinlock.h>
24 #include <linux/completion.h>
25 #include <asm/uaccess.h>
26 
27 #include "internal.h"
28 
29 DEFINE_SPINLOCK(proc_subdir_lock);
30 
31 static int proc_match(unsigned int len, const char *name, struct proc_dir_entry *de)
32 {
33 	if (de->namelen != len)
34 		return 0;
35 	return !memcmp(name, de->name, len);
36 }
37 
38 /* buffer size is one page but our output routines use some slack for overruns */
39 #define PROC_BLOCK_SIZE	(PAGE_SIZE - 1024)
40 
41 static ssize_t
42 __proc_file_read(struct file *file, char __user *buf, size_t nbytes,
43 	       loff_t *ppos)
44 {
45 	struct inode * inode = file->f_path.dentry->d_inode;
46 	char 	*page;
47 	ssize_t	retval=0;
48 	int	eof=0;
49 	ssize_t	n, count;
50 	char	*start;
51 	struct proc_dir_entry * dp;
52 	unsigned long long pos;
53 
54 	/*
55 	 * Gaah, please just use "seq_file" instead. The legacy /proc
56 	 * interfaces cut loff_t down to off_t for reads, and ignore
57 	 * the offset entirely for writes..
58 	 */
59 	pos = *ppos;
60 	if (pos > MAX_NON_LFS)
61 		return 0;
62 	if (nbytes > MAX_NON_LFS - pos)
63 		nbytes = MAX_NON_LFS - pos;
64 
65 	dp = PDE(inode);
66 	if (!(page = (char*) __get_free_page(GFP_TEMPORARY)))
67 		return -ENOMEM;
68 
69 	while ((nbytes > 0) && !eof) {
70 		count = min_t(size_t, PROC_BLOCK_SIZE, nbytes);
71 
72 		start = NULL;
73 		if (dp->read_proc) {
74 			/*
75 			 * How to be a proc read function
76 			 * ------------------------------
77 			 * Prototype:
78 			 *    int f(char *buffer, char **start, off_t offset,
79 			 *          int count, int *peof, void *dat)
80 			 *
81 			 * Assume that the buffer is "count" bytes in size.
82 			 *
83 			 * If you know you have supplied all the data you
84 			 * have, set *peof.
85 			 *
86 			 * You have three ways to return data:
87 			 * 0) Leave *start = NULL.  (This is the default.)
88 			 *    Put the data of the requested offset at that
89 			 *    offset within the buffer.  Return the number (n)
90 			 *    of bytes there are from the beginning of the
91 			 *    buffer up to the last byte of data.  If the
92 			 *    number of supplied bytes (= n - offset) is
93 			 *    greater than zero and you didn't signal eof
94 			 *    and the reader is prepared to take more data
95 			 *    you will be called again with the requested
96 			 *    offset advanced by the number of bytes
97 			 *    absorbed.  This interface is useful for files
98 			 *    no larger than the buffer.
99 			 * 1) Set *start = an unsigned long value less than
100 			 *    the buffer address but greater than zero.
101 			 *    Put the data of the requested offset at the
102 			 *    beginning of the buffer.  Return the number of
103 			 *    bytes of data placed there.  If this number is
104 			 *    greater than zero and you didn't signal eof
105 			 *    and the reader is prepared to take more data
106 			 *    you will be called again with the requested
107 			 *    offset advanced by *start.  This interface is
108 			 *    useful when you have a large file consisting
109 			 *    of a series of blocks which you want to count
110 			 *    and return as wholes.
111 			 *    (Hack by Paul.Russell@rustcorp.com.au)
112 			 * 2) Set *start = an address within the buffer.
113 			 *    Put the data of the requested offset at *start.
114 			 *    Return the number of bytes of data placed there.
115 			 *    If this number is greater than zero and you
116 			 *    didn't signal eof and the reader is prepared to
117 			 *    take more data you will be called again with the
118 			 *    requested offset advanced by the number of bytes
119 			 *    absorbed.
120 			 */
121 			n = dp->read_proc(page, &start, *ppos,
122 					  count, &eof, dp->data);
123 		} else
124 			break;
125 
126 		if (n == 0)   /* end of file */
127 			break;
128 		if (n < 0) {  /* error */
129 			if (retval == 0)
130 				retval = n;
131 			break;
132 		}
133 
134 		if (start == NULL) {
135 			if (n > PAGE_SIZE) {
136 				printk(KERN_ERR
137 				       "proc_file_read: Apparent buffer overflow!\n");
138 				n = PAGE_SIZE;
139 			}
140 			n -= *ppos;
141 			if (n <= 0)
142 				break;
143 			if (n > count)
144 				n = count;
145 			start = page + *ppos;
146 		} else if (start < page) {
147 			if (n > PAGE_SIZE) {
148 				printk(KERN_ERR
149 				       "proc_file_read: Apparent buffer overflow!\n");
150 				n = PAGE_SIZE;
151 			}
152 			if (n > count) {
153 				/*
154 				 * Don't reduce n because doing so might
155 				 * cut off part of a data block.
156 				 */
157 				printk(KERN_WARNING
158 				       "proc_file_read: Read count exceeded\n");
159 			}
160 		} else /* start >= page */ {
161 			unsigned long startoff = (unsigned long)(start - page);
162 			if (n > (PAGE_SIZE - startoff)) {
163 				printk(KERN_ERR
164 				       "proc_file_read: Apparent buffer overflow!\n");
165 				n = PAGE_SIZE - startoff;
166 			}
167 			if (n > count)
168 				n = count;
169 		}
170 
171  		n -= copy_to_user(buf, start < page ? page : start, n);
172 		if (n == 0) {
173 			if (retval == 0)
174 				retval = -EFAULT;
175 			break;
176 		}
177 
178 		*ppos += start < page ? (unsigned long)start : n;
179 		nbytes -= n;
180 		buf += n;
181 		retval += n;
182 	}
183 	free_page((unsigned long) page);
184 	return retval;
185 }
186 
187 static ssize_t
188 proc_file_read(struct file *file, char __user *buf, size_t nbytes,
189 	       loff_t *ppos)
190 {
191 	struct proc_dir_entry *pde = PDE(file->f_path.dentry->d_inode);
192 	ssize_t rv = -EIO;
193 
194 	spin_lock(&pde->pde_unload_lock);
195 	if (!pde->proc_fops) {
196 		spin_unlock(&pde->pde_unload_lock);
197 		return rv;
198 	}
199 	pde->pde_users++;
200 	spin_unlock(&pde->pde_unload_lock);
201 
202 	rv = __proc_file_read(file, buf, nbytes, ppos);
203 
204 	pde_users_dec(pde);
205 	return rv;
206 }
207 
208 static ssize_t
209 proc_file_write(struct file *file, const char __user *buffer,
210 		size_t count, loff_t *ppos)
211 {
212 	struct proc_dir_entry *pde = PDE(file->f_path.dentry->d_inode);
213 	ssize_t rv = -EIO;
214 
215 	if (pde->write_proc) {
216 		spin_lock(&pde->pde_unload_lock);
217 		if (!pde->proc_fops) {
218 			spin_unlock(&pde->pde_unload_lock);
219 			return rv;
220 		}
221 		pde->pde_users++;
222 		spin_unlock(&pde->pde_unload_lock);
223 
224 		/* FIXME: does this routine need ppos?  probably... */
225 		rv = pde->write_proc(file, buffer, count, pde->data);
226 		pde_users_dec(pde);
227 	}
228 	return rv;
229 }
230 
231 
232 static loff_t
233 proc_file_lseek(struct file *file, loff_t offset, int orig)
234 {
235 	loff_t retval = -EINVAL;
236 	switch (orig) {
237 	case 1:
238 		offset += file->f_pos;
239 	/* fallthrough */
240 	case 0:
241 		if (offset < 0 || offset > MAX_NON_LFS)
242 			break;
243 		file->f_pos = retval = offset;
244 	}
245 	return retval;
246 }
247 
248 static const struct file_operations proc_file_operations = {
249 	.llseek		= proc_file_lseek,
250 	.read		= proc_file_read,
251 	.write		= proc_file_write,
252 };
253 
254 static int proc_notify_change(struct dentry *dentry, struct iattr *iattr)
255 {
256 	struct inode *inode = dentry->d_inode;
257 	struct proc_dir_entry *de = PDE(inode);
258 	int error;
259 
260 	error = inode_change_ok(inode, iattr);
261 	if (error)
262 		return error;
263 
264 	if ((iattr->ia_valid & ATTR_SIZE) &&
265 	    iattr->ia_size != i_size_read(inode)) {
266 		error = vmtruncate(inode, iattr->ia_size);
267 		if (error)
268 			return error;
269 	}
270 
271 	setattr_copy(inode, iattr);
272 	mark_inode_dirty(inode);
273 
274 	de->uid = inode->i_uid;
275 	de->gid = inode->i_gid;
276 	de->mode = inode->i_mode;
277 	return 0;
278 }
279 
280 static int proc_getattr(struct vfsmount *mnt, struct dentry *dentry,
281 			struct kstat *stat)
282 {
283 	struct inode *inode = dentry->d_inode;
284 	struct proc_dir_entry *de = PROC_I(inode)->pde;
285 	if (de && de->nlink)
286 		inode->i_nlink = de->nlink;
287 
288 	generic_fillattr(inode, stat);
289 	return 0;
290 }
291 
292 static const struct inode_operations proc_file_inode_operations = {
293 	.setattr	= proc_notify_change,
294 };
295 
296 /*
297  * This function parses a name such as "tty/driver/serial", and
298  * returns the struct proc_dir_entry for "/proc/tty/driver", and
299  * returns "serial" in residual.
300  */
301 static int __xlate_proc_name(const char *name, struct proc_dir_entry **ret,
302 			     const char **residual)
303 {
304 	const char     		*cp = name, *next;
305 	struct proc_dir_entry	*de;
306 	unsigned int		len;
307 
308 	de = *ret;
309 	if (!de)
310 		de = &proc_root;
311 
312 	while (1) {
313 		next = strchr(cp, '/');
314 		if (!next)
315 			break;
316 
317 		len = next - cp;
318 		for (de = de->subdir; de ; de = de->next) {
319 			if (proc_match(len, cp, de))
320 				break;
321 		}
322 		if (!de) {
323 			WARN(1, "name '%s'\n", name);
324 			return -ENOENT;
325 		}
326 		cp += len + 1;
327 	}
328 	*residual = cp;
329 	*ret = de;
330 	return 0;
331 }
332 
333 static int xlate_proc_name(const char *name, struct proc_dir_entry **ret,
334 			   const char **residual)
335 {
336 	int rv;
337 
338 	spin_lock(&proc_subdir_lock);
339 	rv = __xlate_proc_name(name, ret, residual);
340 	spin_unlock(&proc_subdir_lock);
341 	return rv;
342 }
343 
344 static DEFINE_IDA(proc_inum_ida);
345 static DEFINE_SPINLOCK(proc_inum_lock); /* protects the above */
346 
347 #define PROC_DYNAMIC_FIRST 0xF0000000U
348 
349 /*
350  * Return an inode number between PROC_DYNAMIC_FIRST and
351  * 0xffffffff, or zero on failure.
352  */
353 static unsigned int get_inode_number(void)
354 {
355 	unsigned int i;
356 	int error;
357 
358 retry:
359 	if (ida_pre_get(&proc_inum_ida, GFP_KERNEL) == 0)
360 		return 0;
361 
362 	spin_lock(&proc_inum_lock);
363 	error = ida_get_new(&proc_inum_ida, &i);
364 	spin_unlock(&proc_inum_lock);
365 	if (error == -EAGAIN)
366 		goto retry;
367 	else if (error)
368 		return 0;
369 
370 	if (i > UINT_MAX - PROC_DYNAMIC_FIRST) {
371 		spin_lock(&proc_inum_lock);
372 		ida_remove(&proc_inum_ida, i);
373 		spin_unlock(&proc_inum_lock);
374 		return 0;
375 	}
376 	return PROC_DYNAMIC_FIRST + i;
377 }
378 
379 static void release_inode_number(unsigned int inum)
380 {
381 	spin_lock(&proc_inum_lock);
382 	ida_remove(&proc_inum_ida, inum - PROC_DYNAMIC_FIRST);
383 	spin_unlock(&proc_inum_lock);
384 }
385 
386 static void *proc_follow_link(struct dentry *dentry, struct nameidata *nd)
387 {
388 	nd_set_link(nd, PDE(dentry->d_inode)->data);
389 	return NULL;
390 }
391 
392 static const struct inode_operations proc_link_inode_operations = {
393 	.readlink	= generic_readlink,
394 	.follow_link	= proc_follow_link,
395 };
396 
397 /*
398  * As some entries in /proc are volatile, we want to
399  * get rid of unused dentries.  This could be made
400  * smarter: we could keep a "volatile" flag in the
401  * inode to indicate which ones to keep.
402  */
403 static int proc_delete_dentry(const struct dentry * dentry)
404 {
405 	return 1;
406 }
407 
408 static const struct dentry_operations proc_dentry_operations =
409 {
410 	.d_delete	= proc_delete_dentry,
411 };
412 
413 /*
414  * Don't create negative dentries here, return -ENOENT by hand
415  * instead.
416  */
417 struct dentry *proc_lookup_de(struct proc_dir_entry *de, struct inode *dir,
418 		struct dentry *dentry)
419 {
420 	struct inode *inode = NULL;
421 	int error = -ENOENT;
422 
423 	spin_lock(&proc_subdir_lock);
424 	for (de = de->subdir; de ; de = de->next) {
425 		if (de->namelen != dentry->d_name.len)
426 			continue;
427 		if (!memcmp(dentry->d_name.name, de->name, de->namelen)) {
428 			pde_get(de);
429 			spin_unlock(&proc_subdir_lock);
430 			error = -EINVAL;
431 			inode = proc_get_inode(dir->i_sb, de);
432 			goto out_unlock;
433 		}
434 	}
435 	spin_unlock(&proc_subdir_lock);
436 out_unlock:
437 
438 	if (inode) {
439 		d_set_d_op(dentry, &proc_dentry_operations);
440 		d_add(dentry, inode);
441 		return NULL;
442 	}
443 	if (de)
444 		pde_put(de);
445 	return ERR_PTR(error);
446 }
447 
448 struct dentry *proc_lookup(struct inode *dir, struct dentry *dentry,
449 		struct nameidata *nd)
450 {
451 	return proc_lookup_de(PDE(dir), dir, dentry);
452 }
453 
454 /*
455  * This returns non-zero if at EOF, so that the /proc
456  * root directory can use this and check if it should
457  * continue with the <pid> entries..
458  *
459  * Note that the VFS-layer doesn't care about the return
460  * value of the readdir() call, as long as it's non-negative
461  * for success..
462  */
463 int proc_readdir_de(struct proc_dir_entry *de, struct file *filp, void *dirent,
464 		filldir_t filldir)
465 {
466 	unsigned int ino;
467 	int i;
468 	struct inode *inode = filp->f_path.dentry->d_inode;
469 	int ret = 0;
470 
471 	ino = inode->i_ino;
472 	i = filp->f_pos;
473 	switch (i) {
474 		case 0:
475 			if (filldir(dirent, ".", 1, i, ino, DT_DIR) < 0)
476 				goto out;
477 			i++;
478 			filp->f_pos++;
479 			/* fall through */
480 		case 1:
481 			if (filldir(dirent, "..", 2, i,
482 				    parent_ino(filp->f_path.dentry),
483 				    DT_DIR) < 0)
484 				goto out;
485 			i++;
486 			filp->f_pos++;
487 			/* fall through */
488 		default:
489 			spin_lock(&proc_subdir_lock);
490 			de = de->subdir;
491 			i -= 2;
492 			for (;;) {
493 				if (!de) {
494 					ret = 1;
495 					spin_unlock(&proc_subdir_lock);
496 					goto out;
497 				}
498 				if (!i)
499 					break;
500 				de = de->next;
501 				i--;
502 			}
503 
504 			do {
505 				struct proc_dir_entry *next;
506 
507 				/* filldir passes info to user space */
508 				pde_get(de);
509 				spin_unlock(&proc_subdir_lock);
510 				if (filldir(dirent, de->name, de->namelen, filp->f_pos,
511 					    de->low_ino, de->mode >> 12) < 0) {
512 					pde_put(de);
513 					goto out;
514 				}
515 				spin_lock(&proc_subdir_lock);
516 				filp->f_pos++;
517 				next = de->next;
518 				pde_put(de);
519 				de = next;
520 			} while (de);
521 			spin_unlock(&proc_subdir_lock);
522 	}
523 	ret = 1;
524 out:
525 	return ret;
526 }
527 
528 int proc_readdir(struct file *filp, void *dirent, filldir_t filldir)
529 {
530 	struct inode *inode = filp->f_path.dentry->d_inode;
531 
532 	return proc_readdir_de(PDE(inode), filp, dirent, filldir);
533 }
534 
535 /*
536  * These are the generic /proc directory operations. They
537  * use the in-memory "struct proc_dir_entry" tree to parse
538  * the /proc directory.
539  */
540 static const struct file_operations proc_dir_operations = {
541 	.llseek			= generic_file_llseek,
542 	.read			= generic_read_dir,
543 	.readdir		= proc_readdir,
544 };
545 
546 /*
547  * proc directories can do almost nothing..
548  */
549 static const struct inode_operations proc_dir_inode_operations = {
550 	.lookup		= proc_lookup,
551 	.getattr	= proc_getattr,
552 	.setattr	= proc_notify_change,
553 };
554 
555 static int proc_register(struct proc_dir_entry * dir, struct proc_dir_entry * dp)
556 {
557 	unsigned int i;
558 	struct proc_dir_entry *tmp;
559 
560 	i = get_inode_number();
561 	if (i == 0)
562 		return -EAGAIN;
563 	dp->low_ino = i;
564 
565 	if (S_ISDIR(dp->mode)) {
566 		if (dp->proc_iops == NULL) {
567 			dp->proc_fops = &proc_dir_operations;
568 			dp->proc_iops = &proc_dir_inode_operations;
569 		}
570 		dir->nlink++;
571 	} else if (S_ISLNK(dp->mode)) {
572 		if (dp->proc_iops == NULL)
573 			dp->proc_iops = &proc_link_inode_operations;
574 	} else if (S_ISREG(dp->mode)) {
575 		if (dp->proc_fops == NULL)
576 			dp->proc_fops = &proc_file_operations;
577 		if (dp->proc_iops == NULL)
578 			dp->proc_iops = &proc_file_inode_operations;
579 	}
580 
581 	spin_lock(&proc_subdir_lock);
582 
583 	for (tmp = dir->subdir; tmp; tmp = tmp->next)
584 		if (strcmp(tmp->name, dp->name) == 0) {
585 			WARN(1, KERN_WARNING "proc_dir_entry '%s/%s' already registered\n",
586 				dir->name, dp->name);
587 			break;
588 		}
589 
590 	dp->next = dir->subdir;
591 	dp->parent = dir;
592 	dir->subdir = dp;
593 	spin_unlock(&proc_subdir_lock);
594 
595 	return 0;
596 }
597 
598 static struct proc_dir_entry *__proc_create(struct proc_dir_entry **parent,
599 					  const char *name,
600 					  mode_t mode,
601 					  nlink_t nlink)
602 {
603 	struct proc_dir_entry *ent = NULL;
604 	const char *fn = name;
605 	unsigned int len;
606 
607 	/* make sure name is valid */
608 	if (!name || !strlen(name)) goto out;
609 
610 	if (xlate_proc_name(name, parent, &fn) != 0)
611 		goto out;
612 
613 	/* At this point there must not be any '/' characters beyond *fn */
614 	if (strchr(fn, '/'))
615 		goto out;
616 
617 	len = strlen(fn);
618 
619 	ent = kmalloc(sizeof(struct proc_dir_entry) + len + 1, GFP_KERNEL);
620 	if (!ent) goto out;
621 
622 	memset(ent, 0, sizeof(struct proc_dir_entry));
623 	memcpy(ent->name, fn, len + 1);
624 	ent->namelen = len;
625 	ent->mode = mode;
626 	ent->nlink = nlink;
627 	atomic_set(&ent->count, 1);
628 	ent->pde_users = 0;
629 	spin_lock_init(&ent->pde_unload_lock);
630 	ent->pde_unload_completion = NULL;
631 	INIT_LIST_HEAD(&ent->pde_openers);
632  out:
633 	return ent;
634 }
635 
636 struct proc_dir_entry *proc_symlink(const char *name,
637 		struct proc_dir_entry *parent, const char *dest)
638 {
639 	struct proc_dir_entry *ent;
640 
641 	ent = __proc_create(&parent, name,
642 			  (S_IFLNK | S_IRUGO | S_IWUGO | S_IXUGO),1);
643 
644 	if (ent) {
645 		ent->data = kmalloc((ent->size=strlen(dest))+1, GFP_KERNEL);
646 		if (ent->data) {
647 			strcpy((char*)ent->data,dest);
648 			if (proc_register(parent, ent) < 0) {
649 				kfree(ent->data);
650 				kfree(ent);
651 				ent = NULL;
652 			}
653 		} else {
654 			kfree(ent);
655 			ent = NULL;
656 		}
657 	}
658 	return ent;
659 }
660 EXPORT_SYMBOL(proc_symlink);
661 
662 struct proc_dir_entry *proc_mkdir_mode(const char *name, mode_t mode,
663 		struct proc_dir_entry *parent)
664 {
665 	struct proc_dir_entry *ent;
666 
667 	ent = __proc_create(&parent, name, S_IFDIR | mode, 2);
668 	if (ent) {
669 		if (proc_register(parent, ent) < 0) {
670 			kfree(ent);
671 			ent = NULL;
672 		}
673 	}
674 	return ent;
675 }
676 EXPORT_SYMBOL(proc_mkdir_mode);
677 
678 struct proc_dir_entry *proc_net_mkdir(struct net *net, const char *name,
679 		struct proc_dir_entry *parent)
680 {
681 	struct proc_dir_entry *ent;
682 
683 	ent = __proc_create(&parent, name, S_IFDIR | S_IRUGO | S_IXUGO, 2);
684 	if (ent) {
685 		ent->data = net;
686 		if (proc_register(parent, ent) < 0) {
687 			kfree(ent);
688 			ent = NULL;
689 		}
690 	}
691 	return ent;
692 }
693 EXPORT_SYMBOL_GPL(proc_net_mkdir);
694 
695 struct proc_dir_entry *proc_mkdir(const char *name,
696 		struct proc_dir_entry *parent)
697 {
698 	return proc_mkdir_mode(name, S_IRUGO | S_IXUGO, parent);
699 }
700 EXPORT_SYMBOL(proc_mkdir);
701 
702 struct proc_dir_entry *create_proc_entry(const char *name, mode_t mode,
703 					 struct proc_dir_entry *parent)
704 {
705 	struct proc_dir_entry *ent;
706 	nlink_t nlink;
707 
708 	if (S_ISDIR(mode)) {
709 		if ((mode & S_IALLUGO) == 0)
710 			mode |= S_IRUGO | S_IXUGO;
711 		nlink = 2;
712 	} else {
713 		if ((mode & S_IFMT) == 0)
714 			mode |= S_IFREG;
715 		if ((mode & S_IALLUGO) == 0)
716 			mode |= S_IRUGO;
717 		nlink = 1;
718 	}
719 
720 	ent = __proc_create(&parent, name, mode, nlink);
721 	if (ent) {
722 		if (proc_register(parent, ent) < 0) {
723 			kfree(ent);
724 			ent = NULL;
725 		}
726 	}
727 	return ent;
728 }
729 EXPORT_SYMBOL(create_proc_entry);
730 
731 struct proc_dir_entry *proc_create_data(const char *name, mode_t mode,
732 					struct proc_dir_entry *parent,
733 					const struct file_operations *proc_fops,
734 					void *data)
735 {
736 	struct proc_dir_entry *pde;
737 	nlink_t nlink;
738 
739 	if (S_ISDIR(mode)) {
740 		if ((mode & S_IALLUGO) == 0)
741 			mode |= S_IRUGO | S_IXUGO;
742 		nlink = 2;
743 	} else {
744 		if ((mode & S_IFMT) == 0)
745 			mode |= S_IFREG;
746 		if ((mode & S_IALLUGO) == 0)
747 			mode |= S_IRUGO;
748 		nlink = 1;
749 	}
750 
751 	pde = __proc_create(&parent, name, mode, nlink);
752 	if (!pde)
753 		goto out;
754 	pde->proc_fops = proc_fops;
755 	pde->data = data;
756 	if (proc_register(parent, pde) < 0)
757 		goto out_free;
758 	return pde;
759 out_free:
760 	kfree(pde);
761 out:
762 	return NULL;
763 }
764 EXPORT_SYMBOL(proc_create_data);
765 
766 static void free_proc_entry(struct proc_dir_entry *de)
767 {
768 	release_inode_number(de->low_ino);
769 
770 	if (S_ISLNK(de->mode))
771 		kfree(de->data);
772 	kfree(de);
773 }
774 
775 void pde_put(struct proc_dir_entry *pde)
776 {
777 	if (atomic_dec_and_test(&pde->count))
778 		free_proc_entry(pde);
779 }
780 
781 /*
782  * Remove a /proc entry and free it if it's not currently in use.
783  */
784 void remove_proc_entry(const char *name, struct proc_dir_entry *parent)
785 {
786 	struct proc_dir_entry **p;
787 	struct proc_dir_entry *de = NULL;
788 	const char *fn = name;
789 	unsigned int len;
790 
791 	spin_lock(&proc_subdir_lock);
792 	if (__xlate_proc_name(name, &parent, &fn) != 0) {
793 		spin_unlock(&proc_subdir_lock);
794 		return;
795 	}
796 	len = strlen(fn);
797 
798 	for (p = &parent->subdir; *p; p=&(*p)->next ) {
799 		if (proc_match(len, fn, *p)) {
800 			de = *p;
801 			*p = de->next;
802 			de->next = NULL;
803 			break;
804 		}
805 	}
806 	spin_unlock(&proc_subdir_lock);
807 	if (!de) {
808 		WARN(1, "name '%s'\n", name);
809 		return;
810 	}
811 
812 	spin_lock(&de->pde_unload_lock);
813 	/*
814 	 * Stop accepting new callers into module. If you're
815 	 * dynamically allocating ->proc_fops, save a pointer somewhere.
816 	 */
817 	de->proc_fops = NULL;
818 	/* Wait until all existing callers into module are done. */
819 	if (de->pde_users > 0) {
820 		DECLARE_COMPLETION_ONSTACK(c);
821 
822 		if (!de->pde_unload_completion)
823 			de->pde_unload_completion = &c;
824 
825 		spin_unlock(&de->pde_unload_lock);
826 
827 		wait_for_completion(de->pde_unload_completion);
828 
829 		spin_lock(&de->pde_unload_lock);
830 	}
831 
832 	while (!list_empty(&de->pde_openers)) {
833 		struct pde_opener *pdeo;
834 
835 		pdeo = list_first_entry(&de->pde_openers, struct pde_opener, lh);
836 		list_del(&pdeo->lh);
837 		spin_unlock(&de->pde_unload_lock);
838 		pdeo->release(pdeo->inode, pdeo->file);
839 		kfree(pdeo);
840 		spin_lock(&de->pde_unload_lock);
841 	}
842 	spin_unlock(&de->pde_unload_lock);
843 
844 	if (S_ISDIR(de->mode))
845 		parent->nlink--;
846 	de->nlink = 0;
847 	WARN(de->subdir, KERN_WARNING "%s: removing non-empty directory "
848 			"'%s/%s', leaking at least '%s'\n", __func__,
849 			de->parent->name, de->name, de->subdir->name);
850 	pde_put(de);
851 }
852 EXPORT_SYMBOL(remove_proc_entry);
853