xref: /openbmc/linux/fs/fs_struct.c (revision 3ddc8b84)
1 // SPDX-License-Identifier: GPL-2.0-only
2 #include <linux/export.h>
3 #include <linux/sched/signal.h>
4 #include <linux/sched/task.h>
5 #include <linux/fs.h>
6 #include <linux/path.h>
7 #include <linux/slab.h>
8 #include <linux/fs_struct.h>
9 #include "internal.h"
10 
11 /*
12  * Replace the fs->{rootmnt,root} with {mnt,dentry}. Put the old values.
13  * It can block.
14  */
15 void set_fs_root(struct fs_struct *fs, const struct path *path)
16 {
17 	struct path old_root;
18 
19 	path_get(path);
20 	spin_lock(&fs->lock);
21 	write_seqcount_begin(&fs->seq);
22 	old_root = fs->root;
23 	fs->root = *path;
24 	write_seqcount_end(&fs->seq);
25 	spin_unlock(&fs->lock);
26 	if (old_root.dentry)
27 		path_put(&old_root);
28 }
29 
30 /*
31  * Replace the fs->{pwdmnt,pwd} with {mnt,dentry}. Put the old values.
32  * It can block.
33  */
34 void set_fs_pwd(struct fs_struct *fs, const struct path *path)
35 {
36 	struct path old_pwd;
37 
38 	path_get(path);
39 	spin_lock(&fs->lock);
40 	write_seqcount_begin(&fs->seq);
41 	old_pwd = fs->pwd;
42 	fs->pwd = *path;
43 	write_seqcount_end(&fs->seq);
44 	spin_unlock(&fs->lock);
45 
46 	if (old_pwd.dentry)
47 		path_put(&old_pwd);
48 }
49 
50 static inline int replace_path(struct path *p, const struct path *old, const struct path *new)
51 {
52 	if (likely(p->dentry != old->dentry || p->mnt != old->mnt))
53 		return 0;
54 	*p = *new;
55 	return 1;
56 }
57 
58 void chroot_fs_refs(const struct path *old_root, const struct path *new_root)
59 {
60 	struct task_struct *g, *p;
61 	struct fs_struct *fs;
62 	int count = 0;
63 
64 	read_lock(&tasklist_lock);
65 	for_each_process_thread(g, p) {
66 		task_lock(p);
67 		fs = p->fs;
68 		if (fs) {
69 			int hits = 0;
70 			spin_lock(&fs->lock);
71 			write_seqcount_begin(&fs->seq);
72 			hits += replace_path(&fs->root, old_root, new_root);
73 			hits += replace_path(&fs->pwd, old_root, new_root);
74 			write_seqcount_end(&fs->seq);
75 			while (hits--) {
76 				count++;
77 				path_get(new_root);
78 			}
79 			spin_unlock(&fs->lock);
80 		}
81 		task_unlock(p);
82 	}
83 	read_unlock(&tasklist_lock);
84 	while (count--)
85 		path_put(old_root);
86 }
87 
88 void free_fs_struct(struct fs_struct *fs)
89 {
90 	path_put(&fs->root);
91 	path_put(&fs->pwd);
92 	kmem_cache_free(fs_cachep, fs);
93 }
94 
95 void exit_fs(struct task_struct *tsk)
96 {
97 	struct fs_struct *fs = tsk->fs;
98 
99 	if (fs) {
100 		int kill;
101 		task_lock(tsk);
102 		spin_lock(&fs->lock);
103 		tsk->fs = NULL;
104 		kill = !--fs->users;
105 		spin_unlock(&fs->lock);
106 		task_unlock(tsk);
107 		if (kill)
108 			free_fs_struct(fs);
109 	}
110 }
111 
112 struct fs_struct *copy_fs_struct(struct fs_struct *old)
113 {
114 	struct fs_struct *fs = kmem_cache_alloc(fs_cachep, GFP_KERNEL);
115 	/* We don't need to lock fs - think why ;-) */
116 	if (fs) {
117 		fs->users = 1;
118 		fs->in_exec = 0;
119 		spin_lock_init(&fs->lock);
120 		seqcount_spinlock_init(&fs->seq, &fs->lock);
121 		fs->umask = old->umask;
122 
123 		spin_lock(&old->lock);
124 		fs->root = old->root;
125 		path_get(&fs->root);
126 		fs->pwd = old->pwd;
127 		path_get(&fs->pwd);
128 		spin_unlock(&old->lock);
129 	}
130 	return fs;
131 }
132 
133 int unshare_fs_struct(void)
134 {
135 	struct fs_struct *fs = current->fs;
136 	struct fs_struct *new_fs = copy_fs_struct(fs);
137 	int kill;
138 
139 	if (!new_fs)
140 		return -ENOMEM;
141 
142 	task_lock(current);
143 	spin_lock(&fs->lock);
144 	kill = !--fs->users;
145 	current->fs = new_fs;
146 	spin_unlock(&fs->lock);
147 	task_unlock(current);
148 
149 	if (kill)
150 		free_fs_struct(fs);
151 
152 	return 0;
153 }
154 EXPORT_SYMBOL_GPL(unshare_fs_struct);
155 
156 int current_umask(void)
157 {
158 	return current->fs->umask;
159 }
160 EXPORT_SYMBOL(current_umask);
161 
162 /* to be mentioned only in INIT_TASK */
163 struct fs_struct init_fs = {
164 	.users		= 1,
165 	.lock		= __SPIN_LOCK_UNLOCKED(init_fs.lock),
166 	.seq		= SEQCNT_SPINLOCK_ZERO(init_fs.seq, &init_fs.lock),
167 	.umask		= 0022,
168 };
169