1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Copyright 2020-21 IBM Corp.
4  */
5 
6 #define pr_fmt(fmt) "vas: " fmt
7 
8 #include <linux/module.h>
9 #include <linux/kernel.h>
10 #include <linux/export.h>
11 #include <linux/types.h>
12 #include <linux/delay.h>
13 #include <linux/slab.h>
14 #include <linux/interrupt.h>
15 #include <linux/irqdomain.h>
16 #include <asm/machdep.h>
17 #include <asm/hvcall.h>
18 #include <asm/plpar_wrappers.h>
19 #include <asm/vas.h>
20 #include "vas.h"
21 
22 #define VAS_INVALID_WIN_ADDRESS	0xFFFFFFFFFFFFFFFFul
23 #define VAS_DEFAULT_DOMAIN_ID	0xFFFFFFFFFFFFFFFFul
24 /* The hypervisor allows one credit per window right now */
25 #define DEF_WIN_CREDS		1
26 
27 static struct vas_all_caps caps_all;
28 static bool copypaste_feat;
29 static struct hv_vas_cop_feat_caps hv_cop_caps;
30 
31 static struct vas_caps vascaps[VAS_MAX_FEAT_TYPE];
32 static DEFINE_MUTEX(vas_pseries_mutex);
33 static bool migration_in_progress;
34 
35 static long hcall_return_busy_check(long rc)
36 {
37 	/* Check if we are stalled for some time */
38 	if (H_IS_LONG_BUSY(rc)) {
39 		msleep(get_longbusy_msecs(rc));
40 		rc = H_BUSY;
41 	} else if (rc == H_BUSY) {
42 		cond_resched();
43 	}
44 
45 	return rc;
46 }
47 
48 /*
49  * Allocate VAS window hcall
50  */
51 static int h_allocate_vas_window(struct pseries_vas_window *win, u64 *domain,
52 				     u8 wintype, u16 credits)
53 {
54 	long retbuf[PLPAR_HCALL9_BUFSIZE] = {0};
55 	long rc;
56 
57 	do {
58 		rc = plpar_hcall9(H_ALLOCATE_VAS_WINDOW, retbuf, wintype,
59 				  credits, domain[0], domain[1], domain[2],
60 				  domain[3], domain[4], domain[5]);
61 
62 		rc = hcall_return_busy_check(rc);
63 	} while (rc == H_BUSY);
64 
65 	if (rc == H_SUCCESS) {
66 		if (win->win_addr == VAS_INVALID_WIN_ADDRESS) {
67 			pr_err("H_ALLOCATE_VAS_WINDOW: COPY/PASTE is not supported\n");
68 			return -ENOTSUPP;
69 		}
70 		win->vas_win.winid = retbuf[0];
71 		win->win_addr = retbuf[1];
72 		win->complete_irq = retbuf[2];
73 		win->fault_irq = retbuf[3];
74 		return 0;
75 	}
76 
77 	pr_err("H_ALLOCATE_VAS_WINDOW error: %ld, wintype: %u, credits: %u\n",
78 		rc, wintype, credits);
79 
80 	return -EIO;
81 }
82 
83 /*
84  * Deallocate VAS window hcall.
85  */
86 static int h_deallocate_vas_window(u64 winid)
87 {
88 	long rc;
89 
90 	do {
91 		rc = plpar_hcall_norets(H_DEALLOCATE_VAS_WINDOW, winid);
92 
93 		rc = hcall_return_busy_check(rc);
94 	} while (rc == H_BUSY);
95 
96 	if (rc == H_SUCCESS)
97 		return 0;
98 
99 	pr_err("H_DEALLOCATE_VAS_WINDOW error: %ld, winid: %llu\n",
100 		rc, winid);
101 	return -EIO;
102 }
103 
104 /*
105  * Modify VAS window.
106  * After the window is opened with allocate window hcall, configure it
107  * with flags and LPAR PID before using.
108  */
109 static int h_modify_vas_window(struct pseries_vas_window *win)
110 {
111 	long rc;
112 
113 	/*
114 	 * AMR value is not supported in Linux VAS implementation.
115 	 * The hypervisor ignores it if 0 is passed.
116 	 */
117 	do {
118 		rc = plpar_hcall_norets(H_MODIFY_VAS_WINDOW,
119 					win->vas_win.winid, win->pid, 0,
120 					VAS_MOD_WIN_FLAGS, 0);
121 
122 		rc = hcall_return_busy_check(rc);
123 	} while (rc == H_BUSY);
124 
125 	if (rc == H_SUCCESS)
126 		return 0;
127 
128 	pr_err("H_MODIFY_VAS_WINDOW error: %ld, winid %u pid %u\n",
129 			rc, win->vas_win.winid, win->pid);
130 	return -EIO;
131 }
132 
133 /*
134  * This hcall is used to determine the capabilities from the hypervisor.
135  * @hcall: H_QUERY_VAS_CAPABILITIES or H_QUERY_NX_CAPABILITIES
136  * @query_type: If 0 is passed, the hypervisor returns the overall
137  *		capabilities which provides all feature(s) that are
138  *		available. Then query the hypervisor to get the
139  *		corresponding capabilities for the specific feature.
140  *		Example: H_QUERY_VAS_CAPABILITIES provides VAS GZIP QoS
141  *			and VAS GZIP Default capabilities.
142  *			H_QUERY_NX_CAPABILITIES provides NX GZIP
143  *			capabilities.
144  * @result: Return buffer to save capabilities.
145  */
146 int h_query_vas_capabilities(const u64 hcall, u8 query_type, u64 result)
147 {
148 	long rc;
149 
150 	rc = plpar_hcall_norets(hcall, query_type, result);
151 
152 	if (rc == H_SUCCESS)
153 		return 0;
154 
155 	/* H_FUNCTION means HV does not support VAS so don't print an error */
156 	if (rc != H_FUNCTION) {
157 		pr_err("%s error %ld, query_type %u, result buffer 0x%llx\n",
158 			(hcall == H_QUERY_VAS_CAPABILITIES) ?
159 				"H_QUERY_VAS_CAPABILITIES" :
160 				"H_QUERY_NX_CAPABILITIES",
161 			rc, query_type, result);
162 	}
163 
164 	return -EIO;
165 }
166 EXPORT_SYMBOL_GPL(h_query_vas_capabilities);
167 
168 /*
169  * hcall to get fault CRB from the hypervisor.
170  */
171 static int h_get_nx_fault(u32 winid, u64 buffer)
172 {
173 	long rc;
174 
175 	rc = plpar_hcall_norets(H_GET_NX_FAULT, winid, buffer);
176 
177 	if (rc == H_SUCCESS)
178 		return 0;
179 
180 	pr_err("H_GET_NX_FAULT error: %ld, winid %u, buffer 0x%llx\n",
181 		rc, winid, buffer);
182 	return -EIO;
183 
184 }
185 
186 /*
187  * Handle the fault interrupt.
188  * When the fault interrupt is received for each window, query the
189  * hypervisor to get the fault CRB on the specific fault. Then
190  * process the CRB by updating CSB or send signal if the user space
191  * CSB is invalid.
192  * Note: The hypervisor forwards an interrupt for each fault request.
193  *	So one fault CRB to process for each H_GET_NX_FAULT hcall.
194  */
195 static irqreturn_t pseries_vas_fault_thread_fn(int irq, void *data)
196 {
197 	struct pseries_vas_window *txwin = data;
198 	struct coprocessor_request_block crb;
199 	struct vas_user_win_ref *tsk_ref;
200 	int rc;
201 
202 	rc = h_get_nx_fault(txwin->vas_win.winid, (u64)virt_to_phys(&crb));
203 	if (!rc) {
204 		tsk_ref = &txwin->vas_win.task_ref;
205 		vas_dump_crb(&crb);
206 		vas_update_csb(&crb, tsk_ref);
207 	}
208 
209 	return IRQ_HANDLED;
210 }
211 
212 /*
213  * Allocate window and setup IRQ mapping.
214  */
215 static int allocate_setup_window(struct pseries_vas_window *txwin,
216 				 u64 *domain, u8 wintype)
217 {
218 	int rc;
219 
220 	rc = h_allocate_vas_window(txwin, domain, wintype, DEF_WIN_CREDS);
221 	if (rc)
222 		return rc;
223 	/*
224 	 * On PowerVM, the hypervisor setup and forwards the fault
225 	 * interrupt per window. So the IRQ setup and fault handling
226 	 * will be done for each open window separately.
227 	 */
228 	txwin->fault_virq = irq_create_mapping(NULL, txwin->fault_irq);
229 	if (!txwin->fault_virq) {
230 		pr_err("Failed irq mapping %d\n", txwin->fault_irq);
231 		rc = -EINVAL;
232 		goto out_win;
233 	}
234 
235 	txwin->name = kasprintf(GFP_KERNEL, "vas-win-%d",
236 				txwin->vas_win.winid);
237 	if (!txwin->name) {
238 		rc = -ENOMEM;
239 		goto out_irq;
240 	}
241 
242 	rc = request_threaded_irq(txwin->fault_virq, NULL,
243 				  pseries_vas_fault_thread_fn, IRQF_ONESHOT,
244 				  txwin->name, txwin);
245 	if (rc) {
246 		pr_err("VAS-Window[%d]: Request IRQ(%u) failed with %d\n",
247 		       txwin->vas_win.winid, txwin->fault_virq, rc);
248 		goto out_free;
249 	}
250 
251 	txwin->vas_win.wcreds_max = DEF_WIN_CREDS;
252 
253 	return 0;
254 out_free:
255 	kfree(txwin->name);
256 out_irq:
257 	irq_dispose_mapping(txwin->fault_virq);
258 out_win:
259 	h_deallocate_vas_window(txwin->vas_win.winid);
260 	return rc;
261 }
262 
263 static inline void free_irq_setup(struct pseries_vas_window *txwin)
264 {
265 	free_irq(txwin->fault_virq, txwin);
266 	kfree(txwin->name);
267 	irq_dispose_mapping(txwin->fault_virq);
268 }
269 
270 static struct vas_window *vas_allocate_window(int vas_id, u64 flags,
271 					      enum vas_cop_type cop_type)
272 {
273 	long domain[PLPAR_HCALL9_BUFSIZE] = {VAS_DEFAULT_DOMAIN_ID};
274 	struct vas_cop_feat_caps *cop_feat_caps;
275 	struct vas_caps *caps;
276 	struct pseries_vas_window *txwin;
277 	int rc;
278 
279 	txwin = kzalloc(sizeof(*txwin), GFP_KERNEL);
280 	if (!txwin)
281 		return ERR_PTR(-ENOMEM);
282 
283 	/*
284 	 * A VAS window can have many credits which means that many
285 	 * requests can be issued simultaneously. But the hypervisor
286 	 * restricts one credit per window.
287 	 * The hypervisor introduces 2 different types of credits:
288 	 * Default credit type (Uses normal priority FIFO):
289 	 *	A limited number of credits are assigned to partitions
290 	 *	based on processor entitlement. But these credits may be
291 	 *	over-committed on a system depends on whether the CPUs
292 	 *	are in shared or dedicated modes - that is, more requests
293 	 *	may be issued across the system than NX can service at
294 	 *	once which can result in paste command failure (RMA_busy).
295 	 *	Then the process has to resend requests or fall-back to
296 	 *	SW compression.
297 	 * Quality of Service (QoS) credit type (Uses high priority FIFO):
298 	 *	To avoid NX HW contention, the system admins can assign
299 	 *	QoS credits for each LPAR so that this partition is
300 	 *	guaranteed access to NX resources. These credits are
301 	 *	assigned to partitions via the HMC.
302 	 *	Refer PAPR for more information.
303 	 *
304 	 * Allocate window with QoS credits if user requested. Otherwise
305 	 * default credits are used.
306 	 */
307 	if (flags & VAS_TX_WIN_FLAG_QOS_CREDIT)
308 		caps = &vascaps[VAS_GZIP_QOS_FEAT_TYPE];
309 	else
310 		caps = &vascaps[VAS_GZIP_DEF_FEAT_TYPE];
311 
312 	cop_feat_caps = &caps->caps;
313 
314 	if (atomic_inc_return(&cop_feat_caps->nr_used_credits) >
315 			atomic_read(&cop_feat_caps->nr_total_credits)) {
316 		pr_err("Credits are not available to allocate window\n");
317 		rc = -EINVAL;
318 		goto out;
319 	}
320 
321 	if (vas_id == -1) {
322 		/*
323 		 * The user space is requesting to allocate a window on
324 		 * a VAS instance where the process is executing.
325 		 * On PowerVM, domain values are passed to the hypervisor
326 		 * to select VAS instance. Useful if the process is
327 		 * affinity to NUMA node.
328 		 * The hypervisor selects VAS instance if
329 		 * VAS_DEFAULT_DOMAIN_ID (-1) is passed for domain values.
330 		 * The h_allocate_vas_window hcall is defined to take a
331 		 * domain values as specified by h_home_node_associativity,
332 		 * So no unpacking needs to be done.
333 		 */
334 		rc = plpar_hcall9(H_HOME_NODE_ASSOCIATIVITY, domain,
335 				  VPHN_FLAG_VCPU, smp_processor_id());
336 		if (rc != H_SUCCESS) {
337 			pr_err("H_HOME_NODE_ASSOCIATIVITY error: %d\n", rc);
338 			goto out;
339 		}
340 	}
341 
342 	txwin->pid = mfspr(SPRN_PID);
343 
344 	/*
345 	 * Allocate / Deallocate window hcalls and setup / free IRQs
346 	 * have to be protected with mutex.
347 	 * Open VAS window: Allocate window hcall and setup IRQ
348 	 * Close VAS window: Deallocate window hcall and free IRQ
349 	 *	The hypervisor waits until all NX requests are
350 	 *	completed before closing the window. So expects OS
351 	 *	to handle NX faults, means IRQ can be freed only
352 	 *	after the deallocate window hcall is returned.
353 	 * So once the window is closed with deallocate hcall before
354 	 * the IRQ is freed, it can be assigned to new allocate
355 	 * hcall with the same fault IRQ by the hypervisor. It can
356 	 * result in setup IRQ fail for the new window since the
357 	 * same fault IRQ is not freed by the OS before.
358 	 */
359 	mutex_lock(&vas_pseries_mutex);
360 	if (migration_in_progress)
361 		rc = -EBUSY;
362 	else
363 		rc = allocate_setup_window(txwin, (u64 *)&domain[0],
364 				   cop_feat_caps->win_type);
365 	mutex_unlock(&vas_pseries_mutex);
366 	if (rc)
367 		goto out;
368 
369 	/*
370 	 * Modify window and it is ready to use.
371 	 */
372 	rc = h_modify_vas_window(txwin);
373 	if (!rc)
374 		rc = get_vas_user_win_ref(&txwin->vas_win.task_ref);
375 	if (rc)
376 		goto out_free;
377 
378 	txwin->win_type = cop_feat_caps->win_type;
379 	mutex_lock(&vas_pseries_mutex);
380 	/*
381 	 * Possible to lose the acquired credit with DLPAR core
382 	 * removal after the window is opened. So if there are any
383 	 * closed windows (means with lost credits), do not give new
384 	 * window to user space. New windows will be opened only
385 	 * after the existing windows are reopened when credits are
386 	 * available.
387 	 */
388 	if (!caps->nr_close_wins) {
389 		list_add(&txwin->win_list, &caps->list);
390 		caps->nr_open_windows++;
391 		mutex_unlock(&vas_pseries_mutex);
392 		vas_user_win_add_mm_context(&txwin->vas_win.task_ref);
393 		return &txwin->vas_win;
394 	}
395 	mutex_unlock(&vas_pseries_mutex);
396 
397 	put_vas_user_win_ref(&txwin->vas_win.task_ref);
398 	rc = -EBUSY;
399 	pr_err("No credit is available to allocate window\n");
400 
401 out_free:
402 	/*
403 	 * Window is not operational. Free IRQ before closing
404 	 * window so that do not have to hold mutex.
405 	 */
406 	free_irq_setup(txwin);
407 	h_deallocate_vas_window(txwin->vas_win.winid);
408 out:
409 	atomic_dec(&cop_feat_caps->nr_used_credits);
410 	kfree(txwin);
411 	return ERR_PTR(rc);
412 }
413 
414 static u64 vas_paste_address(struct vas_window *vwin)
415 {
416 	struct pseries_vas_window *win;
417 
418 	win = container_of(vwin, struct pseries_vas_window, vas_win);
419 	return win->win_addr;
420 }
421 
422 static int deallocate_free_window(struct pseries_vas_window *win)
423 {
424 	int rc = 0;
425 
426 	/*
427 	 * The hypervisor waits for all requests including faults
428 	 * are processed before closing the window - Means all
429 	 * credits have to be returned. In the case of fault
430 	 * request, a credit is returned after OS issues
431 	 * H_GET_NX_FAULT hcall.
432 	 * So free IRQ after executing H_DEALLOCATE_VAS_WINDOW
433 	 * hcall.
434 	 */
435 	rc = h_deallocate_vas_window(win->vas_win.winid);
436 	if (!rc)
437 		free_irq_setup(win);
438 
439 	return rc;
440 }
441 
442 static int vas_deallocate_window(struct vas_window *vwin)
443 {
444 	struct pseries_vas_window *win;
445 	struct vas_cop_feat_caps *caps;
446 	int rc = 0;
447 
448 	if (!vwin)
449 		return -EINVAL;
450 
451 	win = container_of(vwin, struct pseries_vas_window, vas_win);
452 
453 	/* Should not happen */
454 	if (win->win_type >= VAS_MAX_FEAT_TYPE) {
455 		pr_err("Window (%u): Invalid window type %u\n",
456 				vwin->winid, win->win_type);
457 		return -EINVAL;
458 	}
459 
460 	caps = &vascaps[win->win_type].caps;
461 	mutex_lock(&vas_pseries_mutex);
462 	/*
463 	 * VAS window is already closed in the hypervisor when
464 	 * lost the credit or with migration. So just remove the entry
465 	 * from the list, remove task references and free vas_window
466 	 * struct.
467 	 */
468 	if (!(win->vas_win.status & VAS_WIN_NO_CRED_CLOSE) &&
469 		!(win->vas_win.status & VAS_WIN_MIGRATE_CLOSE)) {
470 		rc = deallocate_free_window(win);
471 		if (rc) {
472 			mutex_unlock(&vas_pseries_mutex);
473 			return rc;
474 		}
475 	} else
476 		vascaps[win->win_type].nr_close_wins--;
477 
478 	list_del(&win->win_list);
479 	atomic_dec(&caps->nr_used_credits);
480 	vascaps[win->win_type].nr_open_windows--;
481 	mutex_unlock(&vas_pseries_mutex);
482 
483 	put_vas_user_win_ref(&vwin->task_ref);
484 	mm_context_remove_vas_window(vwin->task_ref.mm);
485 
486 	kfree(win);
487 	return 0;
488 }
489 
490 static const struct vas_user_win_ops vops_pseries = {
491 	.open_win	= vas_allocate_window,	/* Open and configure window */
492 	.paste_addr	= vas_paste_address,	/* To do copy/paste */
493 	.close_win	= vas_deallocate_window, /* Close window */
494 };
495 
496 /*
497  * Supporting only nx-gzip coprocessor type now, but this API code
498  * extended to other coprocessor types later.
499  */
500 int vas_register_api_pseries(struct module *mod, enum vas_cop_type cop_type,
501 			     const char *name)
502 {
503 	int rc;
504 
505 	if (!copypaste_feat)
506 		return -ENOTSUPP;
507 
508 	rc = vas_register_coproc_api(mod, cop_type, name, &vops_pseries);
509 
510 	return rc;
511 }
512 EXPORT_SYMBOL_GPL(vas_register_api_pseries);
513 
514 void vas_unregister_api_pseries(void)
515 {
516 	vas_unregister_coproc_api();
517 }
518 EXPORT_SYMBOL_GPL(vas_unregister_api_pseries);
519 
520 /*
521  * Get the specific capabilities based on the feature type.
522  * Right now supports GZIP default and GZIP QoS capabilities.
523  */
524 static int __init get_vas_capabilities(u8 feat, enum vas_cop_feat_type type,
525 				struct hv_vas_cop_feat_caps *hv_caps)
526 {
527 	struct vas_cop_feat_caps *caps;
528 	struct vas_caps *vcaps;
529 	int rc = 0;
530 
531 	vcaps = &vascaps[type];
532 	memset(vcaps, 0, sizeof(*vcaps));
533 	INIT_LIST_HEAD(&vcaps->list);
534 
535 	vcaps->feat = feat;
536 	caps = &vcaps->caps;
537 
538 	rc = h_query_vas_capabilities(H_QUERY_VAS_CAPABILITIES, feat,
539 					  (u64)virt_to_phys(hv_caps));
540 	if (rc)
541 		return rc;
542 
543 	caps->user_mode = hv_caps->user_mode;
544 	if (!(caps->user_mode & VAS_COPY_PASTE_USER_MODE)) {
545 		pr_err("User space COPY/PASTE is not supported\n");
546 		return -ENOTSUPP;
547 	}
548 
549 	caps->descriptor = be64_to_cpu(hv_caps->descriptor);
550 	caps->win_type = hv_caps->win_type;
551 	if (caps->win_type >= VAS_MAX_FEAT_TYPE) {
552 		pr_err("Unsupported window type %u\n", caps->win_type);
553 		return -EINVAL;
554 	}
555 	caps->max_lpar_creds = be16_to_cpu(hv_caps->max_lpar_creds);
556 	caps->max_win_creds = be16_to_cpu(hv_caps->max_win_creds);
557 	atomic_set(&caps->nr_total_credits,
558 		   be16_to_cpu(hv_caps->target_lpar_creds));
559 	if (feat == VAS_GZIP_DEF_FEAT) {
560 		caps->def_lpar_creds = be16_to_cpu(hv_caps->def_lpar_creds);
561 
562 		if (caps->max_win_creds < DEF_WIN_CREDS) {
563 			pr_err("Window creds(%u) > max allowed window creds(%u)\n",
564 			       DEF_WIN_CREDS, caps->max_win_creds);
565 			return -EINVAL;
566 		}
567 	}
568 
569 	rc = sysfs_add_vas_caps(caps);
570 	if (rc)
571 		return rc;
572 
573 	copypaste_feat = true;
574 
575 	return 0;
576 }
577 
578 /*
579  * VAS windows can be closed due to lost credits when the core is
580  * removed. So reopen them if credits are available due to DLPAR
581  * core add and set the window active status. When NX sees the page
582  * fault on the unmapped paste address, the kernel handles the fault
583  * by setting the remapping to new paste address if the window is
584  * active.
585  */
586 static int reconfig_open_windows(struct vas_caps *vcaps, int creds,
587 				 bool migrate)
588 {
589 	long domain[PLPAR_HCALL9_BUFSIZE] = {VAS_DEFAULT_DOMAIN_ID};
590 	struct vas_cop_feat_caps *caps = &vcaps->caps;
591 	struct pseries_vas_window *win = NULL, *tmp;
592 	int rc, mv_ents = 0;
593 	int flag;
594 
595 	/*
596 	 * Nothing to do if there are no closed windows.
597 	 */
598 	if (!vcaps->nr_close_wins)
599 		return 0;
600 
601 	/*
602 	 * For the core removal, the hypervisor reduces the credits
603 	 * assigned to the LPAR and the kernel closes VAS windows
604 	 * in the hypervisor depends on reduced credits. The kernel
605 	 * uses LIFO (the last windows that are opened will be closed
606 	 * first) and expects to open in the same order when credits
607 	 * are available.
608 	 * For example, 40 windows are closed when the LPAR lost 2 cores
609 	 * (dedicated). If 1 core is added, this LPAR can have 20 more
610 	 * credits. It means the kernel can reopen 20 windows. So move
611 	 * 20 entries in the VAS windows lost and reopen next 20 windows.
612 	 * For partition migration, reopen all windows that are closed
613 	 * during resume.
614 	 */
615 	if ((vcaps->nr_close_wins > creds) && !migrate)
616 		mv_ents = vcaps->nr_close_wins - creds;
617 
618 	list_for_each_entry_safe(win, tmp, &vcaps->list, win_list) {
619 		if (!mv_ents)
620 			break;
621 
622 		mv_ents--;
623 	}
624 
625 	/*
626 	 * Open windows if they are closed only with migration or
627 	 * DLPAR (lost credit) before.
628 	 */
629 	if (migrate)
630 		flag = VAS_WIN_MIGRATE_CLOSE;
631 	else
632 		flag = VAS_WIN_NO_CRED_CLOSE;
633 
634 	list_for_each_entry_safe_from(win, tmp, &vcaps->list, win_list) {
635 		/*
636 		 * This window is closed with DLPAR and migration events.
637 		 * So reopen the window with the last event.
638 		 * The user space is not suspended with the current
639 		 * migration notifier. So the user space can issue DLPAR
640 		 * CPU hotplug while migration in progress. In this case
641 		 * this window will be opened with the last event.
642 		 */
643 		if ((win->vas_win.status & VAS_WIN_NO_CRED_CLOSE) &&
644 			(win->vas_win.status & VAS_WIN_MIGRATE_CLOSE)) {
645 			win->vas_win.status &= ~flag;
646 			continue;
647 		}
648 
649 		/*
650 		 * Nothing to do on this window if it is not closed
651 		 * with this flag
652 		 */
653 		if (!(win->vas_win.status & flag))
654 			continue;
655 
656 		rc = allocate_setup_window(win, (u64 *)&domain[0],
657 					   caps->win_type);
658 		if (rc)
659 			return rc;
660 
661 		rc = h_modify_vas_window(win);
662 		if (rc)
663 			goto out;
664 
665 		mutex_lock(&win->vas_win.task_ref.mmap_mutex);
666 		/*
667 		 * Set window status to active
668 		 */
669 		win->vas_win.status &= ~flag;
670 		mutex_unlock(&win->vas_win.task_ref.mmap_mutex);
671 		win->win_type = caps->win_type;
672 		if (!--vcaps->nr_close_wins)
673 			break;
674 	}
675 
676 	return 0;
677 out:
678 	/*
679 	 * Window modify HCALL failed. So close the window to the
680 	 * hypervisor and return.
681 	 */
682 	free_irq_setup(win);
683 	h_deallocate_vas_window(win->vas_win.winid);
684 	return rc;
685 }
686 
687 /*
688  * The hypervisor reduces the available credits if the LPAR lost core. It
689  * means the excessive windows should not be active and the user space
690  * should not be using these windows to send compression requests to NX.
691  * So the kernel closes the excessive windows and unmap the paste address
692  * such that the user space receives paste instruction failure. Then up to
693  * the user space to fall back to SW compression and manage with the
694  * existing windows.
695  */
696 static int reconfig_close_windows(struct vas_caps *vcap, int excess_creds,
697 									bool migrate)
698 {
699 	struct pseries_vas_window *win, *tmp;
700 	struct vas_user_win_ref *task_ref;
701 	struct vm_area_struct *vma;
702 	int rc = 0, flag;
703 
704 	if (migrate)
705 		flag = VAS_WIN_MIGRATE_CLOSE;
706 	else
707 		flag = VAS_WIN_NO_CRED_CLOSE;
708 
709 	list_for_each_entry_safe(win, tmp, &vcap->list, win_list) {
710 		/*
711 		 * This window is already closed due to lost credit
712 		 * or for migration before. Go for next window.
713 		 * For migration, nothing to do since this window
714 		 * closed for DLPAR and will be reopened even on
715 		 * the destination system with other DLPAR operation.
716 		 */
717 		if ((win->vas_win.status & VAS_WIN_MIGRATE_CLOSE) ||
718 			(win->vas_win.status & VAS_WIN_NO_CRED_CLOSE)) {
719 			win->vas_win.status |= flag;
720 			continue;
721 		}
722 
723 		task_ref = &win->vas_win.task_ref;
724 		mutex_lock(&task_ref->mmap_mutex);
725 		vma = task_ref->vma;
726 		/*
727 		 * Number of available credits are reduced, So select
728 		 * and close windows.
729 		 */
730 		win->vas_win.status |= flag;
731 
732 		mmap_write_lock(task_ref->mm);
733 		/*
734 		 * vma is set in the original mapping. But this mapping
735 		 * is done with mmap() after the window is opened with ioctl.
736 		 * so we may not see the original mapping if the core remove
737 		 * is done before the original mmap() and after the ioctl.
738 		 */
739 		if (vma)
740 			zap_page_range(vma, vma->vm_start,
741 					vma->vm_end - vma->vm_start);
742 
743 		mmap_write_unlock(task_ref->mm);
744 		mutex_unlock(&task_ref->mmap_mutex);
745 		/*
746 		 * Close VAS window in the hypervisor, but do not
747 		 * free vas_window struct since it may be reused
748 		 * when the credit is available later (DLPAR with
749 		 * adding cores). This struct will be used
750 		 * later when the process issued with close(FD).
751 		 */
752 		rc = deallocate_free_window(win);
753 		/*
754 		 * This failure is from the hypervisor.
755 		 * No way to stop migration for these failures.
756 		 * So ignore error and continue closing other windows.
757 		 */
758 		if (rc && !migrate)
759 			return rc;
760 
761 		vcap->nr_close_wins++;
762 
763 		/*
764 		 * For migration, do not depend on lpar_creds in case if
765 		 * mismatch with the hypervisor value (should not happen).
766 		 * So close all active windows in the list and will be
767 		 * reopened windows based on the new lpar_creds on the
768 		 * destination system during resume.
769 		 */
770 		if (!migrate && !--excess_creds)
771 			break;
772 	}
773 
774 	return 0;
775 }
776 
777 /*
778  * Get new VAS capabilities when the core add/removal configuration
779  * changes. Reconfig window configurations based on the credits
780  * availability from this new capabilities.
781  */
782 int vas_reconfig_capabilties(u8 type, int new_nr_creds)
783 {
784 	struct vas_cop_feat_caps *caps;
785 	int old_nr_creds;
786 	struct vas_caps *vcaps;
787 	int rc = 0, nr_active_wins;
788 
789 	if (type >= VAS_MAX_FEAT_TYPE) {
790 		pr_err("Invalid credit type %d\n", type);
791 		return -EINVAL;
792 	}
793 
794 	vcaps = &vascaps[type];
795 	caps = &vcaps->caps;
796 
797 	mutex_lock(&vas_pseries_mutex);
798 
799 	old_nr_creds = atomic_read(&caps->nr_total_credits);
800 
801 	atomic_set(&caps->nr_total_credits, new_nr_creds);
802 	/*
803 	 * The total number of available credits may be decreased or
804 	 * inceased with DLPAR operation. Means some windows have to be
805 	 * closed / reopened. Hold the vas_pseries_mutex so that the
806 	 * the user space can not open new windows.
807 	 */
808 	if (old_nr_creds <  new_nr_creds) {
809 		/*
810 		 * If the existing target credits is less than the new
811 		 * target, reopen windows if they are closed due to
812 		 * the previous DLPAR (core removal).
813 		 */
814 		rc = reconfig_open_windows(vcaps, new_nr_creds - old_nr_creds,
815 					   false);
816 	} else {
817 		/*
818 		 * # active windows is more than new LPAR available
819 		 * credits. So close the excessive windows.
820 		 * On pseries, each window will have 1 credit.
821 		 */
822 		nr_active_wins = vcaps->nr_open_windows - vcaps->nr_close_wins;
823 		if (nr_active_wins > new_nr_creds)
824 			rc = reconfig_close_windows(vcaps,
825 					nr_active_wins - new_nr_creds,
826 					false);
827 	}
828 
829 	mutex_unlock(&vas_pseries_mutex);
830 	return rc;
831 }
832 /*
833  * Total number of default credits available (target_credits)
834  * in LPAR depends on number of cores configured. It varies based on
835  * whether processors are in shared mode or dedicated mode.
836  * Get the notifier when CPU configuration is changed with DLPAR
837  * operation so that get the new target_credits (vas default capabilities)
838  * and then update the existing windows usage if needed.
839  */
840 static int pseries_vas_notifier(struct notifier_block *nb,
841 				unsigned long action, void *data)
842 {
843 	struct of_reconfig_data *rd = data;
844 	struct device_node *dn = rd->dn;
845 	const __be32 *intserv = NULL;
846 	int new_nr_creds, len, rc = 0;
847 
848 	if ((action == OF_RECONFIG_ATTACH_NODE) ||
849 		(action == OF_RECONFIG_DETACH_NODE))
850 		intserv = of_get_property(dn, "ibm,ppc-interrupt-server#s",
851 					  &len);
852 	/*
853 	 * Processor config is not changed
854 	 */
855 	if (!intserv)
856 		return NOTIFY_OK;
857 
858 	rc = h_query_vas_capabilities(H_QUERY_VAS_CAPABILITIES,
859 					vascaps[VAS_GZIP_DEF_FEAT_TYPE].feat,
860 					(u64)virt_to_phys(&hv_cop_caps));
861 	if (!rc) {
862 		new_nr_creds = be16_to_cpu(hv_cop_caps.target_lpar_creds);
863 		rc = vas_reconfig_capabilties(VAS_GZIP_DEF_FEAT_TYPE,
864 						new_nr_creds);
865 	}
866 
867 	if (rc)
868 		pr_err("Failed reconfig VAS capabilities with DLPAR\n");
869 
870 	return rc;
871 }
872 
873 static struct notifier_block pseries_vas_nb = {
874 	.notifier_call = pseries_vas_notifier,
875 };
876 
877 /*
878  * For LPM, all windows have to be closed on the source partition
879  * before migration and reopen them on the destination partition
880  * after migration. So closing windows during suspend and
881  * reopen them during resume.
882  */
883 int vas_migration_handler(int action)
884 {
885 	struct vas_cop_feat_caps *caps;
886 	int old_nr_creds, new_nr_creds = 0;
887 	struct vas_caps *vcaps;
888 	int i, rc = 0;
889 
890 	/*
891 	 * NX-GZIP is not enabled. Nothing to do for migration.
892 	 */
893 	if (!copypaste_feat)
894 		return rc;
895 
896 	mutex_lock(&vas_pseries_mutex);
897 
898 	if (action == VAS_SUSPEND)
899 		migration_in_progress = true;
900 	else
901 		migration_in_progress = false;
902 
903 	for (i = 0; i < VAS_MAX_FEAT_TYPE; i++) {
904 		vcaps = &vascaps[i];
905 		caps = &vcaps->caps;
906 		old_nr_creds = atomic_read(&caps->nr_total_credits);
907 
908 		rc = h_query_vas_capabilities(H_QUERY_VAS_CAPABILITIES,
909 					      vcaps->feat,
910 					      (u64)virt_to_phys(&hv_cop_caps));
911 		if (!rc) {
912 			new_nr_creds = be16_to_cpu(hv_cop_caps.target_lpar_creds);
913 			/*
914 			 * Should not happen. But incase print messages, close
915 			 * all windows in the list during suspend and reopen
916 			 * windows based on new lpar_creds on the destination
917 			 * system.
918 			 */
919 			if (old_nr_creds != new_nr_creds) {
920 				pr_err("Target credits mismatch with the hypervisor\n");
921 				pr_err("state(%d): lpar creds: %d HV lpar creds: %d\n",
922 					action, old_nr_creds, new_nr_creds);
923 				pr_err("Used creds: %d, Active creds: %d\n",
924 					atomic_read(&caps->nr_used_credits),
925 					vcaps->nr_open_windows - vcaps->nr_close_wins);
926 			}
927 		} else {
928 			pr_err("state(%d): Get VAS capabilities failed with %d\n",
929 				action, rc);
930 			/*
931 			 * We can not stop migration with the current lpm
932 			 * implementation. So continue closing all windows in
933 			 * the list (during suspend) and return without
934 			 * opening windows (during resume) if VAS capabilities
935 			 * HCALL failed.
936 			 */
937 			if (action == VAS_RESUME)
938 				goto out;
939 		}
940 
941 		switch (action) {
942 		case VAS_SUSPEND:
943 			rc = reconfig_close_windows(vcaps, vcaps->nr_open_windows,
944 							true);
945 			break;
946 		case VAS_RESUME:
947 			atomic_set(&caps->nr_total_credits, new_nr_creds);
948 			rc = reconfig_open_windows(vcaps, new_nr_creds, true);
949 			break;
950 		default:
951 			/* should not happen */
952 			pr_err("Invalid migration action %d\n", action);
953 			rc = -EINVAL;
954 			goto out;
955 		}
956 
957 		/*
958 		 * Ignore errors during suspend and return for resume.
959 		 */
960 		if (rc && (action == VAS_RESUME))
961 			goto out;
962 	}
963 
964 out:
965 	mutex_unlock(&vas_pseries_mutex);
966 	return rc;
967 }
968 
969 static int __init pseries_vas_init(void)
970 {
971 	struct hv_vas_all_caps *hv_caps;
972 	int rc = 0;
973 
974 	/*
975 	 * Linux supports user space COPY/PASTE only with Radix
976 	 */
977 	if (!radix_enabled()) {
978 		pr_err("API is supported only with radix page tables\n");
979 		return -ENOTSUPP;
980 	}
981 
982 	hv_caps = kmalloc(sizeof(*hv_caps), GFP_KERNEL);
983 	if (!hv_caps)
984 		return -ENOMEM;
985 	/*
986 	 * Get VAS overall capabilities by passing 0 to feature type.
987 	 */
988 	rc = h_query_vas_capabilities(H_QUERY_VAS_CAPABILITIES, 0,
989 					  (u64)virt_to_phys(hv_caps));
990 	if (rc)
991 		goto out;
992 
993 	caps_all.descriptor = be64_to_cpu(hv_caps->descriptor);
994 	caps_all.feat_type = be64_to_cpu(hv_caps->feat_type);
995 
996 	sysfs_pseries_vas_init(&caps_all);
997 
998 	/*
999 	 * QOS capabilities available
1000 	 */
1001 	if (caps_all.feat_type & VAS_GZIP_QOS_FEAT_BIT) {
1002 		rc = get_vas_capabilities(VAS_GZIP_QOS_FEAT,
1003 					  VAS_GZIP_QOS_FEAT_TYPE, &hv_cop_caps);
1004 
1005 		if (rc)
1006 			goto out;
1007 	}
1008 	/*
1009 	 * Default capabilities available
1010 	 */
1011 	if (caps_all.feat_type & VAS_GZIP_DEF_FEAT_BIT)
1012 		rc = get_vas_capabilities(VAS_GZIP_DEF_FEAT,
1013 					  VAS_GZIP_DEF_FEAT_TYPE, &hv_cop_caps);
1014 
1015 	if (!rc && copypaste_feat) {
1016 		if (firmware_has_feature(FW_FEATURE_LPAR))
1017 			of_reconfig_notifier_register(&pseries_vas_nb);
1018 
1019 		pr_info("GZIP feature is available\n");
1020 	} else {
1021 		/*
1022 		 * Should not happen, but only when get default
1023 		 * capabilities HCALL failed. So disable copy paste
1024 		 * feature.
1025 		 */
1026 		copypaste_feat = false;
1027 	}
1028 
1029 out:
1030 	kfree(hv_caps);
1031 	return rc;
1032 }
1033 machine_device_initcall(pseries, pseries_vas_init);
1034