xref: /openbmc/linux/drivers/xen/grant-table.c (revision bc000245)
1 /******************************************************************************
2  * grant_table.c
3  *
4  * Granting foreign access to our memory reservation.
5  *
6  * Copyright (c) 2005-2006, Christopher Clark
7  * Copyright (c) 2004-2005, K A Fraser
8  *
9  * This program is free software; you can redistribute it and/or
10  * modify it under the terms of the GNU General Public License version 2
11  * as published by the Free Software Foundation; or, when distributed
12  * separately from the Linux kernel or incorporated into other
13  * software packages, subject to the following license:
14  *
15  * Permission is hereby granted, free of charge, to any person obtaining a copy
16  * of this source file (the "Software"), to deal in the Software without
17  * restriction, including without limitation the rights to use, copy, modify,
18  * merge, publish, distribute, sublicense, and/or sell copies of the Software,
19  * and to permit persons to whom the Software is furnished to do so, subject to
20  * the following conditions:
21  *
22  * The above copyright notice and this permission notice shall be included in
23  * all copies or substantial portions of the Software.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
26  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
27  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
28  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
29  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
30  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
31  * IN THE SOFTWARE.
32  */
33 
34 #define pr_fmt(fmt) "xen:" KBUILD_MODNAME ": " fmt
35 
36 #include <linux/module.h>
37 #include <linux/sched.h>
38 #include <linux/mm.h>
39 #include <linux/slab.h>
40 #include <linux/vmalloc.h>
41 #include <linux/uaccess.h>
42 #include <linux/io.h>
43 #include <linux/delay.h>
44 #include <linux/hardirq.h>
45 
46 #include <xen/xen.h>
47 #include <xen/interface/xen.h>
48 #include <xen/page.h>
49 #include <xen/grant_table.h>
50 #include <xen/interface/memory.h>
51 #include <xen/hvc-console.h>
52 #include <xen/swiotlb-xen.h>
53 #include <asm/xen/hypercall.h>
54 #include <asm/xen/interface.h>
55 
56 #include <asm/pgtable.h>
57 #include <asm/sync_bitops.h>
58 
59 /* External tools reserve first few grant table entries. */
60 #define NR_RESERVED_ENTRIES 8
61 #define GNTTAB_LIST_END 0xffffffff
62 
63 static grant_ref_t **gnttab_list;
64 static unsigned int nr_grant_frames;
65 static unsigned int boot_max_nr_grant_frames;
66 static int gnttab_free_count;
67 static grant_ref_t gnttab_free_head;
68 static DEFINE_SPINLOCK(gnttab_list_lock);
69 unsigned long xen_hvm_resume_frames;
70 EXPORT_SYMBOL_GPL(xen_hvm_resume_frames);
71 
72 static union {
73 	struct grant_entry_v1 *v1;
74 	union grant_entry_v2 *v2;
75 	void *addr;
76 } gnttab_shared;
77 
78 /*This is a structure of function pointers for grant table*/
79 struct gnttab_ops {
80 	/*
81 	 * Mapping a list of frames for storing grant entries. Frames parameter
82 	 * is used to store grant table address when grant table being setup,
83 	 * nr_gframes is the number of frames to map grant table. Returning
84 	 * GNTST_okay means success and negative value means failure.
85 	 */
86 	int (*map_frames)(xen_pfn_t *frames, unsigned int nr_gframes);
87 	/*
88 	 * Release a list of frames which are mapped in map_frames for grant
89 	 * entry status.
90 	 */
91 	void (*unmap_frames)(void);
92 	/*
93 	 * Introducing a valid entry into the grant table, granting the frame of
94 	 * this grant entry to domain for accessing or transfering. Ref
95 	 * parameter is reference of this introduced grant entry, domid is id of
96 	 * granted domain, frame is the page frame to be granted, and flags is
97 	 * status of the grant entry to be updated.
98 	 */
99 	void (*update_entry)(grant_ref_t ref, domid_t domid,
100 			     unsigned long frame, unsigned flags);
101 	/*
102 	 * Stop granting a grant entry to domain for accessing. Ref parameter is
103 	 * reference of a grant entry whose grant access will be stopped,
104 	 * readonly is not in use in this function. If the grant entry is
105 	 * currently mapped for reading or writing, just return failure(==0)
106 	 * directly and don't tear down the grant access. Otherwise, stop grant
107 	 * access for this entry and return success(==1).
108 	 */
109 	int (*end_foreign_access_ref)(grant_ref_t ref, int readonly);
110 	/*
111 	 * Stop granting a grant entry to domain for transfer. Ref parameter is
112 	 * reference of a grant entry whose grant transfer will be stopped. If
113 	 * tranfer has not started, just reclaim the grant entry and return
114 	 * failure(==0). Otherwise, wait for the transfer to complete and then
115 	 * return the frame.
116 	 */
117 	unsigned long (*end_foreign_transfer_ref)(grant_ref_t ref);
118 	/*
119 	 * Query the status of a grant entry. Ref parameter is reference of
120 	 * queried grant entry, return value is the status of queried entry.
121 	 * Detailed status(writing/reading) can be gotten from the return value
122 	 * by bit operations.
123 	 */
124 	int (*query_foreign_access)(grant_ref_t ref);
125 	/*
126 	 * Grant a domain to access a range of bytes within the page referred by
127 	 * an available grant entry. Ref parameter is reference of a grant entry
128 	 * which will be sub-page accessed, domid is id of grantee domain, frame
129 	 * is frame address of subpage grant, flags is grant type and flag
130 	 * information, page_off is offset of the range of bytes, and length is
131 	 * length of bytes to be accessed.
132 	 */
133 	void (*update_subpage_entry)(grant_ref_t ref, domid_t domid,
134 				     unsigned long frame, int flags,
135 				     unsigned page_off, unsigned length);
136 	/*
137 	 * Redirect an available grant entry on domain A to another grant
138 	 * reference of domain B, then allow domain C to use grant reference
139 	 * of domain B transitively. Ref parameter is an available grant entry
140 	 * reference on domain A, domid is id of domain C which accesses grant
141 	 * entry transitively, flags is grant type and flag information,
142 	 * trans_domid is id of domain B whose grant entry is finally accessed
143 	 * transitively, trans_gref is grant entry transitive reference of
144 	 * domain B.
145 	 */
146 	void (*update_trans_entry)(grant_ref_t ref, domid_t domid, int flags,
147 				   domid_t trans_domid, grant_ref_t trans_gref);
148 };
149 
150 static struct gnttab_ops *gnttab_interface;
151 
152 /*This reflects status of grant entries, so act as a global value*/
153 static grant_status_t *grstatus;
154 
155 static int grant_table_version;
156 static int grefs_per_grant_frame;
157 
158 static struct gnttab_free_callback *gnttab_free_callback_list;
159 
160 static int gnttab_expand(unsigned int req_entries);
161 
162 #define RPP (PAGE_SIZE / sizeof(grant_ref_t))
163 #define SPP (PAGE_SIZE / sizeof(grant_status_t))
164 
165 static inline grant_ref_t *__gnttab_entry(grant_ref_t entry)
166 {
167 	return &gnttab_list[(entry) / RPP][(entry) % RPP];
168 }
169 /* This can be used as an l-value */
170 #define gnttab_entry(entry) (*__gnttab_entry(entry))
171 
172 static int get_free_entries(unsigned count)
173 {
174 	unsigned long flags;
175 	int ref, rc = 0;
176 	grant_ref_t head;
177 
178 	spin_lock_irqsave(&gnttab_list_lock, flags);
179 
180 	if ((gnttab_free_count < count) &&
181 	    ((rc = gnttab_expand(count - gnttab_free_count)) < 0)) {
182 		spin_unlock_irqrestore(&gnttab_list_lock, flags);
183 		return rc;
184 	}
185 
186 	ref = head = gnttab_free_head;
187 	gnttab_free_count -= count;
188 	while (count-- > 1)
189 		head = gnttab_entry(head);
190 	gnttab_free_head = gnttab_entry(head);
191 	gnttab_entry(head) = GNTTAB_LIST_END;
192 
193 	spin_unlock_irqrestore(&gnttab_list_lock, flags);
194 
195 	return ref;
196 }
197 
198 static void do_free_callbacks(void)
199 {
200 	struct gnttab_free_callback *callback, *next;
201 
202 	callback = gnttab_free_callback_list;
203 	gnttab_free_callback_list = NULL;
204 
205 	while (callback != NULL) {
206 		next = callback->next;
207 		if (gnttab_free_count >= callback->count) {
208 			callback->next = NULL;
209 			callback->fn(callback->arg);
210 		} else {
211 			callback->next = gnttab_free_callback_list;
212 			gnttab_free_callback_list = callback;
213 		}
214 		callback = next;
215 	}
216 }
217 
218 static inline void check_free_callbacks(void)
219 {
220 	if (unlikely(gnttab_free_callback_list))
221 		do_free_callbacks();
222 }
223 
224 static void put_free_entry(grant_ref_t ref)
225 {
226 	unsigned long flags;
227 	spin_lock_irqsave(&gnttab_list_lock, flags);
228 	gnttab_entry(ref) = gnttab_free_head;
229 	gnttab_free_head = ref;
230 	gnttab_free_count++;
231 	check_free_callbacks();
232 	spin_unlock_irqrestore(&gnttab_list_lock, flags);
233 }
234 
235 /*
236  * Following applies to gnttab_update_entry_v1 and gnttab_update_entry_v2.
237  * Introducing a valid entry into the grant table:
238  *  1. Write ent->domid.
239  *  2. Write ent->frame:
240  *      GTF_permit_access:   Frame to which access is permitted.
241  *      GTF_accept_transfer: Pseudo-phys frame slot being filled by new
242  *                           frame, or zero if none.
243  *  3. Write memory barrier (WMB).
244  *  4. Write ent->flags, inc. valid type.
245  */
246 static void gnttab_update_entry_v1(grant_ref_t ref, domid_t domid,
247 				   unsigned long frame, unsigned flags)
248 {
249 	gnttab_shared.v1[ref].domid = domid;
250 	gnttab_shared.v1[ref].frame = frame;
251 	wmb();
252 	gnttab_shared.v1[ref].flags = flags;
253 }
254 
255 static void gnttab_update_entry_v2(grant_ref_t ref, domid_t domid,
256 				   unsigned long frame, unsigned flags)
257 {
258 	gnttab_shared.v2[ref].hdr.domid = domid;
259 	gnttab_shared.v2[ref].full_page.frame = frame;
260 	wmb();
261 	gnttab_shared.v2[ref].hdr.flags = GTF_permit_access | flags;
262 }
263 
264 /*
265  * Public grant-issuing interface functions
266  */
267 void gnttab_grant_foreign_access_ref(grant_ref_t ref, domid_t domid,
268 				     unsigned long frame, int readonly)
269 {
270 	gnttab_interface->update_entry(ref, domid, frame,
271 			   GTF_permit_access | (readonly ? GTF_readonly : 0));
272 }
273 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_access_ref);
274 
275 int gnttab_grant_foreign_access(domid_t domid, unsigned long frame,
276 				int readonly)
277 {
278 	int ref;
279 
280 	ref = get_free_entries(1);
281 	if (unlikely(ref < 0))
282 		return -ENOSPC;
283 
284 	gnttab_grant_foreign_access_ref(ref, domid, frame, readonly);
285 
286 	return ref;
287 }
288 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_access);
289 
290 static void gnttab_update_subpage_entry_v2(grant_ref_t ref, domid_t domid,
291 					   unsigned long frame, int flags,
292 					   unsigned page_off, unsigned length)
293 {
294 	gnttab_shared.v2[ref].sub_page.frame = frame;
295 	gnttab_shared.v2[ref].sub_page.page_off = page_off;
296 	gnttab_shared.v2[ref].sub_page.length = length;
297 	gnttab_shared.v2[ref].hdr.domid = domid;
298 	wmb();
299 	gnttab_shared.v2[ref].hdr.flags =
300 				GTF_permit_access | GTF_sub_page | flags;
301 }
302 
303 int gnttab_grant_foreign_access_subpage_ref(grant_ref_t ref, domid_t domid,
304 					    unsigned long frame, int flags,
305 					    unsigned page_off,
306 					    unsigned length)
307 {
308 	if (flags & (GTF_accept_transfer | GTF_reading |
309 		     GTF_writing | GTF_transitive))
310 		return -EPERM;
311 
312 	if (gnttab_interface->update_subpage_entry == NULL)
313 		return -ENOSYS;
314 
315 	gnttab_interface->update_subpage_entry(ref, domid, frame, flags,
316 					       page_off, length);
317 
318 	return 0;
319 }
320 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_access_subpage_ref);
321 
322 int gnttab_grant_foreign_access_subpage(domid_t domid, unsigned long frame,
323 					int flags, unsigned page_off,
324 					unsigned length)
325 {
326 	int ref, rc;
327 
328 	ref = get_free_entries(1);
329 	if (unlikely(ref < 0))
330 		return -ENOSPC;
331 
332 	rc = gnttab_grant_foreign_access_subpage_ref(ref, domid, frame, flags,
333 						     page_off, length);
334 	if (rc < 0) {
335 		put_free_entry(ref);
336 		return rc;
337 	}
338 
339 	return ref;
340 }
341 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_access_subpage);
342 
343 bool gnttab_subpage_grants_available(void)
344 {
345 	return gnttab_interface->update_subpage_entry != NULL;
346 }
347 EXPORT_SYMBOL_GPL(gnttab_subpage_grants_available);
348 
349 static void gnttab_update_trans_entry_v2(grant_ref_t ref, domid_t domid,
350 					 int flags, domid_t trans_domid,
351 					 grant_ref_t trans_gref)
352 {
353 	gnttab_shared.v2[ref].transitive.trans_domid = trans_domid;
354 	gnttab_shared.v2[ref].transitive.gref = trans_gref;
355 	gnttab_shared.v2[ref].hdr.domid = domid;
356 	wmb();
357 	gnttab_shared.v2[ref].hdr.flags =
358 				GTF_permit_access | GTF_transitive | flags;
359 }
360 
361 int gnttab_grant_foreign_access_trans_ref(grant_ref_t ref, domid_t domid,
362 					  int flags, domid_t trans_domid,
363 					  grant_ref_t trans_gref)
364 {
365 	if (flags & (GTF_accept_transfer | GTF_reading |
366 		     GTF_writing | GTF_sub_page))
367 		return -EPERM;
368 
369 	if (gnttab_interface->update_trans_entry == NULL)
370 		return -ENOSYS;
371 
372 	gnttab_interface->update_trans_entry(ref, domid, flags, trans_domid,
373 					     trans_gref);
374 
375 	return 0;
376 }
377 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_access_trans_ref);
378 
379 int gnttab_grant_foreign_access_trans(domid_t domid, int flags,
380 				      domid_t trans_domid,
381 				      grant_ref_t trans_gref)
382 {
383 	int ref, rc;
384 
385 	ref = get_free_entries(1);
386 	if (unlikely(ref < 0))
387 		return -ENOSPC;
388 
389 	rc = gnttab_grant_foreign_access_trans_ref(ref, domid, flags,
390 						   trans_domid, trans_gref);
391 	if (rc < 0) {
392 		put_free_entry(ref);
393 		return rc;
394 	}
395 
396 	return ref;
397 }
398 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_access_trans);
399 
400 bool gnttab_trans_grants_available(void)
401 {
402 	return gnttab_interface->update_trans_entry != NULL;
403 }
404 EXPORT_SYMBOL_GPL(gnttab_trans_grants_available);
405 
406 static int gnttab_query_foreign_access_v1(grant_ref_t ref)
407 {
408 	return gnttab_shared.v1[ref].flags & (GTF_reading|GTF_writing);
409 }
410 
411 static int gnttab_query_foreign_access_v2(grant_ref_t ref)
412 {
413 	return grstatus[ref] & (GTF_reading|GTF_writing);
414 }
415 
416 int gnttab_query_foreign_access(grant_ref_t ref)
417 {
418 	return gnttab_interface->query_foreign_access(ref);
419 }
420 EXPORT_SYMBOL_GPL(gnttab_query_foreign_access);
421 
422 static int gnttab_end_foreign_access_ref_v1(grant_ref_t ref, int readonly)
423 {
424 	u16 flags, nflags;
425 	u16 *pflags;
426 
427 	pflags = &gnttab_shared.v1[ref].flags;
428 	nflags = *pflags;
429 	do {
430 		flags = nflags;
431 		if (flags & (GTF_reading|GTF_writing))
432 			return 0;
433 	} while ((nflags = sync_cmpxchg(pflags, flags, 0)) != flags);
434 
435 	return 1;
436 }
437 
438 static int gnttab_end_foreign_access_ref_v2(grant_ref_t ref, int readonly)
439 {
440 	gnttab_shared.v2[ref].hdr.flags = 0;
441 	mb();
442 	if (grstatus[ref] & (GTF_reading|GTF_writing)) {
443 		return 0;
444 	} else {
445 		/* The read of grstatus needs to have acquire
446 		semantics.  On x86, reads already have
447 		that, and we just need to protect against
448 		compiler reorderings.  On other
449 		architectures we may need a full
450 		barrier. */
451 #ifdef CONFIG_X86
452 		barrier();
453 #else
454 		mb();
455 #endif
456 	}
457 
458 	return 1;
459 }
460 
461 static inline int _gnttab_end_foreign_access_ref(grant_ref_t ref, int readonly)
462 {
463 	return gnttab_interface->end_foreign_access_ref(ref, readonly);
464 }
465 
466 int gnttab_end_foreign_access_ref(grant_ref_t ref, int readonly)
467 {
468 	if (_gnttab_end_foreign_access_ref(ref, readonly))
469 		return 1;
470 	pr_warn("WARNING: g.e. %#x still in use!\n", ref);
471 	return 0;
472 }
473 EXPORT_SYMBOL_GPL(gnttab_end_foreign_access_ref);
474 
475 struct deferred_entry {
476 	struct list_head list;
477 	grant_ref_t ref;
478 	bool ro;
479 	uint16_t warn_delay;
480 	struct page *page;
481 };
482 static LIST_HEAD(deferred_list);
483 static void gnttab_handle_deferred(unsigned long);
484 static DEFINE_TIMER(deferred_timer, gnttab_handle_deferred, 0, 0);
485 
486 static void gnttab_handle_deferred(unsigned long unused)
487 {
488 	unsigned int nr = 10;
489 	struct deferred_entry *first = NULL;
490 	unsigned long flags;
491 
492 	spin_lock_irqsave(&gnttab_list_lock, flags);
493 	while (nr--) {
494 		struct deferred_entry *entry
495 			= list_first_entry(&deferred_list,
496 					   struct deferred_entry, list);
497 
498 		if (entry == first)
499 			break;
500 		list_del(&entry->list);
501 		spin_unlock_irqrestore(&gnttab_list_lock, flags);
502 		if (_gnttab_end_foreign_access_ref(entry->ref, entry->ro)) {
503 			put_free_entry(entry->ref);
504 			if (entry->page) {
505 				pr_debug("freeing g.e. %#x (pfn %#lx)\n",
506 					 entry->ref, page_to_pfn(entry->page));
507 				__free_page(entry->page);
508 			} else
509 				pr_info("freeing g.e. %#x\n", entry->ref);
510 			kfree(entry);
511 			entry = NULL;
512 		} else {
513 			if (!--entry->warn_delay)
514 				pr_info("g.e. %#x still pending\n", entry->ref);
515 			if (!first)
516 				first = entry;
517 		}
518 		spin_lock_irqsave(&gnttab_list_lock, flags);
519 		if (entry)
520 			list_add_tail(&entry->list, &deferred_list);
521 		else if (list_empty(&deferred_list))
522 			break;
523 	}
524 	if (!list_empty(&deferred_list) && !timer_pending(&deferred_timer)) {
525 		deferred_timer.expires = jiffies + HZ;
526 		add_timer(&deferred_timer);
527 	}
528 	spin_unlock_irqrestore(&gnttab_list_lock, flags);
529 }
530 
531 static void gnttab_add_deferred(grant_ref_t ref, bool readonly,
532 				struct page *page)
533 {
534 	struct deferred_entry *entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
535 	const char *what = KERN_WARNING "leaking";
536 
537 	if (entry) {
538 		unsigned long flags;
539 
540 		entry->ref = ref;
541 		entry->ro = readonly;
542 		entry->page = page;
543 		entry->warn_delay = 60;
544 		spin_lock_irqsave(&gnttab_list_lock, flags);
545 		list_add_tail(&entry->list, &deferred_list);
546 		if (!timer_pending(&deferred_timer)) {
547 			deferred_timer.expires = jiffies + HZ;
548 			add_timer(&deferred_timer);
549 		}
550 		spin_unlock_irqrestore(&gnttab_list_lock, flags);
551 		what = KERN_DEBUG "deferring";
552 	}
553 	printk("%s g.e. %#x (pfn %#lx)\n",
554 	       what, ref, page ? page_to_pfn(page) : -1);
555 }
556 
557 void gnttab_end_foreign_access(grant_ref_t ref, int readonly,
558 			       unsigned long page)
559 {
560 	if (gnttab_end_foreign_access_ref(ref, readonly)) {
561 		put_free_entry(ref);
562 		if (page != 0)
563 			free_page(page);
564 	} else
565 		gnttab_add_deferred(ref, readonly,
566 				    page ? virt_to_page(page) : NULL);
567 }
568 EXPORT_SYMBOL_GPL(gnttab_end_foreign_access);
569 
570 int gnttab_grant_foreign_transfer(domid_t domid, unsigned long pfn)
571 {
572 	int ref;
573 
574 	ref = get_free_entries(1);
575 	if (unlikely(ref < 0))
576 		return -ENOSPC;
577 	gnttab_grant_foreign_transfer_ref(ref, domid, pfn);
578 
579 	return ref;
580 }
581 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_transfer);
582 
583 void gnttab_grant_foreign_transfer_ref(grant_ref_t ref, domid_t domid,
584 				       unsigned long pfn)
585 {
586 	gnttab_interface->update_entry(ref, domid, pfn, GTF_accept_transfer);
587 }
588 EXPORT_SYMBOL_GPL(gnttab_grant_foreign_transfer_ref);
589 
590 static unsigned long gnttab_end_foreign_transfer_ref_v1(grant_ref_t ref)
591 {
592 	unsigned long frame;
593 	u16           flags;
594 	u16          *pflags;
595 
596 	pflags = &gnttab_shared.v1[ref].flags;
597 
598 	/*
599 	 * If a transfer is not even yet started, try to reclaim the grant
600 	 * reference and return failure (== 0).
601 	 */
602 	while (!((flags = *pflags) & GTF_transfer_committed)) {
603 		if (sync_cmpxchg(pflags, flags, 0) == flags)
604 			return 0;
605 		cpu_relax();
606 	}
607 
608 	/* If a transfer is in progress then wait until it is completed. */
609 	while (!(flags & GTF_transfer_completed)) {
610 		flags = *pflags;
611 		cpu_relax();
612 	}
613 
614 	rmb();	/* Read the frame number /after/ reading completion status. */
615 	frame = gnttab_shared.v1[ref].frame;
616 	BUG_ON(frame == 0);
617 
618 	return frame;
619 }
620 
621 static unsigned long gnttab_end_foreign_transfer_ref_v2(grant_ref_t ref)
622 {
623 	unsigned long frame;
624 	u16           flags;
625 	u16          *pflags;
626 
627 	pflags = &gnttab_shared.v2[ref].hdr.flags;
628 
629 	/*
630 	 * If a transfer is not even yet started, try to reclaim the grant
631 	 * reference and return failure (== 0).
632 	 */
633 	while (!((flags = *pflags) & GTF_transfer_committed)) {
634 		if (sync_cmpxchg(pflags, flags, 0) == flags)
635 			return 0;
636 		cpu_relax();
637 	}
638 
639 	/* If a transfer is in progress then wait until it is completed. */
640 	while (!(flags & GTF_transfer_completed)) {
641 		flags = *pflags;
642 		cpu_relax();
643 	}
644 
645 	rmb();  /* Read the frame number /after/ reading completion status. */
646 	frame = gnttab_shared.v2[ref].full_page.frame;
647 	BUG_ON(frame == 0);
648 
649 	return frame;
650 }
651 
652 unsigned long gnttab_end_foreign_transfer_ref(grant_ref_t ref)
653 {
654 	return gnttab_interface->end_foreign_transfer_ref(ref);
655 }
656 EXPORT_SYMBOL_GPL(gnttab_end_foreign_transfer_ref);
657 
658 unsigned long gnttab_end_foreign_transfer(grant_ref_t ref)
659 {
660 	unsigned long frame = gnttab_end_foreign_transfer_ref(ref);
661 	put_free_entry(ref);
662 	return frame;
663 }
664 EXPORT_SYMBOL_GPL(gnttab_end_foreign_transfer);
665 
666 void gnttab_free_grant_reference(grant_ref_t ref)
667 {
668 	put_free_entry(ref);
669 }
670 EXPORT_SYMBOL_GPL(gnttab_free_grant_reference);
671 
672 void gnttab_free_grant_references(grant_ref_t head)
673 {
674 	grant_ref_t ref;
675 	unsigned long flags;
676 	int count = 1;
677 	if (head == GNTTAB_LIST_END)
678 		return;
679 	spin_lock_irqsave(&gnttab_list_lock, flags);
680 	ref = head;
681 	while (gnttab_entry(ref) != GNTTAB_LIST_END) {
682 		ref = gnttab_entry(ref);
683 		count++;
684 	}
685 	gnttab_entry(ref) = gnttab_free_head;
686 	gnttab_free_head = head;
687 	gnttab_free_count += count;
688 	check_free_callbacks();
689 	spin_unlock_irqrestore(&gnttab_list_lock, flags);
690 }
691 EXPORT_SYMBOL_GPL(gnttab_free_grant_references);
692 
693 int gnttab_alloc_grant_references(u16 count, grant_ref_t *head)
694 {
695 	int h = get_free_entries(count);
696 
697 	if (h < 0)
698 		return -ENOSPC;
699 
700 	*head = h;
701 
702 	return 0;
703 }
704 EXPORT_SYMBOL_GPL(gnttab_alloc_grant_references);
705 
706 int gnttab_empty_grant_references(const grant_ref_t *private_head)
707 {
708 	return (*private_head == GNTTAB_LIST_END);
709 }
710 EXPORT_SYMBOL_GPL(gnttab_empty_grant_references);
711 
712 int gnttab_claim_grant_reference(grant_ref_t *private_head)
713 {
714 	grant_ref_t g = *private_head;
715 	if (unlikely(g == GNTTAB_LIST_END))
716 		return -ENOSPC;
717 	*private_head = gnttab_entry(g);
718 	return g;
719 }
720 EXPORT_SYMBOL_GPL(gnttab_claim_grant_reference);
721 
722 void gnttab_release_grant_reference(grant_ref_t *private_head,
723 				    grant_ref_t release)
724 {
725 	gnttab_entry(release) = *private_head;
726 	*private_head = release;
727 }
728 EXPORT_SYMBOL_GPL(gnttab_release_grant_reference);
729 
730 void gnttab_request_free_callback(struct gnttab_free_callback *callback,
731 				  void (*fn)(void *), void *arg, u16 count)
732 {
733 	unsigned long flags;
734 	struct gnttab_free_callback *cb;
735 
736 	spin_lock_irqsave(&gnttab_list_lock, flags);
737 
738 	/* Check if the callback is already on the list */
739 	cb = gnttab_free_callback_list;
740 	while (cb) {
741 		if (cb == callback)
742 			goto out;
743 		cb = cb->next;
744 	}
745 
746 	callback->fn = fn;
747 	callback->arg = arg;
748 	callback->count = count;
749 	callback->next = gnttab_free_callback_list;
750 	gnttab_free_callback_list = callback;
751 	check_free_callbacks();
752 out:
753 	spin_unlock_irqrestore(&gnttab_list_lock, flags);
754 }
755 EXPORT_SYMBOL_GPL(gnttab_request_free_callback);
756 
757 void gnttab_cancel_free_callback(struct gnttab_free_callback *callback)
758 {
759 	struct gnttab_free_callback **pcb;
760 	unsigned long flags;
761 
762 	spin_lock_irqsave(&gnttab_list_lock, flags);
763 	for (pcb = &gnttab_free_callback_list; *pcb; pcb = &(*pcb)->next) {
764 		if (*pcb == callback) {
765 			*pcb = callback->next;
766 			break;
767 		}
768 	}
769 	spin_unlock_irqrestore(&gnttab_list_lock, flags);
770 }
771 EXPORT_SYMBOL_GPL(gnttab_cancel_free_callback);
772 
773 static int grow_gnttab_list(unsigned int more_frames)
774 {
775 	unsigned int new_nr_grant_frames, extra_entries, i;
776 	unsigned int nr_glist_frames, new_nr_glist_frames;
777 
778 	BUG_ON(grefs_per_grant_frame == 0);
779 
780 	new_nr_grant_frames = nr_grant_frames + more_frames;
781 	extra_entries       = more_frames * grefs_per_grant_frame;
782 
783 	nr_glist_frames = (nr_grant_frames * grefs_per_grant_frame + RPP - 1) / RPP;
784 	new_nr_glist_frames =
785 		(new_nr_grant_frames * grefs_per_grant_frame + RPP - 1) / RPP;
786 	for (i = nr_glist_frames; i < new_nr_glist_frames; i++) {
787 		gnttab_list[i] = (grant_ref_t *)__get_free_page(GFP_ATOMIC);
788 		if (!gnttab_list[i])
789 			goto grow_nomem;
790 	}
791 
792 
793 	for (i = grefs_per_grant_frame * nr_grant_frames;
794 	     i < grefs_per_grant_frame * new_nr_grant_frames - 1; i++)
795 		gnttab_entry(i) = i + 1;
796 
797 	gnttab_entry(i) = gnttab_free_head;
798 	gnttab_free_head = grefs_per_grant_frame * nr_grant_frames;
799 	gnttab_free_count += extra_entries;
800 
801 	nr_grant_frames = new_nr_grant_frames;
802 
803 	check_free_callbacks();
804 
805 	return 0;
806 
807 grow_nomem:
808 	for ( ; i >= nr_glist_frames; i--)
809 		free_page((unsigned long) gnttab_list[i]);
810 	return -ENOMEM;
811 }
812 
813 static unsigned int __max_nr_grant_frames(void)
814 {
815 	struct gnttab_query_size query;
816 	int rc;
817 
818 	query.dom = DOMID_SELF;
819 
820 	rc = HYPERVISOR_grant_table_op(GNTTABOP_query_size, &query, 1);
821 	if ((rc < 0) || (query.status != GNTST_okay))
822 		return 4; /* Legacy max supported number of frames */
823 
824 	return query.max_nr_frames;
825 }
826 
827 unsigned int gnttab_max_grant_frames(void)
828 {
829 	unsigned int xen_max = __max_nr_grant_frames();
830 
831 	if (xen_max > boot_max_nr_grant_frames)
832 		return boot_max_nr_grant_frames;
833 	return xen_max;
834 }
835 EXPORT_SYMBOL_GPL(gnttab_max_grant_frames);
836 
837 /* Handling of paged out grant targets (GNTST_eagain) */
838 #define MAX_DELAY 256
839 static inline void
840 gnttab_retry_eagain_gop(unsigned int cmd, void *gop, int16_t *status,
841 						const char *func)
842 {
843 	unsigned delay = 1;
844 
845 	do {
846 		BUG_ON(HYPERVISOR_grant_table_op(cmd, gop, 1));
847 		if (*status == GNTST_eagain)
848 			msleep(delay++);
849 	} while ((*status == GNTST_eagain) && (delay < MAX_DELAY));
850 
851 	if (delay >= MAX_DELAY) {
852 		pr_err("%s: %s eagain grant\n", func, current->comm);
853 		*status = GNTST_bad_page;
854 	}
855 }
856 
857 void gnttab_batch_map(struct gnttab_map_grant_ref *batch, unsigned count)
858 {
859 	struct gnttab_map_grant_ref *op;
860 
861 	if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, batch, count))
862 		BUG();
863 	for (op = batch; op < batch + count; op++)
864 		if (op->status == GNTST_eagain)
865 			gnttab_retry_eagain_gop(GNTTABOP_map_grant_ref, op,
866 						&op->status, __func__);
867 }
868 EXPORT_SYMBOL_GPL(gnttab_batch_map);
869 
870 void gnttab_batch_copy(struct gnttab_copy *batch, unsigned count)
871 {
872 	struct gnttab_copy *op;
873 
874 	if (HYPERVISOR_grant_table_op(GNTTABOP_copy, batch, count))
875 		BUG();
876 	for (op = batch; op < batch + count; op++)
877 		if (op->status == GNTST_eagain)
878 			gnttab_retry_eagain_gop(GNTTABOP_copy, op,
879 						&op->status, __func__);
880 }
881 EXPORT_SYMBOL_GPL(gnttab_batch_copy);
882 
883 int gnttab_map_refs(struct gnttab_map_grant_ref *map_ops,
884 		    struct gnttab_map_grant_ref *kmap_ops,
885 		    struct page **pages, unsigned int count)
886 {
887 	int i, ret;
888 	bool lazy = false;
889 	pte_t *pte;
890 	unsigned long mfn;
891 
892 	ret = HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, map_ops, count);
893 	if (ret)
894 		return ret;
895 
896 	/* Retry eagain maps */
897 	for (i = 0; i < count; i++)
898 		if (map_ops[i].status == GNTST_eagain)
899 			gnttab_retry_eagain_gop(GNTTABOP_map_grant_ref, map_ops + i,
900 						&map_ops[i].status, __func__);
901 
902 	/* this is basically a nop on x86 */
903 	if (xen_feature(XENFEAT_auto_translated_physmap)) {
904 		for (i = 0; i < count; i++) {
905 			if (map_ops[i].status)
906 				continue;
907 			set_phys_to_machine(map_ops[i].host_addr >> PAGE_SHIFT,
908 					map_ops[i].dev_bus_addr >> PAGE_SHIFT);
909 		}
910 		return ret;
911 	}
912 
913 	if (!in_interrupt() && paravirt_get_lazy_mode() == PARAVIRT_LAZY_NONE) {
914 		arch_enter_lazy_mmu_mode();
915 		lazy = true;
916 	}
917 
918 	for (i = 0; i < count; i++) {
919 		/* Do not add to override if the map failed. */
920 		if (map_ops[i].status)
921 			continue;
922 
923 		if (map_ops[i].flags & GNTMAP_contains_pte) {
924 			pte = (pte_t *) (mfn_to_virt(PFN_DOWN(map_ops[i].host_addr)) +
925 				(map_ops[i].host_addr & ~PAGE_MASK));
926 			mfn = pte_mfn(*pte);
927 		} else {
928 			mfn = PFN_DOWN(map_ops[i].dev_bus_addr);
929 		}
930 		ret = m2p_add_override(mfn, pages[i], kmap_ops ?
931 				       &kmap_ops[i] : NULL);
932 		if (ret)
933 			goto out;
934 	}
935 
936  out:
937 	if (lazy)
938 		arch_leave_lazy_mmu_mode();
939 
940 	return ret;
941 }
942 EXPORT_SYMBOL_GPL(gnttab_map_refs);
943 
944 int gnttab_unmap_refs(struct gnttab_unmap_grant_ref *unmap_ops,
945 		      struct gnttab_map_grant_ref *kmap_ops,
946 		      struct page **pages, unsigned int count)
947 {
948 	int i, ret;
949 	bool lazy = false;
950 
951 	ret = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, unmap_ops, count);
952 	if (ret)
953 		return ret;
954 
955 	/* this is basically a nop on x86 */
956 	if (xen_feature(XENFEAT_auto_translated_physmap)) {
957 		for (i = 0; i < count; i++) {
958 			set_phys_to_machine(unmap_ops[i].host_addr >> PAGE_SHIFT,
959 					INVALID_P2M_ENTRY);
960 		}
961 		return ret;
962 	}
963 
964 	if (!in_interrupt() && paravirt_get_lazy_mode() == PARAVIRT_LAZY_NONE) {
965 		arch_enter_lazy_mmu_mode();
966 		lazy = true;
967 	}
968 
969 	for (i = 0; i < count; i++) {
970 		ret = m2p_remove_override(pages[i], kmap_ops ?
971 				       &kmap_ops[i] : NULL);
972 		if (ret)
973 			goto out;
974 	}
975 
976  out:
977 	if (lazy)
978 		arch_leave_lazy_mmu_mode();
979 
980 	return ret;
981 }
982 EXPORT_SYMBOL_GPL(gnttab_unmap_refs);
983 
984 static unsigned nr_status_frames(unsigned nr_grant_frames)
985 {
986 	BUG_ON(grefs_per_grant_frame == 0);
987 	return (nr_grant_frames * grefs_per_grant_frame + SPP - 1) / SPP;
988 }
989 
990 static int gnttab_map_frames_v1(xen_pfn_t *frames, unsigned int nr_gframes)
991 {
992 	int rc;
993 
994 	rc = arch_gnttab_map_shared(frames, nr_gframes,
995 				    gnttab_max_grant_frames(),
996 				    &gnttab_shared.addr);
997 	BUG_ON(rc);
998 
999 	return 0;
1000 }
1001 
1002 static void gnttab_unmap_frames_v1(void)
1003 {
1004 	arch_gnttab_unmap(gnttab_shared.addr, nr_grant_frames);
1005 }
1006 
1007 static int gnttab_map_frames_v2(xen_pfn_t *frames, unsigned int nr_gframes)
1008 {
1009 	uint64_t *sframes;
1010 	unsigned int nr_sframes;
1011 	struct gnttab_get_status_frames getframes;
1012 	int rc;
1013 
1014 	nr_sframes = nr_status_frames(nr_gframes);
1015 
1016 	/* No need for kzalloc as it is initialized in following hypercall
1017 	 * GNTTABOP_get_status_frames.
1018 	 */
1019 	sframes = kmalloc(nr_sframes  * sizeof(uint64_t), GFP_ATOMIC);
1020 	if (!sframes)
1021 		return -ENOMEM;
1022 
1023 	getframes.dom        = DOMID_SELF;
1024 	getframes.nr_frames  = nr_sframes;
1025 	set_xen_guest_handle(getframes.frame_list, sframes);
1026 
1027 	rc = HYPERVISOR_grant_table_op(GNTTABOP_get_status_frames,
1028 				       &getframes, 1);
1029 	if (rc == -ENOSYS) {
1030 		kfree(sframes);
1031 		return -ENOSYS;
1032 	}
1033 
1034 	BUG_ON(rc || getframes.status);
1035 
1036 	rc = arch_gnttab_map_status(sframes, nr_sframes,
1037 				    nr_status_frames(gnttab_max_grant_frames()),
1038 				    &grstatus);
1039 	BUG_ON(rc);
1040 	kfree(sframes);
1041 
1042 	rc = arch_gnttab_map_shared(frames, nr_gframes,
1043 				    gnttab_max_grant_frames(),
1044 				    &gnttab_shared.addr);
1045 	BUG_ON(rc);
1046 
1047 	return 0;
1048 }
1049 
1050 static void gnttab_unmap_frames_v2(void)
1051 {
1052 	arch_gnttab_unmap(gnttab_shared.addr, nr_grant_frames);
1053 	arch_gnttab_unmap(grstatus, nr_status_frames(nr_grant_frames));
1054 }
1055 
1056 static int gnttab_map(unsigned int start_idx, unsigned int end_idx)
1057 {
1058 	struct gnttab_setup_table setup;
1059 	xen_pfn_t *frames;
1060 	unsigned int nr_gframes = end_idx + 1;
1061 	int rc;
1062 
1063 	if (xen_hvm_domain()) {
1064 		struct xen_add_to_physmap xatp;
1065 		unsigned int i = end_idx;
1066 		rc = 0;
1067 		/*
1068 		 * Loop backwards, so that the first hypercall has the largest
1069 		 * index, ensuring that the table will grow only once.
1070 		 */
1071 		do {
1072 			xatp.domid = DOMID_SELF;
1073 			xatp.idx = i;
1074 			xatp.space = XENMAPSPACE_grant_table;
1075 			xatp.gpfn = (xen_hvm_resume_frames >> PAGE_SHIFT) + i;
1076 			rc = HYPERVISOR_memory_op(XENMEM_add_to_physmap, &xatp);
1077 			if (rc != 0) {
1078 				pr_warn("grant table add_to_physmap failed, err=%d\n",
1079 					rc);
1080 				break;
1081 			}
1082 		} while (i-- > start_idx);
1083 
1084 		return rc;
1085 	}
1086 
1087 	/* No need for kzalloc as it is initialized in following hypercall
1088 	 * GNTTABOP_setup_table.
1089 	 */
1090 	frames = kmalloc(nr_gframes * sizeof(unsigned long), GFP_ATOMIC);
1091 	if (!frames)
1092 		return -ENOMEM;
1093 
1094 	setup.dom        = DOMID_SELF;
1095 	setup.nr_frames  = nr_gframes;
1096 	set_xen_guest_handle(setup.frame_list, frames);
1097 
1098 	rc = HYPERVISOR_grant_table_op(GNTTABOP_setup_table, &setup, 1);
1099 	if (rc == -ENOSYS) {
1100 		kfree(frames);
1101 		return -ENOSYS;
1102 	}
1103 
1104 	BUG_ON(rc || setup.status);
1105 
1106 	rc = gnttab_interface->map_frames(frames, nr_gframes);
1107 
1108 	kfree(frames);
1109 
1110 	return rc;
1111 }
1112 
1113 static struct gnttab_ops gnttab_v1_ops = {
1114 	.map_frames			= gnttab_map_frames_v1,
1115 	.unmap_frames			= gnttab_unmap_frames_v1,
1116 	.update_entry			= gnttab_update_entry_v1,
1117 	.end_foreign_access_ref		= gnttab_end_foreign_access_ref_v1,
1118 	.end_foreign_transfer_ref	= gnttab_end_foreign_transfer_ref_v1,
1119 	.query_foreign_access		= gnttab_query_foreign_access_v1,
1120 };
1121 
1122 static struct gnttab_ops gnttab_v2_ops = {
1123 	.map_frames			= gnttab_map_frames_v2,
1124 	.unmap_frames			= gnttab_unmap_frames_v2,
1125 	.update_entry			= gnttab_update_entry_v2,
1126 	.end_foreign_access_ref		= gnttab_end_foreign_access_ref_v2,
1127 	.end_foreign_transfer_ref	= gnttab_end_foreign_transfer_ref_v2,
1128 	.query_foreign_access		= gnttab_query_foreign_access_v2,
1129 	.update_subpage_entry		= gnttab_update_subpage_entry_v2,
1130 	.update_trans_entry		= gnttab_update_trans_entry_v2,
1131 };
1132 
1133 static void gnttab_request_version(void)
1134 {
1135 	int rc;
1136 	struct gnttab_set_version gsv;
1137 
1138 	if (xen_hvm_domain())
1139 		gsv.version = 1;
1140 	else
1141 		gsv.version = 2;
1142 	rc = HYPERVISOR_grant_table_op(GNTTABOP_set_version, &gsv, 1);
1143 	if (rc == 0 && gsv.version == 2) {
1144 		grant_table_version = 2;
1145 		grefs_per_grant_frame = PAGE_SIZE / sizeof(union grant_entry_v2);
1146 		gnttab_interface = &gnttab_v2_ops;
1147 	} else if (grant_table_version == 2) {
1148 		/*
1149 		 * If we've already used version 2 features,
1150 		 * but then suddenly discover that they're not
1151 		 * available (e.g. migrating to an older
1152 		 * version of Xen), almost unbounded badness
1153 		 * can happen.
1154 		 */
1155 		panic("we need grant tables version 2, but only version 1 is available");
1156 	} else {
1157 		grant_table_version = 1;
1158 		grefs_per_grant_frame = PAGE_SIZE / sizeof(struct grant_entry_v1);
1159 		gnttab_interface = &gnttab_v1_ops;
1160 	}
1161 	pr_info("Grant tables using version %d layout\n", grant_table_version);
1162 }
1163 
1164 static int gnttab_setup(void)
1165 {
1166 	unsigned int max_nr_gframes;
1167 
1168 	max_nr_gframes = gnttab_max_grant_frames();
1169 	if (max_nr_gframes < nr_grant_frames)
1170 		return -ENOSYS;
1171 
1172 	if (xen_pv_domain())
1173 		return gnttab_map(0, nr_grant_frames - 1);
1174 
1175 	if (gnttab_shared.addr == NULL) {
1176 		gnttab_shared.addr = xen_remap(xen_hvm_resume_frames,
1177 						PAGE_SIZE * max_nr_gframes);
1178 		if (gnttab_shared.addr == NULL) {
1179 			pr_warn("Failed to ioremap gnttab share frames!\n");
1180 			return -ENOMEM;
1181 		}
1182 	}
1183 
1184 	gnttab_map(0, nr_grant_frames - 1);
1185 
1186 	return 0;
1187 }
1188 
1189 int gnttab_resume(void)
1190 {
1191 	gnttab_request_version();
1192 	return gnttab_setup();
1193 }
1194 
1195 int gnttab_suspend(void)
1196 {
1197 	gnttab_interface->unmap_frames();
1198 	return 0;
1199 }
1200 
1201 static int gnttab_expand(unsigned int req_entries)
1202 {
1203 	int rc;
1204 	unsigned int cur, extra;
1205 
1206 	BUG_ON(grefs_per_grant_frame == 0);
1207 	cur = nr_grant_frames;
1208 	extra = ((req_entries + (grefs_per_grant_frame-1)) /
1209 		 grefs_per_grant_frame);
1210 	if (cur + extra > gnttab_max_grant_frames())
1211 		return -ENOSPC;
1212 
1213 	rc = gnttab_map(cur, cur + extra - 1);
1214 	if (rc == 0)
1215 		rc = grow_gnttab_list(extra);
1216 
1217 	return rc;
1218 }
1219 
1220 int gnttab_init(void)
1221 {
1222 	int i;
1223 	unsigned int max_nr_glist_frames, nr_glist_frames;
1224 	unsigned int nr_init_grefs;
1225 	int ret;
1226 
1227 	gnttab_request_version();
1228 	nr_grant_frames = 1;
1229 	boot_max_nr_grant_frames = __max_nr_grant_frames();
1230 
1231 	/* Determine the maximum number of frames required for the
1232 	 * grant reference free list on the current hypervisor.
1233 	 */
1234 	BUG_ON(grefs_per_grant_frame == 0);
1235 	max_nr_glist_frames = (boot_max_nr_grant_frames *
1236 			       grefs_per_grant_frame / RPP);
1237 
1238 	gnttab_list = kmalloc(max_nr_glist_frames * sizeof(grant_ref_t *),
1239 			      GFP_KERNEL);
1240 	if (gnttab_list == NULL)
1241 		return -ENOMEM;
1242 
1243 	nr_glist_frames = (nr_grant_frames * grefs_per_grant_frame + RPP - 1) / RPP;
1244 	for (i = 0; i < nr_glist_frames; i++) {
1245 		gnttab_list[i] = (grant_ref_t *)__get_free_page(GFP_KERNEL);
1246 		if (gnttab_list[i] == NULL) {
1247 			ret = -ENOMEM;
1248 			goto ini_nomem;
1249 		}
1250 	}
1251 
1252 	if (gnttab_setup() < 0) {
1253 		ret = -ENODEV;
1254 		goto ini_nomem;
1255 	}
1256 
1257 	nr_init_grefs = nr_grant_frames * grefs_per_grant_frame;
1258 
1259 	for (i = NR_RESERVED_ENTRIES; i < nr_init_grefs - 1; i++)
1260 		gnttab_entry(i) = i + 1;
1261 
1262 	gnttab_entry(nr_init_grefs - 1) = GNTTAB_LIST_END;
1263 	gnttab_free_count = nr_init_grefs - NR_RESERVED_ENTRIES;
1264 	gnttab_free_head  = NR_RESERVED_ENTRIES;
1265 
1266 	printk("Grant table initialized\n");
1267 	return 0;
1268 
1269  ini_nomem:
1270 	for (i--; i >= 0; i--)
1271 		free_page((unsigned long)gnttab_list[i]);
1272 	kfree(gnttab_list);
1273 	return ret;
1274 }
1275 EXPORT_SYMBOL_GPL(gnttab_init);
1276 
1277 static int __gnttab_init(void)
1278 {
1279 	/* Delay grant-table initialization in the PV on HVM case */
1280 	if (xen_hvm_domain())
1281 		return 0;
1282 
1283 	if (!xen_pv_domain())
1284 		return -ENODEV;
1285 
1286 	return gnttab_init();
1287 }
1288 
1289 core_initcall(__gnttab_init);
1290