1*e65e175bSOded Gabbay // SPDX-License-Identifier: GPL-2.0
2*e65e175bSOded Gabbay 
3*e65e175bSOded Gabbay /*
4*e65e175bSOded Gabbay  * Copyright 2020-2022 HabanaLabs, Ltd.
5*e65e175bSOded Gabbay  * All Rights Reserved.
6*e65e175bSOded Gabbay  */
7*e65e175bSOded Gabbay 
8*e65e175bSOded Gabbay #include "../habanalabs.h"
9*e65e175bSOded Gabbay #include "../../include/hw_ip/mmu/mmu_general.h"
10*e65e175bSOded Gabbay 
11*e65e175bSOded Gabbay #include <linux/slab.h>
12*e65e175bSOded Gabbay 
hl_mmu_v2_hr_get_pgt_info(struct hl_ctx * ctx,u64 phys_hop_addr)13*e65e175bSOded Gabbay static struct pgt_info *hl_mmu_v2_hr_get_pgt_info(struct hl_ctx *ctx, u64 phys_hop_addr)
14*e65e175bSOded Gabbay {
15*e65e175bSOded Gabbay 	struct pgt_info *pgt_info = NULL;
16*e65e175bSOded Gabbay 
17*e65e175bSOded Gabbay 	hash_for_each_possible(ctx->hr_mmu_phys_hash, pgt_info, node,
18*e65e175bSOded Gabbay 				(unsigned long) phys_hop_addr)
19*e65e175bSOded Gabbay 		if (phys_hop_addr == pgt_info->phys_addr)
20*e65e175bSOded Gabbay 			break;
21*e65e175bSOded Gabbay 
22*e65e175bSOded Gabbay 	return pgt_info;
23*e65e175bSOded Gabbay }
24*e65e175bSOded Gabbay 
hl_mmu_v2_hr_add_pgt_info(struct hl_ctx * ctx,struct pgt_info * pgt_info,dma_addr_t phys_addr)25*e65e175bSOded Gabbay static void hl_mmu_v2_hr_add_pgt_info(struct hl_ctx *ctx, struct pgt_info *pgt_info,
26*e65e175bSOded Gabbay 					dma_addr_t phys_addr)
27*e65e175bSOded Gabbay {
28*e65e175bSOded Gabbay 	hash_add(ctx->hr_mmu_phys_hash, &pgt_info->node, phys_addr);
29*e65e175bSOded Gabbay }
30*e65e175bSOded Gabbay 
hl_mmu_v2_hr_get_hop0_pgt_info(struct hl_ctx * ctx)31*e65e175bSOded Gabbay static struct pgt_info *hl_mmu_v2_hr_get_hop0_pgt_info(struct hl_ctx *ctx)
32*e65e175bSOded Gabbay {
33*e65e175bSOded Gabbay 	return &ctx->hdev->mmu_priv.hr.mmu_asid_hop0[ctx->asid];
34*e65e175bSOded Gabbay }
35*e65e175bSOded Gabbay 
36*e65e175bSOded Gabbay /**
37*e65e175bSOded Gabbay  * hl_mmu_v2_hr_init() - initialize the MMU module.
38*e65e175bSOded Gabbay  * @hdev: habanalabs device structure.
39*e65e175bSOded Gabbay  *
40*e65e175bSOded Gabbay  * This function does the following:
41*e65e175bSOded Gabbay  * - Create a pool of pages for pgt_infos.
42*e65e175bSOded Gabbay  * - Create a shadow table for pgt
43*e65e175bSOded Gabbay  *
44*e65e175bSOded Gabbay  * Return: 0 for success, non-zero for failure.
45*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_init(struct hl_device * hdev)46*e65e175bSOded Gabbay static inline int hl_mmu_v2_hr_init(struct hl_device *hdev)
47*e65e175bSOded Gabbay {
48*e65e175bSOded Gabbay 	struct asic_fixed_properties *prop = &hdev->asic_prop;
49*e65e175bSOded Gabbay 
50*e65e175bSOded Gabbay 	return hl_mmu_hr_init(hdev, &hdev->mmu_priv.hr, prop->mmu_hop_table_size,
51*e65e175bSOded Gabbay 				prop->mmu_pgt_size);
52*e65e175bSOded Gabbay }
53*e65e175bSOded Gabbay 
54*e65e175bSOded Gabbay /**
55*e65e175bSOded Gabbay  * hl_mmu_v2_hr_fini() - release the MMU module.
56*e65e175bSOded Gabbay  * @hdev: habanalabs device structure.
57*e65e175bSOded Gabbay  *
58*e65e175bSOded Gabbay  * This function does the following:
59*e65e175bSOded Gabbay  * - Disable MMU in H/W.
60*e65e175bSOded Gabbay  * - Free the pgt_infos pool.
61*e65e175bSOded Gabbay  *
62*e65e175bSOded Gabbay  * All contexts should be freed before calling this function.
63*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_fini(struct hl_device * hdev)64*e65e175bSOded Gabbay static inline void hl_mmu_v2_hr_fini(struct hl_device *hdev)
65*e65e175bSOded Gabbay {
66*e65e175bSOded Gabbay 	struct asic_fixed_properties *prop = &hdev->asic_prop;
67*e65e175bSOded Gabbay 
68*e65e175bSOded Gabbay 	hl_mmu_hr_fini(hdev, &hdev->mmu_priv.hr, prop->mmu_hop_table_size);
69*e65e175bSOded Gabbay }
70*e65e175bSOded Gabbay 
71*e65e175bSOded Gabbay /**
72*e65e175bSOded Gabbay  * hl_mmu_v2_hr_ctx_init() - initialize a context for using the MMU module.
73*e65e175bSOded Gabbay  * @ctx: pointer to the context structure to initialize.
74*e65e175bSOded Gabbay  *
75*e65e175bSOded Gabbay  * Initialize a mutex to protect the concurrent mapping flow, a hash to hold all
76*e65e175bSOded Gabbay  * page tables hops related to this context.
77*e65e175bSOded Gabbay  * Return: 0 on success, non-zero otherwise.
78*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_ctx_init(struct hl_ctx * ctx)79*e65e175bSOded Gabbay static int hl_mmu_v2_hr_ctx_init(struct hl_ctx *ctx)
80*e65e175bSOded Gabbay {
81*e65e175bSOded Gabbay 	hash_init(ctx->hr_mmu_phys_hash);
82*e65e175bSOded Gabbay 	return 0;
83*e65e175bSOded Gabbay }
84*e65e175bSOded Gabbay 
85*e65e175bSOded Gabbay /*
86*e65e175bSOded Gabbay  * hl_mmu_v2_hr_ctx_fini - disable a ctx from using the mmu module
87*e65e175bSOded Gabbay  *
88*e65e175bSOded Gabbay  * @ctx: pointer to the context structure
89*e65e175bSOded Gabbay  *
90*e65e175bSOded Gabbay  * This function does the following:
91*e65e175bSOded Gabbay  * - Free any pgts which were not freed yet
92*e65e175bSOded Gabbay  * - Free the mutex
93*e65e175bSOded Gabbay  * - Free DRAM default page mapping hops
94*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_ctx_fini(struct hl_ctx * ctx)95*e65e175bSOded Gabbay static void hl_mmu_v2_hr_ctx_fini(struct hl_ctx *ctx)
96*e65e175bSOded Gabbay {
97*e65e175bSOded Gabbay 	struct hl_device *hdev = ctx->hdev;
98*e65e175bSOded Gabbay 	struct pgt_info *pgt_info;
99*e65e175bSOded Gabbay 	struct hlist_node *tmp;
100*e65e175bSOded Gabbay 	int i;
101*e65e175bSOded Gabbay 
102*e65e175bSOded Gabbay 	if (!hash_empty(ctx->hr_mmu_phys_hash))
103*e65e175bSOded Gabbay 		dev_err(hdev->dev, "ctx %d is freed while it has pgts in use\n",
104*e65e175bSOded Gabbay 			ctx->asid);
105*e65e175bSOded Gabbay 
106*e65e175bSOded Gabbay 	hash_for_each_safe(ctx->hr_mmu_phys_hash, i, tmp, pgt_info, node) {
107*e65e175bSOded Gabbay 		dev_err_ratelimited(hdev->dev,
108*e65e175bSOded Gabbay 			"pgt_info of addr 0x%llx of asid %d was not destroyed, num_ptes: %d\n",
109*e65e175bSOded Gabbay 			pgt_info->phys_addr, ctx->asid, pgt_info->num_of_ptes);
110*e65e175bSOded Gabbay 		hl_mmu_hr_free_hop_remove_pgt(pgt_info, &ctx->hdev->mmu_priv.hr,
111*e65e175bSOded Gabbay 							ctx->hdev->asic_prop.mmu_hop_table_size);
112*e65e175bSOded Gabbay 	}
113*e65e175bSOded Gabbay }
114*e65e175bSOded Gabbay 
_hl_mmu_v2_hr_unmap(struct hl_ctx * ctx,u64 virt_addr,bool is_dram_addr)115*e65e175bSOded Gabbay static int _hl_mmu_v2_hr_unmap(struct hl_ctx *ctx,
116*e65e175bSOded Gabbay 				u64 virt_addr, bool is_dram_addr)
117*e65e175bSOded Gabbay {
118*e65e175bSOded Gabbay 	u64 curr_pte, scrambled_virt_addr, hop_pte_phys_addr[MMU_ARCH_6_HOPS] = { 0 };
119*e65e175bSOded Gabbay 	struct pgt_info *hops_pgt_info[MMU_ARCH_6_HOPS] = { NULL };
120*e65e175bSOded Gabbay 	struct hl_device *hdev = ctx->hdev;
121*e65e175bSOded Gabbay 	struct asic_fixed_properties *prop;
122*e65e175bSOded Gabbay 	struct hl_mmu_properties *mmu_prop;
123*e65e175bSOded Gabbay 	bool is_huge = false;
124*e65e175bSOded Gabbay 	int i, hop_last;
125*e65e175bSOded Gabbay 
126*e65e175bSOded Gabbay 	prop = &hdev->asic_prop;
127*e65e175bSOded Gabbay 
128*e65e175bSOded Gabbay 	/* shifts and masks are the same in PMMU and HMMU, use one of them */
129*e65e175bSOded Gabbay 	mmu_prop = is_dram_addr ? &prop->dmmu : &prop->pmmu;
130*e65e175bSOded Gabbay 	hop_last = mmu_prop->num_hops - 1;
131*e65e175bSOded Gabbay 
132*e65e175bSOded Gabbay 	scrambled_virt_addr = hdev->asic_funcs->scramble_addr(hdev, virt_addr);
133*e65e175bSOded Gabbay 	curr_pte = 0;
134*e65e175bSOded Gabbay 
135*e65e175bSOded Gabbay 	for (i = 0 ; i < mmu_prop->num_hops ; i++) {
136*e65e175bSOded Gabbay 		/* we get HOP0 differently, it doesn't need curr_pte */
137*e65e175bSOded Gabbay 		if (i == 0)
138*e65e175bSOded Gabbay 			hops_pgt_info[i] = hl_mmu_v2_hr_get_hop0_pgt_info(ctx);
139*e65e175bSOded Gabbay 		else
140*e65e175bSOded Gabbay 			hops_pgt_info[i] = hl_mmu_hr_get_next_hop_pgt_info(ctx,
141*e65e175bSOded Gabbay 					&ctx->hdev->mmu_func[MMU_HR_PGT].hr_funcs, curr_pte);
142*e65e175bSOded Gabbay 		if (!hops_pgt_info[i])
143*e65e175bSOded Gabbay 			goto not_mapped;
144*e65e175bSOded Gabbay 
145*e65e175bSOded Gabbay 		hop_pte_phys_addr[i] = hl_mmu_get_hop_pte_phys_addr(ctx, mmu_prop, i,
146*e65e175bSOded Gabbay 									hops_pgt_info[i]->phys_addr,
147*e65e175bSOded Gabbay 									scrambled_virt_addr);
148*e65e175bSOded Gabbay 		if (hop_pte_phys_addr[i] == U64_MAX)
149*e65e175bSOded Gabbay 			return -EFAULT;
150*e65e175bSOded Gabbay 
151*e65e175bSOded Gabbay 		curr_pte = *(u64 *) (uintptr_t) hl_mmu_hr_pte_phys_to_virt(ctx, hops_pgt_info[i],
152*e65e175bSOded Gabbay 							hop_pte_phys_addr[i],
153*e65e175bSOded Gabbay 							ctx->hdev->asic_prop.mmu_hop_table_size);
154*e65e175bSOded Gabbay 
155*e65e175bSOded Gabbay 		if ((i < hop_last) && (curr_pte & mmu_prop->last_mask)) {
156*e65e175bSOded Gabbay 			hop_last = i;
157*e65e175bSOded Gabbay 			is_huge = true;
158*e65e175bSOded Gabbay 			break;
159*e65e175bSOded Gabbay 		}
160*e65e175bSOded Gabbay 	}
161*e65e175bSOded Gabbay 
162*e65e175bSOded Gabbay 	if (is_dram_addr && !is_huge) {
163*e65e175bSOded Gabbay 		dev_err(hdev->dev, "DRAM unmapping should use huge pages only\n");
164*e65e175bSOded Gabbay 		return -EFAULT;
165*e65e175bSOded Gabbay 	}
166*e65e175bSOded Gabbay 
167*e65e175bSOded Gabbay 	if (!(curr_pte & PAGE_PRESENT_MASK))
168*e65e175bSOded Gabbay 		goto not_mapped;
169*e65e175bSOded Gabbay 
170*e65e175bSOded Gabbay 	for (i = hop_last ; i > 0 ; i--) {
171*e65e175bSOded Gabbay 		hl_mmu_hr_clear_pte(ctx, hops_pgt_info[i], hop_pte_phys_addr[i],
172*e65e175bSOded Gabbay 						ctx->hdev->asic_prop.mmu_hop_table_size);
173*e65e175bSOded Gabbay 
174*e65e175bSOded Gabbay 		if (hl_mmu_hr_put_pte(ctx, hops_pgt_info[i], &ctx->hdev->mmu_priv.hr,
175*e65e175bSOded Gabbay 						ctx->hdev->asic_prop.mmu_hop_table_size))
176*e65e175bSOded Gabbay 			goto mapped;
177*e65e175bSOded Gabbay 	}
178*e65e175bSOded Gabbay 	hl_mmu_hr_clear_pte(ctx, hops_pgt_info[0], hop_pte_phys_addr[0],
179*e65e175bSOded Gabbay 						ctx->hdev->asic_prop.mmu_hop_table_size);
180*e65e175bSOded Gabbay 
181*e65e175bSOded Gabbay mapped:
182*e65e175bSOded Gabbay 	return 0;
183*e65e175bSOded Gabbay 
184*e65e175bSOded Gabbay not_mapped:
185*e65e175bSOded Gabbay 	dev_err(hdev->dev, "virt addr 0x%llx is not mapped to phys addr\n", virt_addr);
186*e65e175bSOded Gabbay 
187*e65e175bSOded Gabbay 	return -EINVAL;
188*e65e175bSOded Gabbay }
189*e65e175bSOded Gabbay 
hl_mmu_v2_get_last_hop(struct hl_mmu_properties * mmu_prop,u32 page_size)190*e65e175bSOded Gabbay static int hl_mmu_v2_get_last_hop(struct hl_mmu_properties *mmu_prop, u32 page_size)
191*e65e175bSOded Gabbay {
192*e65e175bSOded Gabbay 	int hop;
193*e65e175bSOded Gabbay 
194*e65e175bSOded Gabbay 	for (hop = (mmu_prop->num_hops - 1); hop; hop--) {
195*e65e175bSOded Gabbay 		if (mmu_prop->hop_shifts[hop] == 0)
196*e65e175bSOded Gabbay 			continue;
197*e65e175bSOded Gabbay 
198*e65e175bSOded Gabbay 		if (page_size <= (1 << mmu_prop->hop_shifts[hop]))
199*e65e175bSOded Gabbay 			break;
200*e65e175bSOded Gabbay 	}
201*e65e175bSOded Gabbay 
202*e65e175bSOded Gabbay 	return hop;
203*e65e175bSOded Gabbay }
204*e65e175bSOded Gabbay 
_hl_mmu_v2_hr_map(struct hl_ctx * ctx,u64 virt_addr,u64 phys_addr,u32 page_size,bool is_dram_addr)205*e65e175bSOded Gabbay static int _hl_mmu_v2_hr_map(struct hl_ctx *ctx,
206*e65e175bSOded Gabbay 			u64 virt_addr, u64 phys_addr,
207*e65e175bSOded Gabbay 			u32 page_size, bool is_dram_addr)
208*e65e175bSOded Gabbay {
209*e65e175bSOded Gabbay 	u64 hop_pte_phys_addr[MMU_ARCH_6_HOPS] = { 0 },
210*e65e175bSOded Gabbay 		curr_pte = 0, scrambled_virt_addr, scrambled_phys_addr;
211*e65e175bSOded Gabbay 	struct pgt_info *hops_pgt_info[MMU_ARCH_6_HOPS] = { NULL };
212*e65e175bSOded Gabbay 	bool hop_new[MMU_ARCH_6_HOPS] = { false };
213*e65e175bSOded Gabbay 	struct hl_device *hdev = ctx->hdev;
214*e65e175bSOded Gabbay 	struct asic_fixed_properties *prop = &hdev->asic_prop;
215*e65e175bSOded Gabbay 	struct hl_mmu_properties *mmu_prop;
216*e65e175bSOded Gabbay 	int i, hop_last, rc = -ENOMEM;
217*e65e175bSOded Gabbay 
218*e65e175bSOded Gabbay 	/*
219*e65e175bSOded Gabbay 	 * This mapping function can map a page or a huge page. For huge page
220*e65e175bSOded Gabbay 	 * there are only 4 hops rather than 5. Currently the DRAM allocation
221*e65e175bSOded Gabbay 	 * uses huge pages only but user memory could have been allocated with
222*e65e175bSOded Gabbay 	 * one of the two page sizes. Since this is a common code for all the
223*e65e175bSOded Gabbay 	 * three cases, we need this hugs page check.
224*e65e175bSOded Gabbay 	 */
225*e65e175bSOded Gabbay 	if (is_dram_addr)
226*e65e175bSOded Gabbay 		mmu_prop = &prop->dmmu;
227*e65e175bSOded Gabbay 	else if (page_size == prop->pmmu_huge.page_size)
228*e65e175bSOded Gabbay 		mmu_prop = &prop->pmmu_huge;
229*e65e175bSOded Gabbay 	else
230*e65e175bSOded Gabbay 		mmu_prop = &prop->pmmu;
231*e65e175bSOded Gabbay 
232*e65e175bSOded Gabbay 	hop_last = hl_mmu_v2_get_last_hop(mmu_prop, page_size);
233*e65e175bSOded Gabbay 	if (hop_last <= 0) {
234*e65e175bSOded Gabbay 		dev_err(ctx->hdev->dev, "Invalid last HOP %d\n", hop_last);
235*e65e175bSOded Gabbay 		return -EFAULT;
236*e65e175bSOded Gabbay 	}
237*e65e175bSOded Gabbay 
238*e65e175bSOded Gabbay 	scrambled_virt_addr = hdev->asic_funcs->scramble_addr(hdev, virt_addr);
239*e65e175bSOded Gabbay 	scrambled_phys_addr = hdev->asic_funcs->scramble_addr(hdev, phys_addr);
240*e65e175bSOded Gabbay 
241*e65e175bSOded Gabbay 	for (i = 0 ; i <= hop_last ; i++) {
242*e65e175bSOded Gabbay 
243*e65e175bSOded Gabbay 		if (i == 0)
244*e65e175bSOded Gabbay 			hops_pgt_info[i] = hl_mmu_v2_hr_get_hop0_pgt_info(ctx);
245*e65e175bSOded Gabbay 		else
246*e65e175bSOded Gabbay 			hops_pgt_info[i] = hl_mmu_hr_get_alloc_next_hop(ctx,
247*e65e175bSOded Gabbay 							&ctx->hdev->mmu_priv.hr,
248*e65e175bSOded Gabbay 							&ctx->hdev->mmu_func[MMU_HR_PGT].hr_funcs,
249*e65e175bSOded Gabbay 							mmu_prop, curr_pte, &hop_new[i]);
250*e65e175bSOded Gabbay 		if (!hops_pgt_info[i])
251*e65e175bSOded Gabbay 			goto err;
252*e65e175bSOded Gabbay 
253*e65e175bSOded Gabbay 		hop_pte_phys_addr[i] = hl_mmu_get_hop_pte_phys_addr(ctx, mmu_prop, i,
254*e65e175bSOded Gabbay 									hops_pgt_info[i]->phys_addr,
255*e65e175bSOded Gabbay 									scrambled_virt_addr);
256*e65e175bSOded Gabbay 		curr_pte = *(u64 *) (uintptr_t) hl_mmu_hr_pte_phys_to_virt(ctx, hops_pgt_info[i],
257*e65e175bSOded Gabbay 							hop_pte_phys_addr[i],
258*e65e175bSOded Gabbay 							ctx->hdev->asic_prop.mmu_hop_table_size);
259*e65e175bSOded Gabbay 	}
260*e65e175bSOded Gabbay 
261*e65e175bSOded Gabbay 	if (curr_pte & PAGE_PRESENT_MASK) {
262*e65e175bSOded Gabbay 		dev_err(hdev->dev, "mapping already exists for virt_addr 0x%llx\n",
263*e65e175bSOded Gabbay 									scrambled_virt_addr);
264*e65e175bSOded Gabbay 
265*e65e175bSOded Gabbay 		for (i = 0 ; i <= hop_last ; i++)
266*e65e175bSOded Gabbay 			dev_dbg(hdev->dev, "hop%d pte: 0x%llx (0x%llx)\n",
267*e65e175bSOded Gabbay 					i,
268*e65e175bSOded Gabbay 					*(u64 *) (uintptr_t)
269*e65e175bSOded Gabbay 					hl_mmu_hr_pte_phys_to_virt(ctx, hops_pgt_info[i],
270*e65e175bSOded Gabbay 							hop_pte_phys_addr[i],
271*e65e175bSOded Gabbay 							ctx->hdev->asic_prop.mmu_hop_table_size),
272*e65e175bSOded Gabbay 					hop_pte_phys_addr[i]);
273*e65e175bSOded Gabbay 		rc = -EINVAL;
274*e65e175bSOded Gabbay 		goto err;
275*e65e175bSOded Gabbay 	}
276*e65e175bSOded Gabbay 
277*e65e175bSOded Gabbay 	curr_pte = (scrambled_phys_addr & HOP_PHYS_ADDR_MASK) | mmu_prop->last_mask
278*e65e175bSOded Gabbay 			| PAGE_PRESENT_MASK;
279*e65e175bSOded Gabbay 
280*e65e175bSOded Gabbay 	/* Write the PTEs */
281*e65e175bSOded Gabbay 	hl_mmu_hr_write_pte(ctx, hops_pgt_info[hop_last], hop_pte_phys_addr[hop_last], curr_pte,
282*e65e175bSOded Gabbay 							ctx->hdev->asic_prop.mmu_hop_table_size);
283*e65e175bSOded Gabbay 
284*e65e175bSOded Gabbay 	/* for each new hop, add its address to the table of previous-hop */
285*e65e175bSOded Gabbay 	for (i = 1 ; i <= hop_last ; i++) {
286*e65e175bSOded Gabbay 		if (hop_new[i]) {
287*e65e175bSOded Gabbay 			curr_pte = (hops_pgt_info[i]->phys_addr & HOP_PHYS_ADDR_MASK) |
288*e65e175bSOded Gabbay 							PAGE_PRESENT_MASK;
289*e65e175bSOded Gabbay 			hl_mmu_hr_write_pte(ctx, hops_pgt_info[i - 1], hop_pte_phys_addr[i - 1],
290*e65e175bSOded Gabbay 						curr_pte, ctx->hdev->asic_prop.mmu_hop_table_size);
291*e65e175bSOded Gabbay 			if (i - 1)
292*e65e175bSOded Gabbay 				hl_mmu_hr_get_pte(ctx, &ctx->hdev->mmu_func[MMU_HR_PGT].hr_funcs,
293*e65e175bSOded Gabbay 								hops_pgt_info[i - 1]->phys_addr);
294*e65e175bSOded Gabbay 		}
295*e65e175bSOded Gabbay 	}
296*e65e175bSOded Gabbay 
297*e65e175bSOded Gabbay 	hl_mmu_hr_get_pte(ctx, &ctx->hdev->mmu_func[MMU_HR_PGT].hr_funcs,
298*e65e175bSOded Gabbay 						hops_pgt_info[hop_last]->phys_addr);
299*e65e175bSOded Gabbay 
300*e65e175bSOded Gabbay 	return 0;
301*e65e175bSOded Gabbay 
302*e65e175bSOded Gabbay err:
303*e65e175bSOded Gabbay 	for (i = 1 ; i <= hop_last ; i++)
304*e65e175bSOded Gabbay 		if (hop_new[i] && hops_pgt_info[i])
305*e65e175bSOded Gabbay 			hl_mmu_hr_free_hop_remove_pgt(hops_pgt_info[i], &ctx->hdev->mmu_priv.hr,
306*e65e175bSOded Gabbay 							ctx->hdev->asic_prop.mmu_hop_table_size);
307*e65e175bSOded Gabbay 
308*e65e175bSOded Gabbay 	return rc;
309*e65e175bSOded Gabbay }
310*e65e175bSOded Gabbay 
311*e65e175bSOded Gabbay /*
312*e65e175bSOded Gabbay  * hl_mmu_v2_swap_out - marks all mapping of the given ctx as swapped out
313*e65e175bSOded Gabbay  *
314*e65e175bSOded Gabbay  * @ctx: pointer to the context structure
315*e65e175bSOded Gabbay  *
316*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_swap_out(struct hl_ctx * ctx)317*e65e175bSOded Gabbay static void hl_mmu_v2_hr_swap_out(struct hl_ctx *ctx)
318*e65e175bSOded Gabbay {
319*e65e175bSOded Gabbay 
320*e65e175bSOded Gabbay }
321*e65e175bSOded Gabbay 
322*e65e175bSOded Gabbay /*
323*e65e175bSOded Gabbay  * hl_mmu_v2_swap_in - marks all mapping of the given ctx as swapped in
324*e65e175bSOded Gabbay  *
325*e65e175bSOded Gabbay  * @ctx: pointer to the context structure
326*e65e175bSOded Gabbay  *
327*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_swap_in(struct hl_ctx * ctx)328*e65e175bSOded Gabbay static void hl_mmu_v2_hr_swap_in(struct hl_ctx *ctx)
329*e65e175bSOded Gabbay {
330*e65e175bSOded Gabbay 
331*e65e175bSOded Gabbay }
332*e65e175bSOded Gabbay 
hl_mmu_v2_hr_get_tlb_mapping_params(struct hl_device * hdev,struct hl_mmu_properties ** mmu_prop,struct hl_mmu_hop_info * hops,u64 virt_addr,bool * is_huge)333*e65e175bSOded Gabbay static int hl_mmu_v2_hr_get_tlb_mapping_params(struct hl_device *hdev,
334*e65e175bSOded Gabbay 							struct hl_mmu_properties **mmu_prop,
335*e65e175bSOded Gabbay 							struct hl_mmu_hop_info *hops,
336*e65e175bSOded Gabbay 							u64 virt_addr, bool *is_huge)
337*e65e175bSOded Gabbay {
338*e65e175bSOded Gabbay 	struct asic_fixed_properties *prop = &hdev->asic_prop;
339*e65e175bSOded Gabbay 	bool is_dram_addr, is_pmmu_addr, is_pmmu_h_addr;
340*e65e175bSOded Gabbay 
341*e65e175bSOded Gabbay 	is_dram_addr = hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size,
342*e65e175bSOded Gabbay 						prop->dmmu.start_addr,
343*e65e175bSOded Gabbay 						prop->dmmu.end_addr);
344*e65e175bSOded Gabbay 	is_pmmu_addr = hl_mem_area_inside_range(virt_addr, prop->pmmu.page_size,
345*e65e175bSOded Gabbay 						prop->pmmu.start_addr,
346*e65e175bSOded Gabbay 						prop->pmmu.end_addr);
347*e65e175bSOded Gabbay 	is_pmmu_h_addr = hl_mem_area_inside_range(virt_addr,
348*e65e175bSOded Gabbay 						prop->pmmu_huge.page_size,
349*e65e175bSOded Gabbay 						prop->pmmu_huge.start_addr,
350*e65e175bSOded Gabbay 						prop->pmmu_huge.end_addr);
351*e65e175bSOded Gabbay 	if (is_dram_addr) {
352*e65e175bSOded Gabbay 		*mmu_prop = &prop->dmmu;
353*e65e175bSOded Gabbay 		*is_huge = true;
354*e65e175bSOded Gabbay 		hops->range_type = HL_VA_RANGE_TYPE_DRAM;
355*e65e175bSOded Gabbay 	} else if (is_pmmu_addr) {
356*e65e175bSOded Gabbay 		*mmu_prop = &prop->pmmu;
357*e65e175bSOded Gabbay 		*is_huge = false;
358*e65e175bSOded Gabbay 		hops->range_type = HL_VA_RANGE_TYPE_HOST;
359*e65e175bSOded Gabbay 	} else if (is_pmmu_h_addr) {
360*e65e175bSOded Gabbay 		*mmu_prop = &prop->pmmu_huge;
361*e65e175bSOded Gabbay 		*is_huge = true;
362*e65e175bSOded Gabbay 		hops->range_type = HL_VA_RANGE_TYPE_HOST_HUGE;
363*e65e175bSOded Gabbay 	} else {
364*e65e175bSOded Gabbay 		return -EINVAL;
365*e65e175bSOded Gabbay 	}
366*e65e175bSOded Gabbay 
367*e65e175bSOded Gabbay 	return 0;
368*e65e175bSOded Gabbay }
369*e65e175bSOded Gabbay 
hl_mmu_v2_hr_get_tlb_info(struct hl_ctx * ctx,u64 virt_addr,struct hl_mmu_hop_info * hops)370*e65e175bSOded Gabbay static int hl_mmu_v2_hr_get_tlb_info(struct hl_ctx *ctx, u64 virt_addr,
371*e65e175bSOded Gabbay 					struct hl_mmu_hop_info *hops)
372*e65e175bSOded Gabbay {
373*e65e175bSOded Gabbay 	return hl_mmu_hr_get_tlb_info(ctx, virt_addr, hops,
374*e65e175bSOded Gabbay 					&ctx->hdev->mmu_func[MMU_HR_PGT].hr_funcs);
375*e65e175bSOded Gabbay }
376*e65e175bSOded Gabbay 
377*e65e175bSOded Gabbay /*
378*e65e175bSOded Gabbay  * hl_mmu_v2_prepare - prepare mmu_if for working with mmu v2
379*e65e175bSOded Gabbay  *
380*e65e175bSOded Gabbay  * @hdev: pointer to the device structure
381*e65e175bSOded Gabbay  * @mmu_if: pointer to the mmu interface structure
382*e65e175bSOded Gabbay  */
hl_mmu_v2_hr_set_funcs(struct hl_device * hdev,struct hl_mmu_funcs * mmu)383*e65e175bSOded Gabbay void hl_mmu_v2_hr_set_funcs(struct hl_device *hdev, struct hl_mmu_funcs *mmu)
384*e65e175bSOded Gabbay {
385*e65e175bSOded Gabbay 	mmu->init = hl_mmu_v2_hr_init;
386*e65e175bSOded Gabbay 	mmu->fini = hl_mmu_v2_hr_fini;
387*e65e175bSOded Gabbay 	mmu->ctx_init = hl_mmu_v2_hr_ctx_init;
388*e65e175bSOded Gabbay 	mmu->ctx_fini = hl_mmu_v2_hr_ctx_fini;
389*e65e175bSOded Gabbay 	mmu->map = _hl_mmu_v2_hr_map;
390*e65e175bSOded Gabbay 	mmu->unmap = _hl_mmu_v2_hr_unmap;
391*e65e175bSOded Gabbay 	mmu->flush = hl_mmu_hr_flush;
392*e65e175bSOded Gabbay 	mmu->swap_out = hl_mmu_v2_hr_swap_out;
393*e65e175bSOded Gabbay 	mmu->swap_in = hl_mmu_v2_hr_swap_in;
394*e65e175bSOded Gabbay 	mmu->get_tlb_info = hl_mmu_v2_hr_get_tlb_info;
395*e65e175bSOded Gabbay 	mmu->hr_funcs.get_hop0_pgt_info = hl_mmu_v2_hr_get_hop0_pgt_info;
396*e65e175bSOded Gabbay 	mmu->hr_funcs.get_pgt_info = hl_mmu_v2_hr_get_pgt_info;
397*e65e175bSOded Gabbay 	mmu->hr_funcs.add_pgt_info = hl_mmu_v2_hr_add_pgt_info;
398*e65e175bSOded Gabbay 	mmu->hr_funcs.get_tlb_mapping_params = hl_mmu_v2_hr_get_tlb_mapping_params;
399*e65e175bSOded Gabbay }
400