1caab277bSThomas Gleixner // SPDX-License-Identifier: GPL-2.0-only 2667ce33eSRob Clark /* 3667ce33eSRob Clark * Copyright (C) 2016 Red Hat 4667ce33eSRob Clark * Author: Rob Clark <robdclark@gmail.com> 5667ce33eSRob Clark */ 6667ce33eSRob Clark 7667ce33eSRob Clark #include "msm_drv.h" 8667ce33eSRob Clark #include "msm_gem.h" 9667ce33eSRob Clark #include "msm_mmu.h" 10667ce33eSRob Clark 11ee546cd3SJordan Crouse static void 12ee546cd3SJordan Crouse msm_gem_address_space_destroy(struct kref *kref) 13ee546cd3SJordan Crouse { 14ee546cd3SJordan Crouse struct msm_gem_address_space *aspace = container_of(kref, 15ee546cd3SJordan Crouse struct msm_gem_address_space, kref); 16ee546cd3SJordan Crouse 17ee546cd3SJordan Crouse drm_mm_takedown(&aspace->mm); 18ee546cd3SJordan Crouse if (aspace->mmu) 19ee546cd3SJordan Crouse aspace->mmu->funcs->destroy(aspace->mmu); 20ee546cd3SJordan Crouse kfree(aspace); 21ee546cd3SJordan Crouse } 22ee546cd3SJordan Crouse 23ee546cd3SJordan Crouse 24ee546cd3SJordan Crouse void msm_gem_address_space_put(struct msm_gem_address_space *aspace) 25ee546cd3SJordan Crouse { 26ee546cd3SJordan Crouse if (aspace) 27ee546cd3SJordan Crouse kref_put(&aspace->kref, msm_gem_address_space_destroy); 28ee546cd3SJordan Crouse } 29ee546cd3SJordan Crouse 307ad0e8cfSJordan Crouse /* Actually unmap memory for the vma */ 317ad0e8cfSJordan Crouse void msm_gem_purge_vma(struct msm_gem_address_space *aspace, 3270dc51b4SJordan Crouse struct msm_gem_vma *vma) 33667ce33eSRob Clark { 347ad0e8cfSJordan Crouse unsigned size = vma->node.size << PAGE_SHIFT; 357ad0e8cfSJordan Crouse 367ad0e8cfSJordan Crouse /* Print a message if we try to purge a vma in use */ 377ad0e8cfSJordan Crouse if (WARN_ON(vma->inuse > 0)) 38667ce33eSRob Clark return; 39667ce33eSRob Clark 407ad0e8cfSJordan Crouse /* Don't do anything if the memory isn't mapped */ 417ad0e8cfSJordan Crouse if (!vma->mapped) 427ad0e8cfSJordan Crouse return; 437ad0e8cfSJordan Crouse 447ad0e8cfSJordan Crouse if (aspace->mmu) 4570dc51b4SJordan Crouse aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size); 467ad0e8cfSJordan Crouse 477ad0e8cfSJordan Crouse vma->mapped = false; 48667ce33eSRob Clark } 49667ce33eSRob Clark 507ad0e8cfSJordan Crouse /* Remove reference counts for the mapping */ 517ad0e8cfSJordan Crouse void msm_gem_unmap_vma(struct msm_gem_address_space *aspace, 527ad0e8cfSJordan Crouse struct msm_gem_vma *vma) 537ad0e8cfSJordan Crouse { 547ad0e8cfSJordan Crouse if (!WARN_ON(!vma->iova)) 557ad0e8cfSJordan Crouse vma->inuse--; 56667ce33eSRob Clark } 57667ce33eSRob Clark 58667ce33eSRob Clark int 59667ce33eSRob Clark msm_gem_map_vma(struct msm_gem_address_space *aspace, 60bbc2cd07SRob Clark struct msm_gem_vma *vma, int prot, 61bbc2cd07SRob Clark struct sg_table *sgt, int npages) 62667ce33eSRob Clark { 63c0ee9794SJordan Crouse unsigned size = npages << PAGE_SHIFT; 64c0ee9794SJordan Crouse int ret = 0; 65667ce33eSRob Clark 66c0ee9794SJordan Crouse if (WARN_ON(!vma->iova)) 67c0ee9794SJordan Crouse return -EINVAL; 68c0ee9794SJordan Crouse 697ad0e8cfSJordan Crouse /* Increase the usage counter */ 707ad0e8cfSJordan Crouse vma->inuse++; 717ad0e8cfSJordan Crouse 72c0ee9794SJordan Crouse if (vma->mapped) 73667ce33eSRob Clark return 0; 74c0ee9794SJordan Crouse 75c0ee9794SJordan Crouse vma->mapped = true; 76c0ee9794SJordan Crouse 777603df38SLuca Weiss if (aspace && aspace->mmu) 78c0ee9794SJordan Crouse ret = aspace->mmu->funcs->map(aspace->mmu, vma->iova, sgt, 79bbc2cd07SRob Clark size, prot); 80c0ee9794SJordan Crouse 81c0ee9794SJordan Crouse if (ret) 82c0ee9794SJordan Crouse vma->mapped = false; 83c0ee9794SJordan Crouse 84c0ee9794SJordan Crouse return ret; 850e08270aSSushmita Susheelendra } 86667ce33eSRob Clark 877ad0e8cfSJordan Crouse /* Close an iova. Warn if it is still in use */ 887ad0e8cfSJordan Crouse void msm_gem_close_vma(struct msm_gem_address_space *aspace, 897ad0e8cfSJordan Crouse struct msm_gem_vma *vma) 907ad0e8cfSJordan Crouse { 917ad0e8cfSJordan Crouse if (WARN_ON(vma->inuse > 0 || vma->mapped)) 927ad0e8cfSJordan Crouse return; 937ad0e8cfSJordan Crouse 947ad0e8cfSJordan Crouse spin_lock(&aspace->lock); 957ad0e8cfSJordan Crouse if (vma->iova) 967ad0e8cfSJordan Crouse drm_mm_remove_node(&vma->node); 977ad0e8cfSJordan Crouse spin_unlock(&aspace->lock); 987ad0e8cfSJordan Crouse 997ad0e8cfSJordan Crouse vma->iova = 0; 1007ad0e8cfSJordan Crouse 1017ad0e8cfSJordan Crouse msm_gem_address_space_put(aspace); 1027ad0e8cfSJordan Crouse } 1037ad0e8cfSJordan Crouse 104c0ee9794SJordan Crouse /* Initialize a new vma and allocate an iova for it */ 105c0ee9794SJordan Crouse int msm_gem_init_vma(struct msm_gem_address_space *aspace, 106d3b8877eSJonathan Marek struct msm_gem_vma *vma, int npages, 107d3b8877eSJonathan Marek u64 range_start, u64 range_end) 108c0ee9794SJordan Crouse { 109c0ee9794SJordan Crouse int ret; 110c0ee9794SJordan Crouse 111c0ee9794SJordan Crouse if (WARN_ON(vma->iova)) 112c0ee9794SJordan Crouse return -EBUSY; 113c0ee9794SJordan Crouse 114c0ee9794SJordan Crouse spin_lock(&aspace->lock); 115d3b8877eSJonathan Marek ret = drm_mm_insert_node_in_range(&aspace->mm, &vma->node, npages, 0, 116d3b8877eSJonathan Marek 0, range_start, range_end, 0); 1170e08270aSSushmita Susheelendra spin_unlock(&aspace->lock); 1180e08270aSSushmita Susheelendra 119667ce33eSRob Clark if (ret) 120667ce33eSRob Clark return ret; 121667ce33eSRob Clark 122667ce33eSRob Clark vma->iova = vma->node.start << PAGE_SHIFT; 123c0ee9794SJordan Crouse vma->mapped = false; 124667ce33eSRob Clark 125ee546cd3SJordan Crouse kref_get(&aspace->kref); 126667ce33eSRob Clark 127c0ee9794SJordan Crouse return 0; 128667ce33eSRob Clark } 129667ce33eSRob Clark 1307ad0e8cfSJordan Crouse 131667ce33eSRob Clark struct msm_gem_address_space * 132667ce33eSRob Clark msm_gem_address_space_create(struct device *dev, struct iommu_domain *domain, 133667ce33eSRob Clark const char *name) 134667ce33eSRob Clark { 135667ce33eSRob Clark struct msm_gem_address_space *aspace; 136edf5ceacSJordan Crouse u64 size = domain->geometry.aperture_end - 137edf5ceacSJordan Crouse domain->geometry.aperture_start; 138667ce33eSRob Clark 139667ce33eSRob Clark aspace = kzalloc(sizeof(*aspace), GFP_KERNEL); 140667ce33eSRob Clark if (!aspace) 141667ce33eSRob Clark return ERR_PTR(-ENOMEM); 142667ce33eSRob Clark 1430e08270aSSushmita Susheelendra spin_lock_init(&aspace->lock); 144667ce33eSRob Clark aspace->name = name; 145667ce33eSRob Clark aspace->mmu = msm_iommu_new(dev, domain); 146667ce33eSRob Clark 147667ce33eSRob Clark drm_mm_init(&aspace->mm, (domain->geometry.aperture_start >> PAGE_SHIFT), 148edf5ceacSJordan Crouse size >> PAGE_SHIFT); 149667ce33eSRob Clark 150ee546cd3SJordan Crouse kref_init(&aspace->kref); 151ee546cd3SJordan Crouse 152667ce33eSRob Clark return aspace; 153667ce33eSRob Clark } 154c2052a4eSJonathan Marek 155c2052a4eSJonathan Marek struct msm_gem_address_space * 156c2052a4eSJonathan Marek msm_gem_address_space_create_a2xx(struct device *dev, struct msm_gpu *gpu, 157c2052a4eSJonathan Marek const char *name, uint64_t va_start, uint64_t va_end) 158c2052a4eSJonathan Marek { 159c2052a4eSJonathan Marek struct msm_gem_address_space *aspace; 160c2052a4eSJonathan Marek u64 size = va_end - va_start; 161c2052a4eSJonathan Marek 162c2052a4eSJonathan Marek aspace = kzalloc(sizeof(*aspace), GFP_KERNEL); 163c2052a4eSJonathan Marek if (!aspace) 164c2052a4eSJonathan Marek return ERR_PTR(-ENOMEM); 165c2052a4eSJonathan Marek 166c2052a4eSJonathan Marek spin_lock_init(&aspace->lock); 167c2052a4eSJonathan Marek aspace->name = name; 168c2052a4eSJonathan Marek aspace->mmu = msm_gpummu_new(dev, gpu); 169c2052a4eSJonathan Marek 170c2052a4eSJonathan Marek drm_mm_init(&aspace->mm, (va_start >> PAGE_SHIFT), 171c2052a4eSJonathan Marek size >> PAGE_SHIFT); 172c2052a4eSJonathan Marek 173c2052a4eSJonathan Marek kref_init(&aspace->kref); 174c2052a4eSJonathan Marek 175c2052a4eSJonathan Marek return aspace; 176c2052a4eSJonathan Marek } 177