Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 | // SPDX-License-Identifier: GPL-2.0-only OR MIT /* Copyright (c) 2023 Imagination Technologies Ltd. */ #include "pvr_device.h" #include "pvr_gem.h" #include "pvr_vm.h" #include <drm/drm_gem.h> #include <drm/drm_prime.h> #include <linux/compiler.h> #include <linux/compiler_attributes.h> #include <linux/dma-buf.h> #include <linux/dma-direction.h> #include <linux/dma-mapping.h> #include <linux/err.h> #include <linux/gfp.h> #include <linux/iosys-map.h> #include <linux/log2.h> #include <linux/mutex.h> #include <linux/pagemap.h> #include <linux/refcount.h> #include <linux/scatterlist.h> static void pvr_gem_object_free(struct drm_gem_object *obj) { drm_gem_shmem_object_free(obj); } static int pvr_gem_mmap(struct drm_gem_object *gem_obj, struct vm_area_struct *vma) { struct pvr_gem_object *pvr_obj = gem_to_pvr_gem(gem_obj); struct drm_gem_shmem_object *shmem_obj = shmem_gem_from_pvr_gem(pvr_obj); if (!(pvr_obj->flags & DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS)) return -EINVAL; return drm_gem_shmem_mmap(shmem_obj, vma); } static const struct drm_gem_object_funcs pvr_gem_object_funcs = { .free = pvr_gem_object_free, .print_info = drm_gem_shmem_object_print_info, .pin = drm_gem_shmem_object_pin, .unpin = drm_gem_shmem_object_unpin, .get_sg_table = drm_gem_shmem_object_get_sg_table, .vmap = drm_gem_shmem_object_vmap, .vunmap = drm_gem_shmem_object_vunmap, .mmap = pvr_gem_mmap, .vm_ops = &drm_gem_shmem_vm_ops, }; /** * pvr_gem_object_flags_validate() - Verify that a collection of PowerVR GEM * mapping and/or creation flags form a valid combination. * @flags: PowerVR GEM mapping/creation flags to validate. * * This function explicitly allows kernel-only flags. All ioctl entrypoints * should do their own validation as well as relying on this function. * * Return: * * %true if @flags contains valid mapping and/or creation flags, or * * %false otherwise. */ static bool pvr_gem_object_flags_validate(u64 flags) { static const u64 invalid_combinations[] = { /* * Memory flagged as PM/FW-protected cannot be mapped to * userspace. To make this explicit, we require that the two * flags allowing each of these respective features are never * specified together. */ (DRM_PVR_BO_PM_FW_PROTECT | DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS), }; int i; /* * Check for bits set in undefined regions. Reserved regions refer to * options that can only be set by the kernel. These are explicitly * allowed in most cases, and must be checked specifically in IOCTL * callback code. */ if ((flags & PVR_BO_UNDEFINED_MASK) != 0) return false; /* * Check for all combinations of flags marked as invalid in the array * above. */ for (i = 0; i < ARRAY_SIZE(invalid_combinations); ++i) { u64 combo = invalid_combinations[i]; if ((flags & combo) == combo) return false; } return true; } /** * pvr_gem_object_into_handle() - Convert a reference to an object into a * userspace-accessible handle. * @pvr_obj: [IN] Target PowerVR-specific object. * @pvr_file: [IN] File to associate the handle with. * @handle: [OUT] Pointer to store the created handle in. Remains unmodified if * an error is encountered. * * If an error is encountered, ownership of @pvr_obj will not have been * transferred. If this function succeeds, however, further use of @pvr_obj is * considered undefined behaviour unless another reference to it is explicitly * held. * * Return: * * 0 on success, or * * Any error encountered while attempting to allocate a handle on @pvr_file. */ int pvr_gem_object_into_handle(struct pvr_gem_object *pvr_obj, struct pvr_file *pvr_file, u32 *handle) { struct drm_gem_object *gem_obj = gem_from_pvr_gem(pvr_obj); struct drm_file *file = from_pvr_file(pvr_file); u32 new_handle; int err; err = drm_gem_handle_create(file, gem_obj, &new_handle); if (err) return err; /* * Release our reference to @pvr_obj, effectively transferring * ownership to the handle. */ pvr_gem_object_put(pvr_obj); /* * Do not store the new handle in @handle until no more errors can * occur. */ *handle = new_handle; return 0; } /** * pvr_gem_object_from_handle() - Obtain a reference to an object from a * userspace handle. * @pvr_file: PowerVR-specific file to which @handle is associated. * @handle: Userspace handle referencing the target object. * * On return, @handle always maintains its reference to the requested object * (if it had one in the first place). If this function succeeds, the returned * object will hold an additional reference. When the caller is finished with * the returned object, they should call pvr_gem_object_put() on it to release * this reference. * * Return: * * A pointer to the requested PowerVR-specific object on success, or * * %NULL otherwise. */ struct pvr_gem_object * pvr_gem_object_from_handle(struct pvr_file *pvr_file, u32 handle) { struct drm_file *file = from_pvr_file(pvr_file); struct drm_gem_object *gem_obj; gem_obj = drm_gem_object_lookup(file, handle); if (!gem_obj) return NULL; return gem_to_pvr_gem(gem_obj); } /** * pvr_gem_object_vmap() - Map a PowerVR GEM object into CPU virtual address * space. * @pvr_obj: Target PowerVR GEM object. * * Once the caller is finished with the CPU mapping, they must call * pvr_gem_object_vunmap() on @pvr_obj. * * If @pvr_obj is CPU-cached, dma_sync_sgtable_for_cpu() is called to make * sure the CPU mapping is consistent. * * Return: * * A pointer to the CPU mapping on success, * * -%ENOMEM if the mapping fails, or * * Any error encountered while attempting to acquire a reference to the * backing pages for @pvr_obj. */ void * pvr_gem_object_vmap(struct pvr_gem_object *pvr_obj) { struct drm_gem_shmem_object *shmem_obj = shmem_gem_from_pvr_gem(pvr_obj); struct drm_gem_object *obj = gem_from_pvr_gem(pvr_obj); struct iosys_map map; int err; dma_resv_lock(obj->resv, NULL); err = drm_gem_shmem_vmap(shmem_obj, &map); if (err) goto err_unlock; if (pvr_obj->flags & PVR_BO_CPU_CACHED) { struct device *dev = shmem_obj->base.dev->dev; /* If shmem_obj->sgt is NULL, that means the buffer hasn't been mapped * in GPU space yet. */ if (shmem_obj->sgt) dma_sync_sgtable_for_cpu(dev, shmem_obj->sgt, DMA_BIDIRECTIONAL); } dma_resv_unlock(obj->resv); return map.vaddr; err_unlock: dma_resv_unlock(obj->resv); return ERR_PTR(err); } /** * pvr_gem_object_vunmap() - Unmap a PowerVR memory object from CPU virtual * address space. * @pvr_obj: Target PowerVR GEM object. * * If @pvr_obj is CPU-cached, dma_sync_sgtable_for_device() is called to make * sure the GPU mapping is consistent. */ void pvr_gem_object_vunmap(struct pvr_gem_object *pvr_obj) { struct drm_gem_shmem_object *shmem_obj = shmem_gem_from_pvr_gem(pvr_obj); struct iosys_map map = IOSYS_MAP_INIT_VADDR(shmem_obj->vaddr); struct drm_gem_object *obj = gem_from_pvr_gem(pvr_obj); if (WARN_ON(!map.vaddr)) return; dma_resv_lock(obj->resv, NULL); if (pvr_obj->flags & PVR_BO_CPU_CACHED) { struct device *dev = shmem_obj->base.dev->dev; /* If shmem_obj->sgt is NULL, that means the buffer hasn't been mapped * in GPU space yet. */ if (shmem_obj->sgt) dma_sync_sgtable_for_device(dev, shmem_obj->sgt, DMA_BIDIRECTIONAL); } drm_gem_shmem_vunmap(shmem_obj, &map); dma_resv_unlock(obj->resv); } /** * pvr_gem_object_zero() - Zeroes the physical memory behind an object. * @pvr_obj: Target PowerVR GEM object. * * Return: * * 0 on success, or * * Any error encountered while attempting to map @pvr_obj to the CPU (see * pvr_gem_object_vmap()). */ static int pvr_gem_object_zero(struct pvr_gem_object *pvr_obj) { void *cpu_ptr; cpu_ptr = pvr_gem_object_vmap(pvr_obj); if (IS_ERR(cpu_ptr)) return PTR_ERR(cpu_ptr); memset(cpu_ptr, 0, pvr_gem_object_size(pvr_obj)); /* Make sure the zero-ing is done before vumap-ing the object. */ wmb(); pvr_gem_object_vunmap(pvr_obj); return 0; } /** * pvr_gem_create_object() - Allocate and pre-initializes a pvr_gem_object * @drm_dev: DRM device creating this object. * @size: Size of the object to allocate in bytes. * * Return: * * The new pre-initialized GEM object on success, * * -ENOMEM if the allocation failed. */ struct drm_gem_object *pvr_gem_create_object(struct drm_device *drm_dev, size_t size) { struct drm_gem_object *gem_obj; struct pvr_gem_object *pvr_obj; pvr_obj = kzalloc(sizeof(*pvr_obj), GFP_KERNEL); if (!pvr_obj) return ERR_PTR(-ENOMEM); gem_obj = gem_from_pvr_gem(pvr_obj); gem_obj->funcs = &pvr_gem_object_funcs; return gem_obj; } /** * pvr_gem_object_create() - Creates a PowerVR-specific buffer object. * @pvr_dev: Target PowerVR device. * @size: Size of the object to allocate in bytes. Must be greater than zero. * Any value which is not an exact multiple of the system page size will be * rounded up to satisfy this condition. * @flags: Options which affect both this operation and future mapping * operations performed on the returned object. Must be a combination of * DRM_PVR_BO_* and/or PVR_BO_* flags. * * The created object may be larger than @size, but can never be smaller. To * get the exact size, call pvr_gem_object_size() on the returned pointer. * * Return: * * The newly-minted PowerVR-specific buffer object on success, * * -%EINVAL if @size is zero or @flags is not valid, * * -%ENOMEM if sufficient physical memory cannot be allocated, or * * Any other error returned by drm_gem_create_mmap_offset(). */ struct pvr_gem_object * pvr_gem_object_create(struct pvr_device *pvr_dev, size_t size, u64 flags) { struct drm_gem_shmem_object *shmem_obj; struct pvr_gem_object *pvr_obj; struct sg_table *sgt; int err; /* Verify @size and @flags before continuing. */ if (size == 0 || !pvr_gem_object_flags_validate(flags)) return ERR_PTR(-EINVAL); shmem_obj = drm_gem_shmem_create(from_pvr_device(pvr_dev), size); if (IS_ERR(shmem_obj)) return ERR_CAST(shmem_obj); shmem_obj->pages_mark_dirty_on_put = true; shmem_obj->map_wc = !(flags & PVR_BO_CPU_CACHED); pvr_obj = shmem_gem_to_pvr_gem(shmem_obj); pvr_obj->flags = flags; sgt = drm_gem_shmem_get_pages_sgt(shmem_obj); if (IS_ERR(sgt)) { err = PTR_ERR(sgt); goto err_shmem_object_free; } dma_sync_sgtable_for_device(shmem_obj->base.dev->dev, sgt, DMA_BIDIRECTIONAL); /* * Do this last because pvr_gem_object_zero() requires a fully * configured instance of struct pvr_gem_object. */ pvr_gem_object_zero(pvr_obj); return pvr_obj; err_shmem_object_free: drm_gem_shmem_free(shmem_obj); return ERR_PTR(err); } /** * pvr_gem_get_dma_addr() - Get DMA address for given offset in object * @pvr_obj: Pointer to object to lookup address in. * @offset: Offset within object to lookup address at. * @dma_addr_out: Pointer to location to store DMA address. * * Returns: * * 0 on success, or * * -%EINVAL if object is not currently backed, or if @offset is out of valid * range for this object. */ int pvr_gem_get_dma_addr(struct pvr_gem_object *pvr_obj, u32 offset, dma_addr_t *dma_addr_out) { struct drm_gem_shmem_object *shmem_obj = shmem_gem_from_pvr_gem(pvr_obj); u32 accumulated_offset = 0; struct scatterlist *sgl; unsigned int sgt_idx; WARN_ON(!shmem_obj->sgt); for_each_sgtable_dma_sg(shmem_obj->sgt, sgl, sgt_idx) { u32 new_offset = accumulated_offset + sg_dma_len(sgl); if (offset >= accumulated_offset && offset < new_offset) { *dma_addr_out = sg_dma_address(sgl) + (offset - accumulated_offset); return 0; } accumulated_offset = new_offset; } return -EINVAL; } |