mirror of
				https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
				synced 2025-10-31 20:42:39 +00:00 
			
		
		
		
	 135cba0dc3
			
		
	
	
		135cba0dc3
		
	
	
	
	
		
			
			Use Ben's new range manager hooks to implement a manager for GMRs that manages ids rather than ranges. This means we can use the standard TTM code for binding, unbinding and eviction. Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com> Signed-off-by: Dave Airlie <airlied@redhat.com>
		
			
				
	
	
		
			329 lines
		
	
	
		
			8.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			329 lines
		
	
	
		
			8.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /**************************************************************************
 | |
|  *
 | |
|  * Copyright © 2009 VMware, Inc., Palo Alto, CA., USA
 | |
|  * All Rights Reserved.
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a
 | |
|  * copy of this software and associated documentation files (the
 | |
|  * "Software"), to deal in the Software without restriction, including
 | |
|  * without limitation the rights to use, copy, modify, merge, publish,
 | |
|  * distribute, sub license, and/or sell copies of the Software, and to
 | |
|  * permit persons to whom the Software is furnished to do so, subject to
 | |
|  * the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice (including the
 | |
|  * next paragraph) shall be included in all copies or substantial portions
 | |
|  * of the Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
 | |
|  * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
 | |
|  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
 | |
|  * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
 | |
|  * USE OR OTHER DEALINGS IN THE SOFTWARE.
 | |
|  *
 | |
|  **************************************************************************/
 | |
| 
 | |
| #include "vmwgfx_drv.h"
 | |
| #include "ttm/ttm_bo_driver.h"
 | |
| #include "ttm/ttm_placement.h"
 | |
| 
 | |
| static uint32_t vram_placement_flags = TTM_PL_FLAG_VRAM |
 | |
| 	TTM_PL_FLAG_CACHED;
 | |
| 
 | |
| static uint32_t vram_ne_placement_flags = TTM_PL_FLAG_VRAM |
 | |
| 	TTM_PL_FLAG_CACHED |
 | |
| 	TTM_PL_FLAG_NO_EVICT;
 | |
| 
 | |
| static uint32_t sys_placement_flags = TTM_PL_FLAG_SYSTEM |
 | |
| 	TTM_PL_FLAG_CACHED;
 | |
| 
 | |
| static uint32_t gmr_placement_flags = VMW_PL_FLAG_GMR |
 | |
| 	TTM_PL_FLAG_CACHED;
 | |
| 
 | |
| struct ttm_placement vmw_vram_placement = {
 | |
| 	.fpfn = 0,
 | |
| 	.lpfn = 0,
 | |
| 	.num_placement = 1,
 | |
| 	.placement = &vram_placement_flags,
 | |
| 	.num_busy_placement = 1,
 | |
| 	.busy_placement = &vram_placement_flags
 | |
| };
 | |
| 
 | |
| static uint32_t vram_gmr_placement_flags[] = {
 | |
| 	TTM_PL_FLAG_VRAM | TTM_PL_FLAG_CACHED,
 | |
| 	VMW_PL_FLAG_GMR | TTM_PL_FLAG_CACHED
 | |
| };
 | |
| 
 | |
| struct ttm_placement vmw_vram_gmr_placement = {
 | |
| 	.fpfn = 0,
 | |
| 	.lpfn = 0,
 | |
| 	.num_placement = 2,
 | |
| 	.placement = vram_gmr_placement_flags,
 | |
| 	.num_busy_placement = 1,
 | |
| 	.busy_placement = &gmr_placement_flags
 | |
| };
 | |
| 
 | |
| struct ttm_placement vmw_vram_sys_placement = {
 | |
| 	.fpfn = 0,
 | |
| 	.lpfn = 0,
 | |
| 	.num_placement = 1,
 | |
| 	.placement = &vram_placement_flags,
 | |
| 	.num_busy_placement = 1,
 | |
| 	.busy_placement = &sys_placement_flags
 | |
| };
 | |
| 
 | |
| struct ttm_placement vmw_vram_ne_placement = {
 | |
| 	.fpfn = 0,
 | |
| 	.lpfn = 0,
 | |
| 	.num_placement = 1,
 | |
| 	.placement = &vram_ne_placement_flags,
 | |
| 	.num_busy_placement = 1,
 | |
| 	.busy_placement = &vram_ne_placement_flags
 | |
| };
 | |
| 
 | |
| struct ttm_placement vmw_sys_placement = {
 | |
| 	.fpfn = 0,
 | |
| 	.lpfn = 0,
 | |
| 	.num_placement = 1,
 | |
| 	.placement = &sys_placement_flags,
 | |
| 	.num_busy_placement = 1,
 | |
| 	.busy_placement = &sys_placement_flags
 | |
| };
 | |
| 
 | |
| struct vmw_ttm_backend {
 | |
| 	struct ttm_backend backend;
 | |
| 	struct page **pages;
 | |
| 	unsigned long num_pages;
 | |
| 	struct vmw_private *dev_priv;
 | |
| 	int gmr_id;
 | |
| };
 | |
| 
 | |
| static int vmw_ttm_populate(struct ttm_backend *backend,
 | |
| 			    unsigned long num_pages, struct page **pages,
 | |
| 			    struct page *dummy_read_page)
 | |
| {
 | |
| 	struct vmw_ttm_backend *vmw_be =
 | |
| 	    container_of(backend, struct vmw_ttm_backend, backend);
 | |
| 
 | |
| 	vmw_be->pages = pages;
 | |
| 	vmw_be->num_pages = num_pages;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int vmw_ttm_bind(struct ttm_backend *backend, struct ttm_mem_reg *bo_mem)
 | |
| {
 | |
| 	struct vmw_ttm_backend *vmw_be =
 | |
| 	    container_of(backend, struct vmw_ttm_backend, backend);
 | |
| 
 | |
| 	vmw_be->gmr_id = bo_mem->start;
 | |
| 
 | |
| 	return vmw_gmr_bind(vmw_be->dev_priv, vmw_be->pages,
 | |
| 			    vmw_be->num_pages, vmw_be->gmr_id);
 | |
| }
 | |
| 
 | |
| static int vmw_ttm_unbind(struct ttm_backend *backend)
 | |
| {
 | |
| 	struct vmw_ttm_backend *vmw_be =
 | |
| 	    container_of(backend, struct vmw_ttm_backend, backend);
 | |
| 
 | |
| 	vmw_gmr_unbind(vmw_be->dev_priv, vmw_be->gmr_id);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void vmw_ttm_clear(struct ttm_backend *backend)
 | |
| {
 | |
| 	struct vmw_ttm_backend *vmw_be =
 | |
| 		container_of(backend, struct vmw_ttm_backend, backend);
 | |
| 
 | |
| 	vmw_be->pages = NULL;
 | |
| 	vmw_be->num_pages = 0;
 | |
| }
 | |
| 
 | |
| static void vmw_ttm_destroy(struct ttm_backend *backend)
 | |
| {
 | |
| 	struct vmw_ttm_backend *vmw_be =
 | |
| 	    container_of(backend, struct vmw_ttm_backend, backend);
 | |
| 
 | |
| 	kfree(vmw_be);
 | |
| }
 | |
| 
 | |
| static struct ttm_backend_func vmw_ttm_func = {
 | |
| 	.populate = vmw_ttm_populate,
 | |
| 	.clear = vmw_ttm_clear,
 | |
| 	.bind = vmw_ttm_bind,
 | |
| 	.unbind = vmw_ttm_unbind,
 | |
| 	.destroy = vmw_ttm_destroy,
 | |
| };
 | |
| 
 | |
| struct ttm_backend *vmw_ttm_backend_init(struct ttm_bo_device *bdev)
 | |
| {
 | |
| 	struct vmw_ttm_backend *vmw_be;
 | |
| 
 | |
| 	vmw_be = kmalloc(sizeof(*vmw_be), GFP_KERNEL);
 | |
| 	if (!vmw_be)
 | |
| 		return NULL;
 | |
| 
 | |
| 	vmw_be->backend.func = &vmw_ttm_func;
 | |
| 	vmw_be->dev_priv = container_of(bdev, struct vmw_private, bdev);
 | |
| 
 | |
| 	return &vmw_be->backend;
 | |
| }
 | |
| 
 | |
| int vmw_invalidate_caches(struct ttm_bo_device *bdev, uint32_t flags)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int vmw_init_mem_type(struct ttm_bo_device *bdev, uint32_t type,
 | |
| 		      struct ttm_mem_type_manager *man)
 | |
| {
 | |
| 	switch (type) {
 | |
| 	case TTM_PL_SYSTEM:
 | |
| 		/* System memory */
 | |
| 
 | |
| 		man->flags = TTM_MEMTYPE_FLAG_MAPPABLE;
 | |
| 		man->available_caching = TTM_PL_FLAG_CACHED;
 | |
| 		man->default_caching = TTM_PL_FLAG_CACHED;
 | |
| 		break;
 | |
| 	case TTM_PL_VRAM:
 | |
| 		/* "On-card" video ram */
 | |
| 		man->func = &ttm_bo_manager_func;
 | |
| 		man->gpu_offset = 0;
 | |
| 		man->flags = TTM_MEMTYPE_FLAG_FIXED | TTM_MEMTYPE_FLAG_MAPPABLE;
 | |
| 		man->available_caching = TTM_PL_FLAG_CACHED;
 | |
| 		man->default_caching = TTM_PL_FLAG_CACHED;
 | |
| 		break;
 | |
| 	case VMW_PL_GMR:
 | |
| 		/*
 | |
| 		 * "Guest Memory Regions" is an aperture like feature with
 | |
| 		 *  one slot per bo. There is an upper limit of the number of
 | |
| 		 *  slots as well as the bo size.
 | |
| 		 */
 | |
| 		man->func = &vmw_gmrid_manager_func;
 | |
| 		man->gpu_offset = 0;
 | |
| 		man->flags = TTM_MEMTYPE_FLAG_CMA | TTM_MEMTYPE_FLAG_MAPPABLE;
 | |
| 		man->available_caching = TTM_PL_FLAG_CACHED;
 | |
| 		man->default_caching = TTM_PL_FLAG_CACHED;
 | |
| 		break;
 | |
| 	default:
 | |
| 		DRM_ERROR("Unsupported memory type %u\n", (unsigned)type);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void vmw_evict_flags(struct ttm_buffer_object *bo,
 | |
| 		     struct ttm_placement *placement)
 | |
| {
 | |
| 	*placement = vmw_sys_placement;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * FIXME: Proper access checks on buffers.
 | |
|  */
 | |
| 
 | |
| static int vmw_verify_access(struct ttm_buffer_object *bo, struct file *filp)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int vmw_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem)
 | |
| {
 | |
| 	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
 | |
| 	struct vmw_private *dev_priv = container_of(bdev, struct vmw_private, bdev);
 | |
| 
 | |
| 	mem->bus.addr = NULL;
 | |
| 	mem->bus.is_iomem = false;
 | |
| 	mem->bus.offset = 0;
 | |
| 	mem->bus.size = mem->num_pages << PAGE_SHIFT;
 | |
| 	mem->bus.base = 0;
 | |
| 	if (!(man->flags & TTM_MEMTYPE_FLAG_MAPPABLE))
 | |
| 		return -EINVAL;
 | |
| 	switch (mem->mem_type) {
 | |
| 	case TTM_PL_SYSTEM:
 | |
| 	case VMW_PL_GMR:
 | |
| 		return 0;
 | |
| 	case TTM_PL_VRAM:
 | |
| 		mem->bus.offset = mem->start << PAGE_SHIFT;
 | |
| 		mem->bus.base = dev_priv->vram_start;
 | |
| 		mem->bus.is_iomem = true;
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void vmw_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem)
 | |
| {
 | |
| }
 | |
| 
 | |
| static int vmw_ttm_fault_reserve_notify(struct ttm_buffer_object *bo)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * FIXME: We're using the old vmware polling method to sync.
 | |
|  * Do this with fences instead.
 | |
|  */
 | |
| 
 | |
| static void *vmw_sync_obj_ref(void *sync_obj)
 | |
| {
 | |
| 	return sync_obj;
 | |
| }
 | |
| 
 | |
| static void vmw_sync_obj_unref(void **sync_obj)
 | |
| {
 | |
| 	*sync_obj = NULL;
 | |
| }
 | |
| 
 | |
| static int vmw_sync_obj_flush(void *sync_obj, void *sync_arg)
 | |
| {
 | |
| 	struct vmw_private *dev_priv = (struct vmw_private *)sync_arg;
 | |
| 
 | |
| 	mutex_lock(&dev_priv->hw_mutex);
 | |
| 	vmw_write(dev_priv, SVGA_REG_SYNC, SVGA_SYNC_GENERIC);
 | |
| 	mutex_unlock(&dev_priv->hw_mutex);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static bool vmw_sync_obj_signaled(void *sync_obj, void *sync_arg)
 | |
| {
 | |
| 	struct vmw_private *dev_priv = (struct vmw_private *)sync_arg;
 | |
| 	uint32_t sequence = (unsigned long) sync_obj;
 | |
| 
 | |
| 	return vmw_fence_signaled(dev_priv, sequence);
 | |
| }
 | |
| 
 | |
| static int vmw_sync_obj_wait(void *sync_obj, void *sync_arg,
 | |
| 			     bool lazy, bool interruptible)
 | |
| {
 | |
| 	struct vmw_private *dev_priv = (struct vmw_private *)sync_arg;
 | |
| 	uint32_t sequence = (unsigned long) sync_obj;
 | |
| 
 | |
| 	return vmw_wait_fence(dev_priv, false, sequence, false, 3*HZ);
 | |
| }
 | |
| 
 | |
| struct ttm_bo_driver vmw_bo_driver = {
 | |
| 	.create_ttm_backend_entry = vmw_ttm_backend_init,
 | |
| 	.invalidate_caches = vmw_invalidate_caches,
 | |
| 	.init_mem_type = vmw_init_mem_type,
 | |
| 	.evict_flags = vmw_evict_flags,
 | |
| 	.move = NULL,
 | |
| 	.verify_access = vmw_verify_access,
 | |
| 	.sync_obj_signaled = vmw_sync_obj_signaled,
 | |
| 	.sync_obj_wait = vmw_sync_obj_wait,
 | |
| 	.sync_obj_flush = vmw_sync_obj_flush,
 | |
| 	.sync_obj_unref = vmw_sync_obj_unref,
 | |
| 	.sync_obj_ref = vmw_sync_obj_ref,
 | |
| 	.move_notify = NULL,
 | |
| 	.swap_notify = NULL,
 | |
| 	.fault_reserve_notify = &vmw_ttm_fault_reserve_notify,
 | |
| 	.io_mem_reserve = &vmw_ttm_io_mem_reserve,
 | |
| 	.io_mem_free = &vmw_ttm_io_mem_free,
 | |
| };
 |