/************************************************************************** * * Copyright © 2009 VMware, Inc., Palo Alto, CA., USA * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE * USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #include "vmwgfx_drv.h" #include "ttm/ttm_bo_driver.h" #include "ttm/ttm_placement.h" static uint32_t vram_placement_flags = TTM_PL_FLAG_VRAM | TTM_PL_FLAG_CACHED; static uint32_t vram_ne_placement_flags = TTM_PL_FLAG_VRAM | TTM_PL_FLAG_CACHED | TTM_PL_FLAG_NO_EVICT; static uint32_t sys_placement_flags = TTM_PL_FLAG_SYSTEM | TTM_PL_FLAG_CACHED; static uint32_t gmr_placement_flags = VMW_PL_FLAG_GMR | TTM_PL_FLAG_CACHED; struct ttm_placement vmw_vram_placement = { .fpfn = 0, .lpfn = 0, .num_placement = 1, .placement = &vram_placement_flags, .num_busy_placement = 1, .busy_placement = &vram_placement_flags }; static uint32_t vram_gmr_placement_flags[] = { TTM_PL_FLAG_VRAM | TTM_PL_FLAG_CACHED, VMW_PL_FLAG_GMR | TTM_PL_FLAG_CACHED }; struct ttm_placement vmw_vram_gmr_placement = { .fpfn = 0, .lpfn = 0, .num_placement = 2, .placement = vram_gmr_placement_flags, .num_busy_placement = 1, .busy_placement = &gmr_placement_flags }; struct ttm_placement vmw_vram_sys_placement = { .fpfn = 0, .lpfn = 0, .num_placement = 1, .placement = &vram_placement_flags, .num_busy_placement = 1, .busy_placement = &sys_placement_flags }; struct ttm_placement vmw_vram_ne_placement = { .fpfn = 0, .lpfn = 0, .num_placement = 1, .placement = &vram_ne_placement_flags, .num_busy_placement = 1, .busy_placement = &vram_ne_placement_flags }; struct ttm_placement vmw_sys_placement = { .fpfn = 0, .lpfn = 0, .num_placement = 1, .placement = &sys_placement_flags, .num_busy_placement = 1, .busy_placement = &sys_placement_flags }; struct vmw_ttm_backend { struct ttm_backend backend; struct page **pages; unsigned long num_pages; struct vmw_private *dev_priv; int gmr_id; }; static int vmw_ttm_populate(struct ttm_backend *backend, unsigned long num_pages, struct page **pages, struct page *dummy_read_page) { struct vmw_ttm_backend *vmw_be = container_of(backend, struct vmw_ttm_backend, backend); vmw_be->pages = pages; vmw_be->num_pages = num_pages; return 0; } static int vmw_ttm_bind(struct ttm_backend *backend, struct ttm_mem_reg *bo_mem) { struct vmw_ttm_backend *vmw_be = container_of(backend, struct vmw_ttm_backend, backend); vmw_be->gmr_id = bo_mem->start; return vmw_gmr_bind(vmw_be->dev_priv, vmw_be->pages, vmw_be->num_pages, vmw_be->gmr_id); } static int vmw_ttm_unbind(struct ttm_backend *backend) { struct vmw_ttm_backend *vmw_be = container_of(backend, struct vmw_ttm_backend, backend); vmw_gmr_unbind(vmw_be->dev_priv, vmw_be->gmr_id); return 0; } static void vmw_ttm_clear(struct ttm_backend *backend) { struct vmw_ttm_backend *vmw_be = container_of(backend, struct vmw_ttm_backend, backend); vmw_be->pages = NULL; vmw_be->num_pages = 0; } static void vmw_ttm_destroy(struct ttm_backend *backend) { struct vmw_ttm_backend *vmw_be = container_of(backend, struct vmw_ttm_backend, backend); kfree(vmw_be); } static struct ttm_backend_func vmw_ttm_func = { .populate = vmw_ttm_populate, .clear = vmw_ttm_clear, .bind = vmw_ttm_bind, .unbind = vmw_ttm_unbind, .destroy = vmw_ttm_destroy, }; struct ttm_backend *vmw_ttm_backend_init(struct ttm_bo_device *bdev) { struct vmw_ttm_backend *vmw_be; vmw_be = kmalloc(sizeof(*vmw_be), GFP_KERNEL); if (!vmw_be) return NULL; vmw_be->backend.func = &vmw_ttm_func; vmw_be->dev_priv = container_of(bdev, struct vmw_private, bdev); return &vmw_be->backend; } int vmw_invalidate_caches(struct ttm_bo_device *bdev, uint32_t flags) { return 0; } int vmw_init_mem_type(struct ttm_bo_device *bdev, uint32_t type, struct ttm_mem_type_manager *man) { switch (type) { case TTM_PL_SYSTEM: /* System memory */ man->flags = TTM_MEMTYPE_FLAG_MAPPABLE; man->available_caching = TTM_PL_FLAG_CACHED; man->default_caching = TTM_PL_FLAG_CACHED; break; case TTM_PL_VRAM: /* "On-card" video ram */ man->func = &ttm_bo_manager_func; man->gpu_offset = 0; man->flags = TTM_MEMTYPE_FLAG_FIXED | TTM_MEMTYPE_FLAG_MAPPABLE; man->available_caching = TTM_PL_FLAG_CACHED; man->default_caching = TTM_PL_FLAG_CACHED; break; case VMW_PL_GMR: /* * "Guest Memory Regions" is an aperture like feature with * one slot per bo. There is an upper limit of the number of * slots as well as the bo size. */ man->func = &vmw_gmrid_manager_func; man->gpu_offset = 0; man->flags = TTM_MEMTYPE_FLAG_CMA | TTM_MEMTYPE_FLAG_MAPPABLE; man->available_caching = TTM_PL_FLAG_CACHED; man->default_caching = TTM_PL_FLAG_CACHED; break; default: DRM_ERROR("Unsupported memory type %u\n", (unsigned)type); return -EINVAL; } return 0; } void vmw_evict_flags(struct ttm_buffer_object *bo, struct ttm_placement *placement) { *placement = vmw_sys_placement; } /** * FIXME: Proper access checks on buffers. */ static int vmw_verify_access(struct ttm_buffer_object *bo, struct file *filp) { return 0; } static int vmw_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) { struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; struct vmw_private *dev_priv = container_of(bdev, struct vmw_private, bdev); mem->bus.addr = NULL; mem->bus.is_iomem = false; mem->bus.offset = 0; mem->bus.size = mem->num_pages << PAGE_SHIFT; mem->bus.base = 0; if (!(man->flags & TTM_MEMTYPE_FLAG_MAPPABLE)) return -EINVAL; switch (mem->mem_type) { case TTM_PL_SYSTEM: case VMW_PL_GMR: return 0; case TTM_PL_VRAM: mem->bus.offset = mem->start << PAGE_SHIFT; mem->bus.base = dev_priv->vram_start; mem->bus.is_iomem = true; break; default: return -EINVAL; } return 0; } static void vmw_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) { } static int vmw_ttm_fault_reserve_notify(struct ttm_buffer_object *bo) { return 0; } /** * FIXME: We're using the old vmware polling method to sync. * Do this with fences instead. */ static void *vmw_sync_obj_ref(void *sync_obj) { return sync_obj; } static void vmw_sync_obj_unref(void **sync_obj) { *sync_obj = NULL; } static int vmw_sync_obj_flush(void *sync_obj, void *sync_arg) { struct vmw_private *dev_priv = (struct vmw_private *)sync_arg; mutex_lock(&dev_priv->hw_mutex); vmw_write(dev_priv, SVGA_REG_SYNC, SVGA_SYNC_GENERIC); mutex_unlock(&dev_priv->hw_mutex); return 0; } static bool vmw_sync_obj_signaled(void *sync_obj, void *sync_arg) { struct vmw_private *dev_priv = (struct vmw_private *)sync_arg; uint32_t sequence = (unsigned long) sync_obj; return vmw_fence_signaled(dev_priv, sequence); } static int vmw_sync_obj_wait(void *sync_obj, void *sync_arg, bool lazy, bool interruptible) { struct vmw_private *dev_priv = (struct vmw_private *)sync_arg; uint32_t sequence = (unsigned long) sync_obj; return vmw_wait_fence(dev_priv, false, sequence, false, 3*HZ); } struct ttm_bo_driver vmw_bo_driver = { .create_ttm_backend_entry = vmw_ttm_backend_init, .invalidate_caches = vmw_invalidate_caches, .init_mem_type = vmw_init_mem_type, .evict_flags = vmw_evict_flags, .move = NULL, .verify_access = vmw_verify_access, .sync_obj_signaled = vmw_sync_obj_signaled, .sync_obj_wait = vmw_sync_obj_wait, .sync_obj_flush = vmw_sync_obj_flush, .sync_obj_unref = vmw_sync_obj_unref, .sync_obj_ref = vmw_sync_obj_ref, .move_notify = NULL, .swap_notify = NULL, .fault_reserve_notify = &vmw_ttm_fault_reserve_notify, .io_mem_reserve = &vmw_ttm_io_mem_reserve, .io_mem_free = &vmw_ttm_io_mem_free, };