Merge tag 'drm-misc-next-2019-05-24' of git://anongit.freedesktop.org/drm/drm-misc into drm-next
drm-misc-next for v5.3, try #2: UAPI Changes: - Add HDR source metadata property. - Make drm.h compile on GNU/kFreeBSD by including stdint.h - Clarify how the userspace reviewer has to review new kernel UAPI. - Clarify that for using new UAPI, merging to drm-next or drm-misc-next should be enough. Cross-subsystem Changes: - video/hdmi: Add unpack function for DRM infoframes. - Device tree bindings: * Updating a property for Mali Midgard GPUs * Updating a property for STM32 DSI panel * Adding support for FriendlyELEC HD702E 800x1280 panel * Adding support for Evervision VGG804821 800x480 5.0" WVGA TFT panel * Adding support for the EDT ET035012DM6 3.5" 320x240 QVGA 24-bit RGB TFT. * Adding support for Three Five displays TFC S9700RTWV43TR-01B 800x480 panel with resistive touch found on TI's AM335X-EVM. * Adding support for EDT ETM0430G0DH6 480x272 panel. - Add OSD101T2587-53TS driver with DT bindings. - Add Samsung S6E63M0 panel driver with DT bindings. - Add VXT VL050-8048NT-C01 800x480 panel with DT bindings. - Dma-buf: - Make mmap callback actually optional. - Documentation updates. - Fix debugfs refcount inbalance. - Remove unused sync_dump function. - Fix device tree bindings in drm-misc-next after a botched merge. Core Changes: - Add support for HDR infoframes and related EDID parsing. - Remove prime sg_table caching, now done inside dma-buf. - Add shiny new drm_gem_vram helpers for simple VRAM drivers; with some fixes to the new API on top. - Small fix to job cleanup without timeout handler. - Documentation fixes to drm_fourcc. - Replace lookups of drm_format with struct drm_format_info; remove functions that become obsolete by this conversion. - Remove double include in bridge/panel.c and some drivers. - Remove drmP.h include from drm/edid and drm/dp. - Fix null pointer deref in drm_fb_helper_hotplug_event(). - Remove most members from drm_fb_helper_crtc, only mode_set is kept. - Remove race of fb helpers with userspace; only restore mode when userspace is not master. - Move legacy setup from drm_file.c to drm_legacy_misc.c - Rework scheduler job destruction. - drm/bus was removed, remove from TODO. - Add __drm_atomic_helper_crtc_reset() to subclass crtc_state, and convert some drivers to use it (conversion is not complete yet). - Bump vblank timeout wait to 100 ms for atomic. - Docbook fix for drm_hdmi_infoframe_set_hdr_metadata. Driver Changes: - sun4i: Use DRM_GEM_CMA_VMAP_DRIVER_OPS instead of definining manually. - v3d: Small cleanups, adding support for compute shaders, reservation/synchronization fixes and job management refactoring, fixes MMU and debugfs. - lima: Fix null pointer in irq handler on startup, set default timeout for scheduled jobs. - stm/ltdc: Assorted fixes and adding FB modifier support. - amdgpu: Avoid hw reset if guilty job was already signaled. - virtio: Add seqno to fences, add trace events, use correct flags for fence allocation. - Convert AST, bochs, mgag200, vboxvideo, hisilicon to the new drm_gem_vram API. - sun6i_mipi_dsi: Support DSI GENERIC_SHORT_WRITE_2 transfers. - bochs: Small fix to use PTR_RET_OR_ZERO and driver unload. - gma500: header fixes - cirrus: Remove unused files. - mediatek: Fix compiler warning after merging the HDR series. - vc4: Rework binner bo handling. Signed-off-by: Dave Airlie <airlied@redhat.com> From: Maarten Lankhorst <maarten.lankhorst@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/052875a5-27ba-3832-60c2-193d950afdff@linux.intel.com
This commit is contained in:
@@ -3,7 +3,7 @@ config DRM_HISI_HIBMC
|
||||
tristate "DRM Support for Hisilicon Hibmc"
|
||||
depends on DRM && PCI && MMU
|
||||
select DRM_KMS_HELPER
|
||||
select DRM_TTM
|
||||
select DRM_VRAM_HELPER
|
||||
|
||||
help
|
||||
Choose this option if you have a Hisilicon Hibmc soc chipset.
|
||||
|
@@ -96,27 +96,26 @@ static void hibmc_plane_atomic_update(struct drm_plane *plane,
|
||||
struct drm_plane_state *state = plane->state;
|
||||
u32 reg;
|
||||
int ret;
|
||||
u64 gpu_addr = 0;
|
||||
s64 gpu_addr = 0;
|
||||
unsigned int line_l;
|
||||
struct hibmc_drm_private *priv = plane->dev->dev_private;
|
||||
struct hibmc_framebuffer *hibmc_fb;
|
||||
struct hibmc_bo *bo;
|
||||
struct drm_gem_vram_object *gbo;
|
||||
|
||||
if (!state->fb)
|
||||
return;
|
||||
|
||||
hibmc_fb = to_hibmc_framebuffer(state->fb);
|
||||
bo = gem_to_hibmc_bo(hibmc_fb->obj);
|
||||
ret = ttm_bo_reserve(&bo->bo, true, false, NULL);
|
||||
gbo = drm_gem_vram_of_gem(hibmc_fb->obj);
|
||||
|
||||
ret = drm_gem_vram_pin(gbo, DRM_GEM_VRAM_PL_FLAG_VRAM);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed to reserve ttm_bo: %d", ret);
|
||||
DRM_ERROR("failed to pin bo: %d", ret);
|
||||
return;
|
||||
}
|
||||
|
||||
ret = hibmc_bo_pin(bo, TTM_PL_FLAG_VRAM, &gpu_addr);
|
||||
ttm_bo_unreserve(&bo->bo);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed to pin hibmc_bo: %d", ret);
|
||||
gpu_addr = drm_gem_vram_offset(gbo);
|
||||
if (gpu_addr < 0) {
|
||||
drm_gem_vram_unpin(gbo);
|
||||
return;
|
||||
}
|
||||
|
||||
|
@@ -27,14 +27,7 @@
|
||||
|
||||
static const struct file_operations hibmc_fops = {
|
||||
.owner = THIS_MODULE,
|
||||
.open = drm_open,
|
||||
.release = drm_release,
|
||||
.unlocked_ioctl = drm_ioctl,
|
||||
.compat_ioctl = drm_compat_ioctl,
|
||||
.mmap = hibmc_mmap,
|
||||
.poll = drm_poll,
|
||||
.read = drm_read,
|
||||
.llseek = no_llseek,
|
||||
DRM_VRAM_MM_FILE_OPERATIONS
|
||||
};
|
||||
|
||||
static irqreturn_t hibmc_drm_interrupt(int irq, void *arg)
|
||||
@@ -63,9 +56,10 @@ static struct drm_driver hibmc_driver = {
|
||||
.desc = "hibmc drm driver",
|
||||
.major = 1,
|
||||
.minor = 0,
|
||||
.gem_free_object_unlocked = hibmc_gem_free_object,
|
||||
.gem_free_object_unlocked =
|
||||
drm_gem_vram_driver_gem_free_object_unlocked,
|
||||
.dumb_create = hibmc_dumb_create,
|
||||
.dumb_map_offset = hibmc_dumb_mmap_offset,
|
||||
.dumb_map_offset = drm_gem_vram_driver_dumb_mmap_offset,
|
||||
.irq_handler = hibmc_drm_interrupt,
|
||||
};
|
||||
|
||||
|
@@ -23,7 +23,8 @@
|
||||
#include <drm/drm_atomic.h>
|
||||
#include <drm/drm_fb_helper.h>
|
||||
#include <drm/drm_gem.h>
|
||||
#include <drm/ttm/ttm_bo_driver.h>
|
||||
#include <drm/drm_gem_vram_helper.h>
|
||||
#include <drm/drm_vram_mm_helper.h>
|
||||
|
||||
struct hibmc_framebuffer {
|
||||
struct drm_framebuffer fb;
|
||||
@@ -48,36 +49,12 @@ struct hibmc_drm_private {
|
||||
struct drm_device *dev;
|
||||
bool mode_config_initialized;
|
||||
|
||||
/* ttm */
|
||||
struct ttm_bo_device bdev;
|
||||
bool initialized;
|
||||
|
||||
/* fbdev */
|
||||
struct hibmc_fbdev *fbdev;
|
||||
bool mm_inited;
|
||||
};
|
||||
|
||||
#define to_hibmc_framebuffer(x) container_of(x, struct hibmc_framebuffer, fb)
|
||||
|
||||
struct hibmc_bo {
|
||||
struct ttm_buffer_object bo;
|
||||
struct ttm_placement placement;
|
||||
struct ttm_bo_kmap_obj kmap;
|
||||
struct drm_gem_object gem;
|
||||
struct ttm_place placements[3];
|
||||
int pin_count;
|
||||
};
|
||||
|
||||
static inline struct hibmc_bo *hibmc_bo(struct ttm_buffer_object *bo)
|
||||
{
|
||||
return container_of(bo, struct hibmc_bo, bo);
|
||||
}
|
||||
|
||||
static inline struct hibmc_bo *gem_to_hibmc_bo(struct drm_gem_object *gem)
|
||||
{
|
||||
return container_of(gem, struct hibmc_bo, gem);
|
||||
}
|
||||
|
||||
void hibmc_set_power_mode(struct hibmc_drm_private *priv,
|
||||
unsigned int power_mode);
|
||||
void hibmc_set_current_gate(struct hibmc_drm_private *priv,
|
||||
@@ -97,14 +74,8 @@ hibmc_framebuffer_init(struct drm_device *dev,
|
||||
|
||||
int hibmc_mm_init(struct hibmc_drm_private *hibmc);
|
||||
void hibmc_mm_fini(struct hibmc_drm_private *hibmc);
|
||||
int hibmc_bo_pin(struct hibmc_bo *bo, u32 pl_flag, u64 *gpu_addr);
|
||||
int hibmc_bo_unpin(struct hibmc_bo *bo);
|
||||
void hibmc_gem_free_object(struct drm_gem_object *obj);
|
||||
int hibmc_dumb_create(struct drm_file *file, struct drm_device *dev,
|
||||
struct drm_mode_create_dumb *args);
|
||||
int hibmc_dumb_mmap_offset(struct drm_file *file, struct drm_device *dev,
|
||||
u32 handle, u64 *offset);
|
||||
int hibmc_mmap(struct file *filp, struct vm_area_struct *vma);
|
||||
|
||||
extern const struct drm_mode_config_funcs hibmc_mode_funcs;
|
||||
|
||||
|
@@ -63,10 +63,10 @@ static int hibmc_drm_fb_create(struct drm_fb_helper *helper,
|
||||
struct drm_mode_fb_cmd2 mode_cmd;
|
||||
struct drm_gem_object *gobj = NULL;
|
||||
int ret = 0;
|
||||
int ret1;
|
||||
size_t size;
|
||||
unsigned int bytes_per_pixel;
|
||||
struct hibmc_bo *bo = NULL;
|
||||
struct drm_gem_vram_object *gbo = NULL;
|
||||
void *base;
|
||||
|
||||
DRM_DEBUG_DRIVER("surface width(%d), height(%d) and bpp(%d)\n",
|
||||
sizes->surface_width, sizes->surface_height,
|
||||
@@ -88,26 +88,20 @@ static int hibmc_drm_fb_create(struct drm_fb_helper *helper,
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
bo = gem_to_hibmc_bo(gobj);
|
||||
gbo = drm_gem_vram_of_gem(gobj);
|
||||
|
||||
ret = ttm_bo_reserve(&bo->bo, true, false, NULL);
|
||||
ret = drm_gem_vram_pin(gbo, DRM_GEM_VRAM_PL_FLAG_VRAM);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed to reserve ttm_bo: %d\n", ret);
|
||||
DRM_ERROR("failed to pin fbcon: %d\n", ret);
|
||||
goto out_unref_gem;
|
||||
}
|
||||
|
||||
ret = hibmc_bo_pin(bo, TTM_PL_FLAG_VRAM, NULL);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed to pin fbcon: %d\n", ret);
|
||||
goto out_unreserve_ttm_bo;
|
||||
}
|
||||
|
||||
ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap);
|
||||
if (ret) {
|
||||
base = drm_gem_vram_kmap(gbo, true, NULL);
|
||||
if (IS_ERR(base)) {
|
||||
ret = PTR_ERR(base);
|
||||
DRM_ERROR("failed to kmap fbcon: %d\n", ret);
|
||||
goto out_unpin_bo;
|
||||
}
|
||||
ttm_bo_unreserve(&bo->bo);
|
||||
|
||||
info = drm_fb_helper_alloc_fbi(helper);
|
||||
if (IS_ERR(info)) {
|
||||
@@ -131,24 +125,17 @@ static int hibmc_drm_fb_create(struct drm_fb_helper *helper,
|
||||
|
||||
drm_fb_helper_fill_info(info, &priv->fbdev->helper, sizes);
|
||||
|
||||
info->screen_base = bo->kmap.virtual;
|
||||
info->screen_base = base;
|
||||
info->screen_size = size;
|
||||
|
||||
info->fix.smem_start = bo->bo.mem.bus.offset + bo->bo.mem.bus.base;
|
||||
info->fix.smem_start = gbo->bo.mem.bus.offset + gbo->bo.mem.bus.base;
|
||||
info->fix.smem_len = size;
|
||||
return 0;
|
||||
|
||||
out_release_fbi:
|
||||
ret1 = ttm_bo_reserve(&bo->bo, true, false, NULL);
|
||||
if (ret1) {
|
||||
DRM_ERROR("failed to rsv ttm_bo when release fbi: %d\n", ret1);
|
||||
goto out_unref_gem;
|
||||
}
|
||||
ttm_bo_kunmap(&bo->kmap);
|
||||
drm_gem_vram_kunmap(gbo);
|
||||
out_unpin_bo:
|
||||
hibmc_bo_unpin(bo);
|
||||
out_unreserve_ttm_bo:
|
||||
ttm_bo_unreserve(&bo->bo);
|
||||
drm_gem_vram_unpin(gbo);
|
||||
out_unref_gem:
|
||||
drm_gem_object_put_unlocked(gobj);
|
||||
|
||||
|
@@ -17,335 +17,55 @@
|
||||
*/
|
||||
|
||||
#include <drm/drm_atomic_helper.h>
|
||||
#include <drm/ttm/ttm_page_alloc.h>
|
||||
|
||||
#include "hibmc_drm_drv.h"
|
||||
|
||||
static inline struct hibmc_drm_private *
|
||||
hibmc_bdev(struct ttm_bo_device *bd)
|
||||
{
|
||||
return container_of(bd, struct hibmc_drm_private, bdev);
|
||||
}
|
||||
|
||||
static void hibmc_bo_ttm_destroy(struct ttm_buffer_object *tbo)
|
||||
{
|
||||
struct hibmc_bo *bo = container_of(tbo, struct hibmc_bo, bo);
|
||||
|
||||
drm_gem_object_release(&bo->gem);
|
||||
kfree(bo);
|
||||
}
|
||||
|
||||
static bool hibmc_ttm_bo_is_hibmc_bo(struct ttm_buffer_object *bo)
|
||||
{
|
||||
return bo->destroy == &hibmc_bo_ttm_destroy;
|
||||
}
|
||||
|
||||
static int
|
||||
hibmc_bo_init_mem_type(struct ttm_bo_device *bdev, u32 type,
|
||||
struct ttm_mem_type_manager *man)
|
||||
{
|
||||
switch (type) {
|
||||
case TTM_PL_SYSTEM:
|
||||
man->flags = TTM_MEMTYPE_FLAG_MAPPABLE;
|
||||
man->available_caching = TTM_PL_MASK_CACHING;
|
||||
man->default_caching = TTM_PL_FLAG_CACHED;
|
||||
break;
|
||||
case TTM_PL_VRAM:
|
||||
man->func = &ttm_bo_manager_func;
|
||||
man->flags = TTM_MEMTYPE_FLAG_FIXED |
|
||||
TTM_MEMTYPE_FLAG_MAPPABLE;
|
||||
man->available_caching = TTM_PL_FLAG_UNCACHED |
|
||||
TTM_PL_FLAG_WC;
|
||||
man->default_caching = TTM_PL_FLAG_WC;
|
||||
break;
|
||||
default:
|
||||
DRM_ERROR("unsupported memory type %u\n", type);
|
||||
return -EINVAL;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
void hibmc_ttm_placement(struct hibmc_bo *bo, int domain)
|
||||
{
|
||||
u32 count = 0;
|
||||
u32 i;
|
||||
|
||||
bo->placement.placement = bo->placements;
|
||||
bo->placement.busy_placement = bo->placements;
|
||||
if (domain & TTM_PL_FLAG_VRAM)
|
||||
bo->placements[count++].flags = TTM_PL_FLAG_WC |
|
||||
TTM_PL_FLAG_UNCACHED | TTM_PL_FLAG_VRAM;
|
||||
if (domain & TTM_PL_FLAG_SYSTEM)
|
||||
bo->placements[count++].flags = TTM_PL_MASK_CACHING |
|
||||
TTM_PL_FLAG_SYSTEM;
|
||||
if (!count)
|
||||
bo->placements[count++].flags = TTM_PL_MASK_CACHING |
|
||||
TTM_PL_FLAG_SYSTEM;
|
||||
|
||||
bo->placement.num_placement = count;
|
||||
bo->placement.num_busy_placement = count;
|
||||
for (i = 0; i < count; i++) {
|
||||
bo->placements[i].fpfn = 0;
|
||||
bo->placements[i].lpfn = 0;
|
||||
}
|
||||
}
|
||||
|
||||
static void
|
||||
hibmc_bo_evict_flags(struct ttm_buffer_object *bo, struct ttm_placement *pl)
|
||||
{
|
||||
struct hibmc_bo *hibmcbo = hibmc_bo(bo);
|
||||
|
||||
if (!hibmc_ttm_bo_is_hibmc_bo(bo))
|
||||
return;
|
||||
|
||||
hibmc_ttm_placement(hibmcbo, TTM_PL_FLAG_SYSTEM);
|
||||
*pl = hibmcbo->placement;
|
||||
}
|
||||
|
||||
static int hibmc_bo_verify_access(struct ttm_buffer_object *bo,
|
||||
struct file *filp)
|
||||
{
|
||||
struct hibmc_bo *hibmcbo = hibmc_bo(bo);
|
||||
|
||||
return drm_vma_node_verify_access(&hibmcbo->gem.vma_node,
|
||||
filp->private_data);
|
||||
}
|
||||
|
||||
static int hibmc_ttm_io_mem_reserve(struct ttm_bo_device *bdev,
|
||||
struct ttm_mem_reg *mem)
|
||||
{
|
||||
struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
|
||||
struct hibmc_drm_private *hibmc = hibmc_bdev(bdev);
|
||||
|
||||
mem->bus.addr = NULL;
|
||||
mem->bus.offset = 0;
|
||||
mem->bus.size = mem->num_pages << PAGE_SHIFT;
|
||||
mem->bus.base = 0;
|
||||
mem->bus.is_iomem = false;
|
||||
if (!(man->flags & TTM_MEMTYPE_FLAG_MAPPABLE))
|
||||
return -EINVAL;
|
||||
switch (mem->mem_type) {
|
||||
case TTM_PL_SYSTEM:
|
||||
/* system memory */
|
||||
return 0;
|
||||
case TTM_PL_VRAM:
|
||||
mem->bus.offset = mem->start << PAGE_SHIFT;
|
||||
mem->bus.base = pci_resource_start(hibmc->dev->pdev, 0);
|
||||
mem->bus.is_iomem = true;
|
||||
break;
|
||||
default:
|
||||
return -EINVAL;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void hibmc_ttm_backend_destroy(struct ttm_tt *tt)
|
||||
{
|
||||
ttm_tt_fini(tt);
|
||||
kfree(tt);
|
||||
}
|
||||
|
||||
static struct ttm_backend_func hibmc_tt_backend_func = {
|
||||
.destroy = &hibmc_ttm_backend_destroy,
|
||||
};
|
||||
|
||||
static struct ttm_tt *hibmc_ttm_tt_create(struct ttm_buffer_object *bo,
|
||||
u32 page_flags)
|
||||
{
|
||||
struct ttm_tt *tt;
|
||||
int ret;
|
||||
|
||||
tt = kzalloc(sizeof(*tt), GFP_KERNEL);
|
||||
if (!tt) {
|
||||
DRM_ERROR("failed to allocate ttm_tt\n");
|
||||
return NULL;
|
||||
}
|
||||
tt->func = &hibmc_tt_backend_func;
|
||||
ret = ttm_tt_init(tt, bo, page_flags);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed to initialize ttm_tt: %d\n", ret);
|
||||
kfree(tt);
|
||||
return NULL;
|
||||
}
|
||||
return tt;
|
||||
}
|
||||
|
||||
struct ttm_bo_driver hibmc_bo_driver = {
|
||||
.ttm_tt_create = hibmc_ttm_tt_create,
|
||||
.init_mem_type = hibmc_bo_init_mem_type,
|
||||
.evict_flags = hibmc_bo_evict_flags,
|
||||
.move = NULL,
|
||||
.verify_access = hibmc_bo_verify_access,
|
||||
.io_mem_reserve = &hibmc_ttm_io_mem_reserve,
|
||||
.io_mem_free = NULL,
|
||||
};
|
||||
|
||||
int hibmc_mm_init(struct hibmc_drm_private *hibmc)
|
||||
{
|
||||
struct drm_vram_mm *vmm;
|
||||
int ret;
|
||||
struct drm_device *dev = hibmc->dev;
|
||||
struct ttm_bo_device *bdev = &hibmc->bdev;
|
||||
|
||||
ret = ttm_bo_device_init(&hibmc->bdev,
|
||||
&hibmc_bo_driver,
|
||||
dev->anon_inode->i_mapping,
|
||||
true);
|
||||
if (ret) {
|
||||
DRM_ERROR("error initializing bo driver: %d\n", ret);
|
||||
vmm = drm_vram_helper_alloc_mm(dev,
|
||||
pci_resource_start(dev->pdev, 0),
|
||||
hibmc->fb_size, &drm_gem_vram_mm_funcs);
|
||||
if (IS_ERR(vmm)) {
|
||||
ret = PTR_ERR(vmm);
|
||||
DRM_ERROR("Error initializing VRAM MM; %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = ttm_bo_init_mm(bdev, TTM_PL_VRAM,
|
||||
hibmc->fb_size >> PAGE_SHIFT);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed ttm VRAM init: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
hibmc->mm_inited = true;
|
||||
return 0;
|
||||
}
|
||||
|
||||
void hibmc_mm_fini(struct hibmc_drm_private *hibmc)
|
||||
{
|
||||
if (!hibmc->mm_inited)
|
||||
if (!hibmc->dev->vram_mm)
|
||||
return;
|
||||
|
||||
ttm_bo_device_release(&hibmc->bdev);
|
||||
hibmc->mm_inited = false;
|
||||
}
|
||||
|
||||
static void hibmc_bo_unref(struct hibmc_bo **bo)
|
||||
{
|
||||
struct ttm_buffer_object *tbo;
|
||||
|
||||
if ((*bo) == NULL)
|
||||
return;
|
||||
|
||||
tbo = &((*bo)->bo);
|
||||
ttm_bo_put(tbo);
|
||||
*bo = NULL;
|
||||
}
|
||||
|
||||
int hibmc_bo_create(struct drm_device *dev, int size, int align,
|
||||
u32 flags, struct hibmc_bo **phibmcbo)
|
||||
{
|
||||
struct hibmc_drm_private *hibmc = dev->dev_private;
|
||||
struct hibmc_bo *hibmcbo;
|
||||
size_t acc_size;
|
||||
int ret;
|
||||
|
||||
hibmcbo = kzalloc(sizeof(*hibmcbo), GFP_KERNEL);
|
||||
if (!hibmcbo) {
|
||||
DRM_ERROR("failed to allocate hibmcbo\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
ret = drm_gem_object_init(dev, &hibmcbo->gem, size);
|
||||
if (ret) {
|
||||
DRM_ERROR("failed to initialize drm gem object: %d\n", ret);
|
||||
kfree(hibmcbo);
|
||||
return ret;
|
||||
}
|
||||
|
||||
hibmcbo->bo.bdev = &hibmc->bdev;
|
||||
|
||||
hibmc_ttm_placement(hibmcbo, TTM_PL_FLAG_VRAM | TTM_PL_FLAG_SYSTEM);
|
||||
|
||||
acc_size = ttm_bo_dma_acc_size(&hibmc->bdev, size,
|
||||
sizeof(struct hibmc_bo));
|
||||
|
||||
ret = ttm_bo_init(&hibmc->bdev, &hibmcbo->bo, size,
|
||||
ttm_bo_type_device, &hibmcbo->placement,
|
||||
align >> PAGE_SHIFT, false, acc_size,
|
||||
NULL, NULL, hibmc_bo_ttm_destroy);
|
||||
if (ret) {
|
||||
hibmc_bo_unref(&hibmcbo);
|
||||
DRM_ERROR("failed to initialize ttm_bo: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
*phibmcbo = hibmcbo;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int hibmc_bo_pin(struct hibmc_bo *bo, u32 pl_flag, u64 *gpu_addr)
|
||||
{
|
||||
struct ttm_operation_ctx ctx = { false, false };
|
||||
int i, ret;
|
||||
|
||||
if (bo->pin_count) {
|
||||
bo->pin_count++;
|
||||
if (gpu_addr)
|
||||
*gpu_addr = bo->bo.offset;
|
||||
return 0;
|
||||
}
|
||||
|
||||
hibmc_ttm_placement(bo, pl_flag);
|
||||
for (i = 0; i < bo->placement.num_placement; i++)
|
||||
bo->placements[i].flags |= TTM_PL_FLAG_NO_EVICT;
|
||||
ret = ttm_bo_validate(&bo->bo, &bo->placement, &ctx);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
bo->pin_count = 1;
|
||||
if (gpu_addr)
|
||||
*gpu_addr = bo->bo.offset;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int hibmc_bo_unpin(struct hibmc_bo *bo)
|
||||
{
|
||||
struct ttm_operation_ctx ctx = { false, false };
|
||||
int i, ret;
|
||||
|
||||
if (!bo->pin_count) {
|
||||
DRM_ERROR("unpin bad %p\n", bo);
|
||||
return 0;
|
||||
}
|
||||
bo->pin_count--;
|
||||
if (bo->pin_count)
|
||||
return 0;
|
||||
|
||||
for (i = 0; i < bo->placement.num_placement ; i++)
|
||||
bo->placements[i].flags &= ~TTM_PL_FLAG_NO_EVICT;
|
||||
ret = ttm_bo_validate(&bo->bo, &bo->placement, &ctx);
|
||||
if (ret) {
|
||||
DRM_ERROR("validate failed for unpin: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int hibmc_mmap(struct file *filp, struct vm_area_struct *vma)
|
||||
{
|
||||
struct drm_file *file_priv = filp->private_data;
|
||||
struct hibmc_drm_private *hibmc = file_priv->minor->dev->dev_private;
|
||||
|
||||
return ttm_bo_mmap(filp, vma, &hibmc->bdev);
|
||||
drm_vram_helper_release_mm(hibmc->dev);
|
||||
}
|
||||
|
||||
int hibmc_gem_create(struct drm_device *dev, u32 size, bool iskernel,
|
||||
struct drm_gem_object **obj)
|
||||
{
|
||||
struct hibmc_bo *hibmcbo;
|
||||
struct drm_gem_vram_object *gbo;
|
||||
int ret;
|
||||
|
||||
*obj = NULL;
|
||||
|
||||
size = PAGE_ALIGN(size);
|
||||
if (size == 0) {
|
||||
DRM_ERROR("error: zero size\n");
|
||||
size = roundup(size, PAGE_SIZE);
|
||||
if (size == 0)
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
ret = hibmc_bo_create(dev, size, 0, 0, &hibmcbo);
|
||||
if (ret) {
|
||||
gbo = drm_gem_vram_create(dev, &dev->vram_mm->bdev, size, 0, false);
|
||||
if (IS_ERR(gbo)) {
|
||||
ret = PTR_ERR(gbo);
|
||||
if (ret != -ERESTARTSYS)
|
||||
DRM_ERROR("failed to allocate GEM object: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
*obj = &hibmcbo->gem;
|
||||
*obj = &gbo->gem;
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -377,35 +97,6 @@ int hibmc_dumb_create(struct drm_file *file, struct drm_device *dev,
|
||||
return 0;
|
||||
}
|
||||
|
||||
void hibmc_gem_free_object(struct drm_gem_object *obj)
|
||||
{
|
||||
struct hibmc_bo *hibmcbo = gem_to_hibmc_bo(obj);
|
||||
|
||||
hibmc_bo_unref(&hibmcbo);
|
||||
}
|
||||
|
||||
static u64 hibmc_bo_mmap_offset(struct hibmc_bo *bo)
|
||||
{
|
||||
return drm_vma_node_offset_addr(&bo->bo.vma_node);
|
||||
}
|
||||
|
||||
int hibmc_dumb_mmap_offset(struct drm_file *file, struct drm_device *dev,
|
||||
u32 handle, u64 *offset)
|
||||
{
|
||||
struct drm_gem_object *obj;
|
||||
struct hibmc_bo *bo;
|
||||
|
||||
obj = drm_gem_object_lookup(file, handle);
|
||||
if (!obj)
|
||||
return -ENOENT;
|
||||
|
||||
bo = gem_to_hibmc_bo(obj);
|
||||
*offset = hibmc_bo_mmap_offset(bo);
|
||||
|
||||
drm_gem_object_put_unlocked(obj);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void hibmc_user_framebuffer_destroy(struct drm_framebuffer *fb)
|
||||
{
|
||||
struct hibmc_framebuffer *hibmc_fb = to_hibmc_framebuffer(fb);
|
||||
|
Reference in New Issue
Block a user