disp: msm: sde: add trusted UI support for SDE
Extends atomic check to validate TUI transition request aginst SDE and non-SDE client states. Triggers HW assignment after making sure current frame is completely flushed out of HW pipeline. Initiate the acquire after TUI end after restoring the software states. Change-Id: I6ceccd44c18526ff818723b5631b66af3ff083c1 Signed-off-by: Jeykumar Sankaran <jsanka@codeaurora.org>
This commit is contained in:
@@ -48,6 +48,7 @@
|
||||
#include "sde_crtc.h"
|
||||
#include "sde_reg_dma.h"
|
||||
#include "sde_connector.h"
|
||||
#include "sde_vm.h"
|
||||
|
||||
#include <linux/qcom_scm.h>
|
||||
#include "soc/qcom/secure_buffer.h"
|
||||
@@ -893,6 +894,85 @@ static int _sde_kms_unmap_all_splash_regions(struct sde_kms *sde_kms)
|
||||
return ret;
|
||||
}
|
||||
|
||||
int sde_kms_vm_primary_prepare_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
struct drm_device *ddev;
|
||||
struct drm_crtc *crtc;
|
||||
struct drm_encoder *encoder;
|
||||
struct drm_connector *connector;
|
||||
struct sde_vm_ops *vm_ops;
|
||||
struct sde_crtc_state *cstate;
|
||||
enum sde_crtc_vm_req vm_req;
|
||||
int rc = 0;
|
||||
|
||||
ddev = sde_kms->dev;
|
||||
|
||||
if (!sde_kms->vm)
|
||||
return -EINVAL;
|
||||
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
|
||||
crtc = state->crtcs[0].ptr;
|
||||
|
||||
cstate = to_sde_crtc_state(state->crtcs[0].new_state);
|
||||
|
||||
vm_req = sde_crtc_get_property(cstate, CRTC_PROP_VM_REQ_STATE);
|
||||
if (vm_req != VM_REQ_ACQUIRE)
|
||||
return 0;
|
||||
|
||||
/* enable MDSS irq line */
|
||||
sde_irq_update(&sde_kms->base, true);
|
||||
|
||||
/* clear the stale IRQ status bits */
|
||||
if (sde_kms->hw_intr && sde_kms->hw_intr->ops.clear_all_irqs)
|
||||
sde_kms->hw_intr->ops.clear_all_irqs(sde_kms->hw_intr);
|
||||
|
||||
/* enable the display path IRQ's */
|
||||
drm_for_each_encoder_mask(encoder, crtc->dev, crtc->state->encoder_mask)
|
||||
sde_encoder_irq_control(encoder, true);
|
||||
|
||||
/* Schedule ESD work */
|
||||
list_for_each_entry(connector, &ddev->mode_config.connector_list, head)
|
||||
if (drm_connector_mask(connector) & crtc->state->connector_mask)
|
||||
sde_connector_schedule_status_work(connector, true);
|
||||
|
||||
/* handle non-SDE pre_acquire */
|
||||
if (vm_ops->vm_client_post_acquire)
|
||||
rc = vm_ops->vm_client_post_acquire(sde_kms);
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
int sde_kms_vm_trusted_prepare_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
struct drm_device *ddev;
|
||||
struct drm_plane *plane;
|
||||
struct sde_crtc_state *cstate;
|
||||
enum sde_crtc_vm_req vm_req;
|
||||
|
||||
ddev = sde_kms->dev;
|
||||
|
||||
cstate = to_sde_crtc_state(state->crtcs[0].new_state);
|
||||
|
||||
vm_req = sde_crtc_get_property(cstate, CRTC_PROP_VM_REQ_STATE);
|
||||
if (vm_req != VM_REQ_ACQUIRE)
|
||||
return 0;
|
||||
|
||||
/* Clear the stale IRQ status bits */
|
||||
if (sde_kms->hw_intr && sde_kms->hw_intr->ops.clear_all_irqs)
|
||||
sde_kms->hw_intr->ops.clear_all_irqs(sde_kms->hw_intr);
|
||||
|
||||
/* Program the SID's for the trusted VM */
|
||||
list_for_each_entry(plane, &ddev->mode_config.plane_list, head)
|
||||
sde_plane_set_sid(plane, 1);
|
||||
|
||||
sde_hw_set_lutdma_sid(sde_kms->hw_sid, 1);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void sde_kms_prepare_commit(struct msm_kms *kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
@@ -902,6 +982,7 @@ static void sde_kms_prepare_commit(struct msm_kms *kms,
|
||||
struct drm_encoder *encoder;
|
||||
struct drm_crtc *crtc;
|
||||
struct drm_crtc_state *crtc_state;
|
||||
struct sde_vm_ops *vm_ops;
|
||||
int i, rc;
|
||||
|
||||
if (!kms)
|
||||
@@ -947,6 +1028,14 @@ static void sde_kms_prepare_commit(struct msm_kms *kms,
|
||||
* transitions prepare below if any transtions is required.
|
||||
*/
|
||||
sde_kms_prepare_secure_transition(kms, state);
|
||||
|
||||
if (!sde_kms->vm)
|
||||
goto end;
|
||||
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
|
||||
if (vm_ops->vm_prepare_commit)
|
||||
vm_ops->vm_prepare_commit(sde_kms, state);
|
||||
end:
|
||||
SDE_ATRACE_END("prepare_commit");
|
||||
}
|
||||
@@ -1039,6 +1128,129 @@ static void _sde_kms_release_splash_resource(struct sde_kms *sde_kms,
|
||||
}
|
||||
}
|
||||
|
||||
int sde_kms_vm_trusted_post_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
struct sde_vm_ops *vm_ops;
|
||||
struct drm_device *ddev;
|
||||
struct drm_crtc *crtc;
|
||||
struct drm_plane *plane;
|
||||
struct drm_encoder *encoder;
|
||||
struct sde_crtc_state *cstate;
|
||||
struct drm_crtc_state *new_cstate;
|
||||
enum sde_crtc_vm_req vm_req;
|
||||
int rc = 0;
|
||||
|
||||
if (!sde_kms || !sde_kms->vm)
|
||||
return -EINVAL;
|
||||
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
ddev = sde_kms->dev;
|
||||
|
||||
crtc = state->crtcs[0].ptr;
|
||||
new_cstate = state->crtcs[0].new_state;
|
||||
cstate = to_sde_crtc_state(new_cstate);
|
||||
|
||||
vm_req = sde_crtc_get_property(cstate, CRTC_PROP_VM_REQ_STATE);
|
||||
if (vm_req != VM_REQ_RELEASE)
|
||||
return rc;
|
||||
|
||||
if (!new_cstate->active && !new_cstate->active_changed)
|
||||
return rc;
|
||||
|
||||
/* if vm_req is enabled, once CRTC on the commit is guaranteed */
|
||||
sde_kms_wait_for_frame_transfer_complete(&sde_kms->base, crtc);
|
||||
|
||||
drm_for_each_encoder_mask(encoder, crtc->dev, crtc->state->encoder_mask)
|
||||
sde_encoder_irq_control(encoder, false);
|
||||
|
||||
sde_irq_update(&sde_kms->base, false);
|
||||
|
||||
list_for_each_entry(plane, &ddev->mode_config.plane_list, head)
|
||||
sde_plane_set_sid(plane, 0);
|
||||
|
||||
sde_hw_set_lutdma_sid(sde_kms->hw_sid, 0);
|
||||
|
||||
if (vm_ops->vm_release)
|
||||
rc = vm_ops->vm_release(sde_kms);
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
int sde_kms_vm_pre_release(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
struct drm_device *ddev;
|
||||
struct drm_crtc *crtc;
|
||||
struct drm_encoder *encoder;
|
||||
struct drm_connector *connector;
|
||||
int rc = 0;
|
||||
|
||||
ddev = sde_kms->dev;
|
||||
|
||||
crtc = state->crtcs[0].ptr;
|
||||
|
||||
/* if vm_req is enabled, once CRTC on the commit is guaranteed */
|
||||
sde_kms_wait_for_frame_transfer_complete(&sde_kms->base, crtc);
|
||||
|
||||
/* disable ESD work */
|
||||
list_for_each_entry(connector,
|
||||
&ddev->mode_config.connector_list, head) {
|
||||
if (drm_connector_mask(connector) & crtc->state->connector_mask)
|
||||
sde_connector_schedule_status_work(connector, false);
|
||||
}
|
||||
|
||||
/* disable SDE irq's */
|
||||
drm_for_each_encoder_mask(encoder, crtc->dev, crtc->state->encoder_mask)
|
||||
sde_encoder_irq_control(encoder, false);
|
||||
|
||||
/* disable IRQ line */
|
||||
sde_irq_update(&sde_kms->base, false);
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
int sde_kms_vm_primary_post_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
struct sde_vm_ops *vm_ops;
|
||||
struct sde_crtc_state *cstate;
|
||||
enum sde_crtc_vm_req vm_req;
|
||||
int rc = 0;
|
||||
|
||||
if (!sde_kms || !sde_kms->vm)
|
||||
return -EINVAL;
|
||||
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
|
||||
cstate = to_sde_crtc_state(state->crtcs[0].new_state);
|
||||
|
||||
vm_req = sde_crtc_get_property(cstate, CRTC_PROP_VM_REQ_STATE);
|
||||
if (vm_req != VM_REQ_RELEASE)
|
||||
goto exit;
|
||||
|
||||
/* handle SDE pre-release */
|
||||
sde_kms_vm_pre_release(sde_kms, state);
|
||||
|
||||
/* handle non-SDE clients pre-release */
|
||||
if (vm_ops->vm_client_pre_release) {
|
||||
rc = vm_ops->vm_client_pre_release(sde_kms);
|
||||
if (rc) {
|
||||
SDE_ERROR("sde vm pre_release failed, rc=%d\n", rc);
|
||||
goto exit;
|
||||
}
|
||||
}
|
||||
|
||||
/* release HW */
|
||||
if (vm_ops->vm_release) {
|
||||
rc = vm_ops->vm_release(sde_kms);
|
||||
if (rc)
|
||||
SDE_ERROR("sde vm assign failed, rc=%d\n", rc);
|
||||
}
|
||||
exit:
|
||||
return rc;
|
||||
}
|
||||
|
||||
static void sde_kms_complete_commit(struct msm_kms *kms,
|
||||
struct drm_atomic_state *old_state)
|
||||
{
|
||||
@@ -1049,6 +1261,7 @@ static void sde_kms_complete_commit(struct msm_kms *kms,
|
||||
struct drm_connector *connector;
|
||||
struct drm_connector_state *old_conn_state;
|
||||
struct msm_display_conn_params params;
|
||||
struct sde_vm_ops *vm_ops;
|
||||
int i, rc = 0;
|
||||
|
||||
if (!kms || !old_state)
|
||||
@@ -1093,6 +1306,17 @@ static void sde_kms_complete_commit(struct msm_kms *kms,
|
||||
}
|
||||
}
|
||||
|
||||
if (sde_kms->vm) {
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
|
||||
if (vm_ops->vm_post_commit) {
|
||||
rc = vm_ops->vm_post_commit(sde_kms, old_state);
|
||||
if (rc)
|
||||
SDE_ERROR("vm post commit failed, rc = %d\n",
|
||||
rc);
|
||||
}
|
||||
}
|
||||
|
||||
pm_runtime_put_sync(sde_kms->dev->dev);
|
||||
|
||||
for_each_old_crtc_in_state(old_state, crtc, old_crtc_state, i)
|
||||
@@ -1794,6 +2018,9 @@ static void _sde_kms_hw_destroy(struct sde_kms *sde_kms,
|
||||
of_genpd_del_provider(pdev->dev.of_node);
|
||||
}
|
||||
|
||||
if (sde_kms->vm && sde_kms->vm->vm_ops.vm_deinit)
|
||||
sde_kms->vm->vm_ops.vm_deinit(sde_kms, &sde_kms->vm->vm_ops);
|
||||
|
||||
if (sde_kms->hw_intr)
|
||||
sde_hw_intr_destroy(sde_kms->hw_intr);
|
||||
sde_kms->hw_intr = NULL;
|
||||
@@ -2061,6 +2288,108 @@ backoff:
|
||||
goto retry;
|
||||
}
|
||||
|
||||
static int sde_kms_check_vm_request(struct msm_kms *kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
struct sde_kms *sde_kms;
|
||||
struct drm_device *dev;
|
||||
struct drm_crtc *crtc;
|
||||
struct drm_crtc_state *new_cstate, *old_cstate;
|
||||
uint32_t i, commit_crtc_cnt = 0, global_crtc_cnt = 0;
|
||||
struct drm_crtc *active_crtc = NULL, *global_active_crtc = NULL;
|
||||
enum sde_crtc_vm_req old_vm_req = VM_REQ_NONE, new_vm_req = VM_REQ_NONE;
|
||||
struct sde_vm_ops *vm_ops;
|
||||
bool vm_req_active = false;
|
||||
enum sde_crtc_idle_pc_state idle_pc_state;
|
||||
int rc = 0;
|
||||
|
||||
if (!kms || !state)
|
||||
return -EINVAL;
|
||||
|
||||
sde_kms = to_sde_kms(kms);
|
||||
dev = sde_kms->dev;
|
||||
|
||||
if (!sde_kms->vm)
|
||||
return 0;
|
||||
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
|
||||
for_each_oldnew_crtc_in_state(state, crtc, old_cstate, new_cstate, i) {
|
||||
struct sde_crtc_state *old_state = NULL, *new_state = NULL;
|
||||
|
||||
new_state = to_sde_crtc_state(new_cstate);
|
||||
|
||||
if (!new_cstate->active && !new_cstate->active_changed)
|
||||
continue;
|
||||
|
||||
new_vm_req = sde_crtc_get_property(new_state,
|
||||
CRTC_PROP_VM_REQ_STATE);
|
||||
|
||||
commit_crtc_cnt++;
|
||||
|
||||
if (old_cstate) {
|
||||
old_state = to_sde_crtc_state(old_cstate);
|
||||
old_vm_req = sde_crtc_get_property(old_state,
|
||||
CRTC_PROP_VM_REQ_STATE);
|
||||
}
|
||||
|
||||
/**
|
||||
* No active request if the transition is from
|
||||
* VM_REQ_NONE to VM_REQ_NONE
|
||||
*/
|
||||
if (new_vm_req || (old_state && old_vm_req))
|
||||
vm_req_active = true;
|
||||
|
||||
idle_pc_state = sde_crtc_get_property(new_state,
|
||||
CRTC_PROP_IDLE_PC_STATE);
|
||||
|
||||
active_crtc = crtc;
|
||||
}
|
||||
|
||||
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
|
||||
if (!crtc->state->active)
|
||||
continue;
|
||||
|
||||
global_crtc_cnt++;
|
||||
global_active_crtc = crtc;
|
||||
}
|
||||
|
||||
/* Check for single crtc commits only on valid VM requests */
|
||||
if (vm_req_active && active_crtc && global_active_crtc &&
|
||||
(commit_crtc_cnt > sde_kms->catalog->max_trusted_vm_displays ||
|
||||
global_crtc_cnt > sde_kms->catalog->max_trusted_vm_displays ||
|
||||
active_crtc != global_active_crtc)) {
|
||||
SDE_ERROR(
|
||||
"failed to switch VM due to CRTC concurrencies: MAX_CNT: %d active_cnt: %d global_cnt: %d active_crtc: %d global_crtc: %d\n",
|
||||
sde_kms->catalog->max_trusted_vm_displays,
|
||||
commit_crtc_cnt, global_crtc_cnt, active_crtc,
|
||||
global_active_crtc);
|
||||
return -E2BIG;
|
||||
}
|
||||
|
||||
if (!vm_req_active)
|
||||
return 0;
|
||||
|
||||
/* disable idle-pc before releasing the HW */
|
||||
if ((new_vm_req == VM_REQ_RELEASE) &&
|
||||
(idle_pc_state == IDLE_PC_ENABLE)) {
|
||||
SDE_ERROR("failed to switch VM since idle-pc is enabled\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
mutex_lock(&sde_kms->vm->vm_res_lock);
|
||||
if (vm_ops->vm_request_valid)
|
||||
rc = vm_ops->vm_request_valid(sde_kms, old_vm_req, new_vm_req);
|
||||
if (rc)
|
||||
SDE_ERROR(
|
||||
"failed to complete vm transition request. old_state = %d, new_state = %d, hw_ownership: %d\n",
|
||||
old_vm_req, new_vm_req, vm_ops->vm_owns_hw(sde_kms));
|
||||
mutex_unlock(&sde_kms->vm->vm_res_lock);
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
||||
static int sde_kms_check_secure_transition(struct msm_kms *kms,
|
||||
struct drm_atomic_state *state)
|
||||
{
|
||||
@@ -2174,6 +2503,13 @@ static int sde_kms_atomic_check(struct msm_kms *kms,
|
||||
* Secure state
|
||||
*/
|
||||
ret = sde_kms_check_secure_transition(kms, state);
|
||||
if (ret)
|
||||
goto end;
|
||||
|
||||
ret = sde_kms_check_vm_request(kms, state);
|
||||
if (ret)
|
||||
SDE_ERROR("vm switch request checks failed\n");
|
||||
|
||||
end:
|
||||
SDE_ATRACE_END("atomic_check");
|
||||
return ret;
|
||||
@@ -3703,8 +4039,16 @@ static int sde_kms_hw_init(struct msm_kms *kms)
|
||||
SDE_DEBUG("Registering for notification of irq_num: %d\n", irq_num);
|
||||
irq_set_affinity_notifier(irq_num, &sde_kms->affinity_notify);
|
||||
|
||||
return 0;
|
||||
if (sde_in_trusted_vm(sde_kms))
|
||||
rc = sde_vm_trusted_init(sde_kms);
|
||||
else
|
||||
rc = sde_vm_primary_init(sde_kms);
|
||||
if (rc) {
|
||||
SDE_ERROR("failed to initialize VM ops, rc: %d\n", rc);
|
||||
goto error;
|
||||
}
|
||||
|
||||
return 0;
|
||||
error:
|
||||
_sde_kms_hw_destroy(sde_kms, platformdev);
|
||||
end:
|
||||
|
@@ -40,7 +40,6 @@
|
||||
#include "sde_power_handle.h"
|
||||
#include "sde_irq.h"
|
||||
#include "sde_core_perf.h"
|
||||
#include "sde_vm.h"
|
||||
|
||||
#define DRMID(x) ((x) ? (x)->base.id : -1)
|
||||
|
||||
@@ -312,7 +311,6 @@ struct sde_kms {
|
||||
struct dev_pm_qos_request pm_qos_irq_req[NR_CPUS];
|
||||
struct irq_affinity_notify affinity_notify;
|
||||
|
||||
struct sde_vm_ops vm_ops;
|
||||
struct sde_vm *vm;
|
||||
};
|
||||
|
||||
@@ -703,4 +701,42 @@ void sde_kms_irq_enable_notify(struct sde_kms *sde_kms, bool enable);
|
||||
*/
|
||||
int sde_kms_get_io_resources(struct sde_kms *kms, struct msm_io_res *io_res);
|
||||
|
||||
/**
|
||||
* sde_kms_vm_trusted_post_commit - function to prepare the VM after the
|
||||
* last commit before releasing the HW
|
||||
* resources from trusted VM
|
||||
* @sde_kms: pointer to sde_kms
|
||||
* @state: current frames atomic commit state
|
||||
*/
|
||||
int sde_kms_vm_trusted_post_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state);
|
||||
/**
|
||||
* sde_kms_vm_primary_post_commit - function to prepare the VM after the
|
||||
* last commit before assign the HW
|
||||
* resources from primary VM
|
||||
* @sde_kms: pointer to sde_kms
|
||||
* @state: current frames atomic commit state
|
||||
*/
|
||||
int sde_kms_vm_primary_post_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state);
|
||||
|
||||
/**
|
||||
* sde_kms_vm_trusted_prepare_commit - function to prepare the VM before the
|
||||
* the first commit after the accepting
|
||||
* the HW resources in trusted VM.
|
||||
* @sde_kms: pointer to sde_kms
|
||||
* @state: current frame's atomic commit state
|
||||
*/
|
||||
int sde_kms_vm_trusted_prepare_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state);
|
||||
/**
|
||||
* sde_kms_vm_primary_prepare_commit - function to prepare the VM before the
|
||||
* the first commit after the reclaming
|
||||
* the HW resources in trusted VM.
|
||||
* @sde_kms: pointer to sde_kms
|
||||
* @state: current frame's atomic commit state
|
||||
*/
|
||||
int sde_kms_vm_primary_prepare_commit(struct sde_kms *sde_kms,
|
||||
struct drm_atomic_state *state);
|
||||
|
||||
#endif /* __sde_kms_H__ */
|
||||
|
@@ -30,43 +30,7 @@ struct sde_vm_irq_desc {
|
||||
struct sde_vm_irq_entry *irq_entries;
|
||||
};
|
||||
|
||||
/**
|
||||
* sde_vm - VM layer descriptor. Abstract for all the VM's
|
||||
* @vm_res_lock - mutex to protect resource updates
|
||||
* @mem_notificaiton_cookie - Hyp RM notification identifier
|
||||
* @n_irq_lent - irq count
|
||||
* @io_mem_handle - RM identifier for the IO range
|
||||
* @sde_kms - handle to sde_kms
|
||||
*/
|
||||
struct sde_vm {
|
||||
struct mutex vm_res_lock;
|
||||
void *mem_notification_cookie;
|
||||
atomic_t n_irq_lent;
|
||||
int io_mem_handle;
|
||||
struct sde_kms *sde_kms;
|
||||
};
|
||||
|
||||
/**
|
||||
* sde_vm_primary - VM layer descriptor for Primary VM
|
||||
* @base - parent struct object
|
||||
* @irq_desc - cache copy of irq list for validating reclaim
|
||||
*/
|
||||
struct sde_vm_primary {
|
||||
struct sde_vm base;
|
||||
struct sde_vm_irq_desc *irq_desc;
|
||||
};
|
||||
|
||||
/**
|
||||
* sde_vm_trusted - VM layer descriptor for Trusted VM
|
||||
* @base - parent struct object
|
||||
* @sgl_desc - hyp RM sgl list descriptor for IO ranges
|
||||
* @irq_desc - irq list
|
||||
*/
|
||||
struct sde_vm_trusted {
|
||||
struct sde_vm base;
|
||||
struct sde_vm_irq_desc *irq_desc;
|
||||
struct hh_sgl_desc *sgl_desc;
|
||||
};
|
||||
enum sde_crtc_vm_req;
|
||||
|
||||
/**
|
||||
* sde_vm_ops - VM specific function hooks
|
||||
@@ -139,15 +103,59 @@ struct sde_vm_ops {
|
||||
* @kms - handle to sde_kms
|
||||
*/
|
||||
int (*vm_client_post_acquire)(struct sde_kms *kms);
|
||||
|
||||
int (*vm_request_valid)(struct sde_kms *sde_kms,
|
||||
enum sde_crtc_vm_req old_state,
|
||||
enum sde_crtc_vm_req new_state);
|
||||
};
|
||||
|
||||
/**
|
||||
* sde_vm - VM layer descriptor. Abstract for all the VM's
|
||||
* @vm_res_lock - mutex to protect resource updates
|
||||
* @mem_notificaiton_cookie - Hyp RM notification identifier
|
||||
* @n_irq_lent - irq count
|
||||
* @io_mem_handle - RM identifier for the IO range
|
||||
* @sde_kms - handle to sde_kms
|
||||
* @vm_ops - VM operation hooks for respective VM type
|
||||
*/
|
||||
struct sde_vm {
|
||||
struct mutex vm_res_lock;
|
||||
void *mem_notification_cookie;
|
||||
atomic_t n_irq_lent;
|
||||
int io_mem_handle;
|
||||
struct sde_kms *sde_kms;
|
||||
struct sde_vm_ops vm_ops;
|
||||
};
|
||||
|
||||
/**
|
||||
* sde_vm_primary - VM layer descriptor for Primary VM
|
||||
* @base - parent struct object
|
||||
* @irq_desc - cache copy of irq list for validating reclaim
|
||||
*/
|
||||
struct sde_vm_primary {
|
||||
struct sde_vm base;
|
||||
struct sde_vm_irq_desc *irq_desc;
|
||||
};
|
||||
|
||||
/**
|
||||
* sde_vm_trusted - VM layer descriptor for Trusted VM
|
||||
* @base - parent struct object
|
||||
* @sgl_desc - hyp RM sgl list descriptor for IO ranges
|
||||
* @irq_desc - irq list
|
||||
*/
|
||||
struct sde_vm_trusted {
|
||||
struct sde_vm base;
|
||||
struct sde_vm_irq_desc *irq_desc;
|
||||
struct hh_sgl_desc *sgl_desc;
|
||||
};
|
||||
|
||||
#if IS_ENABLED(CONFIG_DRM_SDE_VM)
|
||||
/**
|
||||
* sde_vm_primary_init - Initialize primary VM layer
|
||||
* @kms - pointer to sde_kms
|
||||
* @ops - primary VM specific ops functions
|
||||
* @return - 0 on success
|
||||
*/
|
||||
int sde_vm_primary_init(struct sde_kms *kms, struct sde_vm_ops *ops);
|
||||
int sde_vm_primary_init(struct sde_kms *kms);
|
||||
|
||||
/**
|
||||
* sde_vm_trusted_init - Initialize Trusted VM layer
|
||||
@@ -155,6 +163,17 @@ int sde_vm_primary_init(struct sde_kms *kms, struct sde_vm_ops *ops);
|
||||
* @ops - primary VM specific ops functions
|
||||
* @return - 0 on success
|
||||
*/
|
||||
int sde_vm_trusted_init(struct sde_kms *kms, struct sde_vm_ops *ops);
|
||||
int sde_vm_trusted_init(struct sde_kms *kms);
|
||||
#else
|
||||
static inline int sde_vm_primary_init(struct sde_kms *kms)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline int sde_vm_trusted_init(struct sde_kms *kms)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
#endif /* IS_ENABLED(CONFIG_DRM_SDE_VM) */
|
||||
#endif /* __SDE_VM_H__ */
|
||||
|
@@ -9,6 +9,7 @@
|
||||
#include "dp/dp_display.h"
|
||||
#include "sde_kms.h"
|
||||
#include "sde_vm_common.h"
|
||||
#include "sde_crtc.h"
|
||||
|
||||
struct hh_notify_vmid_desc *sde_vm_populate_vmid(hh_vmid_t vmid)
|
||||
{
|
||||
@@ -287,3 +288,45 @@ pre_release_rollback:
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
int sde_vm_request_valid(struct sde_kms *sde_kms,
|
||||
enum sde_crtc_vm_req old_state,
|
||||
enum sde_crtc_vm_req new_state)
|
||||
{
|
||||
struct sde_vm_ops *vm_ops;
|
||||
int rc = 0;
|
||||
|
||||
vm_ops = &sde_kms->vm->vm_ops;
|
||||
|
||||
switch (new_state) {
|
||||
case VM_REQ_RELEASE:
|
||||
if (old_state == VM_REQ_RELEASE)
|
||||
rc = -EINVAL;
|
||||
break;
|
||||
case VM_REQ_NONE:
|
||||
if (old_state == VM_REQ_RELEASE)
|
||||
rc = -EINVAL;
|
||||
break;
|
||||
case VM_REQ_ACQUIRE:
|
||||
/**
|
||||
* Only the display which requested for HW assignment
|
||||
* can reclaim it back
|
||||
*/
|
||||
if (old_state != VM_REQ_RELEASE)
|
||||
rc = -EINVAL;
|
||||
break;
|
||||
default:
|
||||
SDE_ERROR("invalid vm request\n");
|
||||
rc = -EINVAL;
|
||||
};
|
||||
|
||||
if (!rc && !vm_ops->vm_owns_hw(sde_kms))
|
||||
rc = -EINVAL;
|
||||
|
||||
SDE_DEBUG("old req: %d new req: %d owns_hw: %d\n",
|
||||
old_state, new_state,
|
||||
vm_ops->vm_owns_hw(sde_kms));
|
||||
SDE_EVT32(old_state, new_state, vm_ops->vm_owns_hw(sde_kms));
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
@@ -75,4 +75,9 @@ int sde_vm_post_acquire(struct sde_kms *kms);
|
||||
*/
|
||||
int sde_vm_pre_release(struct sde_kms *kms);
|
||||
|
||||
|
||||
int sde_vm_request_valid(struct sde_kms *sde_kms,
|
||||
enum sde_crtc_vm_req old_state,
|
||||
enum sde_crtc_vm_req new_state);
|
||||
|
||||
#endif /* __SDE_VM_COMMON_H__ */
|
||||
|
@@ -259,9 +259,12 @@ static void _sde_vm_set_ops(struct sde_vm_ops *ops)
|
||||
ops->vm_release = _sde_vm_release;
|
||||
ops->vm_owns_hw = sde_vm_owns_hw;
|
||||
ops->vm_deinit = _sde_vm_deinit;
|
||||
ops->vm_prepare_commit = sde_kms_vm_primary_prepare_commit;
|
||||
ops->vm_post_commit = sde_kms_vm_primary_post_commit;
|
||||
ops->vm_request_valid = sde_vm_request_valid;
|
||||
}
|
||||
|
||||
int sde_vm_primary_init(struct sde_kms *kms, struct sde_vm_ops *ops)
|
||||
int sde_vm_primary_init(struct sde_kms *kms)
|
||||
{
|
||||
struct sde_vm_primary *sde_vm;
|
||||
void *cookie;
|
||||
@@ -271,7 +274,7 @@ int sde_vm_primary_init(struct sde_kms *kms, struct sde_vm_ops *ops)
|
||||
if (!sde_vm)
|
||||
return -ENOMEM;
|
||||
|
||||
_sde_vm_set_ops(ops);
|
||||
_sde_vm_set_ops(&sde_vm->base.vm_ops);
|
||||
|
||||
cookie = hh_mem_notifier_register(HH_MEM_NOTIFIER_TAG_DISPLAY,
|
||||
sde_vm_mem_release_notification_handler, sde_vm);
|
||||
@@ -290,7 +293,7 @@ int sde_vm_primary_init(struct sde_kms *kms, struct sde_vm_ops *ops)
|
||||
|
||||
return 0;
|
||||
init_fail:
|
||||
_sde_vm_deinit(kms, ops);
|
||||
_sde_vm_deinit(kms, &sde_vm->base.vm_ops);
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
@@ -337,9 +337,12 @@ static void _sde_vm_set_ops(struct sde_vm_ops *ops)
|
||||
ops->vm_release = _sde_vm_release;
|
||||
ops->vm_owns_hw = sde_vm_owns_hw;
|
||||
ops->vm_deinit = _sde_vm_deinit;
|
||||
ops->vm_prepare_commit = sde_kms_vm_trusted_prepare_commit;
|
||||
ops->vm_post_commit = sde_kms_vm_trusted_post_commit;
|
||||
ops->vm_request_valid = sde_vm_request_valid;
|
||||
}
|
||||
|
||||
int sde_vm_trusted_init(struct sde_kms *kms, struct sde_vm_ops *ops)
|
||||
int sde_vm_trusted_init(struct sde_kms *kms)
|
||||
{
|
||||
struct sde_vm_trusted *sde_vm;
|
||||
void *cookie;
|
||||
@@ -349,7 +352,7 @@ int sde_vm_trusted_init(struct sde_kms *kms, struct sde_vm_ops *ops)
|
||||
if (!sde_vm)
|
||||
return -ENOMEM;
|
||||
|
||||
_sde_vm_set_ops(ops);
|
||||
_sde_vm_set_ops(&sde_vm->base.vm_ops);
|
||||
|
||||
sde_vm->base.sde_kms = kms;
|
||||
|
||||
@@ -390,7 +393,7 @@ int sde_vm_trusted_init(struct sde_kms *kms, struct sde_vm_ops *ops)
|
||||
|
||||
return 0;
|
||||
init_fail:
|
||||
_sde_vm_deinit(kms, ops);
|
||||
_sde_vm_deinit(kms, &sde_vm->base.vm_ops);
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
Reference in New Issue
Block a user