drm: msm: Use DRM_DEV_* instead of dev_*
Use DRM_DEV_INFO/ERROR/WARN instead of dev_info/err/debug to generate drm-formatted specific log messages so that it will be easy to differentiate in case of multiple instances of driver. Signed-off-by: Mamta Shukla <mamtashukla555@gmail.com> Signed-off-by: Rob Clark <robdclark@gmail.com>
This commit is contained in:

committed by
Rob Clark

vanhempi
84511abc47
commit
6a41da17e8
@@ -481,7 +481,7 @@ struct msm_gpu *a3xx_gpu_init(struct drm_device *dev)
|
||||
int ret;
|
||||
|
||||
if (!pdev) {
|
||||
dev_err(dev->dev, "no a3xx device\n");
|
||||
DRM_DEV_ERROR(dev->dev, "no a3xx device\n");
|
||||
ret = -ENXIO;
|
||||
goto fail;
|
||||
}
|
||||
@@ -528,7 +528,7 @@ struct msm_gpu *a3xx_gpu_init(struct drm_device *dev)
|
||||
* to not be possible to restrict access, then we must
|
||||
* implement a cmdstream validator.
|
||||
*/
|
||||
dev_err(dev->dev, "No memory protection without IOMMU\n");
|
||||
DRM_DEV_ERROR(dev->dev, "No memory protection without IOMMU\n");
|
||||
ret = -ENXIO;
|
||||
goto fail;
|
||||
}
|
||||
|
@@ -561,7 +561,7 @@ struct msm_gpu *a4xx_gpu_init(struct drm_device *dev)
|
||||
int ret;
|
||||
|
||||
if (!pdev) {
|
||||
dev_err(dev->dev, "no a4xx device\n");
|
||||
DRM_DEV_ERROR(dev->dev, "no a4xx device\n");
|
||||
ret = -ENXIO;
|
||||
goto fail;
|
||||
}
|
||||
@@ -608,7 +608,7 @@ struct msm_gpu *a4xx_gpu_init(struct drm_device *dev)
|
||||
* to not be possible to restrict access, then we must
|
||||
* implement a cmdstream validator.
|
||||
*/
|
||||
dev_err(dev->dev, "No memory protection without IOMMU\n");
|
||||
DRM_DEV_ERROR(dev->dev, "No memory protection without IOMMU\n");
|
||||
ret = -ENXIO;
|
||||
goto fail;
|
||||
}
|
||||
|
@@ -173,7 +173,7 @@ int a5xx_debugfs_init(struct msm_gpu *gpu, struct drm_minor *minor)
|
||||
minor->debugfs_root, minor);
|
||||
|
||||
if (ret) {
|
||||
dev_err(dev->dev, "could not install a5xx_debugfs_list\n");
|
||||
DRM_DEV_ERROR(dev->dev, "could not install a5xx_debugfs_list\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@@ -514,7 +514,7 @@ static int a5xx_ucode_init(struct msm_gpu *gpu)
|
||||
if (IS_ERR(a5xx_gpu->pm4_bo)) {
|
||||
ret = PTR_ERR(a5xx_gpu->pm4_bo);
|
||||
a5xx_gpu->pm4_bo = NULL;
|
||||
dev_err(gpu->dev->dev, "could not allocate PM4: %d\n",
|
||||
DRM_DEV_ERROR(gpu->dev->dev, "could not allocate PM4: %d\n",
|
||||
ret);
|
||||
return ret;
|
||||
}
|
||||
@@ -527,7 +527,7 @@ static int a5xx_ucode_init(struct msm_gpu *gpu)
|
||||
if (IS_ERR(a5xx_gpu->pfp_bo)) {
|
||||
ret = PTR_ERR(a5xx_gpu->pfp_bo);
|
||||
a5xx_gpu->pfp_bo = NULL;
|
||||
dev_err(gpu->dev->dev, "could not allocate PFP: %d\n",
|
||||
DRM_DEV_ERROR(gpu->dev->dev, "could not allocate PFP: %d\n",
|
||||
ret);
|
||||
return ret;
|
||||
}
|
||||
@@ -1028,7 +1028,7 @@ static void a5xx_fault_detect_irq(struct msm_gpu *gpu)
|
||||
struct msm_drm_private *priv = dev->dev_private;
|
||||
struct msm_ringbuffer *ring = gpu->funcs->active_ring(gpu);
|
||||
|
||||
dev_err(dev->dev, "gpu fault ring %d fence %x status %8.8X rb %4.4x/%4.4x ib1 %16.16llX/%4.4x ib2 %16.16llX/%4.4x\n",
|
||||
DRM_DEV_ERROR(dev->dev, "gpu fault ring %d fence %x status %8.8X rb %4.4x/%4.4x ib1 %16.16llX/%4.4x ib2 %16.16llX/%4.4x\n",
|
||||
ring ? ring->id : -1, ring ? ring->seqno : 0,
|
||||
gpu_read(gpu, REG_A5XX_RBBM_STATUS),
|
||||
gpu_read(gpu, REG_A5XX_CP_RB_RPTR),
|
||||
@@ -1134,7 +1134,7 @@ static const u32 a5xx_registers[] = {
|
||||
|
||||
static void a5xx_dump(struct msm_gpu *gpu)
|
||||
{
|
||||
dev_info(gpu->dev->dev, "status: %08x\n",
|
||||
DRM_DEV_INFO(gpu->dev->dev, "status: %08x\n",
|
||||
gpu_read(gpu, REG_A5XX_RBBM_STATUS));
|
||||
adreno_dump(gpu);
|
||||
}
|
||||
@@ -1505,7 +1505,7 @@ struct msm_gpu *a5xx_gpu_init(struct drm_device *dev)
|
||||
int ret;
|
||||
|
||||
if (!pdev) {
|
||||
dev_err(dev->dev, "No A5XX device is defined\n");
|
||||
DRM_DEV_ERROR(dev->dev, "No A5XX device is defined\n");
|
||||
return ERR_PTR(-ENXIO);
|
||||
}
|
||||
|
||||
|
@@ -92,7 +92,7 @@ static void a5xx_preempt_timer(struct timer_list *t)
|
||||
if (!try_preempt_state(a5xx_gpu, PREEMPT_TRIGGERED, PREEMPT_FAULTED))
|
||||
return;
|
||||
|
||||
dev_err(dev->dev, "%s: preemption timed out\n", gpu->name);
|
||||
DRM_DEV_ERROR(dev->dev, "%s: preemption timed out\n", gpu->name);
|
||||
queue_work(priv->wq, &gpu->recover_work);
|
||||
}
|
||||
|
||||
@@ -188,7 +188,7 @@ void a5xx_preempt_irq(struct msm_gpu *gpu)
|
||||
status = gpu_read(gpu, REG_A5XX_CP_CONTEXT_SWITCH_CNTL);
|
||||
if (unlikely(status)) {
|
||||
set_preempt_state(a5xx_gpu, PREEMPT_FAULTED);
|
||||
dev_err(dev->dev, "%s: Preemption failed to complete\n",
|
||||
DRM_DEV_ERROR(dev->dev, "%s: Preemption failed to complete\n",
|
||||
gpu->name);
|
||||
queue_work(priv->wq, &gpu->recover_work);
|
||||
return;
|
||||
|
@@ -153,7 +153,7 @@ static int a6xx_gmu_start(struct a6xx_gmu *gmu)
|
||||
val == 0xbabeface, 100, 10000);
|
||||
|
||||
if (ret)
|
||||
dev_err(gmu->dev, "GMU firmware initialization timed out\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "GMU firmware initialization timed out\n");
|
||||
|
||||
return ret;
|
||||
}
|
||||
@@ -168,7 +168,7 @@ static int a6xx_gmu_hfi_start(struct a6xx_gmu *gmu)
|
||||
ret = gmu_poll_timeout(gmu, REG_A6XX_GMU_HFI_CTRL_STATUS, val,
|
||||
val & 1, 100, 10000);
|
||||
if (ret)
|
||||
dev_err(gmu->dev, "Unable to start the HFI queues\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to start the HFI queues\n");
|
||||
|
||||
return ret;
|
||||
}
|
||||
@@ -209,7 +209,7 @@ int a6xx_gmu_set_oob(struct a6xx_gmu *gmu, enum a6xx_gmu_oob_state state)
|
||||
val & (1 << ack), 100, 10000);
|
||||
|
||||
if (ret)
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"Timeout waiting for GMU OOB set %s: 0x%x\n",
|
||||
name,
|
||||
gmu_read(gmu, REG_A6XX_GMU_GMU2HOST_INTR_INFO));
|
||||
@@ -251,7 +251,7 @@ static int a6xx_sptprac_enable(struct a6xx_gmu *gmu)
|
||||
(val & 0x38) == 0x28, 1, 100);
|
||||
|
||||
if (ret) {
|
||||
dev_err(gmu->dev, "Unable to power on SPTPRAC: 0x%x\n",
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to power on SPTPRAC: 0x%x\n",
|
||||
gmu_read(gmu, REG_A6XX_GMU_SPTPRAC_PWR_CLK_STATUS));
|
||||
}
|
||||
|
||||
@@ -273,7 +273,7 @@ static void a6xx_sptprac_disable(struct a6xx_gmu *gmu)
|
||||
(val & 0x04), 100, 10000);
|
||||
|
||||
if (ret)
|
||||
dev_err(gmu->dev, "failed to power off SPTPRAC: 0x%x\n",
|
||||
DRM_DEV_ERROR(gmu->dev, "failed to power off SPTPRAC: 0x%x\n",
|
||||
gmu_read(gmu, REG_A6XX_GMU_SPTPRAC_PWR_CLK_STATUS));
|
||||
}
|
||||
|
||||
@@ -317,7 +317,7 @@ static int a6xx_gmu_notify_slumber(struct a6xx_gmu *gmu)
|
||||
/* Check to see if the GMU really did slumber */
|
||||
if (gmu_read(gmu, REG_A6XX_GPU_GMU_CX_GMU_RPMH_POWER_STATE)
|
||||
!= 0x0f) {
|
||||
dev_err(gmu->dev, "The GMU did not go into slumber\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "The GMU did not go into slumber\n");
|
||||
ret = -ETIMEDOUT;
|
||||
}
|
||||
}
|
||||
@@ -339,7 +339,7 @@ static int a6xx_rpmh_start(struct a6xx_gmu *gmu)
|
||||
ret = gmu_poll_timeout(gmu, REG_A6XX_GMU_RSCC_CONTROL_ACK, val,
|
||||
val & (1 << 1), 100, 10000);
|
||||
if (ret) {
|
||||
dev_err(gmu->dev, "Unable to power on the GPU RSC\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to power on the GPU RSC\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -354,7 +354,7 @@ static int a6xx_rpmh_start(struct a6xx_gmu *gmu)
|
||||
return 0;
|
||||
}
|
||||
|
||||
dev_err(gmu->dev, "GPU RSC sequence stuck while waking up the GPU\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "GPU RSC sequence stuck while waking up the GPU\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -368,7 +368,7 @@ static void a6xx_rpmh_stop(struct a6xx_gmu *gmu)
|
||||
ret = gmu_poll_timeout(gmu, REG_A6XX_GPU_RSCC_RSC_STATUS0_DRV0,
|
||||
val, val & (1 << 16), 100, 10000);
|
||||
if (ret)
|
||||
dev_err(gmu->dev, "Unable to power off the GPU RSC\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to power off the GPU RSC\n");
|
||||
|
||||
gmu_write(gmu, REG_A6XX_GMU_RSCC_CONTROL_REQ, 0);
|
||||
}
|
||||
@@ -520,7 +520,7 @@ static int a6xx_gmu_fw_start(struct a6xx_gmu *gmu, unsigned int state)
|
||||
|
||||
/* Sanity check the size of the firmware that was loaded */
|
||||
if (adreno_gpu->fw[ADRENO_FW_GMU]->size > 0x8000) {
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"GMU firmware is bigger than the available region\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
@@ -764,7 +764,7 @@ int a6xx_gmu_stop(struct a6xx_gpu *a6xx_gpu)
|
||||
*/
|
||||
|
||||
if (ret)
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"Unable to slumber GMU: status = 0%x/0%x\n",
|
||||
gmu_read(gmu,
|
||||
REG_A6XX_GPU_GMU_AO_GPU_CX_BUSY_STATUS),
|
||||
@@ -843,7 +843,7 @@ static struct a6xx_gmu_bo *a6xx_gmu_memory_alloc(struct a6xx_gmu *gmu,
|
||||
IOMMU_READ | IOMMU_WRITE);
|
||||
|
||||
if (ret) {
|
||||
dev_err(gmu->dev, "Unable to map GMU buffer object\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to map GMU buffer object\n");
|
||||
|
||||
for (i = i - 1 ; i >= 0; i--)
|
||||
iommu_unmap(gmu->domain,
|
||||
@@ -969,12 +969,12 @@ static int a6xx_gmu_rpmh_arc_votes_init(struct device *dev, u32 *votes,
|
||||
}
|
||||
|
||||
if (j == pri_count) {
|
||||
dev_err(dev,
|
||||
DRM_DEV_ERROR(dev,
|
||||
"Level %u not found in in the RPMh list\n",
|
||||
level);
|
||||
dev_err(dev, "Available levels:\n");
|
||||
DRM_DEV_ERROR(dev, "Available levels:\n");
|
||||
for (j = 0; j < pri_count; j++)
|
||||
dev_err(dev, " %u\n", pri[j]);
|
||||
DRM_DEV_ERROR(dev, " %u\n", pri[j]);
|
||||
|
||||
return -EINVAL;
|
||||
}
|
||||
@@ -1081,7 +1081,7 @@ static int a6xx_gmu_pwrlevels_probe(struct a6xx_gmu *gmu)
|
||||
*/
|
||||
ret = dev_pm_opp_of_add_table(gmu->dev);
|
||||
if (ret) {
|
||||
dev_err(gmu->dev, "Unable to set the OPP table for the GMU\n");
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to set the OPP table for the GMU\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -1122,13 +1122,13 @@ static void __iomem *a6xx_gmu_get_mmio(struct platform_device *pdev,
|
||||
IORESOURCE_MEM, name);
|
||||
|
||||
if (!res) {
|
||||
dev_err(&pdev->dev, "Unable to find the %s registers\n", name);
|
||||
DRM_DEV_ERROR(&pdev->dev, "Unable to find the %s registers\n", name);
|
||||
return ERR_PTR(-EINVAL);
|
||||
}
|
||||
|
||||
ret = devm_ioremap(&pdev->dev, res->start, resource_size(res));
|
||||
if (!ret) {
|
||||
dev_err(&pdev->dev, "Unable to map the %s registers\n", name);
|
||||
DRM_DEV_ERROR(&pdev->dev, "Unable to map the %s registers\n", name);
|
||||
return ERR_PTR(-EINVAL);
|
||||
}
|
||||
|
||||
@@ -1145,7 +1145,7 @@ static int a6xx_gmu_get_irq(struct a6xx_gmu *gmu, struct platform_device *pdev,
|
||||
ret = devm_request_irq(&pdev->dev, irq, handler, IRQF_TRIGGER_HIGH,
|
||||
name, gmu);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev, "Unable to get interrupt %s\n", name);
|
||||
DRM_DEV_ERROR(&pdev->dev, "Unable to get interrupt %s\n", name);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@@ -481,7 +481,7 @@ out:
|
||||
|
||||
static void a6xx_dump(struct msm_gpu *gpu)
|
||||
{
|
||||
dev_info(&gpu->pdev->dev, "status: %08x\n",
|
||||
DRM_DEV_INFO(&gpu->pdev->dev, "status: %08x\n",
|
||||
gpu_read(gpu, REG_A6XX_RBBM_STATUS));
|
||||
adreno_dump(gpu);
|
||||
}
|
||||
@@ -498,7 +498,7 @@ static void a6xx_recover(struct msm_gpu *gpu)
|
||||
adreno_dump_info(gpu);
|
||||
|
||||
for (i = 0; i < 8; i++)
|
||||
dev_info(&gpu->pdev->dev, "CP_SCRATCH_REG%d: %u\n", i,
|
||||
DRM_DEV_INFO(&gpu->pdev->dev, "CP_SCRATCH_REG%d: %u\n", i,
|
||||
gpu_read(gpu, REG_A6XX_CP_SCRATCH_REG(i)));
|
||||
|
||||
if (hang_debug)
|
||||
|
@@ -91,7 +91,7 @@ static int a6xx_hfi_wait_for_ack(struct a6xx_gmu *gmu, u32 id, u32 seqnum,
|
||||
val & A6XX_GMU_GMU2HOST_INTR_INFO_MSGQ, 100, 5000);
|
||||
|
||||
if (ret) {
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"Message %s id %d timed out waiting for response\n",
|
||||
a6xx_hfi_msg_id[id], seqnum);
|
||||
return -ETIMEDOUT;
|
||||
@@ -110,7 +110,7 @@ static int a6xx_hfi_wait_for_ack(struct a6xx_gmu *gmu, u32 id, u32 seqnum,
|
||||
|
||||
/* If the queue is empty our response never made it */
|
||||
if (!ret) {
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"The HFI response queue is unexpectedly empty\n");
|
||||
|
||||
return -ENOENT;
|
||||
@@ -120,20 +120,20 @@ static int a6xx_hfi_wait_for_ack(struct a6xx_gmu *gmu, u32 id, u32 seqnum,
|
||||
struct a6xx_hfi_msg_error *error =
|
||||
(struct a6xx_hfi_msg_error *) &resp;
|
||||
|
||||
dev_err(gmu->dev, "GMU firmware error %d\n",
|
||||
DRM_DEV_ERROR(gmu->dev, "GMU firmware error %d\n",
|
||||
error->code);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (seqnum != HFI_HEADER_SEQNUM(resp.ret_header)) {
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"Unexpected message id %d on the response queue\n",
|
||||
HFI_HEADER_SEQNUM(resp.ret_header));
|
||||
continue;
|
||||
}
|
||||
|
||||
if (resp.error) {
|
||||
dev_err(gmu->dev,
|
||||
DRM_DEV_ERROR(gmu->dev,
|
||||
"Message %s id %d returned error %d\n",
|
||||
a6xx_hfi_msg_id[id], seqnum, resp.error);
|
||||
return -EINVAL;
|
||||
@@ -163,7 +163,7 @@ static int a6xx_hfi_send_msg(struct a6xx_gmu *gmu, int id,
|
||||
|
||||
ret = a6xx_hfi_queue_write(gmu, queue, data, dwords);
|
||||
if (ret) {
|
||||
dev_err(gmu->dev, "Unable to send message %s id %d\n",
|
||||
DRM_DEV_ERROR(gmu->dev, "Unable to send message %s id %d\n",
|
||||
a6xx_hfi_msg_id[id], seqnum);
|
||||
return ret;
|
||||
}
|
||||
@@ -317,7 +317,7 @@ void a6xx_hfi_stop(struct a6xx_gmu *gmu)
|
||||
continue;
|
||||
|
||||
if (queue->header->read_index != queue->header->write_index)
|
||||
dev_err(gmu->dev, "HFI queue %d is not empty\n", i);
|
||||
DRM_DEV_ERROR(gmu->dev, "HFI queue %d is not empty\n", i);
|
||||
|
||||
queue->header->read_index = 0;
|
||||
queue->header->write_index = 0;
|
||||
|
@@ -196,7 +196,7 @@ struct msm_gpu *adreno_load_gpu(struct drm_device *dev)
|
||||
|
||||
ret = pm_runtime_get_sync(&pdev->dev);
|
||||
if (ret < 0) {
|
||||
dev_err(dev->dev, "Couldn't power up the GPU: %d\n", ret);
|
||||
DRM_DEV_ERROR(dev->dev, "Couldn't power up the GPU: %d\n", ret);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
@@ -205,7 +205,7 @@ struct msm_gpu *adreno_load_gpu(struct drm_device *dev)
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
pm_runtime_put_autosuspend(&pdev->dev);
|
||||
if (ret) {
|
||||
dev_err(dev->dev, "gpu hw init failed: %d\n", ret);
|
||||
DRM_DEV_ERROR(dev->dev, "gpu hw init failed: %d\n", ret);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
@@ -253,7 +253,7 @@ static int find_chipid(struct device *dev, struct adreno_rev *rev)
|
||||
/* and if that fails, fall back to legacy "qcom,chipid" property: */
|
||||
ret = of_property_read_u32(node, "qcom,chipid", &chipid);
|
||||
if (ret) {
|
||||
dev_err(dev, "could not parse qcom,chipid: %d\n", ret);
|
||||
DRM_DEV_ERROR(dev, "could not parse qcom,chipid: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@@ -89,12 +89,12 @@ adreno_request_fw(struct adreno_gpu *adreno_gpu, const char *fwname)
|
||||
|
||||
ret = request_firmware_direct(&fw, newname, drm->dev);
|
||||
if (!ret) {
|
||||
dev_info(drm->dev, "loaded %s from new location\n",
|
||||
DRM_DEV_INFO(drm->dev, "loaded %s from new location\n",
|
||||
newname);
|
||||
adreno_gpu->fwloc = FW_LOCATION_NEW;
|
||||
goto out;
|
||||
} else if (adreno_gpu->fwloc != FW_LOCATION_UNKNOWN) {
|
||||
dev_err(drm->dev, "failed to load %s: %d\n",
|
||||
DRM_DEV_ERROR(drm->dev, "failed to load %s: %d\n",
|
||||
newname, ret);
|
||||
fw = ERR_PTR(ret);
|
||||
goto out;
|
||||
@@ -109,12 +109,12 @@ adreno_request_fw(struct adreno_gpu *adreno_gpu, const char *fwname)
|
||||
|
||||
ret = request_firmware_direct(&fw, fwname, drm->dev);
|
||||
if (!ret) {
|
||||
dev_info(drm->dev, "loaded %s from legacy location\n",
|
||||
DRM_DEV_INFO(drm->dev, "loaded %s from legacy location\n",
|
||||
newname);
|
||||
adreno_gpu->fwloc = FW_LOCATION_LEGACY;
|
||||
goto out;
|
||||
} else if (adreno_gpu->fwloc != FW_LOCATION_UNKNOWN) {
|
||||
dev_err(drm->dev, "failed to load %s: %d\n",
|
||||
DRM_DEV_ERROR(drm->dev, "failed to load %s: %d\n",
|
||||
fwname, ret);
|
||||
fw = ERR_PTR(ret);
|
||||
goto out;
|
||||
@@ -130,19 +130,19 @@ adreno_request_fw(struct adreno_gpu *adreno_gpu, const char *fwname)
|
||||
|
||||
ret = request_firmware(&fw, newname, drm->dev);
|
||||
if (!ret) {
|
||||
dev_info(drm->dev, "loaded %s with helper\n",
|
||||
DRM_DEV_INFO(drm->dev, "loaded %s with helper\n",
|
||||
newname);
|
||||
adreno_gpu->fwloc = FW_LOCATION_HELPER;
|
||||
goto out;
|
||||
} else if (adreno_gpu->fwloc != FW_LOCATION_UNKNOWN) {
|
||||
dev_err(drm->dev, "failed to load %s: %d\n",
|
||||
DRM_DEV_ERROR(drm->dev, "failed to load %s: %d\n",
|
||||
newname, ret);
|
||||
fw = ERR_PTR(ret);
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
dev_err(drm->dev, "failed to load %s\n", fwname);
|
||||
DRM_DEV_ERROR(drm->dev, "failed to load %s\n", fwname);
|
||||
fw = ERR_PTR(-ENOENT);
|
||||
out:
|
||||
kfree(newname);
|
||||
@@ -212,7 +212,7 @@ int adreno_hw_init(struct msm_gpu *gpu)
|
||||
ret = msm_gem_get_iova(ring->bo, gpu->aspace, &ring->iova);
|
||||
if (ret) {
|
||||
ring->iova = 0;
|
||||
dev_err(gpu->dev->dev,
|
||||
DRM_DEV_ERROR(gpu->dev->dev,
|
||||
"could not map ringbuffer %d: %d\n", i, ret);
|
||||
return ret;
|
||||
}
|
||||
@@ -277,7 +277,7 @@ void adreno_recover(struct msm_gpu *gpu)
|
||||
|
||||
ret = msm_gpu_hw_init(gpu);
|
||||
if (ret) {
|
||||
dev_err(dev->dev, "gpu hw init failed: %d\n", ret);
|
||||
DRM_DEV_ERROR(dev->dev, "gpu hw init failed: %d\n", ret);
|
||||
/* hmm, oh well? */
|
||||
}
|
||||
}
|
||||
@@ -635,7 +635,7 @@ static int adreno_get_legacy_pwrlevels(struct device *dev)
|
||||
|
||||
node = of_get_compatible_child(dev->of_node, "qcom,gpu-pwrlevels");
|
||||
if (!node) {
|
||||
dev_err(dev, "Could not find the GPU powerlevels\n");
|
||||
DRM_DEV_ERROR(dev, "Could not find the GPU powerlevels\n");
|
||||
return -ENXIO;
|
||||
}
|
||||
|
||||
@@ -674,7 +674,7 @@ static int adreno_get_pwrlevels(struct device *dev,
|
||||
else {
|
||||
ret = dev_pm_opp_of_add_table(dev);
|
||||
if (ret)
|
||||
dev_err(dev, "Unable to set the OPP table\n");
|
||||
DRM_DEV_ERROR(dev, "Unable to set the OPP table\n");
|
||||
}
|
||||
|
||||
if (!ret) {
|
||||
|
Viittaa uudesa ongelmassa
Block a user