123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120 |
- /* SPDX-License-Identifier: GPL-2.0-only */
- /*
- * Copyright (C) 2013 Red Hat
- * Author: Rob Clark <[email protected]>
- */
- #ifndef __MSM_RINGBUFFER_H__
- #define __MSM_RINGBUFFER_H__
- #include "drm/gpu_scheduler.h"
- #include "msm_drv.h"
- #define rbmemptr(ring, member) \
- ((ring)->memptrs_iova + offsetof(struct msm_rbmemptrs, member))
- #define rbmemptr_stats(ring, index, member) \
- (rbmemptr((ring), stats) + \
- ((index) * sizeof(struct msm_gpu_submit_stats)) + \
- offsetof(struct msm_gpu_submit_stats, member))
- struct msm_gpu_submit_stats {
- u64 cpcycles_start;
- u64 cpcycles_end;
- u64 alwayson_start;
- u64 alwayson_end;
- };
- #define MSM_GPU_SUBMIT_STATS_COUNT 64
- struct msm_rbmemptrs {
- volatile uint32_t rptr;
- volatile uint32_t fence;
- volatile struct msm_gpu_submit_stats stats[MSM_GPU_SUBMIT_STATS_COUNT];
- volatile u64 ttbr0;
- };
- struct msm_cp_state {
- uint64_t ib1_base, ib2_base;
- uint32_t ib1_rem, ib2_rem;
- };
- struct msm_ringbuffer {
- struct msm_gpu *gpu;
- int id;
- struct drm_gem_object *bo;
- uint32_t *start, *end, *cur, *next;
- /*
- * The job scheduler for this ring.
- */
- struct drm_gpu_scheduler sched;
- /*
- * List of in-flight submits on this ring. Protected by submit_lock.
- *
- * Currently just submits that are already written into the ring, not
- * submits that are still in drm_gpu_scheduler's queues. At a later
- * step we could probably move to letting drm_gpu_scheduler manage
- * hangcheck detection and keep track of submit jobs that are in-
- * flight.
- */
- struct list_head submits;
- spinlock_t submit_lock;
- uint64_t iova;
- uint32_t hangcheck_fence;
- struct msm_rbmemptrs *memptrs;
- uint64_t memptrs_iova;
- struct msm_fence_context *fctx;
- /**
- * hangcheck_progress_retries:
- *
- * The number of extra hangcheck duration cycles that we have given
- * due to it appearing that the GPU is making forward progress.
- *
- * For GPU generations which support progress detection (see.
- * msm_gpu_funcs::progress()), if the GPU appears to be making progress
- * (ie. the CP has advanced in the command stream, we'll allow up to
- * DRM_MSM_HANGCHECK_PROGRESS_RETRIES expirations of the hangcheck timer
- * before killing the job. But to detect progress we need two sample
- * points, so the duration of the hangcheck timer is halved. In other
- * words we'll let the submit run for up to:
- *
- * (DRM_MSM_HANGCHECK_DEFAULT_PERIOD / 2) * (DRM_MSM_HANGCHECK_PROGRESS_RETRIES + 1)
- */
- int hangcheck_progress_retries;
- /**
- * last_cp_state: The state of the CP at the last call to gpu->progress()
- */
- struct msm_cp_state last_cp_state;
- /*
- * preempt_lock protects preemption and serializes wptr updates against
- * preemption. Can be aquired from irq context.
- */
- spinlock_t preempt_lock;
- };
- struct msm_ringbuffer *msm_ringbuffer_new(struct msm_gpu *gpu, int id,
- void *memptrs, uint64_t memptrs_iova);
- void msm_ringbuffer_destroy(struct msm_ringbuffer *ring);
- /* ringbuffer helpers (the parts that are same for a3xx/a2xx/z180..) */
- static inline void
- OUT_RING(struct msm_ringbuffer *ring, uint32_t data)
- {
- /*
- * ring->next points to the current command being written - it won't be
- * committed as ring->cur until the flush
- */
- if (ring->next == ring->end)
- ring->next = ring->start;
- *(ring->next++) = data;
- }
- #endif /* __MSM_RINGBUFFER_H__ */
|