[media] media: videobuf2: Restructure vb2_buffer
Remove v4l2 stuff - v4l2_buf, v4l2_plane - from struct vb2_buffer. Add new member variables - bytesused, length, offset, userptr, fd, data_offset - to struct vb2_plane in order to cover all information of v4l2_plane. struct vb2_plane { <snip> unsigned int bytesused; unsigned int length; union { unsigned int offset; unsigned long userptr; int fd; } m; unsigned int data_offset; } Replace v4l2_buf with new member variables - index, type, memory - which are common fields for buffer management. struct vb2_buffer { <snip> unsigned int index; unsigned int type; unsigned int memory; unsigned int num_planes; struct vb2_plane planes[VIDEO_MAX_PLANES]; <snip> }; v4l2 specific fields - flags, field, timestamp, timecode, sequence - are moved to vb2_v4l2_buffer in videobuf2-v4l2.c struct vb2_v4l2_buffer { struct vb2_buffer vb2_buf; __u32 flags; __u32 field; struct timeval timestamp; struct v4l2_timecode timecode; __u32 sequence; }; Signed-off-by: Junghak Sung <jh1009.sung@samsung.com> Signed-off-by: Geunyoung Kim <nenggun.kim@samsung.com> Acked-by: Seung-Woo Kim <sw0312.kim@samsung.com> Acked-by: Inki Dae <inki.dae@samsung.com> Signed-off-by: Hans Verkuil <hans.verkuil@cisco.com> Signed-off-by: Mauro Carvalho Chehab <mchehab@osg.samsung.com>
This commit is contained in:

committed by
Mauro Carvalho Chehab

parent
c139990e84
commit
2d7007153f
@@ -35,6 +35,7 @@
|
||||
#include <media/v4l2-ioctl.h>
|
||||
#include <media/v4l2-device.h>
|
||||
#include <media/v4l2-fh.h>
|
||||
#include <media/videobuf2-v4l2.h>
|
||||
#include <media/videobuf2-dma-sg.h>
|
||||
|
||||
#include "m00233_video_measure_memmap_package.h"
|
||||
@@ -206,11 +207,12 @@ struct sg_dma_desc_info {
|
||||
#define COBALT_STREAM_FL_ADV_IRQ 1
|
||||
|
||||
struct cobalt_buffer {
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head list;
|
||||
};
|
||||
|
||||
static inline struct cobalt_buffer *to_cobalt_buffer(struct vb2_buffer *vb2)
|
||||
static inline
|
||||
struct cobalt_buffer *to_cobalt_buffer(struct vb2_v4l2_buffer *vb2)
|
||||
{
|
||||
return container_of(vb2, struct cobalt_buffer, vb);
|
||||
}
|
||||
|
@@ -134,11 +134,12 @@ done:
|
||||
skip = true;
|
||||
s->skip_first_frames--;
|
||||
}
|
||||
v4l2_get_timestamp(&cb->vb.v4l2_buf.timestamp);
|
||||
v4l2_get_timestamp(&cb->vb.timestamp);
|
||||
/* TODO: the sequence number should be read from the FPGA so we
|
||||
also know about dropped frames. */
|
||||
cb->vb.v4l2_buf.sequence = s->sequence++;
|
||||
vb2_buffer_done(&cb->vb, (skip || s->unstable_frame) ?
|
||||
cb->vb.sequence = s->sequence++;
|
||||
vb2_buffer_done(&cb->vb.vb2_buf,
|
||||
(skip || s->unstable_frame) ?
|
||||
VB2_BUF_STATE_REQUEUEING : VB2_BUF_STATE_DONE);
|
||||
}
|
||||
|
||||
|
@@ -75,7 +75,7 @@ static int cobalt_buf_init(struct vb2_buffer *vb)
|
||||
const size_t bytes =
|
||||
COBALT_MAX_HEIGHT * max_pages_per_line * 0x20;
|
||||
const size_t audio_bytes = ((1920 * 4) / PAGE_SIZE + 1) * 0x20;
|
||||
struct sg_dma_desc_info *desc = &s->dma_desc_info[vb->v4l2_buf.index];
|
||||
struct sg_dma_desc_info *desc = &s->dma_desc_info[vb->index];
|
||||
struct sg_table *sg_desc = vb2_dma_sg_plane_desc(vb, 0);
|
||||
unsigned size;
|
||||
int ret;
|
||||
@@ -105,17 +105,18 @@ static int cobalt_buf_init(struct vb2_buffer *vb)
|
||||
static void cobalt_buf_cleanup(struct vb2_buffer *vb)
|
||||
{
|
||||
struct cobalt_stream *s = vb->vb2_queue->drv_priv;
|
||||
struct sg_dma_desc_info *desc = &s->dma_desc_info[vb->v4l2_buf.index];
|
||||
struct sg_dma_desc_info *desc = &s->dma_desc_info[vb->index];
|
||||
|
||||
descriptor_list_free(desc);
|
||||
}
|
||||
|
||||
static int cobalt_buf_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cobalt_stream *s = vb->vb2_queue->drv_priv;
|
||||
|
||||
vb2_set_plane_payload(vb, 0, s->stride * s->height);
|
||||
vb->v4l2_buf.field = V4L2_FIELD_NONE;
|
||||
vbuf->field = V4L2_FIELD_NONE;
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -128,7 +129,7 @@ static void chain_all_buffers(struct cobalt_stream *s)
|
||||
|
||||
list_for_each(p, &s->bufs) {
|
||||
cb = list_entry(p, struct cobalt_buffer, list);
|
||||
desc[i] = &s->dma_desc_info[cb->vb.v4l2_buf.index];
|
||||
desc[i] = &s->dma_desc_info[cb->vb.vb2_buf.index];
|
||||
if (i > 0)
|
||||
descriptor_list_chain(desc[i-1], desc[i]);
|
||||
i++;
|
||||
@@ -137,10 +138,11 @@ static void chain_all_buffers(struct cobalt_stream *s)
|
||||
|
||||
static void cobalt_buf_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vb2_queue *q = vb->vb2_queue;
|
||||
struct cobalt_stream *s = q->drv_priv;
|
||||
struct cobalt_buffer *cb = to_cobalt_buffer(vb);
|
||||
struct sg_dma_desc_info *desc = &s->dma_desc_info[vb->v4l2_buf.index];
|
||||
struct cobalt_buffer *cb = to_cobalt_buffer(vbuf);
|
||||
struct sg_dma_desc_info *desc = &s->dma_desc_info[vb->index];
|
||||
unsigned long flags;
|
||||
|
||||
/* Prepare new buffer */
|
||||
@@ -284,7 +286,7 @@ static void cobalt_dma_start_streaming(struct cobalt_stream *s)
|
||||
&vo->control);
|
||||
}
|
||||
cb = list_first_entry(&s->bufs, struct cobalt_buffer, list);
|
||||
omni_sg_dma_start(s, &s->dma_desc_info[cb->vb.v4l2_buf.index]);
|
||||
omni_sg_dma_start(s, &s->dma_desc_info[cb->vb.vb2_buf.index]);
|
||||
spin_unlock_irqrestore(&s->irqlock, flags);
|
||||
}
|
||||
|
||||
@@ -381,7 +383,7 @@ static void cobalt_dma_stop_streaming(struct cobalt_stream *s)
|
||||
spin_lock_irqsave(&s->irqlock, flags);
|
||||
list_for_each(p, &s->bufs) {
|
||||
cb = list_entry(p, struct cobalt_buffer, list);
|
||||
desc = &s->dma_desc_info[cb->vb.v4l2_buf.index];
|
||||
desc = &s->dma_desc_info[cb->vb.vb2_buf.index];
|
||||
/* Stop DMA after this descriptor chain */
|
||||
descriptor_list_end_of_chain(desc);
|
||||
}
|
||||
@@ -416,7 +418,7 @@ static void cobalt_stop_streaming(struct vb2_queue *q)
|
||||
list_for_each_safe(p, safe, &s->bufs) {
|
||||
cb = list_entry(p, struct cobalt_buffer, list);
|
||||
list_del(&cb->list);
|
||||
vb2_buffer_done(&cb->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&cb->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&s->irqlock, flags);
|
||||
|
||||
|
@@ -1155,17 +1155,19 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf =
|
||||
container_of(vb, struct cx23885_buffer, vb);
|
||||
container_of(vbuf, struct cx23885_buffer, vb);
|
||||
|
||||
return cx23885_buf_prepare(buf, &dev->ts1);
|
||||
}
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
|
||||
cx23885_free_buffer(dev, buf);
|
||||
@@ -1173,8 +1175,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
|
||||
cx23885_buf_queue(&dev->ts1, buf);
|
||||
@@ -1201,7 +1204,7 @@ static int cx23885_start_streaming(struct vb2_queue *q, unsigned int count)
|
||||
struct cx23885_buffer, queue);
|
||||
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_QUEUED);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
return ret;
|
||||
|
@@ -427,12 +427,13 @@ static void cx23885_wakeup(struct cx23885_tsport *port,
|
||||
buf = list_entry(q->active.next,
|
||||
struct cx23885_buffer, queue);
|
||||
|
||||
v4l2_get_timestamp(&buf->vb.v4l2_buf.timestamp);
|
||||
buf->vb.v4l2_buf.sequence = q->count++;
|
||||
dprintk(1, "[%p/%d] wakeup reg=%d buf=%d\n", buf, buf->vb.v4l2_buf.index,
|
||||
v4l2_get_timestamp(&buf->vb.timestamp);
|
||||
buf->vb.sequence = q->count++;
|
||||
dprintk(1, "[%p/%d] wakeup reg=%d buf=%d\n", buf,
|
||||
buf->vb.vb2_buf.index,
|
||||
count, q->count);
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_DONE);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
}
|
||||
|
||||
int cx23885_sram_channel_setup(struct cx23885_dev *dev,
|
||||
@@ -1453,12 +1454,12 @@ int cx23885_buf_prepare(struct cx23885_buffer *buf, struct cx23885_tsport *port)
|
||||
{
|
||||
struct cx23885_dev *dev = port->dev;
|
||||
int size = port->ts_packet_size * port->ts_packet_count;
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb, 0);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0);
|
||||
|
||||
dprintk(1, "%s: %p\n", __func__, buf);
|
||||
if (vb2_plane_size(&buf->vb, 0) < size)
|
||||
if (vb2_plane_size(&buf->vb.vb2_buf, 0) < size)
|
||||
return -EINVAL;
|
||||
vb2_set_plane_payload(&buf->vb, 0, size);
|
||||
vb2_set_plane_payload(&buf->vb.vb2_buf, 0, size);
|
||||
|
||||
cx23885_risc_databuffer(dev->pci, &buf->risc,
|
||||
sgt->sgl,
|
||||
@@ -1503,7 +1504,7 @@ void cx23885_buf_queue(struct cx23885_tsport *port, struct cx23885_buffer *buf)
|
||||
if (list_empty(&cx88q->active)) {
|
||||
list_add_tail(&buf->queue, &cx88q->active);
|
||||
dprintk(1, "[%p/%d] %s - first active\n",
|
||||
buf, buf->vb.v4l2_buf.index, __func__);
|
||||
buf, buf->vb.vb2_buf.index, __func__);
|
||||
} else {
|
||||
buf->risc.cpu[0] |= cpu_to_le32(RISC_IRQ1);
|
||||
prev = list_entry(cx88q->active.prev, struct cx23885_buffer,
|
||||
@@ -1511,7 +1512,7 @@ void cx23885_buf_queue(struct cx23885_tsport *port, struct cx23885_buffer *buf)
|
||||
list_add_tail(&buf->queue, &cx88q->active);
|
||||
prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
|
||||
dprintk(1, "[%p/%d] %s - append to active\n",
|
||||
buf, buf->vb.v4l2_buf.index, __func__);
|
||||
buf, buf->vb.vb2_buf.index, __func__);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
@@ -1530,9 +1531,10 @@ static void do_cancel_buffers(struct cx23885_tsport *port, char *reason)
|
||||
buf = list_entry(q->active.next, struct cx23885_buffer,
|
||||
queue);
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
dprintk(1, "[%p/%d] %s - dma=0x%08lx\n",
|
||||
buf, buf->vb.v4l2_buf.index, reason, (unsigned long)buf->risc.dma);
|
||||
buf, buf->vb.vb2_buf.index, reason,
|
||||
(unsigned long)buf->risc.dma);
|
||||
}
|
||||
spin_unlock_irqrestore(&port->slock, flags);
|
||||
}
|
||||
|
@@ -110,18 +110,20 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_tsport *port = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf =
|
||||
container_of(vb, struct cx23885_buffer, vb);
|
||||
container_of(vbuf, struct cx23885_buffer, vb);
|
||||
|
||||
return cx23885_buf_prepare(buf, port);
|
||||
}
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_tsport *port = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_dev *dev = port->dev;
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
|
||||
cx23885_free_buffer(dev, buf);
|
||||
@@ -129,8 +131,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_tsport *port = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
|
||||
cx23885_buf_queue(port, buf);
|
||||
|
@@ -138,8 +138,9 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
unsigned lines = VBI_PAL_LINE_COUNT;
|
||||
@@ -161,7 +162,8 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
|
||||
cx23885_free_buffer(vb->vb2_queue->drv_priv, buf);
|
||||
@@ -190,8 +192,10 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
*/
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf = container_of(vb, struct cx23885_buffer, vb);
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
struct cx23885_buffer *prev;
|
||||
struct cx23885_dmaqueue *q = &dev->vbiq;
|
||||
unsigned long flags;
|
||||
@@ -206,7 +210,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
list_add_tail(&buf->queue, &q->active);
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
dprintk(2, "[%p/%d] vbi_queue - first active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
|
||||
} else {
|
||||
buf->risc.cpu[0] |= cpu_to_le32(RISC_IRQ1);
|
||||
@@ -217,7 +221,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
|
||||
dprintk(2, "[%p/%d] buffer_queue - append to active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -245,7 +249,7 @@ static void cx23885_stop_streaming(struct vb2_queue *q)
|
||||
struct cx23885_buffer, queue);
|
||||
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -104,12 +104,12 @@ void cx23885_video_wakeup(struct cx23885_dev *dev,
|
||||
buf = list_entry(q->active.next,
|
||||
struct cx23885_buffer, queue);
|
||||
|
||||
buf->vb.v4l2_buf.sequence = q->count++;
|
||||
v4l2_get_timestamp(&buf->vb.v4l2_buf.timestamp);
|
||||
dprintk(2, "[%p/%d] wakeup reg=%d buf=%d\n", buf, buf->vb.v4l2_buf.index,
|
||||
count, q->count);
|
||||
buf->vb.sequence = q->count++;
|
||||
v4l2_get_timestamp(&buf->vb.timestamp);
|
||||
dprintk(2, "[%p/%d] wakeup reg=%d buf=%d\n", buf,
|
||||
buf->vb.vb2_buf.index, count, q->count);
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_DONE);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
}
|
||||
|
||||
int cx23885_set_tvnorm(struct cx23885_dev *dev, v4l2_std_id norm)
|
||||
@@ -329,9 +329,10 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf =
|
||||
container_of(vb, struct cx23885_buffer, vb);
|
||||
container_of(vbuf, struct cx23885_buffer, vb);
|
||||
u32 line0_offset, line1_offset;
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
int field_tff;
|
||||
@@ -401,7 +402,7 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
BUG();
|
||||
}
|
||||
dprintk(2, "[%p/%d] buffer_init - %dx%d %dbpp \"%s\" - dma=0x%08lx\n",
|
||||
buf, buf->vb.v4l2_buf.index,
|
||||
buf, buf->vb.vb2_buf.index,
|
||||
dev->width, dev->height, dev->fmt->depth, dev->fmt->name,
|
||||
(unsigned long)buf->risc.dma);
|
||||
return 0;
|
||||
@@ -409,7 +410,8 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
|
||||
cx23885_free_buffer(vb->vb2_queue->drv_priv, buf);
|
||||
@@ -438,8 +440,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
*/
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx23885_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx23885_buffer *buf = container_of(vb,
|
||||
struct cx23885_buffer *buf = container_of(vbuf,
|
||||
struct cx23885_buffer, vb);
|
||||
struct cx23885_buffer *prev;
|
||||
struct cx23885_dmaqueue *q = &dev->vidq;
|
||||
@@ -455,7 +458,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
if (list_empty(&q->active)) {
|
||||
list_add_tail(&buf->queue, &q->active);
|
||||
dprintk(2, "[%p/%d] buffer_queue - first active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
} else {
|
||||
buf->risc.cpu[0] |= cpu_to_le32(RISC_IRQ1);
|
||||
prev = list_entry(q->active.prev, struct cx23885_buffer,
|
||||
@@ -463,7 +466,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
list_add_tail(&buf->queue, &q->active);
|
||||
prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
|
||||
dprintk(2, "[%p/%d] buffer_queue - append to active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
@@ -492,7 +495,7 @@ static void cx23885_stop_streaming(struct vb2_queue *q)
|
||||
struct cx23885_buffer, queue);
|
||||
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -170,7 +170,7 @@ struct cx23885_riscmem {
|
||||
/* buffer for one video frame */
|
||||
struct cx23885_buffer {
|
||||
/* common v4l buffer stuff -- must be first */
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head queue;
|
||||
|
||||
/* cx23885 specific */
|
||||
|
@@ -130,10 +130,10 @@ int cx25821_video_irq(struct cx25821_dev *dev, int chan_num, u32 status)
|
||||
buf = list_entry(dmaq->active.next,
|
||||
struct cx25821_buffer, queue);
|
||||
|
||||
v4l2_get_timestamp(&buf->vb.v4l2_buf.timestamp);
|
||||
buf->vb.v4l2_buf.sequence = dmaq->count++;
|
||||
v4l2_get_timestamp(&buf->vb.timestamp);
|
||||
buf->vb.sequence = dmaq->count++;
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_DONE);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
}
|
||||
spin_unlock(&dev->slock);
|
||||
handled++;
|
||||
@@ -159,10 +159,11 @@ static int cx25821_queue_setup(struct vb2_queue *q, const struct v4l2_format *fm
|
||||
|
||||
static int cx25821_buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx25821_channel *chan = vb->vb2_queue->drv_priv;
|
||||
struct cx25821_dev *dev = chan->dev;
|
||||
struct cx25821_buffer *buf =
|
||||
container_of(vb, struct cx25821_buffer, vb);
|
||||
container_of(vbuf, struct cx25821_buffer, vb);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
u32 line0_offset;
|
||||
int bpl_local = LINE_SIZE_D1;
|
||||
@@ -176,7 +177,7 @@ static int cx25821_buffer_prepare(struct vb2_buffer *vb)
|
||||
if (vb2_plane_size(vb, 0) < chan->height * buf->bpl)
|
||||
return -EINVAL;
|
||||
vb2_set_plane_payload(vb, 0, chan->height * buf->bpl);
|
||||
buf->vb.v4l2_buf.field = chan->field;
|
||||
buf->vb.field = chan->field;
|
||||
|
||||
if (chan->pixel_formats == PIXEL_FRMT_411) {
|
||||
bpl_local = buf->bpl;
|
||||
@@ -231,7 +232,7 @@ static int cx25821_buffer_prepare(struct vb2_buffer *vb)
|
||||
}
|
||||
|
||||
dprintk(2, "[%p/%d] buffer_prep - %dx%d %dbpp \"%s\" - dma=0x%08lx\n",
|
||||
buf, buf->vb.v4l2_buf.index, chan->width, chan->height,
|
||||
buf, buf->vb.vb2_buf.index, chan->width, chan->height,
|
||||
chan->fmt->depth, chan->fmt->name,
|
||||
(unsigned long)buf->risc.dma);
|
||||
|
||||
@@ -240,8 +241,9 @@ static int cx25821_buffer_prepare(struct vb2_buffer *vb)
|
||||
|
||||
static void cx25821_buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx25821_buffer *buf =
|
||||
container_of(vb, struct cx25821_buffer, vb);
|
||||
container_of(vbuf, struct cx25821_buffer, vb);
|
||||
struct cx25821_channel *chan = vb->vb2_queue->drv_priv;
|
||||
struct cx25821_dev *dev = chan->dev;
|
||||
|
||||
@@ -250,8 +252,9 @@ static void cx25821_buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void cx25821_buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx25821_buffer *buf =
|
||||
container_of(vb, struct cx25821_buffer, vb);
|
||||
container_of(vbuf, struct cx25821_buffer, vb);
|
||||
struct cx25821_channel *chan = vb->vb2_queue->drv_priv;
|
||||
struct cx25821_dev *dev = chan->dev;
|
||||
struct cx25821_buffer *prev;
|
||||
@@ -300,7 +303,7 @@ static void cx25821_stop_streaming(struct vb2_queue *q)
|
||||
struct cx25821_buffer, queue);
|
||||
|
||||
list_del(&buf->queue);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -34,6 +34,7 @@
|
||||
#include <media/v4l2-common.h>
|
||||
#include <media/v4l2-device.h>
|
||||
#include <media/v4l2-ctrls.h>
|
||||
#include <media/videobuf2-v4l2.h>
|
||||
#include <media/videobuf2-dma-sg.h>
|
||||
|
||||
#include "cx25821-reg.h"
|
||||
@@ -127,7 +128,7 @@ struct cx25821_riscmem {
|
||||
/* buffer for one video frame */
|
||||
struct cx25821_buffer {
|
||||
/* common v4l buffer stuff -- must be first */
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head queue;
|
||||
|
||||
/* cx25821 specific */
|
||||
|
@@ -653,16 +653,18 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8802_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
|
||||
return cx8802_buf_prepare(vb->vb2_queue, dev, buf);
|
||||
}
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8802_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct cx88_riscmem *risc = &buf->risc;
|
||||
|
||||
if (risc->cpu)
|
||||
@@ -672,8 +674,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8802_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
|
||||
cx8802_buf_queue(dev, buf);
|
||||
}
|
||||
@@ -721,7 +724,7 @@ fail:
|
||||
struct cx88_buffer, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_QUEUED);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
return err;
|
||||
@@ -749,7 +752,7 @@ static void stop_streaming(struct vb2_queue *q)
|
||||
struct cx88_buffer, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -518,11 +518,11 @@ void cx88_wakeup(struct cx88_core *core,
|
||||
|
||||
buf = list_entry(q->active.next,
|
||||
struct cx88_buffer, list);
|
||||
v4l2_get_timestamp(&buf->vb.v4l2_buf.timestamp);
|
||||
buf->vb.v4l2_buf.field = core->field;
|
||||
buf->vb.v4l2_buf.sequence = q->count++;
|
||||
v4l2_get_timestamp(&buf->vb.timestamp);
|
||||
buf->vb.field = core->field;
|
||||
buf->vb.sequence = q->count++;
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_DONE);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
}
|
||||
|
||||
void cx88_shutdown(struct cx88_core *core)
|
||||
|
@@ -99,16 +99,18 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8802_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
|
||||
return cx8802_buf_prepare(vb->vb2_queue, dev, buf);
|
||||
}
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8802_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct cx88_riscmem *risc = &buf->risc;
|
||||
|
||||
if (risc->cpu)
|
||||
@@ -118,8 +120,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8802_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
|
||||
cx8802_buf_queue(dev, buf);
|
||||
}
|
||||
@@ -149,7 +152,7 @@ static void stop_streaming(struct vb2_queue *q)
|
||||
struct cx88_buffer, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -214,7 +214,7 @@ static int cx8802_restart_queue(struct cx8802_dev *dev,
|
||||
|
||||
buf = list_entry(q->active.next, struct cx88_buffer, list);
|
||||
dprintk(2,"restart_queue [%p/%d]: restart dma\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
cx8802_start_dma(dev, q, buf);
|
||||
return 0;
|
||||
}
|
||||
@@ -225,13 +225,13 @@ int cx8802_buf_prepare(struct vb2_queue *q, struct cx8802_dev *dev,
|
||||
struct cx88_buffer *buf)
|
||||
{
|
||||
int size = dev->ts_packet_size * dev->ts_packet_count;
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb, 0);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0);
|
||||
struct cx88_riscmem *risc = &buf->risc;
|
||||
int rc;
|
||||
|
||||
if (vb2_plane_size(&buf->vb, 0) < size)
|
||||
if (vb2_plane_size(&buf->vb.vb2_buf, 0) < size)
|
||||
return -EINVAL;
|
||||
vb2_set_plane_payload(&buf->vb, 0, size);
|
||||
vb2_set_plane_payload(&buf->vb.vb2_buf, 0, size);
|
||||
|
||||
rc = cx88_risc_databuffer(dev->pci, risc, sgt->sgl,
|
||||
dev->ts_packet_size, dev->ts_packet_count, 0);
|
||||
@@ -259,7 +259,7 @@ void cx8802_buf_queue(struct cx8802_dev *dev, struct cx88_buffer *buf)
|
||||
dprintk( 1, "queue is empty - first active\n" );
|
||||
list_add_tail(&buf->list, &cx88q->active);
|
||||
dprintk(1,"[%p/%d] %s - first active\n",
|
||||
buf, buf->vb.v4l2_buf.index, __func__);
|
||||
buf, buf->vb.vb2_buf.index, __func__);
|
||||
|
||||
} else {
|
||||
buf->risc.cpu[0] |= cpu_to_le32(RISC_IRQ1);
|
||||
@@ -268,7 +268,7 @@ void cx8802_buf_queue(struct cx8802_dev *dev, struct cx88_buffer *buf)
|
||||
list_add_tail(&buf->list, &cx88q->active);
|
||||
prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
|
||||
dprintk( 1, "[%p/%d] %s - append to active\n",
|
||||
buf, buf->vb.v4l2_buf.index, __func__);
|
||||
buf, buf->vb.vb2_buf.index, __func__);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -284,7 +284,7 @@ static void do_cancel_buffers(struct cx8802_dev *dev)
|
||||
while (!list_empty(&q->active)) {
|
||||
buf = list_entry(q->active.next, struct cx88_buffer, list);
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock,flags);
|
||||
}
|
||||
|
@@ -100,7 +100,7 @@ int cx8800_restart_vbi_queue(struct cx8800_dev *dev,
|
||||
|
||||
buf = list_entry(q->active.next, struct cx88_buffer, list);
|
||||
dprintk(2,"restart_queue [%p/%d]: restart dma\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
cx8800_start_vbi_dma(dev, q, buf);
|
||||
return 0;
|
||||
}
|
||||
@@ -125,8 +125,9 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8800_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
unsigned int lines;
|
||||
unsigned int size;
|
||||
@@ -149,8 +150,9 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8800_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct cx88_riscmem *risc = &buf->risc;
|
||||
|
||||
if (risc->cpu)
|
||||
@@ -160,8 +162,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8800_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *prev;
|
||||
struct cx88_dmaqueue *q = &dev->vbiq;
|
||||
|
||||
@@ -174,7 +177,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
list_add_tail(&buf->list, &q->active);
|
||||
cx8800_start_vbi_dma(dev, q, buf);
|
||||
dprintk(2,"[%p/%d] vbi_queue - first active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
|
||||
} else {
|
||||
buf->risc.cpu[0] |= cpu_to_le32(RISC_IRQ1);
|
||||
@@ -182,7 +185,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
list_add_tail(&buf->list, &q->active);
|
||||
prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
|
||||
dprintk(2,"[%p/%d] buffer_queue - append to active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -213,7 +216,7 @@ static void stop_streaming(struct vb2_queue *q)
|
||||
struct cx88_buffer, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -420,7 +420,7 @@ static int restart_video_queue(struct cx8800_dev *dev,
|
||||
if (!list_empty(&q->active)) {
|
||||
buf = list_entry(q->active.next, struct cx88_buffer, list);
|
||||
dprintk(2,"restart_queue [%p/%d]: restart dma\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
start_video_dma(dev, q, buf);
|
||||
}
|
||||
return 0;
|
||||
@@ -444,9 +444,10 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8800_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_core *core = dev->core;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
|
||||
buf->bpl = core->width * dev->fmt->depth >> 3;
|
||||
@@ -489,7 +490,7 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
break;
|
||||
}
|
||||
dprintk(2,"[%p/%d] buffer_prepare - %dx%d %dbpp \"%s\" - dma=0x%08lx\n",
|
||||
buf, buf->vb.v4l2_buf.index,
|
||||
buf, buf->vb.vb2_buf.index,
|
||||
core->width, core->height, dev->fmt->depth, dev->fmt->name,
|
||||
(unsigned long)buf->risc.dma);
|
||||
return 0;
|
||||
@@ -497,8 +498,9 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8800_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct cx88_riscmem *risc = &buf->risc;
|
||||
|
||||
if (risc->cpu)
|
||||
@@ -508,8 +510,9 @@ static void buffer_finish(struct vb2_buffer *vb)
|
||||
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct cx8800_dev *dev = vb->vb2_queue->drv_priv;
|
||||
struct cx88_buffer *buf = container_of(vb, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *buf = container_of(vbuf, struct cx88_buffer, vb);
|
||||
struct cx88_buffer *prev;
|
||||
struct cx88_core *core = dev->core;
|
||||
struct cx88_dmaqueue *q = &dev->vidq;
|
||||
@@ -522,7 +525,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
if (list_empty(&q->active)) {
|
||||
list_add_tail(&buf->list, &q->active);
|
||||
dprintk(2,"[%p/%d] buffer_queue - first active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
|
||||
} else {
|
||||
buf->risc.cpu[0] |= cpu_to_le32(RISC_IRQ1);
|
||||
@@ -530,7 +533,7 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
list_add_tail(&buf->list, &q->active);
|
||||
prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
|
||||
dprintk(2, "[%p/%d] buffer_queue - append to active\n",
|
||||
buf, buf->vb.v4l2_buf.index);
|
||||
buf, buf->vb.vb2_buf.index);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -560,7 +563,7 @@ static void stop_streaming(struct vb2_queue *q)
|
||||
struct cx88_buffer, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dev->slock, flags);
|
||||
}
|
||||
|
@@ -321,7 +321,7 @@ struct cx88_riscmem {
|
||||
/* buffer for one video frame */
|
||||
struct cx88_buffer {
|
||||
/* common v4l buffer stuff -- must be first */
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head list;
|
||||
|
||||
/* cx88 specific */
|
||||
|
@@ -160,7 +160,7 @@ static int dt3155_buf_prepare(struct vb2_buffer *vb)
|
||||
static int dt3155_start_streaming(struct vb2_queue *q, unsigned count)
|
||||
{
|
||||
struct dt3155_priv *pd = vb2_get_drv_priv(q);
|
||||
struct vb2_buffer *vb = pd->curr_buf;
|
||||
struct vb2_buffer *vb = &pd->curr_buf->vb2_buf;
|
||||
dma_addr_t dma_addr;
|
||||
|
||||
pd->sequence = 0;
|
||||
@@ -208,7 +208,7 @@ static void dt3155_stop_streaming(struct vb2_queue *q)
|
||||
|
||||
spin_lock_irq(&pd->lock);
|
||||
if (pd->curr_buf) {
|
||||
vb2_buffer_done(pd->curr_buf, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&pd->curr_buf->vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
pd->curr_buf = NULL;
|
||||
}
|
||||
|
||||
@@ -222,6 +222,7 @@ static void dt3155_stop_streaming(struct vb2_queue *q)
|
||||
|
||||
static void dt3155_buf_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct dt3155_priv *pd = vb2_get_drv_priv(vb->vb2_queue);
|
||||
|
||||
/* pd->vidq.streaming = 1 when dt3155_buf_queue() is invoked */
|
||||
@@ -229,7 +230,7 @@ static void dt3155_buf_queue(struct vb2_buffer *vb)
|
||||
if (pd->curr_buf)
|
||||
list_add_tail(&vb->done_entry, &pd->dmaq);
|
||||
else
|
||||
pd->curr_buf = vb;
|
||||
pd->curr_buf = vbuf;
|
||||
spin_unlock_irq(&pd->lock);
|
||||
}
|
||||
|
||||
@@ -269,14 +270,14 @@ static irqreturn_t dt3155_irq_handler_even(int irq, void *dev_id)
|
||||
|
||||
spin_lock(&ipd->lock);
|
||||
if (ipd->curr_buf && !list_empty(&ipd->dmaq)) {
|
||||
v4l2_get_timestamp(&ipd->curr_buf->v4l2_buf.timestamp);
|
||||
ipd->curr_buf->v4l2_buf.sequence = ipd->sequence++;
|
||||
ipd->curr_buf->v4l2_buf.field = V4L2_FIELD_NONE;
|
||||
vb2_buffer_done(ipd->curr_buf, VB2_BUF_STATE_DONE);
|
||||
v4l2_get_timestamp(&ipd->curr_buf->timestamp);
|
||||
ipd->curr_buf->sequence = ipd->sequence++;
|
||||
ipd->curr_buf->field = V4L2_FIELD_NONE;
|
||||
vb2_buffer_done(&ipd->curr_buf->vb2_buf, VB2_BUF_STATE_DONE);
|
||||
|
||||
ivb = list_first_entry(&ipd->dmaq, typeof(*ivb), done_entry);
|
||||
list_del(&ivb->done_entry);
|
||||
ipd->curr_buf = ivb;
|
||||
ipd->curr_buf = to_vb2_v4l2_buffer(ivb);
|
||||
dma_addr = vb2_dma_contig_plane_dma_addr(ivb, 0);
|
||||
iowrite32(dma_addr, ipd->regs + EVEN_DMA_START);
|
||||
iowrite32(dma_addr + ipd->width, ipd->regs + ODD_DMA_START);
|
||||
|
@@ -22,6 +22,7 @@
|
||||
#include <linux/interrupt.h>
|
||||
#include <media/v4l2-device.h>
|
||||
#include <media/v4l2-dev.h>
|
||||
#include <media/videobuf2-v4l2.h>
|
||||
|
||||
#define DT3155_NAME "dt3155"
|
||||
#define DT3155_VER_MAJ 2
|
||||
@@ -181,7 +182,7 @@ struct dt3155_priv {
|
||||
struct pci_dev *pdev;
|
||||
struct vb2_queue vidq;
|
||||
struct vb2_alloc_ctx *alloc_ctx;
|
||||
struct vb2_buffer *curr_buf;
|
||||
struct vb2_v4l2_buffer *curr_buf;
|
||||
struct mutex mux;
|
||||
struct list_head dmaq;
|
||||
spinlock_t lock;
|
||||
|
@@ -27,6 +27,7 @@
|
||||
#include <linux/interrupt.h>
|
||||
#include <linux/delay.h>
|
||||
#include <linux/list.h>
|
||||
#include <media/videobuf2-v4l2.h>
|
||||
#include <media/videobuf2-vmalloc.h>
|
||||
|
||||
#include "netup_unidvb.h"
|
||||
@@ -110,7 +111,7 @@ struct netup_dma_regs {
|
||||
} __packed __aligned(1);
|
||||
|
||||
struct netup_unidvb_buffer {
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head list;
|
||||
u32 size;
|
||||
};
|
||||
@@ -300,7 +301,8 @@ static int netup_unidvb_queue_setup(struct vb2_queue *vq,
|
||||
static int netup_unidvb_buf_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct netup_dma *dma = vb2_get_drv_priv(vb->vb2_queue);
|
||||
struct netup_unidvb_buffer *buf = container_of(vb,
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct netup_unidvb_buffer *buf = container_of(vbuf,
|
||||
struct netup_unidvb_buffer, vb);
|
||||
|
||||
dev_dbg(&dma->ndev->pci_dev->dev, "%s(): buf 0x%p\n", __func__, buf);
|
||||
@@ -312,7 +314,8 @@ static void netup_unidvb_buf_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
unsigned long flags;
|
||||
struct netup_dma *dma = vb2_get_drv_priv(vb->vb2_queue);
|
||||
struct netup_unidvb_buffer *buf = container_of(vb,
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct netup_unidvb_buffer *buf = container_of(vbuf,
|
||||
struct netup_unidvb_buffer, vb);
|
||||
|
||||
dev_dbg(&dma->ndev->pci_dev->dev, "%s(): %p\n", __func__, buf);
|
||||
@@ -509,7 +512,7 @@ static int netup_unidvb_ring_copy(struct netup_dma *dma,
|
||||
{
|
||||
u32 copy_bytes, ring_bytes;
|
||||
u32 buff_bytes = NETUP_DMA_PACKETS_COUNT * 188 - buf->size;
|
||||
u8 *p = vb2_plane_vaddr(&buf->vb, 0);
|
||||
u8 *p = vb2_plane_vaddr(&buf->vb.vb2_buf, 0);
|
||||
struct netup_unidvb_dev *ndev = dma->ndev;
|
||||
|
||||
if (p == NULL) {
|
||||
@@ -579,9 +582,9 @@ static void netup_unidvb_dma_worker(struct work_struct *work)
|
||||
dev_dbg(&ndev->pci_dev->dev,
|
||||
"%s(): buffer %p done, size %d\n",
|
||||
__func__, buf, buf->size);
|
||||
v4l2_get_timestamp(&buf->vb.v4l2_buf.timestamp);
|
||||
vb2_set_plane_payload(&buf->vb, 0, buf->size);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_DONE);
|
||||
v4l2_get_timestamp(&buf->vb.timestamp);
|
||||
vb2_set_plane_payload(&buf->vb.vb2_buf, 0, buf->size);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
}
|
||||
}
|
||||
work_done:
|
||||
@@ -599,7 +602,7 @@ static void netup_unidvb_queue_cleanup(struct netup_dma *dma)
|
||||
buf = list_first_entry(&dma->free_buffers,
|
||||
struct netup_unidvb_buffer, list);
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&dma->lock, flags);
|
||||
}
|
||||
|
@@ -216,13 +216,14 @@ int saa7134_buffer_count(unsigned int size, unsigned int count)
|
||||
|
||||
int saa7134_buffer_startpage(struct saa7134_buf *buf)
|
||||
{
|
||||
return saa7134_buffer_pages(vb2_plane_size(&buf->vb2, 0)) * buf->vb2.v4l2_buf.index;
|
||||
return saa7134_buffer_pages(vb2_plane_size(&buf->vb2.vb2_buf, 0))
|
||||
* buf->vb2.vb2_buf.index;
|
||||
}
|
||||
|
||||
unsigned long saa7134_buffer_base(struct saa7134_buf *buf)
|
||||
{
|
||||
unsigned long base;
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(&buf->vb2, 0);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(&buf->vb2.vb2_buf, 0);
|
||||
|
||||
base = saa7134_buffer_startpage(buf) * 4096;
|
||||
base += dma->sgl[0].offset;
|
||||
@@ -308,9 +309,9 @@ void saa7134_buffer_finish(struct saa7134_dev *dev,
|
||||
core_dbg("buffer_finish %p\n", q->curr);
|
||||
|
||||
/* finish current buffer */
|
||||
v4l2_get_timestamp(&q->curr->vb2.v4l2_buf.timestamp);
|
||||
q->curr->vb2.v4l2_buf.sequence = q->seq_nr++;
|
||||
vb2_buffer_done(&q->curr->vb2, state);
|
||||
v4l2_get_timestamp(&q->curr->vb2.timestamp);
|
||||
q->curr->vb2.sequence = q->seq_nr++;
|
||||
vb2_buffer_done(&q->curr->vb2.vb2_buf, state);
|
||||
q->curr = NULL;
|
||||
}
|
||||
|
||||
@@ -375,7 +376,8 @@ void saa7134_stop_streaming(struct saa7134_dev *dev, struct saa7134_dmaqueue *q)
|
||||
if (!list_empty(&q->queue)) {
|
||||
list_for_each_safe(pos, n, &q->queue) {
|
||||
tmp = list_entry(pos, struct saa7134_buf, entry);
|
||||
vb2_buffer_done(&tmp->vb2, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&tmp->vb2.vb2_buf,
|
||||
VB2_BUF_STATE_ERROR);
|
||||
list_del(pos);
|
||||
tmp = NULL;
|
||||
}
|
||||
|
@@ -79,8 +79,9 @@ static int buffer_activate(struct saa7134_dev *dev,
|
||||
|
||||
int saa7134_ts_buffer_init(struct vb2_buffer *vb2)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb2);
|
||||
struct saa7134_dmaqueue *dmaq = vb2->vb2_queue->drv_priv;
|
||||
struct saa7134_buf *buf = container_of(vb2, struct saa7134_buf, vb2);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
|
||||
dmaq->curr = NULL;
|
||||
buf->activate = buffer_activate;
|
||||
@@ -91,9 +92,10 @@ EXPORT_SYMBOL_GPL(saa7134_ts_buffer_init);
|
||||
|
||||
int saa7134_ts_buffer_prepare(struct vb2_buffer *vb2)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb2);
|
||||
struct saa7134_dmaqueue *dmaq = vb2->vb2_queue->drv_priv;
|
||||
struct saa7134_dev *dev = dmaq->dev;
|
||||
struct saa7134_buf *buf = container_of(vb2, struct saa7134_buf, vb2);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(vb2, 0);
|
||||
unsigned int lines, llength, size;
|
||||
|
||||
@@ -107,7 +109,7 @@ int saa7134_ts_buffer_prepare(struct vb2_buffer *vb2)
|
||||
return -EINVAL;
|
||||
|
||||
vb2_set_plane_payload(vb2, 0, size);
|
||||
vb2->v4l2_buf.field = dev->field;
|
||||
vbuf->field = dev->field;
|
||||
|
||||
return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
|
||||
saa7134_buffer_startpage(buf));
|
||||
@@ -148,10 +150,12 @@ int saa7134_ts_start_streaming(struct vb2_queue *vq, unsigned int count)
|
||||
|
||||
list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) {
|
||||
list_del(&buf->entry);
|
||||
vb2_buffer_done(&buf->vb2, VB2_BUF_STATE_QUEUED);
|
||||
vb2_buffer_done(&buf->vb2.vb2_buf,
|
||||
VB2_BUF_STATE_QUEUED);
|
||||
}
|
||||
if (dmaq->curr) {
|
||||
vb2_buffer_done(&dmaq->curr->vb2, VB2_BUF_STATE_QUEUED);
|
||||
vb2_buffer_done(&dmaq->curr->vb2.vb2_buf,
|
||||
VB2_BUF_STATE_QUEUED);
|
||||
dmaq->curr = NULL;
|
||||
}
|
||||
return -EBUSY;
|
||||
|
@@ -83,7 +83,7 @@ static int buffer_activate(struct saa7134_dev *dev,
|
||||
struct saa7134_buf *buf,
|
||||
struct saa7134_buf *next)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = buf->vb2.vb2_queue->drv_priv;
|
||||
struct saa7134_dmaqueue *dmaq = buf->vb2.vb2_buf.vb2_queue->drv_priv;
|
||||
unsigned long control, base;
|
||||
|
||||
vbi_dbg("buffer_activate [%p]\n", buf);
|
||||
@@ -119,8 +119,9 @@ static int buffer_prepare(struct vb2_buffer *vb2)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = vb2->vb2_queue->drv_priv;
|
||||
struct saa7134_dev *dev = dmaq->dev;
|
||||
struct saa7134_buf *buf = container_of(vb2, struct saa7134_buf, vb2);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(&buf->vb2, 0);
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb2);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(vb2, 0);
|
||||
unsigned int size;
|
||||
|
||||
if (dma->sgl->offset) {
|
||||
@@ -161,7 +162,8 @@ static int queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
static int buffer_init(struct vb2_buffer *vb2)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = vb2->vb2_queue->drv_priv;
|
||||
struct saa7134_buf *buf = container_of(vb2, struct saa7134_buf, vb2);
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb2);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
|
||||
dmaq->curr = NULL;
|
||||
buf->activate = buffer_activate;
|
||||
|
@@ -791,7 +791,7 @@ static int buffer_activate(struct saa7134_dev *dev,
|
||||
struct saa7134_buf *buf,
|
||||
struct saa7134_buf *next)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = buf->vb2.vb2_queue->drv_priv;
|
||||
struct saa7134_dmaqueue *dmaq = buf->vb2.vb2_buf.vb2_queue->drv_priv;
|
||||
unsigned long base,control,bpl;
|
||||
unsigned long bpl_uv,lines_uv,base2,base3,tmp; /* planar */
|
||||
|
||||
@@ -872,7 +872,8 @@ static int buffer_activate(struct saa7134_dev *dev,
|
||||
static int buffer_init(struct vb2_buffer *vb2)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = vb2->vb2_queue->drv_priv;
|
||||
struct saa7134_buf *buf = container_of(vb2, struct saa7134_buf, vb2);
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb2);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
|
||||
dmaq->curr = NULL;
|
||||
buf->activate = buffer_activate;
|
||||
@@ -883,8 +884,9 @@ static int buffer_prepare(struct vb2_buffer *vb2)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = vb2->vb2_queue->drv_priv;
|
||||
struct saa7134_dev *dev = dmaq->dev;
|
||||
struct saa7134_buf *buf = container_of(vb2, struct saa7134_buf, vb2);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(&buf->vb2, 0);
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb2);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(vb2, 0);
|
||||
unsigned int size;
|
||||
|
||||
if (dma->sgl->offset) {
|
||||
@@ -896,7 +898,7 @@ static int buffer_prepare(struct vb2_buffer *vb2)
|
||||
return -EINVAL;
|
||||
|
||||
vb2_set_plane_payload(vb2, 0, size);
|
||||
vb2->v4l2_buf.field = dev->field;
|
||||
vbuf->field = dev->field;
|
||||
|
||||
return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
|
||||
saa7134_buffer_startpage(buf));
|
||||
@@ -932,7 +934,8 @@ void saa7134_vb2_buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct saa7134_dmaqueue *dmaq = vb->vb2_queue->drv_priv;
|
||||
struct saa7134_dev *dev = dmaq->dev;
|
||||
struct saa7134_buf *buf = container_of(vb, struct saa7134_buf, vb2);
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct saa7134_buf *buf = container_of(vbuf, struct saa7134_buf, vb2);
|
||||
|
||||
saa7134_buffer_queue(dev, dmaq, buf);
|
||||
}
|
||||
@@ -953,10 +956,12 @@ int saa7134_vb2_start_streaming(struct vb2_queue *vq, unsigned int count)
|
||||
|
||||
list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) {
|
||||
list_del(&buf->entry);
|
||||
vb2_buffer_done(&buf->vb2, VB2_BUF_STATE_QUEUED);
|
||||
vb2_buffer_done(&buf->vb2.vb2_buf,
|
||||
VB2_BUF_STATE_QUEUED);
|
||||
}
|
||||
if (dmaq->curr) {
|
||||
vb2_buffer_done(&dmaq->curr->vb2, VB2_BUF_STATE_QUEUED);
|
||||
vb2_buffer_done(&dmaq->curr->vb2.vb2_buf,
|
||||
VB2_BUF_STATE_QUEUED);
|
||||
dmaq->curr = NULL;
|
||||
}
|
||||
return -EBUSY;
|
||||
|
@@ -460,7 +460,7 @@ struct saa7134_thread {
|
||||
/* buffer for one video/vbi/ts frame */
|
||||
struct saa7134_buf {
|
||||
/* common v4l buffer stuff -- must be first */
|
||||
struct vb2_buffer vb2;
|
||||
struct vb2_v4l2_buffer vb2;
|
||||
|
||||
/* saa7134 specific */
|
||||
unsigned int top_seen;
|
||||
|
@@ -458,11 +458,12 @@ static inline u32 vop_usec(const vop_header *vh)
|
||||
static int solo_fill_jpeg(struct solo_enc_dev *solo_enc,
|
||||
struct vb2_buffer *vb, const vop_header *vh)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct solo_dev *solo_dev = solo_enc->solo_dev;
|
||||
struct sg_table *vbuf = vb2_dma_sg_plane_desc(vb, 0);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
int frame_size;
|
||||
|
||||
vb->v4l2_buf.flags |= V4L2_BUF_FLAG_KEYFRAME;
|
||||
vbuf->flags |= V4L2_BUF_FLAG_KEYFRAME;
|
||||
|
||||
if (vb2_plane_size(vb, 0) < vop_jpeg_size(vh) + solo_enc->jpeg_len)
|
||||
return -EIO;
|
||||
@@ -470,7 +471,7 @@ static int solo_fill_jpeg(struct solo_enc_dev *solo_enc,
|
||||
frame_size = ALIGN(vop_jpeg_size(vh) + solo_enc->jpeg_len, DMA_ALIGN);
|
||||
vb2_set_plane_payload(vb, 0, vop_jpeg_size(vh) + solo_enc->jpeg_len);
|
||||
|
||||
return solo_send_desc(solo_enc, solo_enc->jpeg_len, vbuf,
|
||||
return solo_send_desc(solo_enc, solo_enc->jpeg_len, sgt,
|
||||
vop_jpeg_offset(vh) - SOLO_JPEG_EXT_ADDR(solo_dev),
|
||||
frame_size, SOLO_JPEG_EXT_ADDR(solo_dev),
|
||||
SOLO_JPEG_EXT_SIZE(solo_dev));
|
||||
@@ -479,8 +480,9 @@ static int solo_fill_jpeg(struct solo_enc_dev *solo_enc,
|
||||
static int solo_fill_mpeg(struct solo_enc_dev *solo_enc,
|
||||
struct vb2_buffer *vb, const vop_header *vh)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct solo_dev *solo_dev = solo_enc->solo_dev;
|
||||
struct sg_table *vbuf = vb2_dma_sg_plane_desc(vb, 0);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
int frame_off, frame_size;
|
||||
int skip = 0;
|
||||
|
||||
@@ -488,15 +490,15 @@ static int solo_fill_mpeg(struct solo_enc_dev *solo_enc,
|
||||
return -EIO;
|
||||
|
||||
/* If this is a key frame, add extra header */
|
||||
vb->v4l2_buf.flags &= ~(V4L2_BUF_FLAG_KEYFRAME | V4L2_BUF_FLAG_PFRAME |
|
||||
vbuf->flags &= ~(V4L2_BUF_FLAG_KEYFRAME | V4L2_BUF_FLAG_PFRAME |
|
||||
V4L2_BUF_FLAG_BFRAME);
|
||||
if (!vop_type(vh)) {
|
||||
skip = solo_enc->vop_len;
|
||||
vb->v4l2_buf.flags |= V4L2_BUF_FLAG_KEYFRAME;
|
||||
vbuf->flags |= V4L2_BUF_FLAG_KEYFRAME;
|
||||
vb2_set_plane_payload(vb, 0, vop_mpeg_size(vh) +
|
||||
solo_enc->vop_len);
|
||||
} else {
|
||||
vb->v4l2_buf.flags |= V4L2_BUF_FLAG_PFRAME;
|
||||
vbuf->flags |= V4L2_BUF_FLAG_PFRAME;
|
||||
vb2_set_plane_payload(vb, 0, vop_mpeg_size(vh));
|
||||
}
|
||||
|
||||
@@ -505,7 +507,7 @@ static int solo_fill_mpeg(struct solo_enc_dev *solo_enc,
|
||||
sizeof(*vh)) % SOLO_MP4E_EXT_SIZE(solo_dev);
|
||||
frame_size = ALIGN(vop_mpeg_size(vh) + skip, DMA_ALIGN);
|
||||
|
||||
return solo_send_desc(solo_enc, skip, vbuf, frame_off, frame_size,
|
||||
return solo_send_desc(solo_enc, skip, sgt, frame_off, frame_size,
|
||||
SOLO_MP4E_EXT_ADDR(solo_dev),
|
||||
SOLO_MP4E_EXT_SIZE(solo_dev));
|
||||
}
|
||||
@@ -513,6 +515,7 @@ static int solo_fill_mpeg(struct solo_enc_dev *solo_enc,
|
||||
static int solo_enc_fillbuf(struct solo_enc_dev *solo_enc,
|
||||
struct vb2_buffer *vb, struct solo_enc_buf *enc_buf)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
const vop_header *vh = enc_buf->vh;
|
||||
int ret;
|
||||
|
||||
@@ -527,17 +530,18 @@ static int solo_enc_fillbuf(struct solo_enc_dev *solo_enc,
|
||||
}
|
||||
|
||||
if (!ret) {
|
||||
vb->v4l2_buf.sequence = solo_enc->sequence++;
|
||||
vb->v4l2_buf.timestamp.tv_sec = vop_sec(vh);
|
||||
vb->v4l2_buf.timestamp.tv_usec = vop_usec(vh);
|
||||
vbuf->sequence = solo_enc->sequence++;
|
||||
vbuf->timestamp.tv_sec = vop_sec(vh);
|
||||
vbuf->timestamp.tv_usec = vop_usec(vh);
|
||||
|
||||
/* Check for motion flags */
|
||||
if (solo_is_motion_on(solo_enc) && enc_buf->motion) {
|
||||
struct v4l2_event ev = {
|
||||
.type = V4L2_EVENT_MOTION_DET,
|
||||
.u.motion_det = {
|
||||
.flags = V4L2_EVENT_MD_FL_HAVE_FRAME_SEQ,
|
||||
.frame_sequence = vb->v4l2_buf.sequence,
|
||||
.flags
|
||||
= V4L2_EVENT_MD_FL_HAVE_FRAME_SEQ,
|
||||
.frame_sequence = vbuf->sequence,
|
||||
.region_mask = enc_buf->motion ? 1 : 0,
|
||||
},
|
||||
};
|
||||
@@ -571,7 +575,7 @@ static void solo_enc_handle_one(struct solo_enc_dev *solo_enc,
|
||||
list_del(&vb->list);
|
||||
spin_unlock_irqrestore(&solo_enc->av_lock, flags);
|
||||
|
||||
solo_enc_fillbuf(solo_enc, &vb->vb, enc_buf);
|
||||
solo_enc_fillbuf(solo_enc, &vb->vb.vb2_buf, enc_buf);
|
||||
unlock:
|
||||
mutex_unlock(&solo_enc->lock);
|
||||
}
|
||||
@@ -678,10 +682,11 @@ static int solo_enc_queue_setup(struct vb2_queue *q,
|
||||
|
||||
static void solo_enc_buf_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vb2_queue *vq = vb->vb2_queue;
|
||||
struct solo_enc_dev *solo_enc = vb2_get_drv_priv(vq);
|
||||
struct solo_vb2_buf *solo_vb =
|
||||
container_of(vb, struct solo_vb2_buf, vb);
|
||||
container_of(vbuf, struct solo_vb2_buf, vb);
|
||||
|
||||
spin_lock(&solo_enc->av_lock);
|
||||
list_add_tail(&solo_vb->list, &solo_enc->vidq_active);
|
||||
@@ -734,25 +739,26 @@ static void solo_enc_stop_streaming(struct vb2_queue *q)
|
||||
struct solo_vb2_buf, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
spin_unlock_irqrestore(&solo_enc->av_lock, flags);
|
||||
}
|
||||
|
||||
static void solo_enc_buf_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct solo_enc_dev *solo_enc = vb2_get_drv_priv(vb->vb2_queue);
|
||||
struct sg_table *vbuf = vb2_dma_sg_plane_desc(vb, 0);
|
||||
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
||||
|
||||
switch (solo_enc->fmt) {
|
||||
case V4L2_PIX_FMT_MPEG4:
|
||||
case V4L2_PIX_FMT_H264:
|
||||
if (vb->v4l2_buf.flags & V4L2_BUF_FLAG_KEYFRAME)
|
||||
sg_copy_from_buffer(vbuf->sgl, vbuf->nents,
|
||||
if (vbuf->flags & V4L2_BUF_FLAG_KEYFRAME)
|
||||
sg_copy_from_buffer(sgt->sgl, sgt->nents,
|
||||
solo_enc->vop, solo_enc->vop_len);
|
||||
break;
|
||||
default: /* V4L2_PIX_FMT_MJPEG */
|
||||
sg_copy_from_buffer(vbuf->sgl, vbuf->nents,
|
||||
sg_copy_from_buffer(sgt->sgl, sgt->nents,
|
||||
solo_enc->jpeg_header, solo_enc->jpeg_len);
|
||||
break;
|
||||
}
|
||||
|
@@ -26,6 +26,7 @@
|
||||
#include <media/v4l2-ioctl.h>
|
||||
#include <media/v4l2-common.h>
|
||||
#include <media/v4l2-event.h>
|
||||
#include <media/videobuf2-v4l2.h>
|
||||
#include <media/videobuf2-dma-contig.h>
|
||||
|
||||
#include "solo6x10.h"
|
||||
@@ -191,13 +192,14 @@ static int solo_v4l2_set_ch(struct solo_dev *solo_dev, u8 ch)
|
||||
static void solo_fillbuf(struct solo_dev *solo_dev,
|
||||
struct vb2_buffer *vb)
|
||||
{
|
||||
dma_addr_t vbuf;
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
dma_addr_t addr;
|
||||
unsigned int fdma_addr;
|
||||
int error = -1;
|
||||
int i;
|
||||
|
||||
vbuf = vb2_dma_contig_plane_dma_addr(vb, 0);
|
||||
if (!vbuf)
|
||||
addr = vb2_dma_contig_plane_dma_addr(vb, 0);
|
||||
if (!addr)
|
||||
goto finish_buf;
|
||||
|
||||
if (erase_off(solo_dev)) {
|
||||
@@ -213,7 +215,7 @@ static void solo_fillbuf(struct solo_dev *solo_dev,
|
||||
fdma_addr = SOLO_DISP_EXT_ADDR + (solo_dev->old_write *
|
||||
(SOLO_HW_BPL * solo_vlines(solo_dev)));
|
||||
|
||||
error = solo_p2m_dma_t(solo_dev, 0, vbuf, fdma_addr,
|
||||
error = solo_p2m_dma_t(solo_dev, 0, addr, fdma_addr,
|
||||
solo_bytesperline(solo_dev),
|
||||
solo_vlines(solo_dev), SOLO_HW_BPL);
|
||||
}
|
||||
@@ -222,8 +224,8 @@ finish_buf:
|
||||
if (!error) {
|
||||
vb2_set_plane_payload(vb, 0,
|
||||
solo_vlines(solo_dev) * solo_bytesperline(solo_dev));
|
||||
vb->v4l2_buf.sequence = solo_dev->sequence++;
|
||||
v4l2_get_timestamp(&vb->v4l2_buf.timestamp);
|
||||
vbuf->sequence = solo_dev->sequence++;
|
||||
v4l2_get_timestamp(&vbuf->timestamp);
|
||||
}
|
||||
|
||||
vb2_buffer_done(vb, error ? VB2_BUF_STATE_ERROR : VB2_BUF_STATE_DONE);
|
||||
@@ -256,7 +258,7 @@ static void solo_thread_try(struct solo_dev *solo_dev)
|
||||
|
||||
spin_unlock(&solo_dev->slock);
|
||||
|
||||
solo_fillbuf(solo_dev, &vb->vb);
|
||||
solo_fillbuf(solo_dev, &vb->vb.vb2_buf);
|
||||
}
|
||||
|
||||
assert_spin_locked(&solo_dev->slock);
|
||||
@@ -345,10 +347,11 @@ static void solo_stop_streaming(struct vb2_queue *q)
|
||||
|
||||
static void solo_buf_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vb2_queue *vq = vb->vb2_queue;
|
||||
struct solo_dev *solo_dev = vb2_get_drv_priv(vq);
|
||||
struct solo_vb2_buf *solo_vb =
|
||||
container_of(vb, struct solo_vb2_buf, vb);
|
||||
container_of(vbuf, struct solo_vb2_buf, vb);
|
||||
|
||||
spin_lock(&solo_dev->slock);
|
||||
list_add_tail(&solo_vb->list, &solo_dev->vidq_active);
|
||||
|
@@ -135,7 +135,7 @@ struct solo_p2m_dev {
|
||||
#define OSD_TEXT_MAX 44
|
||||
|
||||
struct solo_vb2_buf {
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head list;
|
||||
};
|
||||
|
||||
|
@@ -88,11 +88,11 @@
|
||||
|
||||
|
||||
struct vip_buffer {
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head list;
|
||||
dma_addr_t dma;
|
||||
};
|
||||
static inline struct vip_buffer *to_vip_buffer(struct vb2_buffer *vb2)
|
||||
static inline struct vip_buffer *to_vip_buffer(struct vb2_v4l2_buffer *vb2)
|
||||
{
|
||||
return container_of(vb2, struct vip_buffer, vb);
|
||||
}
|
||||
@@ -287,7 +287,8 @@ static int queue_setup(struct vb2_queue *vq, const struct v4l2_format *fmt,
|
||||
};
|
||||
static int buffer_init(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vb);
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vbuf);
|
||||
|
||||
vip_buf->dma = vb2_dma_contig_plane_dma_addr(vb, 0);
|
||||
INIT_LIST_HEAD(&vip_buf->list);
|
||||
@@ -296,8 +297,9 @@ static int buffer_init(struct vb2_buffer *vb)
|
||||
|
||||
static int buffer_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct sta2x11_vip *vip = vb2_get_drv_priv(vb->vb2_queue);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vb);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vbuf);
|
||||
unsigned long size;
|
||||
|
||||
size = vip->format.sizeimage;
|
||||
@@ -307,14 +309,15 @@ static int buffer_prepare(struct vb2_buffer *vb)
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
vb2_set_plane_payload(&vip_buf->vb, 0, size);
|
||||
vb2_set_plane_payload(&vip_buf->vb.vb2_buf, 0, size);
|
||||
|
||||
return 0;
|
||||
}
|
||||
static void buffer_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct sta2x11_vip *vip = vb2_get_drv_priv(vb->vb2_queue);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vb);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vbuf);
|
||||
|
||||
spin_lock(&vip->lock);
|
||||
list_add_tail(&vip_buf->list, &vip->buffer_list);
|
||||
@@ -329,8 +332,9 @@ static void buffer_queue(struct vb2_buffer *vb)
|
||||
}
|
||||
static void buffer_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct sta2x11_vip *vip = vb2_get_drv_priv(vb->vb2_queue);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vb);
|
||||
struct vip_buffer *vip_buf = to_vip_buffer(vbuf);
|
||||
|
||||
/* Buffer handled, remove it from the list */
|
||||
spin_lock(&vip->lock);
|
||||
@@ -370,7 +374,7 @@ static void stop_streaming(struct vb2_queue *vq)
|
||||
/* Release all active buffers */
|
||||
spin_lock(&vip->lock);
|
||||
list_for_each_entry_safe(vip_buf, node, &vip->buffer_list, list) {
|
||||
vb2_buffer_done(&vip_buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&vip_buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
list_del(&vip_buf->list);
|
||||
}
|
||||
spin_unlock(&vip->lock);
|
||||
@@ -813,9 +817,9 @@ static irqreturn_t vip_irq(int irq, struct sta2x11_vip *vip)
|
||||
/* Disable acquisition */
|
||||
reg_write(vip, DVP_CTL, reg_read(vip, DVP_CTL) & ~DVP_CTL_ENA);
|
||||
/* Remove the active buffer from the list */
|
||||
v4l2_get_timestamp(&vip->active->vb.v4l2_buf.timestamp);
|
||||
vip->active->vb.v4l2_buf.sequence = vip->sequence++;
|
||||
vb2_buffer_done(&vip->active->vb, VB2_BUF_STATE_DONE);
|
||||
v4l2_get_timestamp(&vip->active->vb.timestamp);
|
||||
vip->active->vb.sequence = vip->sequence++;
|
||||
vb2_buffer_done(&vip->active->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
}
|
||||
|
||||
return IRQ_HANDLED;
|
||||
|
@@ -423,9 +423,10 @@ static int tw68_queue_setup(struct vb2_queue *q, const struct v4l2_format *fmt,
|
||||
*/
|
||||
static void tw68_buf_queue(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vb2_queue *vq = vb->vb2_queue;
|
||||
struct tw68_dev *dev = vb2_get_drv_priv(vq);
|
||||
struct tw68_buf *buf = container_of(vb, struct tw68_buf, vb);
|
||||
struct tw68_buf *buf = container_of(vbuf, struct tw68_buf, vb);
|
||||
struct tw68_buf *prev;
|
||||
unsigned long flags;
|
||||
|
||||
@@ -457,9 +458,10 @@ static void tw68_buf_queue(struct vb2_buffer *vb)
|
||||
*/
|
||||
static int tw68_buf_prepare(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vb2_queue *vq = vb->vb2_queue;
|
||||
struct tw68_dev *dev = vb2_get_drv_priv(vq);
|
||||
struct tw68_buf *buf = container_of(vb, struct tw68_buf, vb);
|
||||
struct tw68_buf *buf = container_of(vbuf, struct tw68_buf, vb);
|
||||
struct sg_table *dma = vb2_dma_sg_plane_desc(vb, 0);
|
||||
unsigned size, bpl;
|
||||
|
||||
@@ -499,9 +501,10 @@ static int tw68_buf_prepare(struct vb2_buffer *vb)
|
||||
|
||||
static void tw68_buf_finish(struct vb2_buffer *vb)
|
||||
{
|
||||
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
||||
struct vb2_queue *vq = vb->vb2_queue;
|
||||
struct tw68_dev *dev = vb2_get_drv_priv(vq);
|
||||
struct tw68_buf *buf = container_of(vb, struct tw68_buf, vb);
|
||||
struct tw68_buf *buf = container_of(vbuf, struct tw68_buf, vb);
|
||||
|
||||
pci_free_consistent(dev->pci, buf->size, buf->cpu, buf->dma);
|
||||
}
|
||||
@@ -528,7 +531,7 @@ static void tw68_stop_streaming(struct vb2_queue *q)
|
||||
container_of(dev->active.next, struct tw68_buf, list);
|
||||
|
||||
list_del(&buf->list);
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_ERROR);
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1012,10 +1015,10 @@ void tw68_irq_video_done(struct tw68_dev *dev, unsigned long status)
|
||||
buf = list_entry(dev->active.next, struct tw68_buf, list);
|
||||
list_del(&buf->list);
|
||||
spin_unlock(&dev->slock);
|
||||
v4l2_get_timestamp(&buf->vb.v4l2_buf.timestamp);
|
||||
buf->vb.v4l2_buf.field = dev->field;
|
||||
buf->vb.v4l2_buf.sequence = dev->seqnr++;
|
||||
vb2_buffer_done(&buf->vb, VB2_BUF_STATE_DONE);
|
||||
v4l2_get_timestamp(&buf->vb.timestamp);
|
||||
buf->vb.field = dev->field;
|
||||
buf->vb.sequence = dev->seqnr++;
|
||||
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
|
||||
status &= ~(TW68_DMAPI);
|
||||
if (0 == status)
|
||||
return;
|
||||
|
@@ -36,6 +36,7 @@
|
||||
#include <media/v4l2-ioctl.h>
|
||||
#include <media/v4l2-ctrls.h>
|
||||
#include <media/v4l2-device.h>
|
||||
#include <media/videobuf2-v4l2.h>
|
||||
#include <media/videobuf2-dma-sg.h>
|
||||
|
||||
#include "tw68-reg.h"
|
||||
@@ -118,7 +119,7 @@ struct tw68_dev; /* forward delclaration */
|
||||
|
||||
/* buffer for one video/vbi/ts frame */
|
||||
struct tw68_buf {
|
||||
struct vb2_buffer vb;
|
||||
struct vb2_v4l2_buffer vb;
|
||||
struct list_head list;
|
||||
|
||||
unsigned int size;
|
||||
|
Reference in New Issue
Block a user