|  | /* | 
|  | * Memory-to-memory device framework for Video for Linux 2. | 
|  | * | 
|  | * Helper functions for devices that use memory buffers for both source | 
|  | * and destination. | 
|  | * | 
|  | * Copyright (c) 2009 Samsung Electronics Co., Ltd. | 
|  | * Pawel Osciak, <pawel@osciak.com> | 
|  | * Marek Szyprowski, <m.szyprowski@samsung.com> | 
|  | * | 
|  | * This program is free software; you can redistribute it and/or modify | 
|  | * it under the terms of the GNU General Public License as published by the | 
|  | * Free Software Foundation; either version 2 of the | 
|  | * License, or (at your option) any later version | 
|  | */ | 
|  |  | 
|  | #ifndef _MEDIA_V4L2_MEM2MEM_H | 
|  | #define _MEDIA_V4L2_MEM2MEM_H | 
|  |  | 
|  | #include <media/videobuf2-core.h> | 
|  |  | 
|  | /** | 
|  | * struct v4l2_m2m_ops - mem-to-mem device driver callbacks | 
|  | * @device_run:	required. Begin the actual job (transaction) inside this | 
|  | *		callback. | 
|  | *		The job does NOT have to end before this callback returns | 
|  | *		(and it will be the usual case). When the job finishes, | 
|  | *		v4l2_m2m_job_finish() has to be called. | 
|  | * @job_ready:	optional. Should return 0 if the driver does not have a job | 
|  | *		fully prepared to run yet (i.e. it will not be able to finish a | 
|  | *		transaction without sleeping). If not provided, it will be | 
|  | *		assumed that one source and one destination buffer are all | 
|  | *		that is required for the driver to perform one full transaction. | 
|  | *		This method may not sleep. | 
|  | * @job_abort:	required. Informs the driver that it has to abort the currently | 
|  | *		running transaction as soon as possible (i.e. as soon as it can | 
|  | *		stop the device safely; e.g. in the next interrupt handler), | 
|  | *		even if the transaction would not have been finished by then. | 
|  | *		After the driver performs the necessary steps, it has to call | 
|  | *		v4l2_m2m_job_finish() (as if the transaction ended normally). | 
|  | *		This function does not have to (and will usually not) wait | 
|  | *		until the device enters a state when it can be stopped. | 
|  | */ | 
|  | struct v4l2_m2m_ops { | 
|  | void (*device_run)(void *priv); | 
|  | int (*job_ready)(void *priv); | 
|  | void (*job_abort)(void *priv); | 
|  | void (*lock)(void *priv); | 
|  | void (*unlock)(void *priv); | 
|  | }; | 
|  |  | 
|  | struct v4l2_m2m_dev; | 
|  |  | 
|  | struct v4l2_m2m_queue_ctx { | 
|  | /* private: internal use only */ | 
|  | struct vb2_queue	q; | 
|  |  | 
|  | /* Queue for buffers ready to be processed as soon as this | 
|  | * instance receives access to the device */ | 
|  | struct list_head	rdy_queue; | 
|  | spinlock_t		rdy_spinlock; | 
|  | u8			num_rdy; | 
|  | }; | 
|  |  | 
|  | struct v4l2_m2m_ctx { | 
|  | /* private: internal use only */ | 
|  | struct v4l2_m2m_dev		*m2m_dev; | 
|  |  | 
|  | /* Capture (output to memory) queue context */ | 
|  | struct v4l2_m2m_queue_ctx	cap_q_ctx; | 
|  |  | 
|  | /* Output (input from memory) queue context */ | 
|  | struct v4l2_m2m_queue_ctx	out_q_ctx; | 
|  |  | 
|  | /* For device job queue */ | 
|  | struct list_head		queue; | 
|  | unsigned long			job_flags; | 
|  | wait_queue_head_t		finished; | 
|  |  | 
|  | /* Instance private data */ | 
|  | void				*priv; | 
|  | }; | 
|  |  | 
|  | struct v4l2_m2m_buffer { | 
|  | struct vb2_buffer	vb; | 
|  | struct list_head	list; | 
|  | }; | 
|  |  | 
|  | void *v4l2_m2m_get_curr_priv(struct v4l2_m2m_dev *m2m_dev); | 
|  |  | 
|  | struct vb2_queue *v4l2_m2m_get_vq(struct v4l2_m2m_ctx *m2m_ctx, | 
|  | enum v4l2_buf_type type); | 
|  |  | 
|  | void v4l2_m2m_job_finish(struct v4l2_m2m_dev *m2m_dev, | 
|  | struct v4l2_m2m_ctx *m2m_ctx); | 
|  |  | 
|  | static inline void | 
|  | v4l2_m2m_buf_done(struct vb2_buffer *buf, enum vb2_buffer_state state) | 
|  | { | 
|  | vb2_buffer_done(buf, state); | 
|  | } | 
|  |  | 
|  | int v4l2_m2m_reqbufs(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | struct v4l2_requestbuffers *reqbufs); | 
|  |  | 
|  | int v4l2_m2m_querybuf(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | struct v4l2_buffer *buf); | 
|  |  | 
|  | int v4l2_m2m_qbuf(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | struct v4l2_buffer *buf); | 
|  | int v4l2_m2m_dqbuf(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | struct v4l2_buffer *buf); | 
|  |  | 
|  | int v4l2_m2m_streamon(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | enum v4l2_buf_type type); | 
|  | int v4l2_m2m_streamoff(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | enum v4l2_buf_type type); | 
|  |  | 
|  | unsigned int v4l2_m2m_poll(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | struct poll_table_struct *wait); | 
|  |  | 
|  | int v4l2_m2m_mmap(struct file *file, struct v4l2_m2m_ctx *m2m_ctx, | 
|  | struct vm_area_struct *vma); | 
|  |  | 
|  | struct v4l2_m2m_dev *v4l2_m2m_init(struct v4l2_m2m_ops *m2m_ops); | 
|  | void v4l2_m2m_release(struct v4l2_m2m_dev *m2m_dev); | 
|  |  | 
|  | struct v4l2_m2m_ctx *v4l2_m2m_ctx_init(struct v4l2_m2m_dev *m2m_dev, | 
|  | void *drv_priv, | 
|  | int (*queue_init)(void *priv, struct vb2_queue *src_vq, struct vb2_queue *dst_vq)); | 
|  |  | 
|  | void v4l2_m2m_ctx_release(struct v4l2_m2m_ctx *m2m_ctx); | 
|  |  | 
|  | void v4l2_m2m_buf_queue(struct v4l2_m2m_ctx *m2m_ctx, struct vb2_buffer *vb); | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_num_src_bufs_ready() - return the number of source buffers ready for | 
|  | * use | 
|  | */ | 
|  | static inline | 
|  | unsigned int v4l2_m2m_num_src_bufs_ready(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return m2m_ctx->cap_q_ctx.num_rdy; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_num_src_bufs_ready() - return the number of destination buffers | 
|  | * ready for use | 
|  | */ | 
|  | static inline | 
|  | unsigned int v4l2_m2m_num_dst_bufs_ready(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return m2m_ctx->out_q_ctx.num_rdy; | 
|  | } | 
|  |  | 
|  | void *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx); | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_next_src_buf() - return next source buffer from the list of ready | 
|  | * buffers | 
|  | */ | 
|  | static inline void *v4l2_m2m_next_src_buf(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return v4l2_m2m_next_buf(&m2m_ctx->out_q_ctx); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_next_dst_buf() - return next destination buffer from the list of | 
|  | * ready buffers | 
|  | */ | 
|  | static inline void *v4l2_m2m_next_dst_buf(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return v4l2_m2m_next_buf(&m2m_ctx->cap_q_ctx); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_get_src_vq() - return vb2_queue for source buffers | 
|  | */ | 
|  | static inline | 
|  | struct vb2_queue *v4l2_m2m_get_src_vq(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return &m2m_ctx->out_q_ctx.q; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_get_dst_vq() - return vb2_queue for destination buffers | 
|  | */ | 
|  | static inline | 
|  | struct vb2_queue *v4l2_m2m_get_dst_vq(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return &m2m_ctx->cap_q_ctx.q; | 
|  | } | 
|  |  | 
|  | void *v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx); | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_src_buf_remove() - take off a source buffer from the list of ready | 
|  | * buffers and return it | 
|  | */ | 
|  | static inline void *v4l2_m2m_src_buf_remove(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return v4l2_m2m_buf_remove(&m2m_ctx->out_q_ctx); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * v4l2_m2m_dst_buf_remove() - take off a destination buffer from the list of | 
|  | * ready buffers and return it | 
|  | */ | 
|  | static inline void *v4l2_m2m_dst_buf_remove(struct v4l2_m2m_ctx *m2m_ctx) | 
|  | { | 
|  | return v4l2_m2m_buf_remove(&m2m_ctx->cap_q_ctx); | 
|  | } | 
|  |  | 
|  | #endif /* _MEDIA_V4L2_MEM2MEM_H */ | 
|  |  |