diff options
Diffstat (limited to 'drivers/media/platform/imx-jpeg/mxc-jpeg.c')
| -rw-r--r-- | drivers/media/platform/imx-jpeg/mxc-jpeg.c | 2126 | 
1 files changed, 2126 insertions, 0 deletions
| diff --git a/drivers/media/platform/imx-jpeg/mxc-jpeg.c b/drivers/media/platform/imx-jpeg/mxc-jpeg.c new file mode 100644 index 000000000000..03b9264af068 --- /dev/null +++ b/drivers/media/platform/imx-jpeg/mxc-jpeg.c @@ -0,0 +1,2126 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * V4L2 driver for the JPEG encoder/decoder from i.MX8QXP/i.MX8QM application + * processors. + * + * The multi-planar buffers API is used. + * + * Baseline and extended sequential jpeg decoding is supported. + * Progressive jpeg decoding is not supported by the IP. + * Supports encode and decode of various formats: + *     YUV444, YUV422, YUV420, RGB, ARGB, Gray + * YUV420 is the only multi-planar format supported. + * Minimum resolution is 64 x 64, maximum 8192 x 8192. + * To achieve 8192 x 8192, modify in defconfig: CONFIG_CMA_SIZE_MBYTES=320 + * The alignment requirements for the resolution depend on the format, + * multiple of 16 resolutions should work for all formats. + * Special workarounds are made in the driver to support NV12 1080p. + * When decoding, the driver detects image resolution and pixel format + * from the jpeg stream, by parsing the jpeg markers. + * + * The IP has 4 slots available for context switching, but only slot 0 + * was fully tested to work. Context switching is not used by the driver. + * Each driver instance (context) allocates a slot for itself, but this + * is postponed until device_run, to allow unlimited opens. + * + * The driver submits jobs to the IP by setting up a descriptor for the + * used slot, and then validating it. The encoder has an additional descriptor + * for the configuration phase. The driver expects FRM_DONE interrupt from + * IP to mark the job as finished. + * + * The decoder IP has some limitations regarding the component ID's, + * but the driver works around this by replacing them in the jpeg stream. + * + * A module parameter is available for debug purpose (jpeg_tracing), to enable + * it, enable dynamic debug for this module and: + * echo 1 > /sys/module/mxc_jpeg_encdec/parameters/jpeg_tracing + * + * This is inspired by the drivers/media/platform/s5p-jpeg driver + * + * Copyright 2018-2019 NXP + */ + +#include <linux/kernel.h> +#include <linux/module.h> +#include <linux/io.h> +#include <linux/clk.h> +#include <linux/of_platform.h> +#include <linux/platform_device.h> +#include <linux/slab.h> +#include <linux/irqreturn.h> +#include <linux/interrupt.h> +#include <linux/pm_domain.h> +#include <linux/string.h> + +#include <media/v4l2-jpeg.h> +#include <media/v4l2-mem2mem.h> +#include <media/v4l2-ioctl.h> +#include <media/v4l2-common.h> +#include <media/v4l2-event.h> +#include <media/videobuf2-dma-contig.h> + +#include "mxc-jpeg-hw.h" +#include "mxc-jpeg.h" + +static struct mxc_jpeg_fmt mxc_formats[] = { +	{ +		.name		= "JPEG", +		.fourcc		= V4L2_PIX_FMT_JPEG, +		.subsampling	= -1, +		.nc		= -1, +		.colplanes	= 1, +		.flags		= MXC_JPEG_FMT_TYPE_ENC, +	}, +	{ +		.name		= "RGB", /*RGBRGB packed format*/ +		.fourcc		= V4L2_PIX_FMT_RGB24, +		.subsampling	= V4L2_JPEG_CHROMA_SUBSAMPLING_444, +		.nc		= 3, +		.depth		= 24, +		.colplanes	= 1, +		.h_align	= 3, +		.v_align	= 3, +		.flags		= MXC_JPEG_FMT_TYPE_RAW, +	}, +	{ +		.name		= "ARGB", /* ARGBARGB packed format */ +		.fourcc		= V4L2_PIX_FMT_ARGB32, +		.subsampling	= V4L2_JPEG_CHROMA_SUBSAMPLING_444, +		.nc		= 4, +		.depth		= 32, +		.colplanes	= 1, +		.h_align	= 3, +		.v_align	= 3, +		.flags		= MXC_JPEG_FMT_TYPE_RAW, +	}, +	{ +		.name		= "YUV420", /* 1st plane = Y, 2nd plane = UV */ +		.fourcc		= V4L2_PIX_FMT_NV12, +		.subsampling	= V4L2_JPEG_CHROMA_SUBSAMPLING_420, +		.nc		= 3, +		.depth		= 12, /* 6 bytes (4Y + UV) for 4 pixels */ +		.colplanes	= 2, /* 1 plane Y, 1 plane UV interleaved */ +		.h_align	= 4, +		.v_align	= 4, +		.flags		= MXC_JPEG_FMT_TYPE_RAW, +	}, +	{ +		.name		= "YUV422", /* YUYV */ +		.fourcc		= V4L2_PIX_FMT_YUYV, +		.subsampling	= V4L2_JPEG_CHROMA_SUBSAMPLING_422, +		.nc		= 3, +		.depth		= 16, +		.colplanes	= 1, +		.h_align	= 4, +		.v_align	= 3, +		.flags		= MXC_JPEG_FMT_TYPE_RAW, +	}, +	{ +		.name		= "YUV444", /* YUVYUV */ +		.fourcc		= V4L2_PIX_FMT_YUV24, +		.subsampling	= V4L2_JPEG_CHROMA_SUBSAMPLING_444, +		.nc		= 3, +		.depth		= 24, +		.colplanes	= 1, +		.h_align	= 3, +		.v_align	= 3, +		.flags		= MXC_JPEG_FMT_TYPE_RAW, +	}, +	{ +		.name		= "Gray", /* Gray (Y8/Y12) or Single Comp */ +		.fourcc		= V4L2_PIX_FMT_GREY, +		.subsampling	= V4L2_JPEG_CHROMA_SUBSAMPLING_GRAY, +		.nc		= 1, +		.depth		= 8, +		.colplanes	= 1, +		.h_align	= 3, +		.v_align	= 3, +		.flags		= MXC_JPEG_FMT_TYPE_RAW, +	}, +}; + +#define MXC_JPEG_NUM_FORMATS ARRAY_SIZE(mxc_formats) + +static const int mxc_decode_mode = MXC_JPEG_DECODE; +static const int mxc_encode_mode = MXC_JPEG_ENCODE; + +static const struct of_device_id mxc_jpeg_match[] = { +	{ +		.compatible = "nxp,imx8qxp-jpgdec", +		.data       = &mxc_decode_mode, +	}, +	{ +		.compatible = "nxp,imx8qxp-jpgenc", +		.data       = &mxc_encode_mode, +	}, +	{ }, +}; + +/* + * default configuration stream, 64x64 yuv422 + * split by JPEG marker, so it's easier to modify & use + */ +static const unsigned char jpeg_soi[] = { +	0xFF, 0xD8 +}; + +static const unsigned char jpeg_app0[] = { +	0xFF, 0xE0, +	0x00, 0x10, 0x4A, 0x46, 0x49, 0x46, 0x00, +	0x01, 0x01, 0x00, 0x00, 0x01, 0x00, 0x01, +	0x00, 0x00 +}; + +static const unsigned char jpeg_app14[] = { +	0xFF, 0xEE, +	0x00, 0x0E, 0x41, 0x64, 0x6F, 0x62, 0x65, +	0x00, 0x64, 0x00, 0x00, 0x00, 0x00, 0x00 +}; + +static const unsigned char jpeg_dqt[] = { +	0xFF, 0xDB, +	0x00, 0x84, 0x00, 0x10, 0x0B, 0x0C, 0x0E, +	0x0C, 0x0A, 0x10, 0x0E, 0x0D, 0x0E, 0x12, +	0x11, 0x10, 0x13, 0x18, 0x28, 0x1A, 0x18, +	0x16, 0x16, 0x18, 0x31, 0x23, 0x25, 0x1D, +	0x28, 0x3A, 0x33, 0x3D, 0x3C, 0x39, 0x33, +	0x38, 0x37, 0x40, 0x48, 0x5C, 0x4E, 0x40, +	0x44, 0x57, 0x45, 0x37, 0x38, 0x50, 0x6D, +	0x51, 0x57, 0x5F, 0x62, 0x67, 0x68, 0x67, +	0x3E, 0x4D, 0x71, 0x79, 0x70, 0x64, 0x78, +	0x5C, 0x65, 0x67, 0x63, 0x01, 0x11, 0x12, +	0x12, 0x18, 0x15, 0x18, 0x2F, 0x1A, 0x1A, +	0x2F, 0x63, 0x42, 0x38, 0x42, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63, 0x63, +	0x63, 0x63, 0x63, 0x63, 0x63, 0x63 +}; + +static const unsigned char jpeg_sof_maximal[] = { +	0xFF, 0xC0, +	0x00, 0x14, 0x08, 0x00, 0x40, 0x00, 0x40, +	0x04, 0x01, 0x11, 0x00, 0x02, 0x11, 0x01, +	0x03, 0x11, 0x01, 0x04, 0x11, 0x01 +}; + +static const unsigned char jpeg_dht[] = { +	0xFF, 0xC4, +	0x01, 0xA2, 0x00, 0x00, 0x01, 0x05, 0x01, +	0x01, 0x01, 0x01, 0x01, 0x01, 0x00, 0x00, +	0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, +	0x02, 0x03, 0x04, 0x05, 0x06, 0x07, 0x08, +	0x09, 0x0A, 0x0B, 0x10, 0x00, 0x02, 0x01, +	0x03, 0x03, 0x02, 0x04, 0x03, 0x05, 0x05, +	0x04, 0x04, 0x00, 0x00, 0x01, 0x7D, 0x01, +	0x02, 0x03, 0x00, 0x04, 0x11, 0x05, 0x12, +	0x21, 0x31, 0x41, 0x06, 0x13, 0x51, 0x61, +	0x07, 0x22, 0x71, 0x14, 0x32, 0x81, 0x91, +	0xA1, 0x08, 0x23, 0x42, 0xB1, 0xC1, 0x15, +	0x52, 0xD1, 0xF0, 0x24, 0x33, 0x62, 0x72, +	0x82, 0x09, 0x0A, 0x16, 0x17, 0x18, 0x19, +	0x1A, 0x25, 0x26, 0x27, 0x28, 0x29, 0x2A, +	0x34, 0x35, 0x36, 0x37, 0x38, 0x39, 0x3A, +	0x43, 0x44, 0x45, 0x46, 0x47, 0x48, 0x49, +	0x4A, 0x53, 0x54, 0x55, 0x56, 0x57, 0x58, +	0x59, 0x5A, 0x63, 0x64, 0x65, 0x66, 0x67, +	0x68, 0x69, 0x6A, 0x73, 0x74, 0x75, 0x76, +	0x77, 0x78, 0x79, 0x7A, 0x83, 0x84, 0x85, +	0x86, 0x87, 0x88, 0x89, 0x8A, 0x92, 0x93, +	0x94, 0x95, 0x96, 0x97, 0x98, 0x99, 0x9A, +	0xA2, 0xA3, 0xA4, 0xA5, 0xA6, 0xA7, 0xA8, +	0xA9, 0xAA, 0xB2, 0xB3, 0xB4, 0xB5, 0xB6, +	0xB7, 0xB8, 0xB9, 0xBA, 0xC2, 0xC3, 0xC4, +	0xC5, 0xC6, 0xC7, 0xC8, 0xC9, 0xCA, 0xD2, +	0xD3, 0xD4, 0xD5, 0xD6, 0xD7, 0xD8, 0xD9, +	0xDA, 0xE1, 0xE2, 0xE3, 0xE4, 0xE5, 0xE6, +	0xE7, 0xE8, 0xE9, 0xEA, 0xF1, 0xF2, 0xF3, +	0xF4, 0xF5, 0xF6, 0xF7, 0xF8, 0xF9, 0xFA, +	0x01, 0x00, 0x03, 0x01, 0x01, 0x01, 0x01, +	0x01, 0x01, 0x01, 0x01, 0x01, 0x00, 0x00, +	0x00, 0x00, 0x00, 0x00, 0x01, 0x02, 0x03, +	0x04, 0x05, 0x06, 0x07, 0x08, 0x09, 0x0A, +	0x0B, 0x11, 0x00, 0x02, 0x01, 0x02, 0x04, +	0x04, 0x03, 0x04, 0x07, 0x05, 0x04, 0x04, +	0x00, 0x01, 0x02, 0x77, 0x00, 0x01, 0x02, +	0x03, 0x11, 0x04, 0x05, 0x21, 0x31, 0x06, +	0x12, 0x41, 0x51, 0x07, 0x61, 0x71, 0x13, +	0x22, 0x32, 0x81, 0x08, 0x14, 0x42, 0x91, +	0xA1, 0xB1, 0xC1, 0x09, 0x23, 0x33, 0x52, +	0xF0, 0x15, 0x62, 0x72, 0xD1, 0x0A, 0x16, +	0x24, 0x34, 0xE1, 0x25, 0xF1, 0x17, 0x18, +	0x19, 0x1A, 0x26, 0x27, 0x28, 0x29, 0x2A, +	0x35, 0x36, 0x37, 0x38, 0x39, 0x3A, 0x43, +	0x44, 0x45, 0x46, 0x47, 0x48, 0x49, 0x4A, +	0x53, 0x54, 0x55, 0x56, 0x57, 0x58, 0x59, +	0x5A, 0x63, 0x64, 0x65, 0x66, 0x67, 0x68, +	0x69, 0x6A, 0x73, 0x74, 0x75, 0x76, 0x77, +	0x78, 0x79, 0x7A, 0x82, 0x83, 0x84, 0x85, +	0x86, 0x87, 0x88, 0x89, 0x8A, 0x92, 0x93, +	0x94, 0x95, 0x96, 0x97, 0x98, 0x99, 0x9A, +	0xA2, 0xA3, 0xA4, 0xA5, 0xA6, 0xA7, 0xA8, +	0xA9, 0xAA, 0xB2, 0xB3, 0xB4, 0xB5, 0xB6, +	0xB7, 0xB8, 0xB9, 0xBA, 0xC2, 0xC3, 0xC4, +	0xC5, 0xC6, 0xC7, 0xC8, 0xC9, 0xCA, 0xD2, +	0xD3, 0xD4, 0xD5, 0xD6, 0xD7, 0xD8, 0xD9, +	0xDA, 0xE2, 0xE3, 0xE4, 0xE5, 0xE6, 0xE7, +	0xE8, 0xE9, 0xEA, 0xF2, 0xF3, 0xF4, 0xF5, +	0xF6, 0xF7, 0xF8, 0xF9, 0xFA +}; + +static const unsigned char jpeg_dri[] = { +	0xFF, 0xDD, +	0x00, 0x04, 0x00, 0x20 +}; + +static const unsigned char jpeg_sos_maximal[] = { +	0xFF, 0xDA, +	0x00, 0x0C, 0x04, 0x01, 0x00, 0x02, 0x11, 0x03, +	0x11, 0x04, 0x11, 0x00, 0x3F, 0x00 +}; + +static const unsigned char jpeg_eoi[] = { +	0xFF, 0xD9 +}; + +struct mxc_jpeg_src_buf { +	/* common v4l buffer stuff -- must be first */ +	struct vb2_v4l2_buffer	b; +	struct list_head	list; + +	/* mxc-jpeg specific */ +	bool			dht_needed; +	bool			jpeg_parse_error; +}; + +static inline struct mxc_jpeg_src_buf *vb2_to_mxc_buf(struct vb2_buffer *vb) +{ +	return container_of(to_vb2_v4l2_buffer(vb), +			    struct mxc_jpeg_src_buf, b); +} + +static unsigned int debug; +module_param(debug, int, 0644); +MODULE_PARM_DESC(debug, "Debug level (0-3)"); + +static void _bswap16(u16 *a) +{ +	*a = ((*a & 0x00FF) << 8) | ((*a & 0xFF00) >> 8); +} + +static void print_mxc_buf(struct mxc_jpeg_dev *jpeg, struct vb2_buffer *buf, +			  unsigned long len) +{ +	unsigned int plane_no; +	u32 dma_addr; +	void *vaddr; +	unsigned long payload; + +	if (debug < 3) +		return; + +	for (plane_no = 0; plane_no < buf->num_planes; plane_no++) { +		payload = vb2_get_plane_payload(buf, plane_no); +		if (len == 0) +			len = payload; +		dma_addr = vb2_dma_contig_plane_dma_addr(buf, plane_no); +		vaddr = vb2_plane_vaddr(buf, plane_no); +		v4l2_dbg(3, debug, &jpeg->v4l2_dev, +			 "plane %d (vaddr=%p dma_addr=%x payload=%ld):", +			  plane_no, vaddr, dma_addr, payload); +		print_hex_dump(KERN_DEBUG, "", DUMP_PREFIX_OFFSET, 32, 1, +			       vaddr, len, false); +	} +} + +static inline struct mxc_jpeg_ctx *mxc_jpeg_fh_to_ctx(struct v4l2_fh *fh) +{ +	return container_of(fh, struct mxc_jpeg_ctx, fh); +} + +static int enum_fmt(struct mxc_jpeg_fmt *mxc_formats, int n, +		    struct v4l2_fmtdesc *f, u32 type) +{ +	int i, num = 0; + +	for (i = 0; i < n; ++i) { +		if (mxc_formats[i].flags == type) { +			/* index-th format of searched type found ? */ +			if (num == f->index) +				break; +			/* Correct type but haven't reached our index yet, +			 * just increment per-type index +			 */ +			++num; +		} +	} + +	/* Format not found */ +	if (i >= n) +		return -EINVAL; + +	strscpy(f->description, mxc_formats[i].name, sizeof(f->description)); +	f->pixelformat = mxc_formats[i].fourcc; + +	return 0; +} + +static struct mxc_jpeg_fmt *mxc_jpeg_find_format(struct mxc_jpeg_ctx *ctx, +						 u32 pixelformat) +{ +	unsigned int k; + +	for (k = 0; k < MXC_JPEG_NUM_FORMATS; k++) { +		struct mxc_jpeg_fmt *fmt = &mxc_formats[k]; + +		if (fmt->fourcc == pixelformat) +			return fmt; +	} +	return NULL; +} + +static enum mxc_jpeg_image_format mxc_jpeg_fourcc_to_imgfmt(u32 fourcc) +{ +	switch (fourcc) { +	case V4L2_PIX_FMT_GREY: +		return MXC_JPEG_GRAY; +	case V4L2_PIX_FMT_YUYV: +		return MXC_JPEG_YUV422; +	case V4L2_PIX_FMT_NV12: +		return MXC_JPEG_YUV420; +	case V4L2_PIX_FMT_YUV24: +		return MXC_JPEG_YUV444; +	case V4L2_PIX_FMT_RGB24: +		return MXC_JPEG_RGB; +	case V4L2_PIX_FMT_ARGB32: +		return MXC_JPEG_ARGB; +	default: +		return MXC_JPEG_INVALID; +	} +} + +static struct mxc_jpeg_q_data *mxc_jpeg_get_q_data(struct mxc_jpeg_ctx *ctx, +						   enum v4l2_buf_type type) +{ +	if (V4L2_TYPE_IS_OUTPUT(type)) +		return &ctx->out_q; +	return &ctx->cap_q; +} + +static void mxc_jpeg_addrs(struct mxc_jpeg_desc *desc, +			   struct vb2_buffer *raw_buf, +			   struct vb2_buffer *jpeg_buf, int offset) +{ +	int img_fmt = desc->stm_ctrl & STM_CTRL_IMAGE_FORMAT_MASK; + +	desc->buf_base0 = vb2_dma_contig_plane_dma_addr(raw_buf, 0); +	desc->buf_base1 = 0; +	if (img_fmt == STM_CTRL_IMAGE_FORMAT(MXC_JPEG_YUV420)) { +		WARN_ON(raw_buf->num_planes < 2); +		desc->buf_base1 = vb2_dma_contig_plane_dma_addr(raw_buf, 1); +	} +	desc->stm_bufbase = vb2_dma_contig_plane_dma_addr(jpeg_buf, 0) + +		offset; +} + +static void notify_eos(struct mxc_jpeg_ctx *ctx) +{ +	const struct v4l2_event ev = { +		.type = V4L2_EVENT_EOS +	}; + +	dev_dbg(ctx->mxc_jpeg->dev, "Notify app event EOS reached"); +	v4l2_event_queue_fh(&ctx->fh, &ev); +} + +static void notify_src_chg(struct mxc_jpeg_ctx *ctx) +{ +	const struct v4l2_event ev = { +			.type = V4L2_EVENT_SOURCE_CHANGE, +			.u.src_change.changes = +			V4L2_EVENT_SRC_CH_RESOLUTION, +		}; + +	dev_dbg(ctx->mxc_jpeg->dev, "Notify app event SRC_CH_RESOLUTION"); +	v4l2_event_queue_fh(&ctx->fh, &ev); +} + +static int mxc_get_free_slot(struct mxc_jpeg_slot_data slot_data[], int n) +{ +	int free_slot = 0; + +	while (slot_data[free_slot].used && free_slot < n) +		free_slot++; + +	return free_slot; /* >=n when there are no more free slots */ +} + +static bool mxc_jpeg_alloc_slot_data(struct mxc_jpeg_dev *jpeg, +				     unsigned int slot) +{ +	struct mxc_jpeg_desc *desc; +	struct mxc_jpeg_desc *cfg_desc; +	void *cfg_stm; + +	if (jpeg->slot_data[slot].desc) +		goto skip_alloc; /* already allocated, reuse it */ + +	/* allocate descriptor for decoding/encoding phase */ +	desc = dma_alloc_coherent(jpeg->dev, +				  sizeof(struct mxc_jpeg_desc), +				  &jpeg->slot_data[slot].desc_handle, +				  GFP_ATOMIC); +	if (!desc) +		goto err; +	jpeg->slot_data[slot].desc = desc; + +	/* allocate descriptor for configuration phase (encoder only) */ +	cfg_desc = dma_alloc_coherent(jpeg->dev, +				      sizeof(struct mxc_jpeg_desc), +				      &jpeg->slot_data[slot].cfg_desc_handle, +				      GFP_ATOMIC); +	if (!cfg_desc) +		goto err; +	jpeg->slot_data[slot].cfg_desc = cfg_desc; + +	/* allocate configuration stream */ +	cfg_stm = dma_alloc_coherent(jpeg->dev, +				     MXC_JPEG_MAX_CFG_STREAM, +				     &jpeg->slot_data[slot].cfg_stream_handle, +				     GFP_ATOMIC); +	if (!cfg_stm) +		goto err; +	jpeg->slot_data[slot].cfg_stream_vaddr = cfg_stm; + +skip_alloc: +	jpeg->slot_data[slot].used = true; + +	return true; +err: +	dev_err(jpeg->dev, "Could not allocate descriptors for slot %d", slot); + +	return false; +} + +static void mxc_jpeg_free_slot_data(struct mxc_jpeg_dev *jpeg, +				    unsigned int slot) +{ +	if (slot >= MXC_MAX_SLOTS) { +		dev_err(jpeg->dev, "Invalid slot %d, nothing to free.", slot); +		return; +	} + +	/* free descriptor for decoding/encoding phase */ +	dma_free_coherent(jpeg->dev, sizeof(struct mxc_jpeg_desc), +			  jpeg->slot_data[slot].desc, +			  jpeg->slot_data[slot].desc_handle); + +	/* free descriptor for encoder configuration phase / decoder DHT */ +	dma_free_coherent(jpeg->dev, sizeof(struct mxc_jpeg_desc), +			  jpeg->slot_data[slot].cfg_desc, +			  jpeg->slot_data[slot].cfg_desc_handle); + +	/* free configuration stream */ +	dma_free_coherent(jpeg->dev, MXC_JPEG_MAX_CFG_STREAM, +			  jpeg->slot_data[slot].cfg_stream_vaddr, +			  jpeg->slot_data[slot].cfg_stream_handle); + +	jpeg->slot_data[slot].used = false; +} + +static irqreturn_t mxc_jpeg_dec_irq(int irq, void *priv) +{ +	struct mxc_jpeg_dev *jpeg = priv; +	struct mxc_jpeg_ctx *ctx; +	void __iomem *reg = jpeg->base_reg; +	struct device *dev = jpeg->dev; +	struct vb2_v4l2_buffer *src_buf, *dst_buf; +	struct mxc_jpeg_src_buf *jpeg_src_buf; +	enum vb2_buffer_state buf_state; +	u32 dec_ret, com_status; +	unsigned long payload; +	struct mxc_jpeg_q_data *q_data; +	enum v4l2_buf_type cap_type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; +	unsigned int slot; + +	spin_lock(&jpeg->hw_lock); + +	com_status = readl(reg + COM_STATUS); +	slot = COM_STATUS_CUR_SLOT(com_status); +	dev_dbg(dev, "Irq %d on slot %d.\n", irq, slot); + +	ctx = v4l2_m2m_get_curr_priv(jpeg->m2m_dev); +	if (!ctx) { +		dev_err(dev, +			"Instance released before the end of transaction.\n"); +		/* soft reset only resets internal state, not registers */ +		mxc_jpeg_sw_reset(reg); +		/* clear all interrupts */ +		writel(0xFFFFFFFF, reg + MXC_SLOT_OFFSET(slot, SLOT_STATUS)); +		goto job_unlock; +	} + +	if (slot != ctx->slot) { +		/* TODO investigate when adding multi-instance support */ +		dev_warn(dev, "IRQ slot %d != context slot %d.\n", +			 slot, ctx->slot); +		goto job_unlock; +	} + +	dec_ret = readl(reg + MXC_SLOT_OFFSET(slot, SLOT_STATUS)); +	writel(dec_ret, reg + MXC_SLOT_OFFSET(slot, SLOT_STATUS)); /* w1c */ + +	dst_buf = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); +	src_buf = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); +	jpeg_src_buf = vb2_to_mxc_buf(&src_buf->vb2_buf); + +	if (dec_ret & SLOT_STATUS_ENC_CONFIG_ERR) { +		u32 ret = readl(reg + CAST_STATUS12); + +		dev_err(dev, "Encoder/decoder error, status=0x%08x", ret); +		mxc_jpeg_sw_reset(reg); +		buf_state = VB2_BUF_STATE_ERROR; +		goto buffers_done; +	} + +	if (!(dec_ret & SLOT_STATUS_FRMDONE)) +		goto job_unlock; + +	if (jpeg->mode == MXC_JPEG_ENCODE && +	    ctx->enc_state == MXC_JPEG_ENC_CONF) { +		ctx->enc_state = MXC_JPEG_ENCODING; +		dev_dbg(dev, "Encoder config finished. Start encoding...\n"); +		mxc_jpeg_enc_mode_go(dev, reg); +		goto job_unlock; +	} +	if (jpeg->mode == MXC_JPEG_DECODE && jpeg_src_buf->dht_needed) { +		jpeg_src_buf->dht_needed = false; +		dev_dbg(dev, "Decoder DHT cfg finished. Start decoding...\n"); +		goto job_unlock; +	} +	if (jpeg->mode == MXC_JPEG_ENCODE) { +		payload = readl(reg + MXC_SLOT_OFFSET(slot, SLOT_BUF_PTR)); +		vb2_set_plane_payload(&dst_buf->vb2_buf, 0, payload); +		dev_dbg(dev, "Encoding finished, payload size: %ld\n", +			payload); +	} else { +		q_data = mxc_jpeg_get_q_data(ctx, cap_type); +		payload = q_data->sizeimage[0]; +		vb2_set_plane_payload(&dst_buf->vb2_buf, 0, payload); +		vb2_set_plane_payload(&dst_buf->vb2_buf, 1, 0); +		if (q_data->fmt->colplanes == 2) { +			payload = q_data->sizeimage[1]; +			vb2_set_plane_payload(&dst_buf->vb2_buf, 1, payload); +		} +		dev_dbg(dev, "Decoding finished, payload size: %ld + %ld\n", +			vb2_get_plane_payload(&dst_buf->vb2_buf, 0), +			vb2_get_plane_payload(&dst_buf->vb2_buf, 1)); +	} + +	/* short preview of the results */ +	dev_dbg(dev, "src_buf preview: "); +	print_mxc_buf(jpeg, &src_buf->vb2_buf, 32); +	dev_dbg(dev, "dst_buf preview: "); +	print_mxc_buf(jpeg, &dst_buf->vb2_buf, 32); +	buf_state = VB2_BUF_STATE_DONE; + +buffers_done: +	jpeg->slot_data[slot].used = false; /* unused, but don't free */ +	v4l2_m2m_src_buf_remove(ctx->fh.m2m_ctx); +	v4l2_m2m_dst_buf_remove(ctx->fh.m2m_ctx); +	v4l2_m2m_buf_done(src_buf, buf_state); +	v4l2_m2m_buf_done(dst_buf, buf_state); +	spin_unlock(&jpeg->hw_lock); +	v4l2_m2m_job_finish(jpeg->m2m_dev, ctx->fh.m2m_ctx); +	return IRQ_HANDLED; +job_unlock: +	spin_unlock(&jpeg->hw_lock); +	return IRQ_HANDLED; +} + +static int mxc_jpeg_fixup_sof(struct mxc_jpeg_sof *sof, +			      u32 fourcc, +			      u16 w, u16 h) +{ +	int sof_length; + +	sof->precision = 8; /* TODO allow 8/12 bit precision*/ +	sof->height = h; +	_bswap16(&sof->height); +	sof->width = w; +	_bswap16(&sof->width); + +	switch (fourcc) { +	case V4L2_PIX_FMT_NV12: +		sof->components_no = 3; +		sof->comp[0].v = 0x2; +		sof->comp[0].h = 0x2; +		break; +	case V4L2_PIX_FMT_YUYV: +		sof->components_no = 3; +		sof->comp[0].v = 0x1; +		sof->comp[0].h = 0x2; +		break; +	case V4L2_PIX_FMT_YUV24: +	case V4L2_PIX_FMT_RGB24: +	default: +		sof->components_no = 3; +		break; +	case V4L2_PIX_FMT_ARGB32: +		sof->components_no = 4; +		break; +	case V4L2_PIX_FMT_GREY: +		sof->components_no = 1; +		break; +	} +	sof_length = 8 + 3 * sof->components_no; +	sof->length = sof_length; +	_bswap16(&sof->length); + +	return sof_length; /* not swaped */ +} + +static int mxc_jpeg_fixup_sos(struct mxc_jpeg_sos *sos, +			      u32 fourcc) +{ +	int sos_length; +	u8 *sof_u8 = (u8 *)sos; + +	switch (fourcc) { +	case V4L2_PIX_FMT_NV12: +		sos->components_no = 3; +		break; +	case V4L2_PIX_FMT_YUYV: +		sos->components_no = 3; +		break; +	case V4L2_PIX_FMT_YUV24: +	case V4L2_PIX_FMT_RGB24: +	default: +		sos->components_no = 3; +		break; +	case V4L2_PIX_FMT_ARGB32: +		sos->components_no = 4; +		break; +	case V4L2_PIX_FMT_GREY: +		sos->components_no = 1; +		break; +	} +	sos_length = 6 + 2 * sos->components_no; +	sos->length = sos_length; +	_bswap16(&sos->length); + +	/* SOS ignorable bytes, not so ignorable after all */ +	sof_u8[sos_length - 1] = 0x0; +	sof_u8[sos_length - 2] = 0x3f; +	sof_u8[sos_length - 3] = 0x0; + +	return sos_length; /* not swaped */ +} + +static unsigned int mxc_jpeg_setup_cfg_stream(void *cfg_stream_vaddr, +					      u32 fourcc, +					      u16 w, u16 h) +{ +	unsigned int offset = 0; +	u8 *cfg = (u8 *)cfg_stream_vaddr; +	struct mxc_jpeg_sof *sof; +	struct mxc_jpeg_sos *sos; + +	memcpy(cfg + offset, jpeg_soi, ARRAY_SIZE(jpeg_soi)); +	offset += ARRAY_SIZE(jpeg_soi); + +	if (fourcc == V4L2_PIX_FMT_RGB24 || +	    fourcc == V4L2_PIX_FMT_ARGB32) { +		memcpy(cfg + offset, jpeg_app14, sizeof(jpeg_app14)); +		offset += sizeof(jpeg_app14); +	} else { +		memcpy(cfg + offset, jpeg_app0, sizeof(jpeg_app0)); +		offset += sizeof(jpeg_app0); +	} + +	memcpy(cfg + offset, jpeg_dqt, sizeof(jpeg_dqt)); +	offset += sizeof(jpeg_dqt); + +	memcpy(cfg + offset, jpeg_sof_maximal, sizeof(jpeg_sof_maximal)); +	offset += 2; /* skip marker ID */ +	sof = (struct mxc_jpeg_sof *)(cfg + offset); +	offset += mxc_jpeg_fixup_sof(sof, fourcc, w, h); + +	memcpy(cfg + offset, jpeg_dht, sizeof(jpeg_dht)); +	offset += sizeof(jpeg_dht); + +	memcpy(cfg + offset, jpeg_dri, sizeof(jpeg_dri)); +	offset += sizeof(jpeg_dri); + +	memcpy(cfg + offset, jpeg_sos_maximal, sizeof(jpeg_sos_maximal)); +	offset += 2; /* skip marker ID */ +	sos = (struct mxc_jpeg_sos *)(cfg + offset); +	offset += mxc_jpeg_fixup_sos(sos, fourcc); + +	memcpy(cfg + offset, jpeg_eoi, sizeof(jpeg_eoi)); +	offset += sizeof(jpeg_eoi); + +	return offset; +} + +static void mxc_jpeg_config_dec_desc(struct vb2_buffer *out_buf, +				     struct mxc_jpeg_ctx *ctx, +				     struct vb2_buffer *src_buf, +				     struct vb2_buffer *dst_buf) +{ +	enum v4l2_buf_type cap_type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; +	struct mxc_jpeg_q_data *q_data_cap; +	enum mxc_jpeg_image_format img_fmt; +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	void __iomem *reg = jpeg->base_reg; +	unsigned int slot = ctx->slot; +	struct mxc_jpeg_desc *desc = jpeg->slot_data[slot].desc; +	struct mxc_jpeg_desc *cfg_desc = jpeg->slot_data[slot].cfg_desc; +	dma_addr_t desc_handle = jpeg->slot_data[slot].desc_handle; +	dma_addr_t cfg_desc_handle = jpeg->slot_data[slot].cfg_desc_handle; +	dma_addr_t cfg_stream_handle = jpeg->slot_data[slot].cfg_stream_handle; +	unsigned int *cfg_size = &jpeg->slot_data[slot].cfg_stream_size; +	void *cfg_stream_vaddr = jpeg->slot_data[slot].cfg_stream_vaddr; +	struct mxc_jpeg_src_buf *jpeg_src_buf; + +	jpeg_src_buf = vb2_to_mxc_buf(src_buf); + +	/* setup the decoding descriptor */ +	desc->next_descpt_ptr = 0; /* end of chain */ +	q_data_cap = mxc_jpeg_get_q_data(ctx, cap_type); +	desc->imgsize = q_data_cap->w_adjusted << 16 | q_data_cap->h_adjusted; +	img_fmt = mxc_jpeg_fourcc_to_imgfmt(q_data_cap->fmt->fourcc); +	desc->stm_ctrl &= ~STM_CTRL_IMAGE_FORMAT(0xF); /* clear image format */ +	desc->stm_ctrl |= STM_CTRL_IMAGE_FORMAT(img_fmt); +	desc->line_pitch = q_data_cap->bytesperline[0]; +	mxc_jpeg_addrs(desc, dst_buf, src_buf, 0); +	mxc_jpeg_set_bufsize(desc, ALIGN(vb2_plane_size(src_buf, 0), 1024)); +	print_descriptor_info(jpeg->dev, desc); + +	if (!jpeg_src_buf->dht_needed) { +		/* validate the decoding descriptor */ +		mxc_jpeg_set_desc(desc_handle, reg, slot); +		return; +	} + +	/* +	 * if a default huffman table is needed, use the config descriptor to +	 * inject a DHT, by chaining it before the decoding descriptor +	 */ +	*cfg_size = mxc_jpeg_setup_cfg_stream(cfg_stream_vaddr, +					      V4L2_PIX_FMT_YUYV, +					      MXC_JPEG_MIN_WIDTH, +					      MXC_JPEG_MIN_HEIGHT); +	cfg_desc->next_descpt_ptr = desc_handle | MXC_NXT_DESCPT_EN; +	cfg_desc->buf_base0 = vb2_dma_contig_plane_dma_addr(dst_buf, 0); +	cfg_desc->buf_base1 = 0; +	cfg_desc->imgsize = MXC_JPEG_MIN_WIDTH << 16; +	cfg_desc->imgsize |= MXC_JPEG_MIN_HEIGHT; +	cfg_desc->line_pitch = MXC_JPEG_MIN_WIDTH * 2; +	cfg_desc->stm_ctrl = STM_CTRL_IMAGE_FORMAT(MXC_JPEG_YUV422); +	cfg_desc->stm_bufbase = cfg_stream_handle; +	cfg_desc->stm_bufsize = ALIGN(*cfg_size, 1024); +	print_descriptor_info(jpeg->dev, cfg_desc); + +	/* validate the configuration descriptor */ +	mxc_jpeg_set_desc(cfg_desc_handle, reg, slot); +} + +static void mxc_jpeg_config_enc_desc(struct vb2_buffer *out_buf, +				     struct mxc_jpeg_ctx *ctx, +				     struct vb2_buffer *src_buf, +				     struct vb2_buffer *dst_buf) +{ +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	void __iomem *reg = jpeg->base_reg; +	unsigned int slot = ctx->slot; +	struct mxc_jpeg_desc *desc = jpeg->slot_data[slot].desc; +	struct mxc_jpeg_desc *cfg_desc = jpeg->slot_data[slot].cfg_desc; +	dma_addr_t desc_handle = jpeg->slot_data[slot].desc_handle; +	dma_addr_t cfg_desc_handle = jpeg->slot_data[slot].cfg_desc_handle; +	void *cfg_stream_vaddr = jpeg->slot_data[slot].cfg_stream_vaddr; +	struct mxc_jpeg_q_data *q_data; +	enum mxc_jpeg_image_format img_fmt; +	int w, h; + +	q_data = mxc_jpeg_get_q_data(ctx, src_buf->vb2_queue->type); + +	jpeg->slot_data[slot].cfg_stream_size = +			mxc_jpeg_setup_cfg_stream(cfg_stream_vaddr, +						  q_data->fmt->fourcc, +						  q_data->w_adjusted, +						  q_data->h_adjusted); + +	/* chain the config descriptor with the encoding descriptor */ +	cfg_desc->next_descpt_ptr = desc_handle | MXC_NXT_DESCPT_EN; + +	cfg_desc->buf_base0 = jpeg->slot_data[slot].cfg_stream_handle; +	cfg_desc->buf_base1 = 0; +	cfg_desc->line_pitch = 0; +	cfg_desc->stm_bufbase = 0; /* no output expected */ +	cfg_desc->stm_bufsize = 0x0; +	cfg_desc->imgsize = 0; +	cfg_desc->stm_ctrl = STM_CTRL_CONFIG_MOD(1); + +	desc->next_descpt_ptr = 0; /* end of chain */ + +	/* use adjusted resolution for CAST IP job */ +	w = q_data->w_adjusted; +	h = q_data->h_adjusted; +	mxc_jpeg_set_res(desc, w, h); +	mxc_jpeg_set_line_pitch(desc, w * (q_data->fmt->depth / 8)); +	mxc_jpeg_set_bufsize(desc, desc->line_pitch * h); +	img_fmt = mxc_jpeg_fourcc_to_imgfmt(q_data->fmt->fourcc); +	if (img_fmt == MXC_JPEG_INVALID) +		dev_err(jpeg->dev, "No valid image format detected\n"); +	desc->stm_ctrl = STM_CTRL_CONFIG_MOD(0) | +			 STM_CTRL_IMAGE_FORMAT(img_fmt); +	mxc_jpeg_addrs(desc, src_buf, dst_buf, 0); +	dev_dbg(jpeg->dev, "cfg_desc:\n"); +	print_descriptor_info(jpeg->dev, cfg_desc); +	dev_dbg(jpeg->dev, "enc desc:\n"); +	print_descriptor_info(jpeg->dev, desc); +	print_wrapper_info(jpeg->dev, reg); +	print_cast_status(jpeg->dev, reg, MXC_JPEG_ENCODE); + +	/* validate the configuration descriptor */ +	mxc_jpeg_set_desc(cfg_desc_handle, reg, slot); +} + +static void mxc_jpeg_device_run(void *priv) +{ +	struct mxc_jpeg_ctx *ctx = priv; +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	void __iomem *reg = jpeg->base_reg; +	struct device *dev = jpeg->dev; +	struct vb2_v4l2_buffer *src_buf, *dst_buf; +	unsigned long flags; +	struct mxc_jpeg_q_data *q_data_cap, *q_data_out; +	struct mxc_jpeg_src_buf *jpeg_src_buf; + +	spin_lock_irqsave(&ctx->mxc_jpeg->hw_lock, flags); +	src_buf = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); +	dst_buf = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); +	if (!src_buf || !dst_buf) { +		dev_err(dev, "Null src or dst buf\n"); +		goto end; +	} + +	q_data_cap = mxc_jpeg_get_q_data(ctx, V4L2_BUF_TYPE_VIDEO_CAPTURE); +	if (!q_data_cap) +		goto end; +	q_data_out = mxc_jpeg_get_q_data(ctx, V4L2_BUF_TYPE_VIDEO_OUTPUT); +	if (!q_data_out) +		goto end; +	src_buf->sequence = q_data_out->sequence++; +	dst_buf->sequence = q_data_cap->sequence++; + +	v4l2_m2m_buf_copy_metadata(src_buf, dst_buf, true); + +	jpeg_src_buf = vb2_to_mxc_buf(&src_buf->vb2_buf); +	if (jpeg_src_buf->jpeg_parse_error) { +		jpeg->slot_data[ctx->slot].used = false; +		v4l2_m2m_src_buf_remove(ctx->fh.m2m_ctx); +		v4l2_m2m_dst_buf_remove(ctx->fh.m2m_ctx); +		v4l2_m2m_buf_done(src_buf, VB2_BUF_STATE_ERROR); +		v4l2_m2m_buf_done(dst_buf, VB2_BUF_STATE_ERROR); +		spin_unlock_irqrestore(&ctx->mxc_jpeg->hw_lock, flags); +		v4l2_m2m_job_finish(jpeg->m2m_dev, ctx->fh.m2m_ctx); + +		return; +	} + +	/* +	 * TODO: this reset should be removed, once we figure out +	 * how to overcome hardware issues both on encoder and decoder +	 */ +	mxc_jpeg_sw_reset(reg); +	mxc_jpeg_enable(reg); +	mxc_jpeg_set_l_endian(reg, 1); + +	ctx->slot = mxc_get_free_slot(jpeg->slot_data, MXC_MAX_SLOTS); +	if (ctx->slot >= MXC_MAX_SLOTS) { +		dev_err(dev, "No more free slots\n"); +		goto end; +	} +	if (!mxc_jpeg_alloc_slot_data(jpeg, ctx->slot)) { +		dev_err(dev, "Cannot allocate slot data\n"); +		goto end; +	} + +	mxc_jpeg_enable_slot(reg, ctx->slot); +	mxc_jpeg_enable_irq(reg, ctx->slot); + +	if (jpeg->mode == MXC_JPEG_ENCODE) { +		dev_dbg(dev, "Encoding on slot %d\n", ctx->slot); +		ctx->enc_state = MXC_JPEG_ENC_CONF; +		mxc_jpeg_config_enc_desc(&dst_buf->vb2_buf, ctx, +					 &src_buf->vb2_buf, &dst_buf->vb2_buf); +		mxc_jpeg_enc_mode_conf(dev, reg); /* start config phase */ +	} else { +		dev_dbg(dev, "Decoding on slot %d\n", ctx->slot); +		print_mxc_buf(jpeg, &src_buf->vb2_buf, 0); +		mxc_jpeg_config_dec_desc(&dst_buf->vb2_buf, ctx, +					 &src_buf->vb2_buf, &dst_buf->vb2_buf); +		mxc_jpeg_dec_mode_go(dev, reg); +	} +end: +	spin_unlock_irqrestore(&ctx->mxc_jpeg->hw_lock, flags); +} + +static int mxc_jpeg_decoder_cmd(struct file *file, void *priv, +				struct v4l2_decoder_cmd *cmd) +{ +	struct v4l2_fh *fh = file->private_data; +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(fh); +	struct device *dev = ctx->mxc_jpeg->dev; +	int ret; + +	ret = v4l2_m2m_ioctl_try_decoder_cmd(file, fh, cmd); +	if (ret < 0) +		return ret; + +	if (cmd->cmd == V4L2_DEC_CMD_STOP) { +		dev_dbg(dev, "Received V4L2_DEC_CMD_STOP"); +		if (v4l2_m2m_num_src_bufs_ready(fh->m2m_ctx) == 0) { +			/* No more src bufs, notify app EOS */ +			notify_eos(ctx); +		} else { +			/* will send EOS later*/ +			ctx->stopping = 1; +		} +	} + +	return 0; +} + +static int mxc_jpeg_encoder_cmd(struct file *file, void *priv, +				struct v4l2_encoder_cmd *cmd) +{ +	struct v4l2_fh *fh = file->private_data; +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(fh); +	struct device *dev = ctx->mxc_jpeg->dev; +	int ret; + +	ret = v4l2_m2m_ioctl_try_encoder_cmd(file, fh, cmd); +	if (ret < 0) +		return ret; + +	if (cmd->cmd == V4L2_ENC_CMD_STOP) { +		dev_dbg(dev, "Received V4L2_ENC_CMD_STOP"); +		if (v4l2_m2m_num_src_bufs_ready(fh->m2m_ctx) == 0) { +			/* No more src bufs, notify app EOS */ +			notify_eos(ctx); +		} else { +			/* will send EOS later*/ +			ctx->stopping = 1; +		} +	} + +	return 0; +} + +static int mxc_jpeg_queue_setup(struct vb2_queue *q, +				unsigned int *nbuffers, +				unsigned int *nplanes, +				unsigned int sizes[], +				struct device *alloc_ctxs[]) +{ +	struct mxc_jpeg_ctx *ctx = vb2_get_drv_priv(q); +	struct mxc_jpeg_q_data *q_data = NULL; +	int i; + +	q_data = mxc_jpeg_get_q_data(ctx, q->type); +	if (!q_data) +		return -EINVAL; + +	/* Handle CREATE_BUFS situation - *nplanes != 0 */ +	if (*nplanes) { +		for (i = 0; i < *nplanes; i++) { +			if (sizes[i] < q_data->sizeimage[i]) +				return -EINVAL; +		} +		return 0; +	} + +	/* Handle REQBUFS situation */ +	*nplanes = q_data->fmt->colplanes; +	for (i = 0; i < *nplanes; i++) +		sizes[i] = q_data->sizeimage[i]; + +	return 0; +} + +static int mxc_jpeg_start_streaming(struct vb2_queue *q, unsigned int count) +{ +	struct mxc_jpeg_ctx *ctx = vb2_get_drv_priv(q); +	struct mxc_jpeg_q_data *q_data = mxc_jpeg_get_q_data(ctx, q->type); + +	dev_dbg(ctx->mxc_jpeg->dev, "Start streaming ctx=%p", ctx); +	q_data->sequence = 0; + +	return 0; +} + +static void mxc_jpeg_stop_streaming(struct vb2_queue *q) +{ +	struct mxc_jpeg_ctx *ctx = vb2_get_drv_priv(q); +	struct vb2_v4l2_buffer *vbuf; + +	dev_dbg(ctx->mxc_jpeg->dev, "Stop streaming ctx=%p", ctx); + +	/* Release all active buffers */ +	for (;;) { +		if (V4L2_TYPE_IS_OUTPUT(q->type)) +			vbuf = v4l2_m2m_src_buf_remove(ctx->fh.m2m_ctx); +		else +			vbuf = v4l2_m2m_dst_buf_remove(ctx->fh.m2m_ctx); +		if (!vbuf) +			return; +		v4l2_m2m_buf_done(vbuf, VB2_BUF_STATE_ERROR); +	} +} + +static int mxc_jpeg_valid_comp_id(struct device *dev, +				  struct mxc_jpeg_sof *sof, +				  struct mxc_jpeg_sos *sos) +{ +	int valid = 1; +	int i; + +	/* +	 * there's a limitation in the IP that the component IDs must be +	 * between 0..4, if they are not, let's patch them +	 */ +	for (i = 0; i < sof->components_no; i++) +		if (sof->comp[i].id > MXC_JPEG_MAX_COMPONENTS) { +			valid = 0; +			dev_err(dev, "Component %d has invalid ID: %d", +				i, sof->comp[i].id); +		} +	if (!valid) +		/* patch all comp IDs if at least one is invalid */ +		for (i = 0; i < sof->components_no; i++) { +			dev_warn(dev, "Component %d ID patched to: %d", +				 i, i + 1); +			sof->comp[i].id = i + 1; +			sos->comp[i].id = i + 1; +		} + +	return valid; +} + +static u32 mxc_jpeg_get_image_format(struct device *dev, +				     const struct v4l2_jpeg_header *header) +{ +	int i; +	u32 fourcc = 0; + +	for (i = 0; i < MXC_JPEG_NUM_FORMATS; i++) +		if (mxc_formats[i].subsampling == header->frame.subsampling && +		    mxc_formats[i].nc == header->frame.num_components) { +			fourcc = mxc_formats[i].fourcc; +			break; +		} +	if (fourcc == 0) { +		dev_err(dev, "Could not identify image format nc=%d, subsampling=%d\n", +			header->frame.num_components, +			header->frame.subsampling); +		return fourcc; +	} +	/* +	 * If the transform flag from APP14 marker is 0, images that are +	 * encoded with 3 components have RGB colorspace, see Recommendation +	 * ITU-T T.872 chapter 6.5.3 APP14 marker segment for colour encoding +	 */ +	if (fourcc == V4L2_PIX_FMT_YUV24 || fourcc == V4L2_PIX_FMT_RGB24) { +		if (header->app14_tf == V4L2_JPEG_APP14_TF_CMYK_RGB) +			fourcc = V4L2_PIX_FMT_RGB24; +		else +			fourcc = V4L2_PIX_FMT_YUV24; +	} + +	return fourcc; +} + +static void mxc_jpeg_bytesperline(struct mxc_jpeg_q_data *q, +				  u32 precision) +{ +	/* Bytes distance between the leftmost pixels in two adjacent lines */ +	if (q->fmt->fourcc == V4L2_PIX_FMT_JPEG) { +		/* bytesperline unused for compressed formats */ +		q->bytesperline[0] = 0; +		q->bytesperline[1] = 0; +	} else if (q->fmt->fourcc == V4L2_PIX_FMT_NV12) { +		/* When the image format is planar the bytesperline value +		 * applies to the first plane and is divided by the same factor +		 * as the width field for the other planes +		 */ +		q->bytesperline[0] = q->w * (precision / 8) * +				     (q->fmt->depth / 8); +		q->bytesperline[1] = q->bytesperline[0]; +	} else { +		/* single plane formats */ +		q->bytesperline[0] = q->w * (precision / 8) * +				     (q->fmt->depth / 8); +		q->bytesperline[1] = 0; +	} +} + +static void mxc_jpeg_sizeimage(struct mxc_jpeg_q_data *q) +{ +	if (q->fmt->fourcc == V4L2_PIX_FMT_JPEG) { +		/* if no sizeimage from user, assume worst jpeg compression */ +		if (!q->sizeimage[0]) +			q->sizeimage[0] = 6 * q->w * q->h; +		q->sizeimage[1] = 0; + +		if (q->sizeimage[0] > MXC_JPEG_MAX_SIZEIMAGE) +			q->sizeimage[0] = MXC_JPEG_MAX_SIZEIMAGE; + +		/* jpeg stream size must be multiple of 1K */ +		q->sizeimage[0] = ALIGN(q->sizeimage[0], 1024); +	} else { +		q->sizeimage[0] = q->bytesperline[0] * q->h; +		q->sizeimage[1] = 0; +		if (q->fmt->fourcc == V4L2_PIX_FMT_NV12) +			q->sizeimage[1] = q->sizeimage[0] / 2; +	} +} + +static int mxc_jpeg_parse(struct mxc_jpeg_ctx *ctx, +			  u8 *src_addr, u32 size, bool *dht_needed) +{ +	struct device *dev = ctx->mxc_jpeg->dev; +	struct mxc_jpeg_q_data *q_data_out, *q_data_cap; +	enum v4l2_buf_type cap_type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; +	bool src_chg = false; +	u32 fourcc; +	struct v4l2_jpeg_header header; +	struct mxc_jpeg_sof *psof = NULL; +	struct mxc_jpeg_sos *psos = NULL; +	int ret; + +	memset(&header, 0, sizeof(header)); +	ret = v4l2_jpeg_parse_header((void *)src_addr, size, &header); +	if (ret < 0) { +		dev_err(dev, "Error parsing JPEG stream markers\n"); +		return ret; +	} + +	/* if DHT marker present, no need to inject default one */ +	*dht_needed = (header.num_dht == 0); + +	q_data_out = mxc_jpeg_get_q_data(ctx, +					 V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE); +	if (q_data_out->w == 0 && q_data_out->h == 0) { +		dev_warn(dev, "Invalid user resolution 0x0"); +		dev_warn(dev, "Keeping resolution from JPEG: %dx%d", +			 header.frame.width, header.frame.height); +		q_data_out->w = header.frame.width; +		q_data_out->h = header.frame.height; +	} else if (header.frame.width != q_data_out->w || +		   header.frame.height != q_data_out->h) { +		dev_err(dev, +			"Resolution mismatch: %dx%d (JPEG) versus %dx%d(user)", +			header.frame.width, header.frame.height, +			q_data_out->w, q_data_out->h); +		return -EINVAL; +	} +	if (header.frame.width % 8 != 0 || header.frame.height % 8 != 0) { +		dev_err(dev, "JPEG width or height not multiple of 8: %dx%d\n", +			header.frame.width, header.frame.height); +		return -EINVAL; +	} +	if (header.frame.width > MXC_JPEG_MAX_WIDTH || +	    header.frame.height > MXC_JPEG_MAX_HEIGHT) { +		dev_err(dev, "JPEG width or height should be <= 8192: %dx%d\n", +			header.frame.width, header.frame.height); +		return -EINVAL; +	} +	if (header.frame.width < MXC_JPEG_MIN_WIDTH || +	    header.frame.height < MXC_JPEG_MIN_HEIGHT) { +		dev_err(dev, "JPEG width or height should be > 64: %dx%d\n", +			header.frame.width, header.frame.height); +		return -EINVAL; +	} +	if (header.frame.num_components > V4L2_JPEG_MAX_COMPONENTS) { +		dev_err(dev, "JPEG number of components should be <=%d", +			V4L2_JPEG_MAX_COMPONENTS); +		return -EINVAL; +	} +	/* check and, if necessary, patch component IDs*/ +	psof = (struct mxc_jpeg_sof *)header.sof.start; +	psos = (struct mxc_jpeg_sos *)header.sos.start; +	if (!mxc_jpeg_valid_comp_id(dev, psof, psos)) +		dev_warn(dev, "JPEG component ids should be 0-3 or 1-4"); + +	fourcc = mxc_jpeg_get_image_format(dev, &header); +	if (fourcc == 0) +		return -EINVAL; + +	/* +	 * set-up the capture queue with the pixelformat and resolution +	 * detected from the jpeg output stream +	 */ +	q_data_cap = mxc_jpeg_get_q_data(ctx, cap_type); +	if (q_data_cap->w != header.frame.width || +	    q_data_cap->h != header.frame.height) +		src_chg = true; +	q_data_cap->w = header.frame.width; +	q_data_cap->h = header.frame.height; +	q_data_cap->fmt = mxc_jpeg_find_format(ctx, fourcc); +	q_data_cap->w_adjusted = q_data_cap->w; +	q_data_cap->h_adjusted = q_data_cap->h; +	/* +	 * align up the resolution for CAST IP, +	 * but leave the buffer resolution unchanged +	 */ +	v4l_bound_align_image(&q_data_cap->w_adjusted, +			      q_data_cap->w_adjusted,  /* adjust up */ +			      MXC_JPEG_MAX_WIDTH, +			      q_data_cap->fmt->h_align, +			      &q_data_cap->h_adjusted, +			      q_data_cap->h_adjusted, /* adjust up */ +			      MXC_JPEG_MAX_HEIGHT, +			      q_data_cap->fmt->v_align, +			      0); +	dev_dbg(dev, "Detected jpeg res=(%dx%d)->(%dx%d), pixfmt=%c%c%c%c\n", +		q_data_cap->w, q_data_cap->h, +		q_data_cap->w_adjusted, q_data_cap->h_adjusted, +		(fourcc & 0xff), +		(fourcc >>  8) & 0xff, +		(fourcc >> 16) & 0xff, +		(fourcc >> 24) & 0xff); + +	/* setup bytesperline/sizeimage for capture queue */ +	mxc_jpeg_bytesperline(q_data_cap, header.frame.precision); +	mxc_jpeg_sizeimage(q_data_cap); + +	/* +	 * if the CAPTURE format was updated with new values, regardless of +	 * whether they match the values set by the client or not, signal +	 * a source change event +	 */ +	if (src_chg) +		notify_src_chg(ctx); + +	return 0; +} + +static void mxc_jpeg_buf_queue(struct vb2_buffer *vb) +{ +	int ret; +	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); +	struct mxc_jpeg_ctx *ctx = vb2_get_drv_priv(vb->vb2_queue); +	struct mxc_jpeg_src_buf *jpeg_src_buf; + +	if (vb->vb2_queue->type == V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE) +		goto end; + +	/* for V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE */ +	if (ctx->mxc_jpeg->mode != MXC_JPEG_DECODE) +		goto end; + +	jpeg_src_buf = vb2_to_mxc_buf(vb); +	jpeg_src_buf->jpeg_parse_error = false; +	ret = mxc_jpeg_parse(ctx, +			     (u8 *)vb2_plane_vaddr(vb, 0), +			     vb2_get_plane_payload(vb, 0), +			     &jpeg_src_buf->dht_needed); +	if (ret) +		jpeg_src_buf->jpeg_parse_error = true; + +end: +	v4l2_m2m_buf_queue(ctx->fh.m2m_ctx, vbuf); +} + +static int mxc_jpeg_buf_out_validate(struct vb2_buffer *vb) +{ +	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); + +	vbuf->field = V4L2_FIELD_NONE; + +	return 0; +} + +static int mxc_jpeg_buf_prepare(struct vb2_buffer *vb) +{ +	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); +	struct mxc_jpeg_ctx *ctx = vb2_get_drv_priv(vb->vb2_queue); +	struct mxc_jpeg_q_data *q_data = NULL; +	struct device *dev = ctx->mxc_jpeg->dev; +	unsigned long sizeimage; +	int i; + +	vbuf->field = V4L2_FIELD_NONE; + +	q_data = mxc_jpeg_get_q_data(ctx, vb->vb2_queue->type); +	if (!q_data) +		return -EINVAL; +	for (i = 0; i < q_data->fmt->colplanes; i++) { +		sizeimage = q_data->sizeimage[i]; +		if (vb2_plane_size(vb, i) < sizeimage) { +			dev_err(dev, "plane %d too small (%lu < %lu)", +				i, vb2_plane_size(vb, i), sizeimage); +			return -EINVAL; +		} +		vb2_set_plane_payload(vb, i, sizeimage); +	} +	return 0; +} + +static const struct vb2_ops mxc_jpeg_qops = { +	.queue_setup		= mxc_jpeg_queue_setup, +	.wait_prepare		= vb2_ops_wait_prepare, +	.wait_finish		= vb2_ops_wait_finish, +	.buf_out_validate	= mxc_jpeg_buf_out_validate, +	.buf_prepare		= mxc_jpeg_buf_prepare, +	.start_streaming	= mxc_jpeg_start_streaming, +	.stop_streaming		= mxc_jpeg_stop_streaming, +	.buf_queue		= mxc_jpeg_buf_queue, +}; + +static int mxc_jpeg_queue_init(void *priv, struct vb2_queue *src_vq, +			       struct vb2_queue *dst_vq) +{ +	struct mxc_jpeg_ctx *ctx = priv; +	int ret; + +	src_vq->type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; +	src_vq->io_modes = VB2_MMAP | VB2_DMABUF; +	src_vq->drv_priv = ctx; +	src_vq->buf_struct_size = sizeof(struct mxc_jpeg_src_buf); +	src_vq->ops = &mxc_jpeg_qops; +	src_vq->mem_ops = &vb2_dma_contig_memops; +	src_vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY; +	src_vq->lock = &ctx->mxc_jpeg->lock; +	src_vq->dev = ctx->mxc_jpeg->dev; +	src_vq->allow_zero_bytesused = 1; /* keep old userspace apps working */ + +	ret = vb2_queue_init(src_vq); +	if (ret) +		return ret; + +	dst_vq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; +	dst_vq->io_modes = VB2_MMAP | VB2_DMABUF; +	dst_vq->drv_priv = ctx; +	dst_vq->buf_struct_size = sizeof(struct v4l2_m2m_buffer); +	dst_vq->ops = &mxc_jpeg_qops; +	dst_vq->mem_ops = &vb2_dma_contig_memops; +	dst_vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY; +	dst_vq->lock = &ctx->mxc_jpeg->lock; +	dst_vq->dev = ctx->mxc_jpeg->dev; + +	ret = vb2_queue_init(dst_vq); +	return ret; +} + +static void mxc_jpeg_set_default_params(struct mxc_jpeg_ctx *ctx) +{ +	struct mxc_jpeg_q_data *out_q = &ctx->out_q; +	struct mxc_jpeg_q_data *cap_q = &ctx->cap_q; +	struct mxc_jpeg_q_data *q[2] = {out_q, cap_q}; +	int i; + +	if (ctx->mxc_jpeg->mode == MXC_JPEG_ENCODE) { +		out_q->fmt = mxc_jpeg_find_format(ctx, MXC_JPEG_DEFAULT_PFMT); +		cap_q->fmt = mxc_jpeg_find_format(ctx, V4L2_PIX_FMT_JPEG); +	} else { +		out_q->fmt = mxc_jpeg_find_format(ctx, V4L2_PIX_FMT_JPEG); +		cap_q->fmt = mxc_jpeg_find_format(ctx, MXC_JPEG_DEFAULT_PFMT); +	} + +	for (i = 0; i < 2; i++) { +		q[i]->w = MXC_JPEG_DEFAULT_WIDTH; +		q[i]->h = MXC_JPEG_DEFAULT_HEIGHT; +		q[i]->w_adjusted = MXC_JPEG_DEFAULT_WIDTH; +		q[i]->h_adjusted = MXC_JPEG_DEFAULT_HEIGHT; +		mxc_jpeg_bytesperline(q[i], 8); +		mxc_jpeg_sizeimage(q[i]); +	} +} + +static int mxc_jpeg_open(struct file *file) +{ +	struct mxc_jpeg_dev *mxc_jpeg = video_drvdata(file); +	struct video_device *mxc_vfd = video_devdata(file); +	struct device *dev = mxc_jpeg->dev; +	struct mxc_jpeg_ctx *ctx; +	int ret = 0; + +	ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); +	if (!ctx) +		return -ENOMEM; + +	if (mutex_lock_interruptible(&mxc_jpeg->lock)) { +		ret = -ERESTARTSYS; +		goto free; +	} + +	v4l2_fh_init(&ctx->fh, mxc_vfd); +	file->private_data = &ctx->fh; +	v4l2_fh_add(&ctx->fh); + +	ctx->mxc_jpeg = mxc_jpeg; + +	ctx->fh.m2m_ctx = v4l2_m2m_ctx_init(mxc_jpeg->m2m_dev, ctx, +					    mxc_jpeg_queue_init); + +	if (IS_ERR(ctx->fh.m2m_ctx)) { +		ret = PTR_ERR(ctx->fh.m2m_ctx); +		goto error; +	} + +	mxc_jpeg_set_default_params(ctx); +	ctx->slot = MXC_MAX_SLOTS; /* slot not allocated yet */ + +	if (mxc_jpeg->mode == MXC_JPEG_DECODE) +		dev_dbg(dev, "Opened JPEG decoder instance %p\n", ctx); +	else +		dev_dbg(dev, "Opened JPEG encoder instance %p\n", ctx); +	mutex_unlock(&mxc_jpeg->lock); + +	return 0; + +error: +	v4l2_fh_del(&ctx->fh); +	v4l2_fh_exit(&ctx->fh); +	mutex_unlock(&mxc_jpeg->lock); +free: +	kfree(ctx); +	return ret; +} + +static int mxc_jpeg_querycap(struct file *file, void *priv, +			     struct v4l2_capability *cap) +{ +	struct mxc_jpeg_dev *mxc_jpeg = video_drvdata(file); + +	strscpy(cap->driver, MXC_JPEG_NAME " codec", sizeof(cap->driver)); +	strscpy(cap->card, MXC_JPEG_NAME " codec", sizeof(cap->card)); +	snprintf(cap->bus_info, sizeof(cap->bus_info), "platform:%s", +		 dev_name(mxc_jpeg->dev)); +	cap->device_caps = V4L2_CAP_STREAMING | V4L2_CAP_VIDEO_M2M_MPLANE; +	cap->capabilities = cap->device_caps | V4L2_CAP_DEVICE_CAPS; + +	return 0; +} + +static int mxc_jpeg_enum_fmt_vid_cap(struct file *file, void *priv, +				     struct v4l2_fmtdesc *f) +{ +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(priv); + +	if (ctx->mxc_jpeg->mode == MXC_JPEG_ENCODE) +		return enum_fmt(mxc_formats, MXC_JPEG_NUM_FORMATS, f, +			MXC_JPEG_FMT_TYPE_ENC); +	else +		return enum_fmt(mxc_formats, MXC_JPEG_NUM_FORMATS, f, +			MXC_JPEG_FMT_TYPE_RAW); +} + +static int mxc_jpeg_enum_fmt_vid_out(struct file *file, void *priv, +				     struct v4l2_fmtdesc *f) +{ +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(priv); + +	if (ctx->mxc_jpeg->mode == MXC_JPEG_DECODE) +		return enum_fmt(mxc_formats, MXC_JPEG_NUM_FORMATS, f, +				MXC_JPEG_FMT_TYPE_ENC); +	else +		return enum_fmt(mxc_formats, MXC_JPEG_NUM_FORMATS, f, +				MXC_JPEG_FMT_TYPE_RAW); +} + +static int mxc_jpeg_try_fmt(struct v4l2_format *f, struct mxc_jpeg_fmt *fmt, +			    struct mxc_jpeg_ctx *ctx, int q_type) +{ +	struct v4l2_pix_format_mplane *pix_mp = &f->fmt.pix_mp; +	struct v4l2_plane_pix_format *pfmt; +	u32 w = (pix_mp->width < MXC_JPEG_MAX_WIDTH) ? +		 pix_mp->width : MXC_JPEG_MAX_WIDTH; +	u32 h = (pix_mp->height < MXC_JPEG_MAX_HEIGHT) ? +		 pix_mp->height : MXC_JPEG_MAX_HEIGHT; +	int i; +	struct mxc_jpeg_q_data tmp_q; + +	memset(pix_mp->reserved, 0, sizeof(pix_mp->reserved)); +	pix_mp->field = V4L2_FIELD_NONE; +	pix_mp->num_planes = fmt->colplanes; +	pix_mp->pixelformat = fmt->fourcc; + +	/* +	 * use MXC_JPEG_H_ALIGN instead of fmt->v_align, for vertical +	 * alignment, to loosen up the alignment to multiple of 8, +	 * otherwise NV12-1080p fails as 1080 is not a multiple of 16 +	 */ +	v4l_bound_align_image(&w, +			      MXC_JPEG_MIN_WIDTH, +			      w, /* adjust downwards*/ +			      fmt->h_align, +			      &h, +			      MXC_JPEG_MIN_HEIGHT, +			      h, /* adjust downwards*/ +			      MXC_JPEG_H_ALIGN, +			      0); +	pix_mp->width = w; /* negotiate the width */ +	pix_mp->height = h; /* negotiate the height */ + +	/* get user input into the tmp_q */ +	tmp_q.w = w; +	tmp_q.h = h; +	tmp_q.fmt = fmt; +	for (i = 0; i < pix_mp->num_planes; i++) { +		pfmt = &pix_mp->plane_fmt[i]; +		tmp_q.bytesperline[i] = pfmt->bytesperline; +		tmp_q.sizeimage[i] = pfmt->sizeimage; +	} + +	/* calculate bytesperline & sizeimage into the tmp_q */ +	mxc_jpeg_bytesperline(&tmp_q, 8); +	mxc_jpeg_sizeimage(&tmp_q); + +	/* adjust user format according to our calculations */ +	for (i = 0; i < pix_mp->num_planes; i++) { +		pfmt = &pix_mp->plane_fmt[i]; +		memset(pfmt->reserved, 0, sizeof(pfmt->reserved)); +		pfmt->bytesperline = tmp_q.bytesperline[i]; +		pfmt->sizeimage = tmp_q.sizeimage[i]; +	} + +	/* fix colorspace information to sRGB for both output & capture */ +	pix_mp->colorspace = V4L2_COLORSPACE_SRGB; +	pix_mp->ycbcr_enc = V4L2_YCBCR_ENC_601; +	pix_mp->xfer_func = V4L2_XFER_FUNC_SRGB; +	/* +	 * this hardware does not change the range of the samples +	 * but since inside JPEG the YUV quantization is full-range, +	 * this driver will always use full-range for the raw frames, too +	 */ +	pix_mp->quantization = V4L2_QUANTIZATION_FULL_RANGE; + +	return 0; +} + +static int mxc_jpeg_try_fmt_vid_cap(struct file *file, void *priv, +				    struct v4l2_format *f) +{ +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(priv); +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	struct device *dev = jpeg->dev; +	struct mxc_jpeg_fmt *fmt; +	u32 fourcc = f->fmt.pix_mp.pixelformat; + +	int q_type = (jpeg->mode == MXC_JPEG_DECODE) ? +		     MXC_JPEG_FMT_TYPE_RAW : MXC_JPEG_FMT_TYPE_ENC; + +	if (!V4L2_TYPE_IS_MULTIPLANAR(f->type)) { +		dev_err(dev, "TRY_FMT with Invalid type: %d\n", f->type); +		return -EINVAL; +	} + +	fmt = mxc_jpeg_find_format(ctx, fourcc); +	if (!fmt || fmt->flags != q_type) { +		dev_warn(dev, "Format not supported: %c%c%c%c, use the default.\n", +			 (fourcc & 0xff), +			 (fourcc >>  8) & 0xff, +			 (fourcc >> 16) & 0xff, +			 (fourcc >> 24) & 0xff); +		f->fmt.pix_mp.pixelformat = (jpeg->mode == MXC_JPEG_DECODE) ? +				MXC_JPEG_DEFAULT_PFMT : V4L2_PIX_FMT_JPEG; +		fmt = mxc_jpeg_find_format(ctx, f->fmt.pix_mp.pixelformat); +	} +	return mxc_jpeg_try_fmt(f, fmt, ctx, q_type); +} + +static int mxc_jpeg_try_fmt_vid_out(struct file *file, void *priv, +				    struct v4l2_format *f) +{ +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(priv); +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	struct device *dev = jpeg->dev; +	struct mxc_jpeg_fmt *fmt; +	u32 fourcc = f->fmt.pix_mp.pixelformat; + +	int q_type = (jpeg->mode == MXC_JPEG_ENCODE) ? +		     MXC_JPEG_FMT_TYPE_RAW : MXC_JPEG_FMT_TYPE_ENC; + +	if (!V4L2_TYPE_IS_MULTIPLANAR(f->type)) { +		dev_err(dev, "TRY_FMT with Invalid type: %d\n", f->type); +		return -EINVAL; +	} + +	fmt = mxc_jpeg_find_format(ctx, fourcc); +	if (!fmt || fmt->flags != q_type) { +		dev_warn(dev, "Format not supported: %c%c%c%c, use the default.\n", +			 (fourcc & 0xff), +			 (fourcc >>  8) & 0xff, +			 (fourcc >> 16) & 0xff, +			 (fourcc >> 24) & 0xff); +		f->fmt.pix_mp.pixelformat = (jpeg->mode == MXC_JPEG_ENCODE) ? +				MXC_JPEG_DEFAULT_PFMT : V4L2_PIX_FMT_JPEG; +		fmt = mxc_jpeg_find_format(ctx, f->fmt.pix_mp.pixelformat); +	} +	return mxc_jpeg_try_fmt(f, fmt, ctx, q_type); +} + +static int mxc_jpeg_s_fmt(struct mxc_jpeg_ctx *ctx, +			  struct v4l2_format *f) +{ +	struct vb2_queue *vq; +	struct mxc_jpeg_q_data *q_data = NULL; +	struct v4l2_pix_format_mplane *pix_mp = &f->fmt.pix_mp; +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	int i; + +	vq = v4l2_m2m_get_vq(ctx->fh.m2m_ctx, f->type); +	if (!vq) +		return -EINVAL; + +	q_data = mxc_jpeg_get_q_data(ctx, f->type); + +	if (vb2_is_busy(vq)) { +		v4l2_err(&jpeg->v4l2_dev, "queue busy\n"); +		return -EBUSY; +	} + +	q_data->fmt = mxc_jpeg_find_format(ctx, pix_mp->pixelformat); +	q_data->w = pix_mp->width; +	q_data->h = pix_mp->height; + +	q_data->w_adjusted = q_data->w; +	q_data->h_adjusted = q_data->h; +	if (jpeg->mode == MXC_JPEG_DECODE) { +		/* +		 * align up the resolution for CAST IP, +		 * but leave the buffer resolution unchanged +		 */ +		v4l_bound_align_image(&q_data->w_adjusted, +				      q_data->w_adjusted,  /* adjust upwards */ +				      MXC_JPEG_MAX_WIDTH, +				      q_data->fmt->h_align, +				      &q_data->h_adjusted, +				      q_data->h_adjusted, /* adjust upwards */ +				      MXC_JPEG_MAX_HEIGHT, +				      q_data->fmt->v_align, +				      0); +	} else { +		/* +		 * align down the resolution for CAST IP, +		 * but leave the buffer resolution unchanged +		 */ +		v4l_bound_align_image(&q_data->w_adjusted, +				      MXC_JPEG_MIN_WIDTH, +				      q_data->w_adjusted, /* adjust downwards*/ +				      q_data->fmt->h_align, +				      &q_data->h_adjusted, +				      MXC_JPEG_MIN_HEIGHT, +				      q_data->h_adjusted, /* adjust downwards*/ +				      q_data->fmt->v_align, +				      0); +	} + +	for (i = 0; i < pix_mp->num_planes; i++) { +		q_data->bytesperline[i] = pix_mp->plane_fmt[i].bytesperline; +		q_data->sizeimage[i] = pix_mp->plane_fmt[i].sizeimage; +	} + +	return 0; +} + +static int mxc_jpeg_s_fmt_vid_cap(struct file *file, void *priv, +				  struct v4l2_format *f) +{ +	int ret; + +	ret = mxc_jpeg_try_fmt_vid_cap(file, priv, f); +	if (ret) +		return ret; + +	return mxc_jpeg_s_fmt(mxc_jpeg_fh_to_ctx(priv), f); +} + +static int mxc_jpeg_s_fmt_vid_out(struct file *file, void *priv, +				  struct v4l2_format *f) +{ +	int ret; + +	ret = mxc_jpeg_try_fmt_vid_out(file, priv, f); +	if (ret) +		return ret; + +	return mxc_jpeg_s_fmt(mxc_jpeg_fh_to_ctx(priv), f); +} + +static int mxc_jpeg_g_fmt_vid(struct file *file, void *priv, +			      struct v4l2_format *f) +{ +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(priv); +	struct mxc_jpeg_dev *jpeg = ctx->mxc_jpeg; +	struct device *dev = jpeg->dev; +	struct v4l2_pix_format_mplane   *pix_mp = &f->fmt.pix_mp; +	struct mxc_jpeg_q_data *q_data = mxc_jpeg_get_q_data(ctx, f->type); +	int i; + +	if (!V4L2_TYPE_IS_MULTIPLANAR(f->type)) { +		dev_err(dev, "G_FMT with Invalid type: %d\n", f->type); +		return -EINVAL; +	} + +	pix_mp->pixelformat = q_data->fmt->fourcc; +	pix_mp->width = q_data->w; +	pix_mp->height = q_data->h; +	pix_mp->field = V4L2_FIELD_NONE; + +	/* fix colorspace information to sRGB for both output & capture */ +	pix_mp->colorspace = V4L2_COLORSPACE_SRGB; +	pix_mp->ycbcr_enc = V4L2_YCBCR_ENC_601; +	pix_mp->xfer_func = V4L2_XFER_FUNC_SRGB; +	pix_mp->quantization = V4L2_QUANTIZATION_FULL_RANGE; + +	pix_mp->num_planes = q_data->fmt->colplanes; +	for (i = 0; i < pix_mp->num_planes; i++) { +		pix_mp->plane_fmt[i].bytesperline = q_data->bytesperline[i]; +		pix_mp->plane_fmt[i].sizeimage = q_data->sizeimage[i]; +	} + +	return 0; +} + +static int mxc_jpeg_subscribe_event(struct v4l2_fh *fh, +				    const struct v4l2_event_subscription *sub) +{ +	switch (sub->type) { +	case V4L2_EVENT_EOS: +		return v4l2_event_subscribe(fh, sub, 0, NULL); +	case V4L2_EVENT_SOURCE_CHANGE: +		return v4l2_src_change_event_subscribe(fh, sub); +	default: +		return -EINVAL; +	} +} + +static int mxc_jpeg_dqbuf(struct file *file, void *priv, +			  struct v4l2_buffer *buf) +{ +	struct v4l2_fh *fh = file->private_data; +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(priv); +	struct device *dev = ctx->mxc_jpeg->dev; +	int num_src_ready = v4l2_m2m_num_src_bufs_ready(fh->m2m_ctx); +	int ret; + +	dev_dbg(dev, "DQBUF type=%d, index=%d", buf->type, buf->index); +	if (ctx->stopping == 1	&& num_src_ready == 0) { +		/* No more src bufs, notify app EOS */ +		notify_eos(ctx); +		ctx->stopping = 0; +	} + +	ret = v4l2_m2m_dqbuf(file, fh->m2m_ctx, buf); + +	return ret; +} + +static const struct v4l2_ioctl_ops mxc_jpeg_ioctl_ops = { +	.vidioc_querycap		= mxc_jpeg_querycap, +	.vidioc_enum_fmt_vid_cap	= mxc_jpeg_enum_fmt_vid_cap, +	.vidioc_enum_fmt_vid_out	= mxc_jpeg_enum_fmt_vid_out, + +	.vidioc_try_fmt_vid_cap_mplane	= mxc_jpeg_try_fmt_vid_cap, +	.vidioc_try_fmt_vid_out_mplane	= mxc_jpeg_try_fmt_vid_out, + +	.vidioc_s_fmt_vid_cap_mplane	= mxc_jpeg_s_fmt_vid_cap, +	.vidioc_s_fmt_vid_out_mplane	= mxc_jpeg_s_fmt_vid_out, + +	.vidioc_g_fmt_vid_cap_mplane	= mxc_jpeg_g_fmt_vid, +	.vidioc_g_fmt_vid_out_mplane	= mxc_jpeg_g_fmt_vid, + +	.vidioc_subscribe_event		= mxc_jpeg_subscribe_event, +	.vidioc_unsubscribe_event	= v4l2_event_unsubscribe, + +	.vidioc_try_decoder_cmd		= v4l2_m2m_ioctl_try_decoder_cmd, +	.vidioc_decoder_cmd		= mxc_jpeg_decoder_cmd, +	.vidioc_try_encoder_cmd		= v4l2_m2m_ioctl_try_encoder_cmd, +	.vidioc_encoder_cmd		= mxc_jpeg_encoder_cmd, + +	.vidioc_qbuf			= v4l2_m2m_ioctl_qbuf, +	.vidioc_dqbuf			= mxc_jpeg_dqbuf, + +	.vidioc_create_bufs		= v4l2_m2m_ioctl_create_bufs, +	.vidioc_prepare_buf		= v4l2_m2m_ioctl_prepare_buf, +	.vidioc_reqbufs			= v4l2_m2m_ioctl_reqbufs, +	.vidioc_querybuf		= v4l2_m2m_ioctl_querybuf, +	.vidioc_expbuf			= v4l2_m2m_ioctl_expbuf, +	.vidioc_streamon		= v4l2_m2m_ioctl_streamon, +	.vidioc_streamoff		= v4l2_m2m_ioctl_streamoff, +}; + +static int mxc_jpeg_release(struct file *file) +{ +	struct mxc_jpeg_dev *mxc_jpeg = video_drvdata(file); +	struct mxc_jpeg_ctx *ctx = mxc_jpeg_fh_to_ctx(file->private_data); +	struct device *dev = mxc_jpeg->dev; + +	mutex_lock(&mxc_jpeg->lock); +	if (mxc_jpeg->mode == MXC_JPEG_DECODE) +		dev_dbg(dev, "Release JPEG decoder instance on slot %d.", +			ctx->slot); +	else +		dev_dbg(dev, "Release JPEG encoder instance on slot %d.", +			ctx->slot); +	v4l2_m2m_ctx_release(ctx->fh.m2m_ctx); +	v4l2_fh_del(&ctx->fh); +	v4l2_fh_exit(&ctx->fh); +	kfree(ctx); +	mutex_unlock(&mxc_jpeg->lock); + +	return 0; +} + +static const struct v4l2_file_operations mxc_jpeg_fops = { +	.owner		= THIS_MODULE, +	.open		= mxc_jpeg_open, +	.release	= mxc_jpeg_release, +	.poll		= v4l2_m2m_fop_poll, +	.unlocked_ioctl	= video_ioctl2, +	.mmap		= v4l2_m2m_fop_mmap, +}; + +static struct v4l2_m2m_ops mxc_jpeg_m2m_ops = { +	.device_run	= mxc_jpeg_device_run, +}; + +static void mxc_jpeg_detach_pm_domains(struct mxc_jpeg_dev *jpeg) +{ +	int i; + +	for (i = 0; i < jpeg->num_domains; i++) { +		if (jpeg->pd_link[i] && !IS_ERR(jpeg->pd_link[i])) +			device_link_del(jpeg->pd_link[i]); +		if (jpeg->pd_dev[i] && !IS_ERR(jpeg->pd_dev[i])) +			dev_pm_domain_detach(jpeg->pd_dev[i], true); +		jpeg->pd_dev[i] = NULL; +		jpeg->pd_link[i] = NULL; +	} +} + +static int mxc_jpeg_attach_pm_domains(struct mxc_jpeg_dev *jpeg) +{ +	struct device *dev = jpeg->dev; +	struct device_node *np = jpeg->pdev->dev.of_node; +	int i; +	int ret; + +	jpeg->num_domains = of_count_phandle_with_args(np, "power-domains", +						       "#power-domain-cells"); +	if (jpeg->num_domains < 0) { +		dev_err(dev, "No power domains defined for jpeg node\n"); +		return jpeg->num_domains; +	} + +	jpeg->pd_dev = devm_kmalloc_array(dev, jpeg->num_domains, +					  sizeof(*jpeg->pd_dev), GFP_KERNEL); +	if (!jpeg->pd_dev) +		return -ENOMEM; + +	jpeg->pd_link = devm_kmalloc_array(dev, jpeg->num_domains, +					   sizeof(*jpeg->pd_link), GFP_KERNEL); +	if (!jpeg->pd_link) +		return -ENOMEM; + +	for (i = 0; i < jpeg->num_domains; i++) { +		jpeg->pd_dev[i] = dev_pm_domain_attach_by_id(dev, i); +		if (IS_ERR(jpeg->pd_dev[i])) { +			ret = PTR_ERR(jpeg->pd_dev[i]); +			goto fail; +		} + +		jpeg->pd_link[i] = device_link_add(dev, jpeg->pd_dev[i], +						   DL_FLAG_STATELESS | +						   DL_FLAG_PM_RUNTIME | +						   DL_FLAG_RPM_ACTIVE); +		if (!jpeg->pd_link[i]) { +			ret = -EINVAL; +			goto fail; +		} +	} + +	return 0; +fail: +	mxc_jpeg_detach_pm_domains(jpeg); +	return ret; +} + +static int mxc_jpeg_probe(struct platform_device *pdev) +{ +	struct mxc_jpeg_dev *jpeg; +	struct device *dev = &pdev->dev; +	struct resource *res; +	int dec_irq; +	int ret; +	int mode; +	const struct of_device_id *of_id; +	unsigned int slot; + +	of_id = of_match_node(mxc_jpeg_match, dev->of_node); +	mode = *(const int *)of_id->data; + +	jpeg = devm_kzalloc(dev, sizeof(struct mxc_jpeg_dev), GFP_KERNEL); +	if (!jpeg) +		return -ENOMEM; + +	mutex_init(&jpeg->lock); +	spin_lock_init(&jpeg->hw_lock); + +	ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); +	if (ret) { +		dev_err(&pdev->dev, "No suitable DMA available.\n"); +		goto err_irq; +	} + +	res = platform_get_resource(pdev, IORESOURCE_MEM, 0); +	jpeg->base_reg = devm_ioremap_resource(&pdev->dev, res); +	if (IS_ERR(jpeg->base_reg)) +		return PTR_ERR(jpeg->base_reg); + +	for (slot = 0; slot < MXC_MAX_SLOTS; slot++) { +		dec_irq = platform_get_irq(pdev, slot); +		if (dec_irq < 0) { +			dev_err(&pdev->dev, "Failed to get irq %d\n", dec_irq); +			ret = dec_irq; +			goto err_irq; +		} +		ret = devm_request_irq(&pdev->dev, dec_irq, mxc_jpeg_dec_irq, +				       0, pdev->name, jpeg); +		if (ret) { +			dev_err(&pdev->dev, "Failed to request irq %d (%d)\n", +				dec_irq, ret); +			goto err_irq; +		} +	} + +	jpeg->pdev = pdev; +	jpeg->dev = dev; +	jpeg->mode = mode; + +	ret = mxc_jpeg_attach_pm_domains(jpeg); +	if (ret < 0) { +		dev_err(dev, "failed to attach power domains %d\n", ret); +		return ret; +	} + +	/* v4l2 */ +	ret = v4l2_device_register(dev, &jpeg->v4l2_dev); +	if (ret) { +		dev_err(dev, "failed to register v4l2 device\n"); +		goto err_register; +	} +	jpeg->m2m_dev = v4l2_m2m_init(&mxc_jpeg_m2m_ops); +	if (IS_ERR(jpeg->m2m_dev)) { +		dev_err(dev, "failed to register v4l2 device\n"); +		ret = PTR_ERR(jpeg->m2m_dev); +		goto err_m2m; +	} + +	jpeg->dec_vdev = video_device_alloc(); +	if (!jpeg->dec_vdev) { +		dev_err(dev, "failed to register v4l2 device\n"); +		ret = -ENOMEM; +		goto err_vdev_alloc; +	} +	if (mode == MXC_JPEG_ENCODE) +		snprintf(jpeg->dec_vdev->name, +			 sizeof(jpeg->dec_vdev->name), +			 "%s-enc", MXC_JPEG_NAME); +	else +		snprintf(jpeg->dec_vdev->name, +			 sizeof(jpeg->dec_vdev->name), +			 "%s-dec", MXC_JPEG_NAME); + +	jpeg->dec_vdev->fops = &mxc_jpeg_fops; +	jpeg->dec_vdev->ioctl_ops = &mxc_jpeg_ioctl_ops; +	jpeg->dec_vdev->minor = -1; +	jpeg->dec_vdev->release = video_device_release; +	jpeg->dec_vdev->lock = &jpeg->lock; /* lock for ioctl serialization */ +	jpeg->dec_vdev->v4l2_dev = &jpeg->v4l2_dev; +	jpeg->dec_vdev->vfl_dir = VFL_DIR_M2M; +	jpeg->dec_vdev->device_caps = V4L2_CAP_STREAMING | +					V4L2_CAP_VIDEO_M2M_MPLANE; +	if (mode == MXC_JPEG_ENCODE) { +		v4l2_disable_ioctl(jpeg->dec_vdev, VIDIOC_DECODER_CMD); +		v4l2_disable_ioctl(jpeg->dec_vdev, VIDIOC_TRY_DECODER_CMD); +	} else { +		v4l2_disable_ioctl(jpeg->dec_vdev, VIDIOC_ENCODER_CMD); +		v4l2_disable_ioctl(jpeg->dec_vdev, VIDIOC_TRY_ENCODER_CMD); +	} +	ret = video_register_device(jpeg->dec_vdev, VFL_TYPE_VIDEO, -1); +	if (ret) { +		dev_err(dev, "failed to register video device\n"); +		goto err_vdev_register; +	} +	video_set_drvdata(jpeg->dec_vdev, jpeg); +	if (mode == MXC_JPEG_ENCODE) +		v4l2_info(&jpeg->v4l2_dev, +			  "encoder device registered as /dev/video%d (%d,%d)\n", +			  jpeg->dec_vdev->num, VIDEO_MAJOR, +			  jpeg->dec_vdev->minor); +	else +		v4l2_info(&jpeg->v4l2_dev, +			  "decoder device registered as /dev/video%d (%d,%d)\n", +			  jpeg->dec_vdev->num, VIDEO_MAJOR, +			  jpeg->dec_vdev->minor); + +	platform_set_drvdata(pdev, jpeg); + +	return 0; + +err_vdev_register: +	video_device_release(jpeg->dec_vdev); + +err_vdev_alloc: +	v4l2_m2m_release(jpeg->m2m_dev); + +err_m2m: +	v4l2_device_unregister(&jpeg->v4l2_dev); + +err_register: +err_irq: +	return ret; +} + +static int mxc_jpeg_remove(struct platform_device *pdev) +{ +	unsigned int slot; +	struct mxc_jpeg_dev *jpeg = platform_get_drvdata(pdev); + +	for (slot = 0; slot < MXC_MAX_SLOTS; slot++) +		mxc_jpeg_free_slot_data(jpeg, slot); + +	video_unregister_device(jpeg->dec_vdev); +	v4l2_m2m_release(jpeg->m2m_dev); +	v4l2_device_unregister(&jpeg->v4l2_dev); +	mxc_jpeg_detach_pm_domains(jpeg); + +	return 0; +} + +MODULE_DEVICE_TABLE(of, mxc_jpeg_match); + +static struct platform_driver mxc_jpeg_driver = { +	.probe = mxc_jpeg_probe, +	.remove = mxc_jpeg_remove, +	.driver = { +		.name = "mxc-jpeg", +		.of_match_table = mxc_jpeg_match, +	}, +}; +module_platform_driver(mxc_jpeg_driver); + +MODULE_AUTHOR("Zhengyu Shen <zhengyu.shen_1@nxp.com>"); +MODULE_AUTHOR("Mirela Rabulea <mirela.rabulea@nxp.com>"); +MODULE_DESCRIPTION("V4L2 driver for i.MX8 QXP/QM JPEG encoder/decoder"); +MODULE_LICENSE("GPL v2"); | 
