diff options
author | Daniel Scally <djrscally@gmail.com> | 2021-01-07 14:28:34 +0100 |
---|---|---|
committer | Mauro Carvalho Chehab <mchehab+huawei@kernel.org> | 2021-01-26 19:30:00 +0100 |
commit | acec1ff638a636f95a68271ca9b60c670a023d74 (patch) | |
tree | 6c5768c66a3f74c993d30b7a30e2823be1fef8a6 /drivers/media/pci/intel/ipu3/ipu3-cio2.c | |
parent | 900104c8483736f8e022ff3b084c0928e895d903 (diff) |
media: ipu3-cio2: Rename ipu3-cio2.c
ipu3-cio2 driver needs extending with multiple files; rename the main
source file and specify the renamed file in Makefile to accommodate that.
Suggested-by: Andy Shevchenko <andriy.shevchenko@linux.intel.com>
Reviewed-by: Laurent Pinchart <laurent.pinchart@ideasonboard.com>
Reviewed-by: Andy Shevchenko <andriy.shevchenko@linux.intel.com>
Signed-off-by: Daniel Scally <djrscally@gmail.com>
Signed-off-by: Sakari Ailus <sakari.ailus@linux.intel.com>
Signed-off-by: Mauro Carvalho Chehab <mchehab+huawei@kernel.org>
Diffstat (limited to 'drivers/media/pci/intel/ipu3/ipu3-cio2.c')
-rw-r--r-- | drivers/media/pci/intel/ipu3/ipu3-cio2.c | 2026 |
1 files changed, 0 insertions, 2026 deletions
diff --git a/drivers/media/pci/intel/ipu3/ipu3-cio2.c b/drivers/media/pci/intel/ipu3/ipu3-cio2.c deleted file mode 100644 index e8ea69d30bfd..000000000000 --- a/drivers/media/pci/intel/ipu3/ipu3-cio2.c +++ /dev/null @@ -1,2026 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0 -/* - * Copyright (C) 2017,2020 Intel Corporation - * - * Based partially on Intel IPU4 driver written by - * Sakari Ailus <sakari.ailus@linux.intel.com> - * Samu Onkalo <samu.onkalo@intel.com> - * Jouni Högander <jouni.hogander@intel.com> - * Jouni Ukkonen <jouni.ukkonen@intel.com> - * Antti Laakso <antti.laakso@intel.com> - * et al. - */ - -#include <linux/delay.h> -#include <linux/interrupt.h> -#include <linux/iopoll.h> -#include <linux/mm.h> -#include <linux/module.h> -#include <linux/pci.h> -#include <linux/pfn.h> -#include <linux/pm_runtime.h> -#include <linux/property.h> -#include <linux/vmalloc.h> -#include <media/v4l2-ctrls.h> -#include <media/v4l2-device.h> -#include <media/v4l2-event.h> -#include <media/v4l2-fwnode.h> -#include <media/v4l2-ioctl.h> -#include <media/videobuf2-dma-sg.h> - -#include "ipu3-cio2.h" - -struct ipu3_cio2_fmt { - u32 mbus_code; - u32 fourcc; - u8 mipicode; - u8 bpp; -}; - -/* - * These are raw formats used in Intel's third generation of - * Image Processing Unit known as IPU3. - * 10bit raw bayer packed, 32 bytes for every 25 pixels, - * last LSB 6 bits unused. - */ -static const struct ipu3_cio2_fmt formats[] = { - { /* put default entry at beginning */ - .mbus_code = MEDIA_BUS_FMT_SGRBG10_1X10, - .fourcc = V4L2_PIX_FMT_IPU3_SGRBG10, - .mipicode = 0x2b, - .bpp = 10, - }, { - .mbus_code = MEDIA_BUS_FMT_SGBRG10_1X10, - .fourcc = V4L2_PIX_FMT_IPU3_SGBRG10, - .mipicode = 0x2b, - .bpp = 10, - }, { - .mbus_code = MEDIA_BUS_FMT_SBGGR10_1X10, - .fourcc = V4L2_PIX_FMT_IPU3_SBGGR10, - .mipicode = 0x2b, - .bpp = 10, - }, { - .mbus_code = MEDIA_BUS_FMT_SRGGB10_1X10, - .fourcc = V4L2_PIX_FMT_IPU3_SRGGB10, - .mipicode = 0x2b, - .bpp = 10, - }, -}; - -/* - * cio2_find_format - lookup color format by fourcc or/and media bus code - * @pixelformat: fourcc to match, ignored if null - * @mbus_code: media bus code to match, ignored if null - */ -static const struct ipu3_cio2_fmt *cio2_find_format(const u32 *pixelformat, - const u32 *mbus_code) -{ - unsigned int i; - - for (i = 0; i < ARRAY_SIZE(formats); i++) { - if (pixelformat && *pixelformat != formats[i].fourcc) - continue; - if (mbus_code && *mbus_code != formats[i].mbus_code) - continue; - - return &formats[i]; - } - - return NULL; -} - -static inline u32 cio2_bytesperline(const unsigned int width) -{ - /* - * 64 bytes for every 50 pixels, the line length - * in bytes is multiple of 64 (line end alignment). - */ - return DIV_ROUND_UP(width, 50) * 64; -} - -/**************** FBPT operations ****************/ - -static void cio2_fbpt_exit_dummy(struct cio2_device *cio2) -{ - if (cio2->dummy_lop) { - dma_free_coherent(&cio2->pci_dev->dev, PAGE_SIZE, - cio2->dummy_lop, cio2->dummy_lop_bus_addr); - cio2->dummy_lop = NULL; - } - if (cio2->dummy_page) { - dma_free_coherent(&cio2->pci_dev->dev, PAGE_SIZE, - cio2->dummy_page, cio2->dummy_page_bus_addr); - cio2->dummy_page = NULL; - } -} - -static int cio2_fbpt_init_dummy(struct cio2_device *cio2) -{ - unsigned int i; - - cio2->dummy_page = dma_alloc_coherent(&cio2->pci_dev->dev, PAGE_SIZE, - &cio2->dummy_page_bus_addr, - GFP_KERNEL); - cio2->dummy_lop = dma_alloc_coherent(&cio2->pci_dev->dev, PAGE_SIZE, - &cio2->dummy_lop_bus_addr, - GFP_KERNEL); - if (!cio2->dummy_page || !cio2->dummy_lop) { - cio2_fbpt_exit_dummy(cio2); - return -ENOMEM; - } - /* - * List of Pointers(LOP) contains 1024x32b pointers to 4KB page each - * Initialize each entry to dummy_page bus base address. - */ - for (i = 0; i < CIO2_LOP_ENTRIES; i++) - cio2->dummy_lop[i] = PFN_DOWN(cio2->dummy_page_bus_addr); - - return 0; -} - -static void cio2_fbpt_entry_enable(struct cio2_device *cio2, - struct cio2_fbpt_entry entry[CIO2_MAX_LOPS]) -{ - /* - * The CPU first initializes some fields in fbpt, then sets - * the VALID bit, this barrier is to ensure that the DMA(device) - * does not see the VALID bit enabled before other fields are - * initialized; otherwise it could lead to havoc. - */ - dma_wmb(); - - /* - * Request interrupts for start and completion - * Valid bit is applicable only to 1st entry - */ - entry[0].first_entry.ctrl = CIO2_FBPT_CTRL_VALID | - CIO2_FBPT_CTRL_IOC | CIO2_FBPT_CTRL_IOS; -} - -/* Initialize fpbt entries to point to dummy frame */ -static void cio2_fbpt_entry_init_dummy(struct cio2_device *cio2, - struct cio2_fbpt_entry - entry[CIO2_MAX_LOPS]) -{ - unsigned int i; - - entry[0].first_entry.first_page_offset = 0; - entry[1].second_entry.num_of_pages = CIO2_LOP_ENTRIES * CIO2_MAX_LOPS; - entry[1].second_entry.last_page_available_bytes = PAGE_SIZE - 1; - - for (i = 0; i < CIO2_MAX_LOPS; i++) - entry[i].lop_page_addr = PFN_DOWN(cio2->dummy_lop_bus_addr); - - cio2_fbpt_entry_enable(cio2, entry); -} - -/* Initialize fpbt entries to point to a given buffer */ -static void cio2_fbpt_entry_init_buf(struct cio2_device *cio2, - struct cio2_buffer *b, - struct cio2_fbpt_entry - entry[CIO2_MAX_LOPS]) -{ - struct vb2_buffer *vb = &b->vbb.vb2_buf; - unsigned int length = vb->planes[0].length; - int remaining, i; - - entry[0].first_entry.first_page_offset = b->offset; - remaining = length + entry[0].first_entry.first_page_offset; - entry[1].second_entry.num_of_pages = PFN_UP(remaining); - /* - * last_page_available_bytes has the offset of the last byte in the - * last page which is still accessible by DMA. DMA cannot access - * beyond this point. Valid range for this is from 0 to 4095. - * 0 indicates 1st byte in the page is DMA accessible. - * 4095 (PAGE_SIZE - 1) means every single byte in the last page - * is available for DMA transfer. - */ - remaining = offset_in_page(remaining) ?: PAGE_SIZE; - entry[1].second_entry.last_page_available_bytes = remaining - 1; - /* Fill FBPT */ - remaining = length; - i = 0; - while (remaining > 0) { - entry->lop_page_addr = PFN_DOWN(b->lop_bus_addr[i]); - remaining -= CIO2_LOP_ENTRIES * PAGE_SIZE; - entry++; - i++; - } - - /* - * The first not meaningful FBPT entry should point to a valid LOP - */ - entry->lop_page_addr = PFN_DOWN(cio2->dummy_lop_bus_addr); - - cio2_fbpt_entry_enable(cio2, entry); -} - -static int cio2_fbpt_init(struct cio2_device *cio2, struct cio2_queue *q) -{ - struct device *dev = &cio2->pci_dev->dev; - - q->fbpt = dma_alloc_coherent(dev, CIO2_FBPT_SIZE, &q->fbpt_bus_addr, - GFP_KERNEL); - if (!q->fbpt) - return -ENOMEM; - - return 0; -} - -static void cio2_fbpt_exit(struct cio2_queue *q, struct device *dev) -{ - dma_free_coherent(dev, CIO2_FBPT_SIZE, q->fbpt, q->fbpt_bus_addr); -} - -/**************** CSI2 hardware setup ****************/ - -/* - * The CSI2 receiver has several parameters affecting - * the receiver timings. These depend on the MIPI bus frequency - * F in Hz (sensor transmitter rate) as follows: - * register value = (A/1e9 + B * UI) / COUNT_ACC - * where - * UI = 1 / (2 * F) in seconds - * COUNT_ACC = counter accuracy in seconds - * For IPU3 COUNT_ACC = 0.0625 - * - * A and B are coefficients from the table below, - * depending whether the register minimum or maximum value is - * calculated. - * Minimum Maximum - * Clock lane A B A B - * reg_rx_csi_dly_cnt_termen_clane 0 0 38 0 - * reg_rx_csi_dly_cnt_settle_clane 95 -8 300 -16 - * Data lanes - * reg_rx_csi_dly_cnt_termen_dlane0 0 0 35 4 - * reg_rx_csi_dly_cnt_settle_dlane0 85 -2 145 -6 - * reg_rx_csi_dly_cnt_termen_dlane1 0 0 35 4 - * reg_rx_csi_dly_cnt_settle_dlane1 85 -2 145 -6 - * reg_rx_csi_dly_cnt_termen_dlane2 0 0 35 4 - * reg_rx_csi_dly_cnt_settle_dlane2 85 -2 145 -6 - * reg_rx_csi_dly_cnt_termen_dlane3 0 0 35 4 - * reg_rx_csi_dly_cnt_settle_dlane3 85 -2 145 -6 - * - * We use the minimum values of both A and B. - */ - -/* - * shift for keeping value range suitable for 32-bit integer arithmetic - */ -#define LIMIT_SHIFT 8 - -static s32 cio2_rx_timing(s32 a, s32 b, s64 freq, int def) -{ - const u32 accinv = 16; /* invert of counter resolution */ - const u32 uiinv = 500000000; /* 1e9 / 2 */ - s32 r; - - freq >>= LIMIT_SHIFT; - - if (WARN_ON(freq <= 0 || freq > S32_MAX)) - return def; - /* - * b could be 0, -2 or -8, so |accinv * b| is always - * less than (1 << ds) and thus |r| < 500000000. - */ - r = accinv * b * (uiinv >> LIMIT_SHIFT); - r = r / (s32)freq; - /* max value of a is 95 */ - r += accinv * a; - - return r; -}; - -/* Calculate the delay value for termination enable of clock lane HS Rx */ -static int cio2_csi2_calc_timing(struct cio2_device *cio2, struct cio2_queue *q, - struct cio2_csi2_timing *timing, - unsigned int bpp, unsigned int lanes) -{ - struct device *dev = &cio2->pci_dev->dev; - s64 freq; - - if (!q->sensor) - return -ENODEV; - - freq = v4l2_get_link_rate(q->sensor->ctrl_handler, bpp, lanes); - if (freq < 0) { - dev_err(dev, "error %lld, invalid link_freq\n", freq); - return freq; - } - - timing->clk_termen = cio2_rx_timing(CIO2_CSIRX_DLY_CNT_TERMEN_CLANE_A, - CIO2_CSIRX_DLY_CNT_TERMEN_CLANE_B, - freq, - CIO2_CSIRX_DLY_CNT_TERMEN_DEFAULT); - timing->clk_settle = cio2_rx_timing(CIO2_CSIRX_DLY_CNT_SETTLE_CLANE_A, - CIO2_CSIRX_DLY_CNT_SETTLE_CLANE_B, - freq, - CIO2_CSIRX_DLY_CNT_SETTLE_DEFAULT); - timing->dat_termen = cio2_rx_timing(CIO2_CSIRX_DLY_CNT_TERMEN_DLANE_A, - CIO2_CSIRX_DLY_CNT_TERMEN_DLANE_B, - freq, - CIO2_CSIRX_DLY_CNT_TERMEN_DEFAULT); - timing->dat_settle = cio2_rx_timing(CIO2_CSIRX_DLY_CNT_SETTLE_DLANE_A, - CIO2_CSIRX_DLY_CNT_SETTLE_DLANE_B, - freq, - CIO2_CSIRX_DLY_CNT_SETTLE_DEFAULT); - - dev_dbg(dev, "freq ct value is %d\n", timing->clk_termen); - dev_dbg(dev, "freq cs value is %d\n", timing->clk_settle); - dev_dbg(dev, "freq dt value is %d\n", timing->dat_termen); - dev_dbg(dev, "freq ds value is %d\n", timing->dat_settle); - - return 0; -}; - -static int cio2_hw_init(struct cio2_device *cio2, struct cio2_queue *q) -{ - static const int NUM_VCS = 4; - static const int SID; /* Stream id */ - static const int ENTRY; - static const int FBPT_WIDTH = DIV_ROUND_UP(CIO2_MAX_LOPS, - CIO2_FBPT_SUBENTRY_UNIT); - const u32 num_buffers1 = CIO2_MAX_BUFFERS - 1; - const struct ipu3_cio2_fmt *fmt; - void __iomem *const base = cio2->base; - u8 lanes, csi2bus = q->csi2.port; - u8 sensor_vc = SENSOR_VIR_CH_DFLT; - struct cio2_csi2_timing timing; - int i, r; - - fmt = cio2_find_format(NULL, &q->subdev_fmt.code); - if (!fmt) - return -EINVAL; - - lanes = q->csi2.lanes; - - r = cio2_csi2_calc_timing(cio2, q, &timing, fmt->bpp, lanes); - if (r) - return r; - - writel(timing.clk_termen, q->csi_rx_base + - CIO2_REG_CSIRX_DLY_CNT_TERMEN(CIO2_CSIRX_DLY_CNT_CLANE_IDX)); - writel(timing.clk_settle, q->csi_rx_base + - CIO2_REG_CSIRX_DLY_CNT_SETTLE(CIO2_CSIRX_DLY_CNT_CLANE_IDX)); - - for (i = 0; i < lanes; i++) { - writel(timing.dat_termen, q->csi_rx_base + - CIO2_REG_CSIRX_DLY_CNT_TERMEN(i)); - writel(timing.dat_settle, q->csi_rx_base + - CIO2_REG_CSIRX_DLY_CNT_SETTLE(i)); - } - - writel(CIO2_PBM_WMCTRL1_MIN_2CK | - CIO2_PBM_WMCTRL1_MID1_2CK | - CIO2_PBM_WMCTRL1_MID2_2CK, base + CIO2_REG_PBM_WMCTRL1); - writel(CIO2_PBM_WMCTRL2_HWM_2CK << CIO2_PBM_WMCTRL2_HWM_2CK_SHIFT | - CIO2_PBM_WMCTRL2_LWM_2CK << CIO2_PBM_WMCTRL2_LWM_2CK_SHIFT | - CIO2_PBM_WMCTRL2_OBFFWM_2CK << - CIO2_PBM_WMCTRL2_OBFFWM_2CK_SHIFT | - CIO2_PBM_WMCTRL2_TRANSDYN << CIO2_PBM_WMCTRL2_TRANSDYN_SHIFT | - CIO2_PBM_WMCTRL2_OBFF_MEM_EN, base + CIO2_REG_PBM_WMCTRL2); - writel(CIO2_PBM_ARB_CTRL_LANES_DIV << - CIO2_PBM_ARB_CTRL_LANES_DIV_SHIFT | - CIO2_PBM_ARB_CTRL_LE_EN | - CIO2_PBM_ARB_CTRL_PLL_POST_SHTDN << - CIO2_PBM_ARB_CTRL_PLL_POST_SHTDN_SHIFT | - CIO2_PBM_ARB_CTRL_PLL_AHD_WK_UP << - CIO2_PBM_ARB_CTRL_PLL_AHD_WK_UP_SHIFT, - base + CIO2_REG_PBM_ARB_CTRL); - writel(CIO2_CSIRX_STATUS_DLANE_HS_MASK, - q->csi_rx_base + CIO2_REG_CSIRX_STATUS_DLANE_HS); - writel(CIO2_CSIRX_STATUS_DLANE_LP_MASK, - q->csi_rx_base + CIO2_REG_CSIRX_STATUS_DLANE_LP); - - writel(CIO2_FB_HPLL_FREQ, base + CIO2_REG_FB_HPLL_FREQ); - writel(CIO2_ISCLK_RATIO, base + CIO2_REG_ISCLK_RATIO); - - /* Configure MIPI backend */ - for (i = 0; i < NUM_VCS; i++) - writel(1, q->csi_rx_base + CIO2_REG_MIPIBE_SP_LUT_ENTRY(i)); - - /* There are 16 short packet LUT entry */ - for (i = 0; i < 16; i++) - writel(CIO2_MIPIBE_LP_LUT_ENTRY_DISREGARD, - q->csi_rx_base + CIO2_REG_MIPIBE_LP_LUT_ENTRY(i)); - writel(CIO2_MIPIBE_GLOBAL_LUT_DISREGARD, - q->csi_rx_base + CIO2_REG_MIPIBE_GLOBAL_LUT_DISREGARD); - - writel(CIO2_INT_EN_EXT_IE_MASK, base + CIO2_REG_INT_EN_EXT_IE); - writel(CIO2_IRQCTRL_MASK, q->csi_rx_base + CIO2_REG_IRQCTRL_MASK); - writel(CIO2_IRQCTRL_MASK, q->csi_rx_base + CIO2_REG_IRQCTRL_ENABLE); - writel(0, q->csi_rx_base + CIO2_REG_IRQCTRL_EDGE); - writel(0, q->csi_rx_base + CIO2_REG_IRQCTRL_LEVEL_NOT_PULSE); - writel(CIO2_INT_EN_EXT_OE_MASK, base + CIO2_REG_INT_EN_EXT_OE); - - writel(CIO2_REG_INT_EN_IRQ | CIO2_INT_IOC(CIO2_DMA_CHAN) | - CIO2_REG_INT_EN_IOS(CIO2_DMA_CHAN), - base + CIO2_REG_INT_EN); - - writel((CIO2_PXM_PXF_FMT_CFG_BPP_10 | CIO2_PXM_PXF_FMT_CFG_PCK_64B) - << CIO2_PXM_PXF_FMT_CFG_SID0_SHIFT, - base + CIO2_REG_PXM_PXF_FMT_CFG0(csi2bus)); - writel(SID << CIO2_MIPIBE_LP_LUT_ENTRY_SID_SHIFT | - sensor_vc << CIO2_MIPIBE_LP_LUT_ENTRY_VC_SHIFT | - fmt->mipicode << CIO2_MIPIBE_LP_LUT_ENTRY_FORMAT_TYPE_SHIFT, - q->csi_rx_base + CIO2_REG_MIPIBE_LP_LUT_ENTRY(ENTRY)); - writel(0, q->csi_rx_base + CIO2_REG_MIPIBE_COMP_FORMAT(sensor_vc)); - writel(0, q->csi_rx_base + CIO2_REG_MIPIBE_FORCE_RAW8); - writel(0, base + CIO2_REG_PXM_SID2BID0(csi2bus)); - - writel(lanes, q->csi_rx_base + CIO2_REG_CSIRX_NOF_ENABLED_LANES); - writel(CIO2_CGC_PRIM_TGE | - CIO2_CGC_SIDE_TGE | - CIO2_CGC_XOSC_TGE | - CIO2_CGC_D3I3_TGE | - CIO2_CGC_CSI2_INTERFRAME_TGE | - CIO2_CGC_CSI2_PORT_DCGE | - CIO2_CGC_SIDE_DCGE | - CIO2_CGC_PRIM_DCGE | - CIO2_CGC_ROSC_DCGE | - CIO2_CGC_XOSC_DCGE | - CIO2_CGC_CLKGATE_HOLDOFF << CIO2_CGC_CLKGATE_HOLDOFF_SHIFT | - CIO2_CGC_CSI_CLKGATE_HOLDOFF - << CIO2_CGC_CSI_CLKGATE_HOLDOFF_SHIFT, base + CIO2_REG_CGC); - writel(CIO2_LTRCTRL_LTRDYNEN, base + CIO2_REG_LTRCTRL); - writel(CIO2_LTRVAL0_VAL << CIO2_LTRVAL02_VAL_SHIFT | - CIO2_LTRVAL0_SCALE << CIO2_LTRVAL02_SCALE_SHIFT | - CIO2_LTRVAL1_VAL << CIO2_LTRVAL13_VAL_SHIFT | - CIO2_LTRVAL1_SCALE << CIO2_LTRVAL13_SCALE_SHIFT, - base + CIO2_REG_LTRVAL01); - writel(CIO2_LTRVAL2_VAL << CIO2_LTRVAL02_VAL_SHIFT | - CIO2_LTRVAL2_SCALE << CIO2_LTRVAL02_SCALE_SHIFT | - CIO2_LTRVAL3_VAL << CIO2_LTRVAL13_VAL_SHIFT | - CIO2_LTRVAL3_SCALE << CIO2_LTRVAL13_SCALE_SHIFT, - base + CIO2_REG_LTRVAL23); - - for (i = 0; i < CIO2_NUM_DMA_CHAN; i++) { - writel(0, base + CIO2_REG_CDMABA(i)); - writel(0, base + CIO2_REG_CDMAC0(i)); - writel(0, base + CIO2_REG_CDMAC1(i)); - } - - /* Enable DMA */ - writel(PFN_DOWN(q->fbpt_bus_addr), base + CIO2_REG_CDMABA(CIO2_DMA_CHAN)); - - writel(num_buffers1 << CIO2_CDMAC0_FBPT_LEN_SHIFT | - FBPT_WIDTH << CIO2_CDMAC0_FBPT_WIDTH_SHIFT | - CIO2_CDMAC0_DMA_INTR_ON_FE | - CIO2_CDMAC0_FBPT_UPDATE_FIFO_FULL | - CIO2_CDMAC0_DMA_EN | - CIO2_CDMAC0_DMA_INTR_ON_FS | - CIO2_CDMAC0_DMA_HALTED, base + CIO2_REG_CDMAC0(CIO2_DMA_CHAN)); - - writel(1 << CIO2_CDMAC1_LINENUMUPDATE_SHIFT, - base + CIO2_REG_CDMAC1(CIO2_DMA_CHAN)); - - writel(0, base + CIO2_REG_PBM_FOPN_ABORT); - - writel(CIO2_PXM_FRF_CFG_CRC_TH << CIO2_PXM_FRF_CFG_CRC_TH_SHIFT | - CIO2_PXM_FRF_CFG_MSK_ECC_DPHY_NR | - CIO2_PXM_FRF_CFG_MSK_ECC_RE | - CIO2_PXM_FRF_CFG_MSK_ECC_DPHY_NE, - base + CIO2_REG_PXM_FRF_CFG(q->csi2.port)); - - /* Clear interrupts */ - writel(CIO2_IRQCTRL_MASK, q->csi_rx_base + CIO2_REG_IRQCTRL_CLEAR); - writel(~0, base + CIO2_REG_INT_STS_EXT_OE); - writel(~0, base + CIO2_REG_INT_STS_EXT_IE); - writel(~0, base + CIO2_REG_INT_STS); - - /* Enable devices, starting from the last device in the pipe */ - writel(1, q->csi_rx_base + CIO2_REG_MIPIBE_ENABLE); - writel(1, q->csi_rx_base + CIO2_REG_CSIRX_ENABLE); - - return 0; -} - -static void cio2_hw_exit(struct cio2_device *cio2, struct cio2_queue *q) -{ - void __iomem *const base = cio2->base; - unsigned int i; - u32 value; - int ret; - - /* Disable CSI receiver and MIPI backend devices */ - writel(0, q->csi_rx_base + CIO2_REG_IRQCTRL_MASK); - writel(0, q->csi_rx_base + CIO2_REG_IRQCTRL_ENABLE); - writel(0, q->csi_rx_base + CIO2_REG_CSIRX_ENABLE); - writel(0, q->csi_rx_base + CIO2_REG_MIPIBE_ENABLE); - - /* Halt DMA */ - writel(0, base + CIO2_REG_CDMAC0(CIO2_DMA_CHAN)); - ret = readl_poll_timeout(base + CIO2_REG_CDMAC0(CIO2_DMA_CHAN), - value, value & CIO2_CDMAC0_DMA_HALTED, - 4000, 2000000); - if (ret) - dev_err(&cio2->pci_dev->dev, - "DMA %i can not be halted\n", CIO2_DMA_CHAN); - - for (i = 0; i < CIO2_NUM_PORTS; i++) { - writel(readl(base + CIO2_REG_PXM_FRF_CFG(i)) | - CIO2_PXM_FRF_CFG_ABORT, base + CIO2_REG_PXM_FRF_CFG(i)); - writel(readl(base + CIO2_REG_PBM_FOPN_ABORT) | - CIO2_PBM_FOPN_ABORT(i), base + CIO2_REG_PBM_FOPN_ABORT); - } -} - -static void cio2_buffer_done(struct cio2_device *cio2, unsigned int dma_chan) -{ - struct device *dev = &cio2->pci_dev->dev; - struct cio2_queue *q = cio2->cur_queue; - struct cio2_fbpt_entry *entry; - u64 ns = ktime_get_ns(); - - if (dma_chan >= CIO2_QUEUES) { - dev_err(dev, "bad DMA channel %i\n", dma_chan); - return; - } - - entry = &q->fbpt[q->bufs_first * CIO2_MAX_LOPS]; - if (entry->first_entry.ctrl & CIO2_FBPT_CTRL_VALID) { - dev_warn(&cio2->pci_dev->dev, - "no ready buffers found on DMA channel %u\n", - dma_chan); - return; - } - - /* Find out which buffer(s) are ready */ - do { - struct cio2_buffer *b; - - b = q->bufs[q->bufs_first]; - if (b) { - unsigned int received = entry[1].second_entry.num_of_bytes; - unsigned long payload = - vb2_get_plane_payload(&b->vbb.vb2_buf, 0); - - q->bufs[q->bufs_first] = NULL; - atomic_dec(&q->bufs_queued); - dev_dbg(&cio2->pci_dev->dev, - "buffer %i done\n", b->vbb.vb2_buf.index); - - b->vbb.vb2_buf.timestamp = ns; - b->vbb.field = V4L2_FIELD_NONE; - b->vbb.sequence = atomic_read(&q->frame_sequence); - if (payload != received) - dev_warn(dev, - "payload length is %lu, received %u\n", - payload, received); - vb2_buffer_done(&b->vbb.vb2_buf, VB2_BUF_STATE_DONE); - } - atomic_inc(&q->frame_sequence); - cio2_fbpt_entry_init_dummy(cio2, entry); - q->bufs_first = (q->bufs_first + 1) % CIO2_MAX_BUFFERS; - entry = &q->fbpt[q->bufs_first * CIO2_MAX_LOPS]; - } while (!(entry->first_entry.ctrl & CIO2_FBPT_CTRL_VALID)); -} - -static void cio2_queue_event_sof(struct cio2_device *cio2, struct cio2_queue *q) -{ - /* - * For the user space camera control algorithms it is essential - * to know when the reception of a frame has begun. That's often - * the best timing information to get from the hardware. - */ - struct v4l2_event event = { - .type = V4L2_EVENT_FRAME_SYNC, - .u.frame_sync.frame_sequence = atomic_read(&q->frame_sequence), - }; - - v4l2_event_queue(q->subdev.devnode, &event); -} - -static const char *const cio2_irq_errs[] = { - "single packet header error corrected", - "multiple packet header errors detected", - "payload checksum (CRC) error", - "fifo overflow", - "reserved short packet data type detected", - "reserved long packet data type detected", - "incomplete long packet detected", - "frame sync error", - "line sync error", - "DPHY start of transmission error", - "DPHY synchronization error", - "escape mode error", - "escape mode trigger event", - "escape mode ultra-low power state for data lane(s)", - "escape mode ultra-low power state exit for clock lane", - "inter-frame short packet discarded", - "inter-frame long packet discarded", - "non-matching Long Packet stalled", -}; - -static const char *const cio2_port_errs[] = { - "ECC recoverable", - "DPHY not recoverable", - "ECC not recoverable", - "CRC error", - "INTERFRAMEDATA", - "PKT2SHORT", - "PKT2LONG", -}; - -static void cio2_irq_handle_once(struct cio2_device *cio2, u32 int_status) -{ - void __iomem *const base = cio2->base; - struct device *dev = &cio2->pci_dev->dev; - - if (int_status & CIO2_INT_IOOE) { - /* - * Interrupt on Output Error: - * 1) SRAM is full and FS received, or - * 2) An invalid bit detected by DMA. - */ - u32 oe_status, oe_clear; - - oe_clear = readl(base + CIO2_REG_INT_STS_EXT_OE); - oe_status = oe_clear; - - if (oe_status & CIO2_INT_EXT_OE_DMAOE_MASK) { - dev_err(dev, "DMA output error: 0x%x\n", - (oe_status & CIO2_INT_EXT_OE_DMAOE_MASK) - >> CIO2_INT_EXT_OE_DMAOE_SHIFT); - oe_status &= ~CIO2_INT_EXT_OE_DMAOE_MASK; - } - if (oe_status & CIO2_INT_EXT_OE_OES_MASK) { - dev_err(dev, "DMA output error on CSI2 buses: 0x%x\n", - (oe_status & CIO2_INT_EXT_OE_OES_MASK) - >> CIO2_INT_EXT_OE_OES_SHIFT); - oe_status &= ~CIO2_INT_EXT_OE_OES_MASK; - } - writel(oe_clear, base + CIO2_REG_INT_STS_EXT_OE); - if (oe_status) - dev_warn(dev, "unknown interrupt 0x%x on OE\n", - oe_status); - int_status &= ~CIO2_INT_IOOE; - } - - if (int_status & CIO2_INT_IOC_MASK) { - /* DMA IO done -- frame ready */ - u32 clr = 0; - unsigned int d; - - for (d = 0; d < CIO2_NUM_DMA_CHAN; d++) - if (int_status & CIO2_INT_IOC(d)) { - clr |= CIO2_INT_IOC(d); - cio2_buffer_done(cio2, d); - } - int_status &= ~clr; - } - - if (int_status & CIO2_INT_IOS_IOLN_MASK) { - /* DMA IO starts or reached specified line */ - u32 clr = 0; - unsigned int d; - - for (d = 0; d < CIO2_NUM_DMA_CHAN; d++) - if (int_status & CIO2_INT_IOS_IOLN(d)) { - clr |= CIO2_INT_IOS_IOLN(d); - if (d == CIO2_DMA_CHAN) - cio2_queue_event_sof(cio2, - cio2->cur_queue); - } - int_status &= ~clr; - } - - if (int_status & (CIO2_INT_IOIE | CIO2_INT_IOIRQ)) { - /* CSI2 receiver (error) interrupt */ - u32 ie_status, ie_clear; - unsigned int port; - - ie_clear = readl(base + CIO2_REG_INT_STS_EXT_IE); - ie_status = ie_clear; - - for (port = 0; port < CIO2_NUM_PORTS; port++) { - u32 port_status = (ie_status >> (port * 8)) & 0xff; - u32 err_mask = BIT_MASK(ARRAY_SIZE(cio2_port_errs)) - 1; - void __iomem *const csi_rx_base = - base + CIO2_REG_PIPE_BASE(port); - unsigned int i; - - while (port_status & err_mask) { - i = ffs(port_status) - 1; - dev_err(dev, "port %i error %s\n", - port, cio2_port_errs[i]); - ie_status &= ~BIT(port * 8 + i); - port_status &= ~BIT(i); - } - - if (ie_status & CIO2_INT_EXT_IE_IRQ(port)) { - u32 csi2_status, csi2_clear; - - csi2_status = readl(csi_rx_base + - CIO2_REG_IRQCTRL_STATUS); - csi2_clear = csi2_status; - err_mask = - BIT_MASK(ARRAY_SIZE(cio2_irq_errs)) - 1; - - while (csi2_status & err_mask) { - i = ffs(csi2_status) - 1; - dev_err(dev, - "CSI-2 receiver port %i: %s\n", - port, cio2_irq_errs[i]); - csi2_status &= ~BIT(i); - } - - writel(csi2_clear, - csi_rx_base + CIO2_REG_IRQCTRL_CLEAR); - if (csi2_status) - dev_warn(dev, - "unknown CSI2 error 0x%x on port %i\n", - csi2_status, port); - - ie_status &= ~CIO2_INT_EXT_IE_IRQ(port); - } - } - - writel(ie_clear, base + CIO2_REG_INT_STS_EXT_IE); - if (ie_status) - dev_warn(dev, "unknown interrupt 0x%x on IE\n", - ie_status); - - int_status &= ~(CIO2_INT_IOIE | CIO2_INT_IOIRQ); - } - - if (int_status) - dev_warn(dev, "unknown interrupt 0x%x on INT\n", int_status); -} - -static irqreturn_t cio2_irq(int irq, void *cio2_ptr) -{ - struct cio2_device *cio2 = cio2_ptr; - void __iomem *const base = cio2->base; - struct device *dev = &cio2->pci_dev->dev; - u32 int_status; - - int_status = readl(base + CIO2_REG_INT_STS); - dev_dbg(dev, "isr enter - interrupt status 0x%x\n", int_status); - if (!int_status) - return IRQ_NONE; - - do { - writel(int_status, base + CIO2_REG_INT_STS); - cio2_irq_handle_once(cio2, int_status); - int_status = readl(base + CIO2_REG_INT_STS); - if (int_status) - dev_dbg(dev, "pending status 0x%x\n", int_status); - } while (int_status); - - return IRQ_HANDLED; -} - -/**************** Videobuf2 interface ****************/ - -static void cio2_vb2_return_all_buffers(struct cio2_queue *q, - enum vb2_buffer_state state) -{ - unsigned int i; - - for (i = 0; i < CIO2_MAX_BUFFERS; i++) { - if (q->bufs[i]) { - atomic_dec(&q->bufs_queued); - vb2_buffer_done(&q->bufs[i]->vbb.vb2_buf, - state); - q->bufs[i] = NULL; - } - } -} - -static int cio2_vb2_queue_setup(struct vb2_queue *vq, - unsigned int *num_buffers, - unsigned int *num_planes, - unsigned int sizes[], - struct device *alloc_devs[]) -{ - struct cio2_device *cio2 = vb2_get_drv_priv(vq); - struct cio2_queue *q = vb2q_to_cio2_queue(vq); - unsigned int i; - - *num_planes = q->format.num_planes; - - for (i = 0; i < *num_planes; ++i) { - sizes[i] = q->format.plane_fmt[i].sizeimage; - alloc_devs[i] = &cio2->pci_dev->dev; - } - - *num_buffers = clamp_val(*num_buffers, 1, CIO2_MAX_BUFFERS); - - /* Initialize buffer queue */ - for (i = 0; i < CIO2_MAX_BUFFERS; i++) { - q->bufs[i] = NULL; - cio2_fbpt_entry_init_dummy(cio2, &q->fbpt[i * CIO2_MAX_LOPS]); - } - atomic_set(&q->bufs_queued, 0); - q->bufs_first = 0; - q->bufs_next = 0; - - return 0; -} - -/* Called after each buffer is allocated */ -static int cio2_vb2_buf_init(struct vb2_buffer *vb) -{ - struct cio2_device *cio2 = vb2_get_drv_priv(vb->vb2_queue); - struct device *dev = &cio2->pci_dev->dev; - struct cio2_buffer *b = - container_of(vb, struct cio2_buffer, vbb.vb2_buf); - unsigned int pages = PFN_UP(vb->planes[0].length); - unsigned int lops = DIV_ROUND_UP(pages + 1, CIO2_LOP_ENTRIES); - struct sg_table *sg; - struct sg_dma_page_iter sg_iter; - unsigned int i, j; - - if (lops <= 0 || lops > CIO2_MAX_LOPS) { - dev_err(dev, "%s: bad buffer size (%i)\n", __func__, - vb->planes[0].length); - return -ENOSPC; /* Should never happen */ - } - - memset(b->lop, 0, sizeof(b->lop)); - /* Allocate LOP table */ - for (i = 0; i < lops; i++) { - b->lop[i] = dma_alloc_coherent(dev, PAGE_SIZE, - &b->lop_bus_addr[i], GFP_KERNEL); - if (!b->lop[i]) - goto fail; - } - - /* Fill LOP */ - sg = vb2_dma_sg_plane_desc(vb, 0); - if (!sg) - return -ENOMEM; - - if (sg->nents && sg->sgl) - b->offset = sg->sgl->offset; - - i = j = 0; - for_each_sg_dma_page(sg->sgl, &sg_iter, sg->nents, 0) { - if (!pages--) - break; - b->lop[i][j] = PFN_DOWN(sg_page_iter_dma_address(&sg_iter)); - j++; - if (j == CIO2_LOP_ENTRIES) { - i++; - j = 0; - } - } - - b->lop[i][j] = PFN_DOWN(cio2->dummy_page_bus_addr); - return 0; -fail: - while (i--) - dma_free_coherent(dev, PAGE_SIZE, b->lop[i], b->lop_bus_addr[i]); - return -ENOMEM; -} - -/* Transfer buffer ownership to cio2 */ -static void cio2_vb2_buf_queue(struct vb2_buffer *vb) -{ - struct cio2_device *cio2 = vb2_get_drv_priv(vb->vb2_queue); - struct cio2_queue *q = - container_of(vb->vb2_queue, struct cio2_queue, vbq); - struct cio2_buffer *b = - container_of(vb, struct cio2_buffer, vbb.vb2_buf); - struct cio2_fbpt_entry *entry; - unsigned long flags; - unsigned int i, j, next = q->bufs_next; - int bufs_queued = atomic_inc_return(&q->bufs_queued); - u32 fbpt_rp; - - dev_dbg(&cio2->pci_dev->dev, "queue buffer %d\n", vb->index); - - /* - * This code queues the buffer to the CIO2 DMA engine, which starts - * running once streaming has started. It is possible that this code - * gets pre-empted due to increased CPU load. Upon this, the driver - * does not get an opportunity to queue new buffers to the CIO2 DMA - * engine. When the DMA engine encounters an FBPT entry without the - * VALID bit set, the DMA engine halts, which requires a restart of - * the DMA engine and sensor, to continue streaming. - * This is not desired and is highly unlikely given that there are - * 32 FBPT entries that the DMA engine needs to process, to run into - * an FBPT entry, without the VALID bit set. We try to mitigate this - * by disabling interrupts for the duration of this queueing. - */ - local_irq_save(flags); - - fbpt_rp = (readl(cio2->base + CIO2_REG_CDMARI(CIO2_DMA_CHAN)) - >> CIO2_CDMARI_FBPT_RP_SHIFT) - & CIO2_CDMARI_FBPT_RP_MASK; - - /* - * fbpt_rp is the fbpt entry that the dma is currently working - * on, but since it could jump to next entry at any time, - * assume that we might already be there. - */ - fbpt_rp = (fbpt_rp + 1) % CIO2_MAX_BUFFERS; - - if (bufs_queued <= 1 || fbpt_rp == next) - /* Buffers were drained */ - next = (fbpt_rp + 1) % CIO2_MAX_BUFFERS; - - for (i = 0; i < CIO2_MAX_BUFFERS; i++) { - /* - * We have allocated CIO2_MAX_BUFFERS circularly for the - * hw, the user has requested N buffer queue. The driver - * ensures N <= CIO2_MAX_BUFFERS and guarantees that whenever - * user queues a buffer, there necessarily is a free buffer. - */ - if (!q->bufs[next]) { - q->bufs[next] = b; - entry = &q->fbpt[next * CIO2_MAX_LOPS]; - cio2_fbpt_entry_init_buf(cio2, b, entry); - local_irq_restore(flags); - q->bufs_next = (next + 1) % CIO2_MAX_BUFFERS; - for (j = 0; j < vb->num_planes; j++) - vb2_set_plane_payload(vb, j, - q->format.plane_fmt[j].sizeimage); - return; - } - - dev_dbg(&cio2->pci_dev->dev, "entry %i was full!\n", next); - next = (next + 1) % CIO2_MAX_BUFFERS; - } - - local_irq_restore(flags); - dev_err(&cio2->pci_dev->dev, "error: all cio2 entries were full!\n"); - atomic_dec(&q->bufs_queued); - vb2_buffer_done(vb, VB2_BUF_STATE_ERROR); -} - -/* Called when each buffer is freed */ -static void cio2_vb2_buf_cleanup(struct vb2_buffer *vb) -{ - struct cio2_device *cio2 = vb2_get_drv_priv(vb->vb2_queue); - struct cio2_buffer *b = - container_of(vb, struct cio2_buffer, vbb.vb2_buf); - unsigned int i; - - /* Free LOP table */ - for (i = 0; i < CIO2_MAX_LOPS; i++) { - if (b->lop[i]) - dma_free_coherent(&cio2->pci_dev->dev, PAGE_SIZE, - b->lop[i], b->lop_bus_addr[i]); - } -} - -static int cio2_vb2_start_streaming(struct vb2_queue *vq, unsigned int count) -{ - struct cio2_queue *q = vb2q_to_cio2_queue(vq); - struct cio2_device *cio2 = vb2_get_drv_priv(vq); - int r; - - cio2->cur_queue = q; - atomic_set(&q->frame_sequence, 0); - - r = pm_runtime_get_sync(&cio2->pci_dev->dev); - if (r < 0) { - dev_info(&cio2->pci_dev->dev, "failed to set power %d\n", r); - pm_runtime_put_noidle(&cio2->pci_dev->dev); - return r; - } - - r = media_pipeline_start(&q->vdev.entity, &q->pipe); - if (r) - goto fail_pipeline; - - r = cio2_hw_init(cio2, q); - if (r) - goto fail_hw; - - /* Start streaming on sensor */ - r = v4l2_subdev_call(q->sensor, video, s_stream, 1); - if (r) - goto fail_csi2_subdev; - - cio2->streaming = true; - - return 0; - -fail_csi2_subdev: - cio2_hw_exit(cio2, q); -fail_hw: - media_pipeline_stop(&q->vdev.entity); -fail_pipeline: - dev_dbg(&cio2->pci_dev->dev, "failed to start streaming (%d)\n", r); - cio2_vb2_return_all_buffers(q, VB2_BUF_STATE_QUEUED); - pm_runtime_put(&cio2->pci_dev->dev); - - return r; -} - -static void cio2_vb2_stop_streaming(struct vb2_queue *vq) -{ - struct cio2_queue *q = vb2q_to_cio2_queue(vq); - struct cio2_device *cio2 = vb2_get_drv_priv(vq); - - if (v4l2_subdev_call(q->sensor, video, s_stream, 0)) - dev_err(&cio2->pci_dev->dev, - "failed to stop sensor streaming\n"); - - cio2_hw_exit(cio2, q); - synchronize_irq(cio2->pci_dev->irq); - cio2_vb2_return_all_buffers(q, VB2_BUF_STATE_ERROR); - media_pipeline_stop(&q->vdev.entity); - pm_runtime_put(&cio2->pci_dev->dev); - cio2->streaming = false; -} - -static const struct vb2_ops cio2_vb2_ops = { - .buf_init = cio2_vb2_buf_init, - .buf_queue = cio2_vb2_buf_queue, - .buf_cleanup = cio2_vb2_buf_cleanup, - .queue_setup = cio2_vb2_queue_setup, - .start_streaming = cio2_vb2_start_streaming, - .stop_streaming = cio2_vb2_stop_streaming, - .wait_prepare = vb2_ops_wait_prepare, - .wait_finish = vb2_ops_wait_finish, -}; - -/**************** V4L2 interface ****************/ - -static int cio2_v4l2_querycap(struct file *file, void *fh, - struct v4l2_capability *cap) -{ - struct cio2_device *cio2 = video_drvdata(file); - - strscpy(cap->driver, CIO2_NAME, sizeof(cap->driver)); - strscpy(cap->card, CIO2_DEVICE_NAME, sizeof(cap->card)); - snprintf(cap->bus_info, sizeof(cap->bus_info), - "PCI:%s", pci_name(cio2->pci_dev)); - - return 0; -} - -static int cio2_v4l2_enum_fmt(struct file *file, void *fh, - struct v4l2_fmtdesc *f) -{ - if (f->index >= ARRAY_SIZE(formats)) - return -EINVAL; - - f->pixelformat = formats[f->index].fourcc; - - return 0; -} - -/* The format is validated in cio2_video_link_validate() */ -static int cio2_v4l2_g_fmt(struct file *file, void *fh, struct v4l2_format *f) -{ - struct cio2_queue *q = file_to_cio2_queue(file); - - f->fmt.pix_mp = q->format; - - return 0; -} - -static int cio2_v4l2_try_fmt(struct file *file, void *fh, struct v4l2_format *f) -{ - const struct ipu3_cio2_fmt *fmt; - struct v4l2_pix_format_mplane *mpix = &f->fmt.pix_mp; - - fmt = cio2_find_format(&mpix->pixelformat, NULL); - if (!fmt) - fmt = &formats[0]; - - /* Only supports up to 4224x3136 */ - if (mpix->width > CIO2_IMAGE_MAX_WIDTH) - mpix->width = CIO2_IMAGE_MAX_WIDTH; - if (mpix->height > CIO2_IMAGE_MAX_HEIGHT) - mpix->height = CIO2_IMAGE_MAX_HEIGHT; - - mpix->num_planes = 1; - mpix->pixelformat = fmt->fourcc; - mpix->colorspace = V4L2_COLORSPACE_RAW; - mpix->field = V4L2_FIELD_NONE; - memset(mpix->reserved, 0, sizeof(mpix->reserved)); - mpix->plane_fmt[0].bytesperline = cio2_bytesperline(mpix->width); - mpix->plane_fmt[0].sizeimage = mpix->plane_fmt[0].bytesperline * - mpix->height; - memset(mpix->plane_fmt[0].reserved, 0, - sizeof(mpix->plane_fmt[0].reserved)); - - /* use default */ - mpix->ycbcr_enc = V4L2_YCBCR_ENC_DEFAULT; - mpix->quantization = V4L2_QUANTIZATION_DEFAULT; - mpix->xfer_func = V4L2_XFER_FUNC_DEFAULT; - - return 0; -} - -static int cio2_v4l2_s_fmt(struct file *file, void *fh, struct v4l2_format *f) -{ - struct cio2_queue *q = file_to_cio2_queue(file); - - cio2_v4l2_try_fmt(file, fh, f); - q->format = f->fmt.pix_mp; - - return 0; -} - -static int -cio2_video_enum_input(struct file *file, void *fh, struct v4l2_input *input) -{ - if (input->index > 0) - return -EINVAL; - - strscpy(input->name, "camera", sizeof(input->name)); - input->type = V4L2_INPUT_TYPE_CAMERA; - - return 0; -} - -static int -cio2_video_g_input(struct file *file, void *fh, unsigned int *input) -{ - *input = 0; - - return 0; -} - -static int -cio2_video_s_input(struct file *file, void *fh, unsigned int input) -{ - return input == 0 ? 0 : -EINVAL; -} - -static const struct v4l2_file_operations cio2_v4l2_fops = { - .owner = THIS_MODULE, - .unlocked_ioctl = video_ioctl2, - .open = v4l2_fh_open, - .release = vb2_fop_release, - .poll = vb2_fop_poll, - .mmap = vb2_fop_mmap, -}; - -static const struct v4l2_ioctl_ops cio2_v4l2_ioctl_ops = { - .vidioc_querycap = cio2_v4l2_querycap, - .vidioc_enum_fmt_vid_cap = cio2_v4l2_enum_fmt, - .vidioc_g_fmt_vid_cap_mplane = cio2_v4l2_g_fmt, - .vidioc_s_fmt_vid_cap_mplane = cio2_v4l2_s_fmt, - .vidioc_try_fmt_vid_cap_mplane = cio2_v4l2_try_fmt, - .vidioc_reqbufs = vb2_ioctl_reqbufs, - .vidioc_create_bufs = vb2_ioctl_create_bufs, - .vidioc_prepare_buf = vb2_ioctl_prepare_buf, - .vidioc_querybuf = vb2_ioctl_querybuf, - .vidioc_qbuf = vb2_ioctl_qbuf, - .vidioc_dqbuf = vb2_ioctl_dqbuf, - .vidioc_streamon = vb2_ioctl_streamon, - .vidioc_streamoff = vb2_ioctl_streamoff, - .vidioc_expbuf = vb2_ioctl_expbuf, - .vidioc_enum_input = cio2_video_enum_input, - .vidioc_g_input = cio2_video_g_input, - .vidioc_s_input = cio2_video_s_input, -}; - -static int cio2_subdev_subscribe_event(struct v4l2_subdev *sd, - struct v4l2_fh *fh, - struct v4l2_event_subscription *sub) -{ - if (sub->type != V4L2_EVENT_FRAME_SYNC) - return -EINVAL; - - /* Line number. For now only zero accepted. */ - if (sub->id != 0) - return -EINVAL; - - return v4l2_event_subscribe(fh, sub, 0, NULL); -} - -static int cio2_subdev_open(struct v4l2_subdev *sd, struct v4l2_subdev_fh *fh) -{ - struct v4l2_mbus_framefmt *format; - const struct v4l2_mbus_framefmt fmt_default = { - .width = 1936, - .height = 1096, - .code = formats[0].mbus_code, - .field = V4L2_FIELD_NONE, - .colorspace = V4L2_COLORSPACE_RAW, - .ycbcr_enc = V4L2_YCBCR_ENC_DEFAULT, - .quantization = V4L2_QUANTIZATION_DEFAULT, - .xfer_func = V4L2_XFER_FUNC_DEFAULT, - }; - - /* Initialize try_fmt */ - format = v4l2_subdev_get_try_format(sd, fh->pad, CIO2_PAD_SINK); - *format = fmt_default; - - /* same as sink */ - format = v4l2_subdev_get_try_format(sd, fh->pad, CIO2_PAD_SOURCE); - *format = fmt_default; - - return 0; -} - -/* - * cio2_subdev_get_fmt - Handle get format by pads subdev method - * @sd : pointer to v4l2 subdev structure - * @cfg: V4L2 subdev pad config - * @fmt: pointer to v4l2 subdev format structure - * return -EINVAL or zero on success - */ -static int cio2_subdev_get_fmt(struct v4l2_subdev *sd, - struct v4l2_subdev_pad_config *cfg, - struct v4l2_subdev_format *fmt) -{ - struct cio2_queue *q = container_of(sd, struct cio2_queue, subdev); - - mutex_lock(&q->subdev_lock); - - if (fmt->which == V4L2_SUBDEV_FORMAT_TRY) - fmt->format = *v4l2_subdev_get_try_format(sd, cfg, fmt->pad); - else - fmt->format = q->subdev_fmt; - - mutex_unlock(&q->subdev_lock); - - return 0; -} - -/* - * cio2_subdev_set_fmt - Handle set format by pads subdev method - * @sd : pointer to v4l2 subdev structure - * @cfg: V4L2 subdev pad config - * @fmt: pointer to v4l2 subdev format structure - * return -EINVAL or zero on success - */ -static int cio2_subdev_set_fmt(struct v4l2_subdev *sd, - struct v4l2_subdev_pad_config *cfg, - struct v4l2_subdev_format *fmt) -{ - struct cio2_queue *q = container_of(sd, struct cio2_queue, subdev); - struct v4l2_mbus_framefmt *mbus; - u32 mbus_code = fmt->format.code; - unsigned int i; - - /* - * Only allow setting sink pad format; - * source always propagates from sink - */ - if (fmt->pad == CIO2_PAD_SOURCE) - return cio2_subdev_get_fmt(sd, cfg, fmt); - - if (fmt->which == V4L2_SUBDEV_FORMAT_TRY) - mbus = v4l2_subdev_get_try_format(sd, cfg, fmt->pad); - else - mbus = &q->subdev_fmt; - - fmt->format.code = formats[0].mbus_code; - - for (i = 0; i < ARRAY_SIZE(formats); i++) { - if (formats[i].mbus_code == mbus_code) { - fmt->format.code = mbus_code; - break; - } - } - - fmt->format.width = min(fmt->format.width, CIO2_IMAGE_MAX_WIDTH); - fmt->format.height = min(fmt->format.height, CIO2_IMAGE_MAX_HEIGHT); - fmt->format.field = V4L2_FIELD_NONE; - - mutex_lock(&q->subdev_lock); - *mbus = fmt->format; - mutex_unlock(&q->subdev_lock); - - return 0; -} - -static int cio2_subdev_enum_mbus_code(struct v4l2_subdev *sd, - struct v4l2_subdev_pad_config *cfg, - struct v4l2_subdev_mbus_code_enum *code) -{ - if (code->index >= ARRAY_SIZE(formats)) - return -EINVAL; - - code->code = formats[code->index].mbus_code; - return 0; -} - -static int cio2_subdev_link_validate_get_format(struct media_pad *pad, - struct v4l2_subdev_format *fmt) -{ - if (is_media_entity_v4l2_subdev(pad->entity)) { - struct v4l2_subdev *sd = - media_entity_to_v4l2_subdev(pad->entity); - - fmt->which = V4L2_SUBDEV_FORMAT_ACTIVE; - fmt->pad = pad->index; - return v4l2_subdev_call(sd, pad, get_fmt, NULL, fmt); - } - - return -EINVAL; -} - -static int cio2_video_link_validate(struct media_link *link) -{ - struct video_device *vd = container_of(link->sink->entity, - struct video_device, entity); - struct cio2_queue *q = container_of(vd, struct cio2_queue, vdev); - struct cio2_device *cio2 = video_get_drvdata(vd); - struct v4l2_subdev_format source_fmt; - int ret; - - if (!media_entity_remote_pad(link->sink->entity->pads)) { - dev_info(&cio2->pci_dev->dev, - "video node %s pad not connected\n", vd->name); - return -ENOTCONN; - } - - ret = cio2_subdev_link_validate_get_format(link->source, &source_fmt); - if (ret < 0) - return 0; - - if (source_fmt.format.width != q->format.width || - source_fmt.format.height != q->format.height) { - dev_err(&cio2->pci_dev->dev, - "Wrong width or height %ux%u (%ux%u expected)\n", - q->format.width, q->format.height, - source_fmt.format.width, source_fmt.format.height); - return -EINVAL; - } - - if (!cio2_find_format(&q->format.pixelformat, &source_fmt.format.code)) - return -EINVAL; - - return 0; -} - -static const struct v4l2_subdev_core_ops cio2_subdev_core_ops = { - .subscribe_event = cio2_subdev_subscribe_event, - .unsubscribe_event = v4l2_event_subdev_unsubscribe, -}; - -static const struct v4l2_subdev_internal_ops cio2_subdev_internal_ops = { - .open = cio2_subdev_open, -}; - -static const struct v4l2_subdev_pad_ops cio2_subdev_pad_ops = { - .link_validate = v4l2_subdev_link_validate_default, - .get_fmt = cio2_subdev_get_fmt, - .set_fmt = cio2_subdev_set_fmt, - .enum_mbus_code = cio2_subdev_enum_mbus_code, -}; - -static const struct v4l2_subdev_ops cio2_subdev_ops = { - .core = &cio2_subdev_core_ops, - .pad = &cio2_subdev_pad_ops, -}; - -/******* V4L2 sub-device asynchronous registration callbacks***********/ - -struct sensor_async_subdev { - struct v4l2_async_subdev asd; - struct csi2_bus_info csi2; -}; - -/* The .bound() notifier callback when a match is found */ -static int cio2_notifier_bound(struct v4l2_async_notifier *notifier, - struct v4l2_subdev *sd, - struct v4l2_async_subdev *asd) -{ - struct cio2_device *cio2 = container_of(notifier, - struct cio2_device, notifier); - struct sensor_async_subdev *s_asd = container_of(asd, - struct sensor_async_subdev, asd); - struct cio2_queue *q; - - if (cio2->queue[s_asd->csi2.port].sensor) - return -EBUSY; - - q = &cio2->queue[s_asd->csi2.port]; - - q->csi2 = s_asd->csi2; - q->sensor = sd; - q->csi_rx_base = cio2->base + CIO2_REG_PIPE_BASE(q->csi2.port); - - return 0; -} - -/* The .unbind callback */ -static void cio2_notifier_unbind(struct v4l2_async_notifier *notifier, - struct v4l2_subdev *sd, - struct v4l2_async_subdev *asd) -{ - struct cio2_device *cio2 = container_of(notifier, - struct cio2_device, notifier); - struct sensor_async_subdev *s_asd = container_of(asd, - struct sensor_async_subdev, asd); - - cio2->queue[s_asd->csi2.port].sensor = NULL; -} - -/* .complete() is called after all subdevices have been located */ -static int cio2_notifier_complete(struct v4l2_async_notifier *notifier) -{ - struct cio2_device *cio2 = container_of(notifier, struct cio2_device, - notifier); - struct sensor_async_subdev *s_asd; - struct v4l2_async_subdev *asd; - struct cio2_queue *q; - unsigned int pad; - int ret; - - list_for_each_entry(asd, &cio2->notifier.asd_list, asd_list) { - s_asd = container_of(asd, struct sensor_async_subdev, asd); - q = &cio2->queue[s_asd->csi2.port]; - - for (pad = 0; pad < q->sensor->entity.num_pads; pad++) - if (q->sensor->entity.pads[pad].flags & - MEDIA_PAD_FL_SOURCE) - break; - - if (pad == q->sensor->entity.num_pads) { - dev_err(&cio2->pci_dev->dev, - "failed to find src pad for %s\n", - q->sensor->name); - return -ENXIO; - } - - ret = media_create_pad_link( - &q->sensor->entity, pad, - &q->subdev.entity, CIO2_PAD_SINK, - 0); - if (ret) { - dev_err(&cio2->pci_dev->dev, - "failed to create link for %s\n", - q->sensor->name); - return ret; - } - } - - return v4l2_device_register_subdev_nodes(&cio2->v4l2_dev); -} - -static const struct v4l2_async_notifier_operations cio2_async_ops = { - .bound = cio2_notifier_bound, - .unbind = cio2_notifier_unbind, - .complete = cio2_notifier_complete, -}; - -static int cio2_parse_firmware(struct cio2_device *cio2) -{ - unsigned int i; - int ret; - - for (i = 0; i < CIO2_NUM_PORTS; i++) { - struct v4l2_fwnode_endpoint vep = { - .bus_type = V4L2_MBUS_CSI2_DPHY - }; - struct sensor_async_subdev *s_asd = NULL; - struct fwnode_handle *ep; - - ep = fwnode_graph_get_endpoint_by_id( - dev_fwnode(&cio2->pci_dev->dev), i, 0, - FWNODE_GRAPH_ENDPOINT_NEXT); - - if (!ep) - continue; - - ret = v4l2_fwnode_endpoint_parse(ep, &vep); - if (ret) - goto err_parse; - - s_asd = kzalloc(sizeof(*s_asd), GFP_KERNEL); - if (!s_asd) { - ret = -ENOMEM; - goto err_parse; - } - - s_asd->csi2.port = vep.base.port; - s_asd->csi2.lanes = vep.bus.mipi_csi2.num_data_lanes; - - ret = v4l2_async_notifier_add_fwnode_remote_subdev( - &cio2->notifier, ep, &s_asd->asd); - if (ret) - goto err_parse; - - fwnode_handle_put(ep); - - continue; - -err_parse: - fwnode_handle_put(ep); - kfree(s_asd); - return ret; - } - - /* - * Proceed even without sensors connected to allow the device to - * suspend. - */ - cio2->notifier.ops = &cio2_async_ops; - ret = v4l2_async_notifier_register(&cio2->v4l2_dev, &cio2->notifier); - if (ret) - dev_err(&cio2->pci_dev->dev, - "failed to register async notifier : %d\n", ret); - - return ret; -} - -/**************** Queue initialization ****************/ -static const struct media_entity_operations cio2_media_ops = { - .link_validate = v4l2_subdev_link_validate, -}; - -static const struct media_entity_operations cio2_video_entity_ops = { - .link_validate = cio2_video_link_validate, -}; - -static int cio2_queue_init(struct cio2_device *cio2, struct cio2_queue *q) -{ - static const u32 default_width = 1936; - static const u32 default_height = 1096; - const struct ipu3_cio2_fmt dflt_fmt = formats[0]; - - struct video_device *vdev = &q->vdev; - struct vb2_queue *vbq = &q->vbq; - struct v4l2_subdev *subdev = &q->subdev; - struct v4l2_mbus_framefmt *fmt; - int r; - - /* Initialize miscellaneous variables */ - mutex_init(&q->lock); - mutex_init(&q->subdev_lock); - - /* Initialize formats to default values */ - fmt = &q->subdev_fmt; - fmt->width = default_width; - fmt->height = default_height; - fmt->code = dflt_fmt.mbus_code; - fmt->field = V4L2_FIELD_NONE; - - q->format.width = default_width; - q->format.height = default_height; - q->format.pixelformat = dflt_fmt.fourcc; - q->format.colorspace = V4L2_COLORSPACE_RAW; - q->format.field = V4L2_FIELD_NONE; - q->format.num_planes = 1; - q->format.plane_fmt[0].bytesperline = - cio2_bytesperline(q->format.width); - q->format.plane_fmt[0].sizeimage = q->format.plane_fmt[0].bytesperline * - q->format.height; - - /* Initialize fbpt */ - r = cio2_fbpt_init(cio2, q); - if (r) - goto fail_fbpt; - - /* Initialize media entities */ - q->subdev_pads[CIO2_PAD_SINK].flags = MEDIA_PAD_FL_SINK | - MEDIA_PAD_FL_MUST_CONNECT; - q->subdev_pads[CIO2_PAD_SOURCE].flags = MEDIA_PAD_FL_SOURCE; - subdev->entity.ops = &cio2_media_ops; - subdev->internal_ops = &cio2_subdev_internal_ops; - r = media_entity_pads_init(&subdev->entity, CIO2_PADS, q->subdev_pads); - if (r) { - dev_err(&cio2->pci_dev->dev, - "failed initialize subdev media entity (%d)\n", r); - goto fail_subdev_media_entity; - } - - q->vdev_pad.flags = MEDIA_PAD_FL_SINK | MEDIA_PAD_FL_MUST_CONNECT; - vdev->entity.ops = &cio2_video_entity_ops; - r = media_entity_pads_init(&vdev->entity, 1, &q->vdev_pad); - if (r) { - dev_err(&cio2->pci_dev->dev, - "failed initialize videodev media entity (%d)\n", r); - goto fail_vdev_media_entity; - } - - /* Initialize subdev */ - v4l2_subdev_init(subdev, &cio2_subdev_ops); - subdev->flags = V4L2_SUBDEV_FL_HAS_DEVNODE | V4L2_SUBDEV_FL_HAS_EVENTS; - subdev->owner = THIS_MODULE; - snprintf(subdev->name, sizeof(subdev->name), - CIO2_ENTITY_NAME " %td", q - cio2->queue); - subdev->entity.function = MEDIA_ENT_F_VID_IF_BRIDGE; - v4l2_set_subdevdata(subdev, cio2); - r = v4l2_device_register_subdev(&cio2->v4l2_dev, subdev); - if (r) { - dev_err(&cio2->pci_dev->dev, - "failed initialize subdev (%d)\n", r); - goto fail_subdev; - } - - /* Initialize vbq */ - vbq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; - vbq->io_modes = VB2_USERPTR | VB2_MMAP | VB2_DMABUF; - vbq->ops = &cio2_vb2_ops; - vbq->mem_ops = &vb2_dma_sg_memops; - vbq->buf_struct_size = sizeof(struct cio2_buffer); - vbq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; - vbq->min_buffers_needed = 1; - vbq->drv_priv = cio2; - vbq->lock = &q->lock; - r = vb2_queue_init(vbq); - if (r) { - dev_err(&cio2->pci_dev->dev, - "failed to initialize videobuf2 queue (%d)\n", r); - goto fail_subdev; - } - - /* Initialize vdev */ - snprintf(vdev->name, sizeof(vdev->name), - "%s %td", CIO2_NAME, q - cio2->queue); - vdev->release = video_device_release_empty; - vdev->fops = &cio2_v4l2_fops; - vdev->ioctl_ops = &cio2_v4l2_ioctl_ops; - vdev->lock = &cio2->lock; - vdev->v4l2_dev = &cio2->v4l2_dev; - vdev->queue = &q->vbq; - vdev->device_caps = V4L2_CAP_VIDEO_CAPTURE_MPLANE | V4L2_CAP_STREAMING; - video_set_drvdata(vdev, cio2); - r = video_register_device(vdev, VFL_TYPE_VIDEO, -1); - if (r) { - dev_err(&cio2->pci_dev->dev, - "failed to register video device (%d)\n", r); - goto fail_vdev; - } - - /* Create link from CIO2 subdev to output node */ - r = media_create_pad_link( - &subdev->entity, CIO2_PAD_SOURCE, &vdev->entity, 0, - MEDIA_LNK_FL_ENABLED | MEDIA_LNK_FL_IMMUTABLE); - if (r) - goto fail_link; - - return 0; - -fail_link: - vb2_video_unregister_device(&q->vdev); -fail_vdev: - v4l2_device_unregister_subdev(subdev); -fail_subdev: - media_entity_cleanup(&vdev->entity); -fail_vdev_media_entity: - media_entity_cleanup(&subdev->entity); -fail_subdev_media_entity: - cio2_fbpt_exit(q, &cio2->pci_dev->dev); -fail_fbpt: - mutex_destroy(&q->subdev_lock); - mutex_destroy(&q->lock); - - return r; -} - -static void cio2_queue_exit(struct cio2_device *cio2, struct cio2_queue *q) -{ - vb2_video_unregister_device(&q->vdev); - media_entity_cleanup(&q->vdev.entity); - v4l2_device_unregister_subdev(&q->subdev); - media_entity_cleanup(&q->subdev.entity); - cio2_fbpt_exit(q, &cio2->pci_dev->dev); - mutex_destroy(&q->subdev_lock); - mutex_destroy(&q->lock); -} - -static int cio2_queues_init(struct cio2_device *cio2) -{ - int i, r; - - for (i = 0; i < CIO2_QUEUES; i++) { - r = cio2_queue_init(cio2, &cio2->queue[i]); - if (r) - break; - } - - if (i == CIO2_QUEUES) - return 0; - - for (i--; i >= 0; i--) - cio2_queue_exit(cio2, &cio2->queue[i]); - - return r; -} - -static void cio2_queues_exit(struct cio2_device *cio2) -{ - unsigned int i; - - for (i = 0; i < CIO2_QUEUES; i++) - cio2_queue_exit(cio2, &cio2->queue[i]); -} - -/**************** PCI interface ****************/ - -static int cio2_pci_probe(struct pci_dev *pci_dev, - const struct pci_device_id *id) -{ - struct cio2_device *cio2; - int r; - - cio2 = devm_kzalloc(&pci_dev->dev, sizeof(*cio2), GFP_KERNEL); - if (!cio2) - return -ENOMEM; - cio2->pci_dev = pci_dev; - - r = pcim_enable_device(pci_dev); - if (r) { - dev_err(&pci_dev->dev, "failed to enable device (%d)\n", r); - return r; - } - - dev_info(&pci_dev->dev, "device 0x%x (rev: 0x%x)\n", - pci_dev->device, pci_dev->revision); - - r = pcim_iomap_regions(pci_dev, 1 << CIO2_PCI_BAR, pci_name(pci_dev)); - if (r) { - dev_err(&pci_dev->dev, "failed to remap I/O memory (%d)\n", r); - return -ENODEV; - } - - cio2->base = pcim_iomap_table(pci_dev)[CIO2_PCI_BAR]; - - pci_set_drvdata(pci_dev, cio2); - - pci_set_master(pci_dev); - - r = pci_set_dma_mask(pci_dev, CIO2_DMA_MASK); - if (r) { - dev_err(&pci_dev->dev, "failed to set DMA mask (%d)\n", r); - return -ENODEV; - } - - r = pci_enable_msi(pci_dev); - if (r) { - dev_err(&pci_dev->dev, "failed to enable MSI (%d)\n", r); - return r; - } - - r = cio2_fbpt_init_dummy(cio2); - if (r) - return r; - - mutex_init(&cio2->lock); - - cio2->media_dev.dev = &cio2->pci_dev->dev; - strscpy(cio2->media_dev.model, CIO2_DEVICE_NAME, - sizeof(cio2->media_dev.model)); - snprintf(cio2->media_dev.bus_info, sizeof(cio2->media_dev.bus_info), - "PCI:%s", pci_name(cio2->pci_dev)); - cio2->media_dev.hw_revision = 0; - - media_device_init(&cio2->media_dev); - r = media_device_register(&cio2->media_dev); - if (r < 0) - goto fail_mutex_destroy; - - cio2->v4l2_dev.mdev = &cio2->media_dev; - r = v4l2_device_register(&pci_dev->dev, &cio2->v4l2_dev); - if (r) { - dev_err(&pci_dev->dev, - "failed to register V4L2 device (%d)\n", r); - goto fail_media_device_unregister; - } - - r = cio2_queues_init(cio2); - if (r) - goto fail_v4l2_device_unregister; - - v4l2_async_notifier_init(&cio2->notifier); - - /* Register notifier for subdevices we care */ - r = cio2_parse_firmware(cio2); - if (r) - goto fail_clean_notifier; - - r = devm_request_irq(&pci_dev->dev, pci_dev->irq, cio2_irq, - IRQF_SHARED, CIO2_NAME, cio2); - if (r) { - dev_err(&pci_dev->dev, "failed to request IRQ (%d)\n", r); - goto fail_clean_notifier; - } - - pm_runtime_put_noidle(&pci_dev->dev); - pm_runtime_allow(&pci_dev->dev); - - return 0; - -fail_clean_notifier: - v4l2_async_notifier_unregister(&cio2->notifier); - v4l2_async_notifier_cleanup(&cio2->notifier); - cio2_queues_exit(cio2); -fail_v4l2_device_unregister: - v4l2_device_unregister(&cio2->v4l2_dev); -fail_media_device_unregister: - media_device_unregister(&cio2->media_dev); - media_device_cleanup(&cio2->media_dev); -fail_mutex_destroy: - mutex_destroy(&cio2->lock); - cio2_fbpt_exit_dummy(cio2); - - return r; -} - -static void cio2_pci_remove(struct pci_dev *pci_dev) -{ - struct cio2_device *cio2 = pci_get_drvdata(pci_dev); - - media_device_unregister(&cio2->media_dev); - v4l2_async_notifier_unregister(&cio2->notifier); - v4l2_async_notifier_cleanup(&cio2->notifier); - cio2_queues_exit(cio2); - cio2_fbpt_exit_dummy(cio2); - v4l2_device_unregister(&cio2->v4l2_dev); - media_device_cleanup(&cio2->media_dev); - mutex_destroy(&cio2->lock); -} - -static int __maybe_unused cio2_runtime_suspend(struct device *dev) -{ - struct pci_dev *pci_dev = to_pci_dev(dev); - struct cio2_device *cio2 = pci_get_drvdata(pci_dev); - void __iomem *const base = cio2->base; - u16 pm; - - writel(CIO2_D0I3C_I3, base + CIO2_REG_D0I3C); - dev_dbg(dev, "cio2 runtime suspend.\n"); - - pci_read_config_word(pci_dev, pci_dev->pm_cap + CIO2_PMCSR_OFFSET, &pm); - pm = (pm >> CIO2_PMCSR_D0D3_SHIFT) << CIO2_PMCSR_D0D3_SHIFT; - pm |= CIO2_PMCSR_D3; - pci_write_config_word(pci_dev, pci_dev->pm_cap + CIO2_PMCSR_OFFSET, pm); - - return 0; -} - -static int __maybe_unused cio2_runtime_resume(struct device *dev) -{ - struct pci_dev *pci_dev = to_pci_dev(dev); - struct cio2_device *cio2 = pci_get_drvdata(pci_dev); - void __iomem *const base = cio2->base; - u16 pm; - - writel(CIO2_D0I3C_RR, base + CIO2_REG_D0I3C); - dev_dbg(dev, "cio2 runtime resume.\n"); - - pci_read_config_word(pci_dev, pci_dev->pm_cap + CIO2_PMCSR_OFFSET, &pm); - pm = (pm >> CIO2_PMCSR_D0D3_SHIFT) << CIO2_PMCSR_D0D3_SHIFT; - pci_write_config_word(pci_dev, pci_dev->pm_cap + CIO2_PMCSR_OFFSET, pm); - - return 0; -} - -/* - * Helper function to advance all the elements of a circular buffer by "start" - * positions - */ -static void arrange(void *ptr, size_t elem_size, size_t elems, size_t start) -{ - struct { - size_t begin, end; - } arr[2] = { - { 0, start - 1 }, - { start, elems - 1 }, - }; - -#define CHUNK_SIZE(a) ((a)->end - (a)->begin + 1) - - /* Loop as long as we have out-of-place entries */ - while (CHUNK_SIZE(&arr[0]) && CHUNK_SIZE(&arr[1])) { - size_t size0, i; - - /* - * Find the number of entries that can be arranged on this - * iteration. - */ - size0 = min(CHUNK_SIZE(&arr[0]), CHUNK_SIZE(&arr[1])); - - /* Swap the entries in two parts of the array. */ - for (i = 0; i < size0; i++) { - u8 *d = ptr + elem_size * (arr[1].begin + i); - u8 *s = ptr + elem_size * (arr[0].begin + i); - size_t j; - - for (j = 0; j < elem_size; j++) - swap(d[j], s[j]); - } - - if (CHUNK_SIZE(&arr[0]) > CHUNK_SIZE(&arr[1])) { - /* The end of the first array remains unarranged. */ - arr[0].begin += size0; - } else { - /* - * The first array is fully arranged so we proceed - * handling the next one. - */ - arr[0].begin = arr[1].begin; - arr[0].end = arr[1].begin + size0 - 1; - arr[1].begin += size0; - } - } -} - -static void cio2_fbpt_rearrange(struct cio2_device *cio2, struct cio2_queue *q) -{ - unsigned int i, j; - - for (i = 0, j = q->bufs_first; i < CIO2_MAX_BUFFERS; - i++, j = (j + 1) % CIO2_MAX_BUFFERS) - if (q->bufs[j]) - break; - - if (i == CIO2_MAX_BUFFERS) - return; - - if (j) { - arrange(q->fbpt, sizeof(struct cio2_fbpt_entry) * CIO2_MAX_LOPS, - CIO2_MAX_BUFFERS, j); - arrange(q->bufs, sizeof(struct cio2_buffer *), - CIO2_MAX_BUFFERS, j); - } - - /* - * DMA clears the valid bit when accessing the buffer. - * When stopping stream in suspend callback, some of the buffers - * may be in invalid state. After resume, when DMA meets the invalid - * buffer, it will halt and stop receiving new data. - * To avoid DMA halting, set the valid bit for all buffers in FBPT. - */ - for (i = 0; i < CIO2_MAX_BUFFERS; i++) - cio2_fbpt_entry_enable(cio2, q->fbpt + i * CIO2_MAX_LOPS); -} - -static int __maybe_unused cio2_suspend(struct device *dev) -{ - struct pci_dev *pci_dev = to_pci_dev(dev); - struct cio2_device *cio2 = pci_get_drvdata(pci_dev); - struct cio2_queue *q = cio2->cur_queue; - - dev_dbg(dev, "cio2 suspend\n"); - if (!cio2->streaming) - return 0; - - /* Stop stream */ - cio2_hw_exit(cio2, q); - synchronize_irq(pci_dev->irq); - - pm_runtime_force_suspend(dev); - - /* - * Upon resume, hw starts to process the fbpt entries from beginning, - * so relocate the queued buffs to the fbpt head before suspend. - */ - cio2_fbpt_rearrange(cio2, q); - q->bufs_first = 0; - q->bufs_next = 0; - - return 0; -} - -static int __maybe_unused cio2_resume(struct device *dev) -{ - struct cio2_device *cio2 = dev_get_drvdata(dev); - struct cio2_queue *q = cio2->cur_queue; - int r; - - dev_dbg(dev, "cio2 resume\n"); - if (!cio2->streaming) - return 0; - /* Start stream */ - r = pm_runtime_force_resume(&cio2->pci_dev->dev); - if (r < 0) { - dev_err(&cio2->pci_dev->dev, - "failed to set power %d\n", r); - return r; - } - - r = cio2_hw_init(cio2, q); - if (r) - dev_err(dev, "fail to init cio2 hw\n"); - - return r; -} - -static const struct dev_pm_ops cio2_pm_ops = { - SET_RUNTIME_PM_OPS(&cio2_runtime_suspend, &cio2_runtime_resume, NULL) - SET_SYSTEM_SLEEP_PM_OPS(&cio2_suspend, &cio2_resume) -}; - -static const struct pci_device_id cio2_pci_id_table[] = { - { PCI_DEVICE(PCI_VENDOR_ID_INTEL, CIO2_PCI_ID) }, - { } -}; - -MODULE_DEVICE_TABLE(pci, cio2_pci_id_table); - -static struct pci_driver cio2_pci_driver = { - .name = CIO2_NAME, - .id_table = cio2_pci_id_table, - .probe = cio2_pci_probe, - .remove = cio2_pci_remove, - .driver = { - .pm = &cio2_pm_ops, - }, -}; - -module_pci_driver(cio2_pci_driver); - -MODULE_AUTHOR("Tuukka Toivonen <tuukka.toivonen@intel.com>"); -MODULE_AUTHOR("Tianshu Qiu <tian.shu.qiu@intel.com>"); -MODULE_AUTHOR("Jian Xu Zheng"); -MODULE_AUTHOR("Yuning Pu <yuning.pu@intel.com>"); -MODULE_AUTHOR("Yong Zhi <yong.zhi@intel.com>"); -MODULE_LICENSE("GPL v2"); -MODULE_DESCRIPTION("IPU3 CIO2 driver"); |