|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * Driver for the Texas Instruments DS90UB960-Q1 video deserializer | 
|  | * | 
|  | * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net> | 
|  | * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com> | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * (Possible) TODOs: | 
|  | * | 
|  | * - PM for serializer and remote peripherals. We need to manage: | 
|  | *   - VPOC | 
|  | *     - Power domain? Regulator? Somehow any remote device should be able to | 
|  | *       cause the VPOC to be turned on. | 
|  | *   - Link between the deserializer and the serializer | 
|  | *     - Related to VPOC management. We probably always want to turn on the VPOC | 
|  | *       and then enable the link. | 
|  | *   - Serializer's services: i2c, gpios, power | 
|  | *     - The serializer needs to resume before the remote peripherals can | 
|  | *       e.g. use the i2c. | 
|  | *     - How to handle gpios? Reserving a gpio essentially keeps the provider | 
|  | *       (serializer) always powered on. | 
|  | * - Do we need a new bus for the FPD-Link? At the moment the serializers | 
|  | *   are children of the same i2c-adapter where the deserializer resides. | 
|  | * - i2c-atr could be made embeddable instead of allocatable. | 
|  | */ | 
|  |  | 
|  | #include <linux/bitops.h> | 
|  | #include <linux/cleanup.h> | 
|  | #include <linux/clk.h> | 
|  | #include <linux/delay.h> | 
|  | #include <linux/fwnode.h> | 
|  | #include <linux/gpio/consumer.h> | 
|  | #include <linux/i2c-atr.h> | 
|  | #include <linux/i2c.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/interrupt.h> | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/kthread.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/mutex.h> | 
|  | #include <linux/property.h> | 
|  | #include <linux/regmap.h> | 
|  | #include <linux/regulator/consumer.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/units.h> | 
|  | #include <linux/workqueue.h> | 
|  |  | 
|  | #include <media/i2c/ds90ub9xx.h> | 
|  | #include <media/mipi-csi2.h> | 
|  | #include <media/v4l2-ctrls.h> | 
|  | #include <media/v4l2-fwnode.h> | 
|  | #include <media/v4l2-subdev.h> | 
|  |  | 
|  | #include "ds90ub953.h" | 
|  |  | 
|  | #define MHZ(v) ((u32)((v) * HZ_PER_MHZ)) | 
|  |  | 
|  | /* | 
|  | * If this is defined, the i2c addresses from UB960_DEBUG_I2C_RX_ID to | 
|  | * UB960_DEBUG_I2C_RX_ID + 3 can be used to access the paged RX port registers | 
|  | * directly. | 
|  | * | 
|  | * Only for debug purposes. | 
|  | */ | 
|  | /* #define UB960_DEBUG_I2C_RX_ID	0x40 */ | 
|  |  | 
|  | #define UB960_POLL_TIME_MS	500 | 
|  |  | 
|  | #define UB960_MAX_RX_NPORTS	4 | 
|  | #define UB960_MAX_TX_NPORTS	2 | 
|  | #define UB960_MAX_NPORTS	(UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS) | 
|  |  | 
|  | #define UB960_MAX_PORT_ALIASES	8 | 
|  |  | 
|  | #define UB960_NUM_BC_GPIOS		4 | 
|  |  | 
|  | /* | 
|  | * Register map | 
|  | * | 
|  | * 0x00-0x32   Shared (UB960_SR) | 
|  | * 0x33-0x3a   CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR) | 
|  | * 0x4c        Shared (UB960_SR) | 
|  | * 0x4d-0x7f   FPD-Link RX, per-port paged (UB960_RR) | 
|  | * 0xb0-0xbf   Shared (UB960_SR) | 
|  | * 0xd0-0xdf   FPD-Link RX, per-port paged (UB960_RR) | 
|  | * 0xf0-0xf5   Shared (UB960_SR) | 
|  | * 0xf8-0xfb   Shared (UB960_SR) | 
|  | * All others  Reserved | 
|  | * | 
|  | * Register prefixes: | 
|  | * UB960_SR_* = Shared register | 
|  | * UB960_RR_* = FPD-Link RX, per-port paged register | 
|  | * UB960_TR_* = CSI-2 TX, per-port paged register | 
|  | * UB960_XR_* = Reserved register | 
|  | * UB960_IR_* = Indirect register | 
|  | */ | 
|  |  | 
|  | #define UB960_SR_I2C_DEV_ID			0x00 | 
|  | #define UB960_SR_RESET				0x01 | 
|  | #define UB960_SR_RESET_DIGITAL_RESET1		BIT(1) | 
|  | #define UB960_SR_RESET_DIGITAL_RESET0		BIT(0) | 
|  | #define UB960_SR_RESET_GPIO_LOCK_RELEASE	BIT(5) | 
|  |  | 
|  | #define UB960_SR_GEN_CONFIG			0x02 | 
|  | #define UB960_SR_REV_MASK			0x03 | 
|  | #define UB960_SR_DEVICE_STS			0x04 | 
|  | #define UB960_SR_PAR_ERR_THOLD_HI		0x05 | 
|  | #define UB960_SR_PAR_ERR_THOLD_LO		0x06 | 
|  | #define UB960_SR_BCC_WDOG_CTL			0x07 | 
|  | #define UB960_SR_I2C_CTL1			0x08 | 
|  | #define UB960_SR_I2C_CTL2			0x09 | 
|  | #define UB960_SR_SCL_HIGH_TIME			0x0a | 
|  | #define UB960_SR_SCL_LOW_TIME			0x0b | 
|  | #define UB960_SR_RX_PORT_CTL			0x0c | 
|  | #define UB960_SR_IO_CTL				0x0d | 
|  | #define UB960_SR_GPIO_PIN_STS			0x0e | 
|  | #define UB960_SR_GPIO_INPUT_CTL			0x0f | 
|  | #define UB960_SR_GPIO_PIN_CTL(n)		(0x10 + (n)) /* n < UB960_NUM_GPIOS */ | 
|  | #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL		5 | 
|  | #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT	2 | 
|  | #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN		BIT(0) | 
|  |  | 
|  | #define UB960_SR_FS_CTL				0x18 | 
|  | #define UB960_SR_FS_HIGH_TIME_1			0x19 | 
|  | #define UB960_SR_FS_HIGH_TIME_0			0x1a | 
|  | #define UB960_SR_FS_LOW_TIME_1			0x1b | 
|  | #define UB960_SR_FS_LOW_TIME_0			0x1c | 
|  | #define UB960_SR_MAX_FRM_HI			0x1d | 
|  | #define UB960_SR_MAX_FRM_LO			0x1e | 
|  | #define UB960_SR_CSI_PLL_CTL			0x1f | 
|  |  | 
|  | #define UB960_SR_FWD_CTL1			0x20 | 
|  | #define UB960_SR_FWD_CTL1_PORT_DIS(n)		BIT((n) + 4) | 
|  |  | 
|  | #define UB960_SR_FWD_CTL2			0x21 | 
|  | #define UB960_SR_FWD_STS			0x22 | 
|  |  | 
|  | #define UB960_SR_INTERRUPT_CTL			0x23 | 
|  | #define UB960_SR_INTERRUPT_CTL_INT_EN		BIT(7) | 
|  | #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0	BIT(4) | 
|  | #define UB960_SR_INTERRUPT_CTL_IE_RX(n)		BIT((n)) /* rxport[n] IRQ */ | 
|  |  | 
|  | #define UB960_SR_INTERRUPT_STS			0x24 | 
|  | #define UB960_SR_INTERRUPT_STS_INT		BIT(7) | 
|  | #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n)	BIT(4 + (n)) /* txport[n] IRQ */ | 
|  | #define UB960_SR_INTERRUPT_STS_IS_RX(n)		BIT((n)) /* rxport[n] IRQ */ | 
|  |  | 
|  | #define UB960_SR_TS_CONFIG			0x25 | 
|  | #define UB960_SR_TS_CONTROL			0x26 | 
|  | #define UB960_SR_TS_LINE_HI			0x27 | 
|  | #define UB960_SR_TS_LINE_LO			0x28 | 
|  | #define UB960_SR_TS_STATUS			0x29 | 
|  | #define UB960_SR_TIMESTAMP_P0_HI		0x2a | 
|  | #define UB960_SR_TIMESTAMP_P0_LO		0x2b | 
|  | #define UB960_SR_TIMESTAMP_P1_HI		0x2c | 
|  | #define UB960_SR_TIMESTAMP_P1_LO		0x2d | 
|  |  | 
|  | #define UB960_SR_CSI_PORT_SEL			0x32 | 
|  |  | 
|  | #define UB960_TR_CSI_CTL			0x33 | 
|  | #define UB960_TR_CSI_CTL_CSI_CAL_EN		BIT(6) | 
|  | #define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK	BIT(1) | 
|  | #define UB960_TR_CSI_CTL_CSI_ENABLE		BIT(0) | 
|  |  | 
|  | #define UB960_TR_CSI_CTL2			0x34 | 
|  | #define UB960_TR_CSI_STS			0x35 | 
|  | #define UB960_TR_CSI_TX_ICR			0x36 | 
|  |  | 
|  | #define UB960_TR_CSI_TX_ISR			0x37 | 
|  | #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR	BIT(3) | 
|  | #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR	BIT(1) | 
|  |  | 
|  | #define UB960_TR_CSI_TEST_CTL			0x38 | 
|  | #define UB960_TR_CSI_TEST_PATT_HI		0x39 | 
|  | #define UB960_TR_CSI_TEST_PATT_LO		0x3a | 
|  |  | 
|  | #define UB960_XR_SFILTER_CFG			0x41 | 
|  | #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT	4 | 
|  | #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT	0 | 
|  |  | 
|  | #define UB960_XR_AEQ_CTL1			0x42 | 
|  | #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK	BIT(6) | 
|  | #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING	BIT(5) | 
|  | #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY	BIT(4) | 
|  | #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK        \ | 
|  | (UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK |  \ | 
|  | UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \ | 
|  | UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY) | 
|  | #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN	BIT(0) | 
|  |  | 
|  | #define UB960_XR_AEQ_ERR_THOLD			0x43 | 
|  |  | 
|  | #define UB960_RR_BCC_ERR_CTL			0x46 | 
|  | #define UB960_RR_BCC_STATUS			0x47 | 
|  | #define UB960_RR_BCC_STATUS_SEQ_ERROR		BIT(5) | 
|  | #define UB960_RR_BCC_STATUS_MASTER_ERR		BIT(4) | 
|  | #define UB960_RR_BCC_STATUS_MASTER_TO		BIT(3) | 
|  | #define UB960_RR_BCC_STATUS_SLAVE_ERR		BIT(2) | 
|  | #define UB960_RR_BCC_STATUS_SLAVE_TO		BIT(1) | 
|  | #define UB960_RR_BCC_STATUS_RESP_ERR		BIT(0) | 
|  | #define UB960_RR_BCC_STATUS_ERROR_MASK                                    \ | 
|  | (UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \ | 
|  | UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR |  \ | 
|  | UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR) | 
|  |  | 
|  | #define UB960_RR_FPD3_CAP			0x4a | 
|  | #define UB960_RR_RAW_EMBED_DTYPE		0x4b | 
|  | #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT	6 | 
|  |  | 
|  | #define UB960_SR_FPD3_PORT_SEL			0x4c | 
|  |  | 
|  | #define UB960_RR_RX_PORT_STS1			0x4d | 
|  | #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR	BIT(5) | 
|  | #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG	BIT(4) | 
|  | #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR	BIT(3) | 
|  | #define UB960_RR_RX_PORT_STS1_PARITY_ERROR	BIT(2) | 
|  | #define UB960_RR_RX_PORT_STS1_PORT_PASS		BIT(1) | 
|  | #define UB960_RR_RX_PORT_STS1_LOCK_STS		BIT(0) | 
|  | #define UB960_RR_RX_PORT_STS1_ERROR_MASK       \ | 
|  | (UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \ | 
|  | UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \ | 
|  | UB960_RR_RX_PORT_STS1_PARITY_ERROR) | 
|  |  | 
|  | #define UB960_RR_RX_PORT_STS2			0x4e | 
|  | #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE	BIT(7) | 
|  | #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG	BIT(6) | 
|  | #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR	BIT(5) | 
|  | #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR	BIT(4) | 
|  | #define UB960_RR_RX_PORT_STS2_CSI_ERROR		BIT(3) | 
|  | #define UB960_RR_RX_PORT_STS2_FREQ_STABLE	BIT(2) | 
|  | #define UB960_RR_RX_PORT_STS2_CABLE_FAULT	BIT(1) | 
|  | #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG	BIT(0) | 
|  | #define UB960_RR_RX_PORT_STS2_ERROR_MASK       \ | 
|  | UB960_RR_RX_PORT_STS2_BUFFER_ERROR | 
|  |  | 
|  | #define UB960_RR_RX_FREQ_HIGH			0x4f | 
|  | #define UB960_RR_RX_FREQ_LOW			0x50 | 
|  | #define UB960_RR_SENSOR_STS_0			0x51 | 
|  | #define UB960_RR_SENSOR_STS_1			0x52 | 
|  | #define UB960_RR_SENSOR_STS_2			0x53 | 
|  | #define UB960_RR_SENSOR_STS_3			0x54 | 
|  | #define UB960_RR_RX_PAR_ERR_HI			0x55 | 
|  | #define UB960_RR_RX_PAR_ERR_LO			0x56 | 
|  | #define UB960_RR_BIST_ERR_COUNT			0x57 | 
|  |  | 
|  | #define UB960_RR_BCC_CONFIG			0x58 | 
|  | #define UB960_RR_BCC_CONFIG_BC_ALWAYS_ON	BIT(4) | 
|  | #define UB960_RR_BCC_CONFIG_AUTO_ACK_ALL	BIT(5) | 
|  | #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH	BIT(6) | 
|  | #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK	GENMASK(2, 0) | 
|  |  | 
|  | #define UB960_RR_DATAPATH_CTL1			0x59 | 
|  | #define UB960_RR_DATAPATH_CTL2			0x5a | 
|  | #define UB960_RR_SER_ID				0x5b | 
|  | #define UB960_RR_SER_ID_FREEZE_DEVICE_ID	BIT(0) | 
|  | #define UB960_RR_SER_ALIAS_ID			0x5c | 
|  | #define UB960_RR_SER_ALIAS_ID_AUTO_ACK		BIT(0) | 
|  |  | 
|  | /* For these two register sets: n < UB960_MAX_PORT_ALIASES */ | 
|  | #define UB960_RR_SLAVE_ID(n)			(0x5d + (n)) | 
|  | #define UB960_RR_SLAVE_ALIAS(n)			(0x65 + (n)) | 
|  |  | 
|  | #define UB960_RR_PORT_CONFIG			0x6d | 
|  | #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK	GENMASK(1, 0) | 
|  |  | 
|  | #define UB960_RR_BC_GPIO_CTL(n)			(0x6e + (n)) /* n < 2 */ | 
|  | #define UB960_RR_RAW10_ID			0x70 | 
|  | #define UB960_RR_RAW10_ID_VC_SHIFT		6 | 
|  | #define UB960_RR_RAW10_ID_DT_SHIFT		0 | 
|  |  | 
|  | #define UB960_RR_RAW12_ID			0x71 | 
|  | #define UB960_RR_CSI_VC_MAP			0x72 | 
|  | #define UB960_RR_CSI_VC_MAP_SHIFT(x)		((x) * 2) | 
|  |  | 
|  | #define UB960_RR_LINE_COUNT_HI			0x73 | 
|  | #define UB960_RR_LINE_COUNT_LO			0x74 | 
|  | #define UB960_RR_LINE_LEN_1			0x75 | 
|  | #define UB960_RR_LINE_LEN_0			0x76 | 
|  | #define UB960_RR_FREQ_DET_CTL			0x77 | 
|  | #define UB960_RR_MAILBOX_1			0x78 | 
|  | #define UB960_RR_MAILBOX_2			0x79 | 
|  |  | 
|  | #define UB960_RR_CSI_RX_STS			0x7a | 
|  | #define UB960_RR_CSI_RX_STS_LENGTH_ERR		BIT(3) | 
|  | #define UB960_RR_CSI_RX_STS_CKSUM_ERR		BIT(2) | 
|  | #define UB960_RR_CSI_RX_STS_ECC2_ERR		BIT(1) | 
|  | #define UB960_RR_CSI_RX_STS_ECC1_ERR		BIT(0) | 
|  | #define UB960_RR_CSI_RX_STS_ERROR_MASK                                    \ | 
|  | (UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \ | 
|  | UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR) | 
|  |  | 
|  | #define UB960_RR_CSI_ERR_COUNTER		0x7b | 
|  | #define UB960_RR_PORT_CONFIG2			0x7c | 
|  | #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6) | 
|  | #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6 | 
|  |  | 
|  | #define UB960_RR_PORT_CONFIG2_LV_POL_LOW	BIT(1) | 
|  | #define UB960_RR_PORT_CONFIG2_FV_POL_LOW	BIT(0) | 
|  |  | 
|  | #define UB960_RR_PORT_PASS_CTL			0x7d | 
|  | #define UB960_RR_SEN_INT_RISE_CTL		0x7e | 
|  | #define UB960_RR_SEN_INT_FALL_CTL		0x7f | 
|  |  | 
|  | #define UB960_SR_CSI_FRAME_COUNT_HI(n)		(0x90 + 8 * (n)) | 
|  | #define UB960_SR_CSI_FRAME_COUNT_LO(n)		(0x91 + 8 * (n)) | 
|  | #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n)	(0x92 + 8 * (n)) | 
|  | #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n)	(0x93 + 8 * (n)) | 
|  | #define UB960_SR_CSI_LINE_COUNT_HI(n)		(0x94 + 8 * (n)) | 
|  | #define UB960_SR_CSI_LINE_COUNT_LO(n)		(0x95 + 8 * (n)) | 
|  | #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n)	(0x96 + 8 * (n)) | 
|  | #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n)	(0x97 + 8 * (n)) | 
|  |  | 
|  | #define UB960_XR_REFCLK_FREQ			0xa5	/* UB960 */ | 
|  |  | 
|  | #define UB960_SR_IND_ACC_CTL			0xb0 | 
|  | #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC	BIT(1) | 
|  |  | 
|  | #define UB960_SR_IND_ACC_ADDR			0xb1 | 
|  | #define UB960_SR_IND_ACC_DATA			0xb2 | 
|  | #define UB960_SR_BIST_CONTROL			0xb3 | 
|  | #define UB960_SR_MODE_IDX_STS			0xb8 | 
|  | #define UB960_SR_LINK_ERROR_COUNT		0xb9 | 
|  | #define UB960_SR_FPD3_ENC_CTL			0xba | 
|  | #define UB960_SR_FV_MIN_TIME			0xbc | 
|  | #define UB960_SR_GPIO_PD_CTL			0xbe | 
|  |  | 
|  | #define UB960_RR_PORT_DEBUG			0xd0 | 
|  | #define UB960_RR_AEQ_CTL2			0xd2 | 
|  | #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR		BIT(2) | 
|  |  | 
|  | #define UB960_RR_AEQ_STATUS			0xd3 | 
|  | #define UB960_RR_AEQ_STATUS_STATUS_2		GENMASK(5, 3) | 
|  | #define UB960_RR_AEQ_STATUS_STATUS_1		GENMASK(2, 0) | 
|  |  | 
|  | #define UB960_RR_AEQ_BYPASS			0xd4 | 
|  | #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT	5 | 
|  | #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK	GENMASK(7, 5) | 
|  | #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT	1 | 
|  | #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK	GENMASK(3, 1) | 
|  | #define UB960_RR_AEQ_BYPASS_ENABLE			BIT(0) | 
|  |  | 
|  | #define UB960_RR_AEQ_MIN_MAX			0xd5 | 
|  | #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT	4 | 
|  | #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT	0 | 
|  |  | 
|  | #define UB960_RR_SFILTER_STS_0			0xd6 | 
|  | #define UB960_RR_SFILTER_STS_1			0xd7 | 
|  | #define UB960_RR_PORT_ICR_HI			0xd8 | 
|  | #define UB960_RR_PORT_ICR_LO			0xd9 | 
|  | #define UB960_RR_PORT_ISR_HI			0xda | 
|  | #define UB960_RR_PORT_ISR_LO			0xdb | 
|  | #define UB960_RR_FC_GPIO_STS			0xdc | 
|  | #define UB960_RR_FC_GPIO_ICR			0xdd | 
|  | #define UB960_RR_SEN_INT_RISE_STS		0xde | 
|  | #define UB960_RR_SEN_INT_FALL_STS		0xdf | 
|  |  | 
|  |  | 
|  | #define UB960_SR_FPD3_RX_ID(n)			(0xf0 + (n)) | 
|  | #define UB960_SR_FPD3_RX_ID_LEN			6 | 
|  |  | 
|  | #define UB960_SR_I2C_RX_ID(n)			(0xf8 + (n)) | 
|  |  | 
|  | /* Indirect register blocks */ | 
|  | #define UB960_IND_TARGET_PAT_GEN		0x00 | 
|  | #define UB960_IND_TARGET_RX_ANA(n)		(0x01 + (n)) | 
|  | #define UB960_IND_TARGET_CSI_ANA		0x07 | 
|  |  | 
|  | /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */ | 
|  |  | 
|  | #define UB960_IR_PGEN_CTL			0x01 | 
|  | #define UB960_IR_PGEN_CTL_PGEN_ENABLE		BIT(0) | 
|  |  | 
|  | #define UB960_IR_PGEN_CFG			0x02 | 
|  | #define UB960_IR_PGEN_CSI_DI			0x03 | 
|  | #define UB960_IR_PGEN_LINE_SIZE1		0x04 | 
|  | #define UB960_IR_PGEN_LINE_SIZE0		0x05 | 
|  | #define UB960_IR_PGEN_BAR_SIZE1			0x06 | 
|  | #define UB960_IR_PGEN_BAR_SIZE0			0x07 | 
|  | #define UB960_IR_PGEN_ACT_LPF1			0x08 | 
|  | #define UB960_IR_PGEN_ACT_LPF0			0x09 | 
|  | #define UB960_IR_PGEN_TOT_LPF1			0x0a | 
|  | #define UB960_IR_PGEN_TOT_LPF0			0x0b | 
|  | #define UB960_IR_PGEN_LINE_PD1			0x0c | 
|  | #define UB960_IR_PGEN_LINE_PD0			0x0d | 
|  | #define UB960_IR_PGEN_VBP			0x0e | 
|  | #define UB960_IR_PGEN_VFP			0x0f | 
|  | #define UB960_IR_PGEN_COLOR(n)			(0x10 + (n)) /* n < 15 */ | 
|  |  | 
|  | #define UB960_IR_RX_ANA_STROBE_SET_CLK		0x08 | 
|  | #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY	BIT(3) | 
|  | #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK	GENMASK(2, 0) | 
|  |  | 
|  | #define UB960_IR_RX_ANA_STROBE_SET_DATA		0x09 | 
|  | #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY	BIT(3) | 
|  | #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK	GENMASK(2, 0) | 
|  |  | 
|  | /* UB9702 Registers */ | 
|  |  | 
|  | #define UB9702_SR_CSI_EXCLUSIVE_FWD2		0x3c | 
|  | #define UB9702_SR_REFCLK_FREQ			0x3d | 
|  | #define UB9702_RR_RX_CTL_1			0x80 | 
|  | #define UB9702_RR_RX_CTL_2			0x87 | 
|  | #define UB9702_RR_VC_ID_MAP(x)			(0xa0 + (x)) | 
|  | #define UB9702_SR_FPD_RATE_CFG			0xc2 | 
|  | #define UB9702_SR_CSI_PLL_DIV			0xc9 | 
|  | #define UB9702_RR_RX_SM_SEL_2			0xd4 | 
|  | #define UB9702_RR_CHANNEL_MODE			0xe4 | 
|  |  | 
|  | #define UB9702_IND_TARGET_SAR_ADC		0x0a | 
|  |  | 
|  | #define UB9702_IR_RX_ANA_FPD_BC_CTL0		0x04 | 
|  | #define UB9702_IR_RX_ANA_FPD_BC_CTL1		0x0d | 
|  | #define UB9702_IR_RX_ANA_FPD_BC_CTL2		0x1b | 
|  | #define UB9702_IR_RX_ANA_SYSTEM_INIT_REG0	0x21 | 
|  | #define UB9702_IR_RX_ANA_AEQ_ALP_SEL6		0x27 | 
|  | #define UB9702_IR_RX_ANA_AEQ_ALP_SEL7		0x28 | 
|  | #define UB9702_IR_RX_ANA_AEQ_ALP_SEL10		0x2b | 
|  | #define UB9702_IR_RX_ANA_AEQ_ALP_SEL11		0x2c | 
|  | #define UB9702_IR_RX_ANA_EQ_ADAPT_CTRL		0x2e | 
|  | #define UB9702_IR_RX_ANA_AEQ_CFG_1		0x34 | 
|  | #define UB9702_IR_RX_ANA_AEQ_CFG_2		0x4d | 
|  | #define UB9702_IR_RX_ANA_GAIN_CTRL_0		0x71 | 
|  | #define UB9702_IR_RX_ANA_GAIN_CTRL_0		0x71 | 
|  | #define UB9702_IR_RX_ANA_VGA_CTRL_SEL_1		0x72 | 
|  | #define UB9702_IR_RX_ANA_VGA_CTRL_SEL_2		0x73 | 
|  | #define UB9702_IR_RX_ANA_VGA_CTRL_SEL_3		0x74 | 
|  | #define UB9702_IR_RX_ANA_VGA_CTRL_SEL_6		0x77 | 
|  | #define UB9702_IR_RX_ANA_AEQ_CFG_3		0x79 | 
|  | #define UB9702_IR_RX_ANA_AEQ_CFG_4		0x85 | 
|  | #define UB9702_IR_RX_ANA_EQ_CTRL_SEL_15		0x87 | 
|  | #define UB9702_IR_RX_ANA_EQ_CTRL_SEL_24		0x90 | 
|  | #define UB9702_IR_RX_ANA_EQ_CTRL_SEL_38		0x9e | 
|  | #define UB9702_IR_RX_ANA_FPD3_CDR_CTRL_SEL_5	0xa5 | 
|  | #define UB9702_IR_RX_ANA_FPD3_AEQ_CTRL_SEL_1	0xa8 | 
|  | #define UB9702_IR_RX_ANA_EQ_OVERRIDE_CTRL	0xf0 | 
|  | #define UB9702_IR_RX_ANA_VGA_CTRL_SEL_8		0xf1 | 
|  |  | 
|  | #define UB9702_IR_CSI_ANA_CSIPLL_REG_1		0x92 | 
|  |  | 
|  | /* EQ related */ | 
|  |  | 
|  | #define UB960_MIN_AEQ_STROBE_POS -7 | 
|  | #define UB960_MAX_AEQ_STROBE_POS  7 | 
|  |  | 
|  | #define UB960_MANUAL_STROBE_EXTRA_DELAY 6 | 
|  |  | 
|  | #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY) | 
|  | #define UB960_MAX_MANUAL_STROBE_POS  (7 + UB960_MANUAL_STROBE_EXTRA_DELAY) | 
|  | #define UB960_NUM_MANUAL_STROBE_POS  (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1) | 
|  |  | 
|  | #define UB960_MIN_EQ_LEVEL  0 | 
|  | #define UB960_MAX_EQ_LEVEL  14 | 
|  | #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1) | 
|  |  | 
|  | struct ub960_hw_data { | 
|  | const char *model; | 
|  | u8 num_rxports; | 
|  | u8 num_txports; | 
|  | bool is_ub9702; | 
|  | bool is_fpdlink4; | 
|  | }; | 
|  |  | 
|  | enum ub960_rxport_mode { | 
|  | RXPORT_MODE_RAW10 = 0, | 
|  | RXPORT_MODE_RAW12_HF = 1, | 
|  | RXPORT_MODE_RAW12_LF = 2, | 
|  | RXPORT_MODE_CSI2_SYNC = 3, | 
|  | RXPORT_MODE_CSI2_NONSYNC = 4, | 
|  | RXPORT_MODE_LAST = RXPORT_MODE_CSI2_NONSYNC, | 
|  | }; | 
|  |  | 
|  | enum ub960_rxport_cdr { | 
|  | RXPORT_CDR_FPD3 = 0, | 
|  | RXPORT_CDR_FPD4 = 1, | 
|  | RXPORT_CDR_LAST = RXPORT_CDR_FPD4, | 
|  | }; | 
|  |  | 
|  | struct ub960_rxport { | 
|  | struct ub960_data      *priv; | 
|  | u8                      nport;	/* RX port number, and index in priv->rxport[] */ | 
|  |  | 
|  | struct { | 
|  | struct v4l2_subdev *sd; | 
|  | u16 pad; | 
|  | struct fwnode_handle *ep_fwnode; | 
|  | } source; | 
|  |  | 
|  | /* Serializer */ | 
|  | struct { | 
|  | struct fwnode_handle *fwnode; | 
|  | struct i2c_client *client; | 
|  | unsigned short alias; /* I2C alias (lower 7 bits) */ | 
|  | short addr; /* Local I2C address (lower 7 bits) */ | 
|  | struct ds90ub9xx_platform_data pdata; | 
|  | struct regmap *regmap; | 
|  | } ser; | 
|  |  | 
|  | enum ub960_rxport_mode  rx_mode; | 
|  | enum ub960_rxport_cdr	cdr_mode; | 
|  |  | 
|  | u8			lv_fv_pol;	/* LV and FV polarities */ | 
|  |  | 
|  | struct regulator	*vpoc; | 
|  |  | 
|  | /* EQ settings */ | 
|  | struct { | 
|  | bool manual_eq; | 
|  |  | 
|  | s8 strobe_pos; | 
|  |  | 
|  | union { | 
|  | struct { | 
|  | u8 eq_level_min; | 
|  | u8 eq_level_max; | 
|  | } aeq; | 
|  |  | 
|  | struct { | 
|  | u8 eq_level; | 
|  | } manual; | 
|  | }; | 
|  | } eq; | 
|  |  | 
|  | /* lock for aliased_addrs and associated registers */ | 
|  | struct mutex aliased_addrs_lock; | 
|  | u16 aliased_addrs[UB960_MAX_PORT_ALIASES]; | 
|  | }; | 
|  |  | 
|  | struct ub960_asd { | 
|  | struct v4l2_async_connection base; | 
|  | struct ub960_rxport *rxport; | 
|  | }; | 
|  |  | 
|  | static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd) | 
|  | { | 
|  | return container_of(asd, struct ub960_asd, base); | 
|  | } | 
|  |  | 
|  | struct ub960_txport { | 
|  | struct ub960_data      *priv; | 
|  | u8                      nport;	/* TX port number, and index in priv->txport[] */ | 
|  |  | 
|  | u32 num_data_lanes; | 
|  | bool non_continous_clk; | 
|  | }; | 
|  |  | 
|  | struct ub960_data { | 
|  | const struct ub960_hw_data	*hw_data; | 
|  | struct i2c_client	*client; /* for shared local registers */ | 
|  | struct regmap		*regmap; | 
|  |  | 
|  | /* lock for register access */ | 
|  | struct mutex		reg_lock; | 
|  |  | 
|  | struct clk		*refclk; | 
|  |  | 
|  | struct regulator	*vddio; | 
|  |  | 
|  | struct gpio_desc	*pd_gpio; | 
|  | struct delayed_work	poll_work; | 
|  | struct ub960_rxport	*rxports[UB960_MAX_RX_NPORTS]; | 
|  | struct ub960_txport	*txports[UB960_MAX_TX_NPORTS]; | 
|  |  | 
|  | struct v4l2_subdev	sd; | 
|  | struct media_pad	pads[UB960_MAX_NPORTS]; | 
|  |  | 
|  | struct v4l2_ctrl_handler   ctrl_handler; | 
|  | struct v4l2_async_notifier notifier; | 
|  |  | 
|  | u32 tx_data_rate;		/* Nominal data rate (Gb/s) */ | 
|  | s64 tx_link_freq[1]; | 
|  |  | 
|  | struct i2c_atr *atr; | 
|  |  | 
|  | struct { | 
|  | u8 rxport; | 
|  | u8 txport; | 
|  | u8 indirect_target; | 
|  | } reg_current; | 
|  |  | 
|  | bool streaming; | 
|  |  | 
|  | u8 stored_fwd_ctl; | 
|  |  | 
|  | u64 stream_enable_mask[UB960_MAX_NPORTS]; | 
|  |  | 
|  | /* These are common to all ports */ | 
|  | struct { | 
|  | bool manual; | 
|  |  | 
|  | s8 min; | 
|  | s8 max; | 
|  | } strobe; | 
|  | }; | 
|  |  | 
|  | static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd) | 
|  | { | 
|  | return container_of(sd, struct ub960_data, sd); | 
|  | } | 
|  |  | 
|  | static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad) | 
|  | { | 
|  | return pad < priv->hw_data->num_rxports; | 
|  | } | 
|  |  | 
|  | static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad) | 
|  | { | 
|  | return pad >= priv->hw_data->num_rxports; | 
|  | } | 
|  |  | 
|  | static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad) | 
|  | { | 
|  | if (ub960_pad_is_sink(priv, pad)) | 
|  | return pad; | 
|  | else | 
|  | return pad - priv->hw_data->num_rxports; | 
|  | } | 
|  |  | 
|  | struct ub960_format_info { | 
|  | u32 code; | 
|  | u32 bpp; | 
|  | u8 datatype; | 
|  | bool meta; | 
|  | }; | 
|  |  | 
|  | static const struct ub960_format_info ub960_formats[] = { | 
|  | { .code = MEDIA_BUS_FMT_RGB888_1X24, .bpp = 24, .datatype = MIPI_CSI2_DT_RGB888, }, | 
|  |  | 
|  | { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, }, | 
|  | { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, }, | 
|  | { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, }, | 
|  | { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, }, | 
|  |  | 
|  | { .code = MEDIA_BUS_FMT_SBGGR8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, }, | 
|  | { .code = MEDIA_BUS_FMT_SGBRG8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, }, | 
|  | { .code = MEDIA_BUS_FMT_SGRBG8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, }, | 
|  | { .code = MEDIA_BUS_FMT_SRGGB8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, }, | 
|  |  | 
|  | { .code = MEDIA_BUS_FMT_SBGGR10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, }, | 
|  | { .code = MEDIA_BUS_FMT_SGBRG10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, }, | 
|  | { .code = MEDIA_BUS_FMT_SGRBG10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, }, | 
|  | { .code = MEDIA_BUS_FMT_SRGGB10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, }, | 
|  |  | 
|  | { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, }, | 
|  | { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, }, | 
|  | { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, }, | 
|  | { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, }, | 
|  | }; | 
|  |  | 
|  | static const struct ub960_format_info *ub960_find_format(u32 code) | 
|  | { | 
|  | unsigned int i; | 
|  |  | 
|  | for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) { | 
|  | if (ub960_formats[i].code == code) | 
|  | return &ub960_formats[i]; | 
|  | } | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | struct ub960_rxport_iter { | 
|  | unsigned int nport; | 
|  | struct ub960_rxport *rxport; | 
|  | }; | 
|  |  | 
|  | enum ub960_iter_flags { | 
|  | UB960_ITER_ACTIVE_ONLY = BIT(0), | 
|  | UB960_ITER_FPD4_ONLY = BIT(1), | 
|  | }; | 
|  |  | 
|  | static struct ub960_rxport_iter ub960_iter_rxport(struct ub960_data *priv, | 
|  | struct ub960_rxport_iter it, | 
|  | enum ub960_iter_flags flags) | 
|  | { | 
|  | for (; it.nport < priv->hw_data->num_rxports; it.nport++) { | 
|  | it.rxport = priv->rxports[it.nport]; | 
|  |  | 
|  | if ((flags & UB960_ITER_ACTIVE_ONLY) && !it.rxport) | 
|  | continue; | 
|  |  | 
|  | if ((flags & UB960_ITER_FPD4_ONLY) && | 
|  | it.rxport->cdr_mode != RXPORT_CDR_FPD4) | 
|  | continue; | 
|  |  | 
|  | return it; | 
|  | } | 
|  |  | 
|  | it.rxport = NULL; | 
|  |  | 
|  | return it; | 
|  | } | 
|  |  | 
|  | #define for_each_rxport(priv, it)                                             \ | 
|  | for (struct ub960_rxport_iter it =                                    \ | 
|  | ub960_iter_rxport(priv, (struct ub960_rxport_iter){ 0 }, \ | 
|  | 0);                                    \ | 
|  | it.nport < (priv)->hw_data->num_rxports;                         \ | 
|  | it.nport++, it = ub960_iter_rxport(priv, it, 0)) | 
|  |  | 
|  | #define for_each_active_rxport(priv, it)                                      \ | 
|  | for (struct ub960_rxport_iter it =                                    \ | 
|  | ub960_iter_rxport(priv, (struct ub960_rxport_iter){ 0 }, \ | 
|  | UB960_ITER_ACTIVE_ONLY);               \ | 
|  | it.nport < (priv)->hw_data->num_rxports;                         \ | 
|  | it.nport++, it = ub960_iter_rxport(priv, it,                     \ | 
|  | UB960_ITER_ACTIVE_ONLY)) | 
|  |  | 
|  | #define for_each_active_rxport_fpd4(priv, it)                                 \ | 
|  | for (struct ub960_rxport_iter it =                                    \ | 
|  | ub960_iter_rxport(priv, (struct ub960_rxport_iter){ 0 }, \ | 
|  | UB960_ITER_ACTIVE_ONLY |               \ | 
|  | UB960_ITER_FPD4_ONLY);         \ | 
|  | it.nport < (priv)->hw_data->num_rxports;                         \ | 
|  | it.nport++, it = ub960_iter_rxport(priv, it,                     \ | 
|  | UB960_ITER_ACTIVE_ONLY |      \ | 
|  | UB960_ITER_FPD4_ONLY)) | 
|  |  | 
|  | /* ----------------------------------------------------------------------------- | 
|  | * Basic device access | 
|  | */ | 
|  |  | 
|  | static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int v; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = regmap_read(priv->regmap, reg, &v); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | *val = v; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_write(struct ub960_data *priv, u8 reg, u8 val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = regmap_write(priv->regmap, reg, val); | 
|  | if (ret) | 
|  | dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  |  | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val, | 
|  | int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = regmap_update_bits(priv->regmap, reg, mask, val); | 
|  | if (ret) | 
|  | dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  |  | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | __be16 __v; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v)); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | *val = be16_to_cpu(__v); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_select(struct ub960_data *priv, u8 nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | lockdep_assert_held(&priv->reg_lock); | 
|  |  | 
|  | if (priv->reg_current.rxport == nport) | 
|  | return 0; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL, | 
|  | (nport << 4) | BIT(nport)); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__, | 
|  | nport, ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | priv->reg_current.rxport = nport; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u8 *val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int v; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_rxport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_read(priv->regmap, reg, &v); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | *val = v; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u8 val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_rxport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, reg, val); | 
|  | if (ret) | 
|  | dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u8 mask, u8 val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_rxport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_update_bits(priv->regmap, reg, mask, val); | 
|  | if (ret) | 
|  | dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u16 *val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | __be16 __v; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_rxport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v)); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | *val = be16_to_cpu(__v); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_txport_select(struct ub960_data *priv, u8 nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | lockdep_assert_held(&priv->reg_lock); | 
|  |  | 
|  | if (priv->reg_current.txport == nport) | 
|  | return 0; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL, | 
|  | (nport << 4) | BIT(nport)); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__, | 
|  | nport, ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | priv->reg_current.txport = nport; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u8 *val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int v; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_txport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_read(priv->regmap, reg, &v); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | *val = v; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u8 val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_txport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, reg, val); | 
|  | if (ret) | 
|  | dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg, | 
|  | u8 mask, u8 val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_txport_select(priv, nport); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_update_bits(priv->regmap, reg, mask, val); | 
|  | if (ret) | 
|  | dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n", | 
|  | __func__, reg, ret); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | lockdep_assert_held(&priv->reg_lock); | 
|  |  | 
|  | if (priv->reg_current.indirect_target == block) | 
|  | return 0; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2); | 
|  | if (ret) { | 
|  | dev_err(dev, "%s: cannot select indirect target %u (%d)!\n", | 
|  | __func__, block, ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | priv->reg_current.indirect_target = block; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val, | 
|  | int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int v; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_select_ind_reg_block(priv, block); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n", | 
|  | block, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n", | 
|  | block, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | *val = v; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val, | 
|  | int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_select_ind_reg_block(priv, block); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n", | 
|  | block, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n", | 
|  | block, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg, | 
|  | u8 mask, u8 val, int *err) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = ub960_select_ind_reg_block(priv, block); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n", | 
|  | block, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask, | 
|  | val); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n", | 
|  | block, reg, ret); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_reset(struct ub960_data *priv, bool reset_regs) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int v; | 
|  | int ret; | 
|  | u8 bit; | 
|  |  | 
|  | bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 : | 
|  | UB960_SR_RESET_DIGITAL_RESET0; | 
|  |  | 
|  | ret = ub960_write(priv, UB960_SR_RESET, bit, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | mutex_lock(&priv->reg_lock); | 
|  |  | 
|  | ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v, | 
|  | (v & bit) == 0, 2000, 100000); | 
|  |  | 
|  | mutex_unlock(&priv->reg_lock); | 
|  |  | 
|  | if (ret) | 
|  | dev_err(dev, "reset failed: %d\n", ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* ----------------------------------------------------------------------------- | 
|  | * I2C-ATR (address translator) | 
|  | */ | 
|  |  | 
|  | static int ub960_atr_attach_addr(struct i2c_atr *atr, u32 chan_id, | 
|  | u16 addr, u16 alias) | 
|  | { | 
|  | struct ub960_data *priv = i2c_atr_get_driver_data(atr); | 
|  | struct ub960_rxport *rxport = priv->rxports[chan_id]; | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int reg_idx; | 
|  | int ret = 0; | 
|  |  | 
|  | guard(mutex)(&rxport->aliased_addrs_lock); | 
|  |  | 
|  | for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_addrs); reg_idx++) { | 
|  | if (!rxport->aliased_addrs[reg_idx]) | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (reg_idx == ARRAY_SIZE(rxport->aliased_addrs)) { | 
|  | dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport); | 
|  | return -EADDRNOTAVAIL; | 
|  | } | 
|  |  | 
|  | rxport->aliased_addrs[reg_idx] = addr; | 
|  |  | 
|  | ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx), | 
|  | addr << 1, &ret); | 
|  | ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), | 
|  | alias << 1, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n", | 
|  | rxport->nport, addr, alias, reg_idx); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ub960_atr_detach_addr(struct i2c_atr *atr, u32 chan_id, | 
|  | u16 addr) | 
|  | { | 
|  | struct ub960_data *priv = i2c_atr_get_driver_data(atr); | 
|  | struct ub960_rxport *rxport = priv->rxports[chan_id]; | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int reg_idx; | 
|  | int ret; | 
|  |  | 
|  | guard(mutex)(&rxport->aliased_addrs_lock); | 
|  |  | 
|  | for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_addrs); reg_idx++) { | 
|  | if (rxport->aliased_addrs[reg_idx] == addr) | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (reg_idx == ARRAY_SIZE(rxport->aliased_addrs)) { | 
|  | dev_err(dev, "rx%u: client 0x%02x is not mapped!\n", | 
|  | rxport->nport, addr); | 
|  | return; | 
|  | } | 
|  |  | 
|  | rxport->aliased_addrs[reg_idx] = 0; | 
|  |  | 
|  | ret = ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), | 
|  | 0, NULL); | 
|  | if (ret) { | 
|  | dev_err(dev, "rx%u: unable to fully unmap client 0x%02x: %d\n", | 
|  | rxport->nport, addr, ret); | 
|  | return; | 
|  | } | 
|  |  | 
|  | dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport, | 
|  | addr, reg_idx); | 
|  | } | 
|  |  | 
|  | static const struct i2c_atr_ops ub960_atr_ops = { | 
|  | .attach_addr = ub960_atr_attach_addr, | 
|  | .detach_addr = ub960_atr_detach_addr, | 
|  | }; | 
|  |  | 
|  | static int ub960_init_atr(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct i2c_adapter *parent_adap = priv->client->adapter; | 
|  |  | 
|  | priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops, | 
|  | priv->hw_data->num_rxports, 0); | 
|  | if (IS_ERR(priv->atr)) | 
|  | return PTR_ERR(priv->atr); | 
|  |  | 
|  | i2c_atr_set_driver_data(priv->atr, priv); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ub960_uninit_atr(struct ub960_data *priv) | 
|  | { | 
|  | i2c_atr_delete(priv->atr); | 
|  | priv->atr = NULL; | 
|  | } | 
|  |  | 
|  | /* ----------------------------------------------------------------------------- | 
|  | * TX ports | 
|  | */ | 
|  |  | 
|  | static int ub960_parse_dt_txport(struct ub960_data *priv, | 
|  | struct fwnode_handle *ep_fwnode, | 
|  | u8 nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct v4l2_fwnode_endpoint vep = {}; | 
|  | struct ub960_txport *txport; | 
|  | int ret; | 
|  |  | 
|  | txport = kzalloc(sizeof(*txport), GFP_KERNEL); | 
|  | if (!txport) | 
|  | return -ENOMEM; | 
|  |  | 
|  | txport->priv = priv; | 
|  | txport->nport = nport; | 
|  |  | 
|  | vep.bus_type = V4L2_MBUS_CSI2_DPHY; | 
|  | ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep); | 
|  | if (ret) { | 
|  | dev_err(dev, "tx%u: failed to parse endpoint data\n", nport); | 
|  | goto err_free_txport; | 
|  | } | 
|  |  | 
|  | txport->non_continous_clk = vep.bus.mipi_csi2.flags & | 
|  | V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK; | 
|  |  | 
|  | txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes; | 
|  |  | 
|  | if (vep.nr_of_link_frequencies != 1) { | 
|  | ret = -EINVAL; | 
|  | goto err_free_vep; | 
|  | } | 
|  |  | 
|  | priv->tx_link_freq[0] = vep.link_frequencies[0]; | 
|  | priv->tx_data_rate = priv->tx_link_freq[0] * 2; | 
|  |  | 
|  | if (priv->tx_data_rate != MHZ(1600) && | 
|  | priv->tx_data_rate != MHZ(1200) && | 
|  | priv->tx_data_rate != MHZ(800) && | 
|  | priv->tx_data_rate != MHZ(400)) { | 
|  | dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport); | 
|  | ret = -EINVAL; | 
|  | goto err_free_vep; | 
|  | } | 
|  |  | 
|  | v4l2_fwnode_endpoint_free(&vep); | 
|  |  | 
|  | priv->txports[nport] = txport; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_free_vep: | 
|  | v4l2_fwnode_endpoint_free(&vep); | 
|  | err_free_txport: | 
|  | kfree(txport); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int  ub960_csi_handle_events(struct ub960_data *priv, u8 nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 csi_tx_isr; | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR) | 
|  | dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport); | 
|  |  | 
|  | if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR) | 
|  | dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* ----------------------------------------------------------------------------- | 
|  | * RX ports | 
|  | */ | 
|  |  | 
|  | static int ub960_rxport_enable_vpocs(struct ub960_data *priv) | 
|  | { | 
|  | unsigned int failed_nport; | 
|  | int ret; | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | if (!it.rxport->vpoc) | 
|  | continue; | 
|  |  | 
|  | ret = regulator_enable(it.rxport->vpoc); | 
|  | if (ret) { | 
|  | failed_nport = it.nport; | 
|  | goto err_disable_vpocs; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_disable_vpocs: | 
|  | while (failed_nport--) { | 
|  | struct ub960_rxport *rxport = priv->rxports[failed_nport]; | 
|  |  | 
|  | if (!rxport || !rxport->vpoc) | 
|  | continue; | 
|  |  | 
|  | regulator_disable(rxport->vpoc); | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void ub960_rxport_disable_vpocs(struct ub960_data *priv) | 
|  | { | 
|  | for_each_active_rxport(priv, it) { | 
|  | if (!it.rxport->vpoc) | 
|  | continue; | 
|  |  | 
|  | regulator_disable(it.rxport->vpoc); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_clear_errors(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | int ret = 0; | 
|  | u8 v; | 
|  |  | 
|  | ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v, &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v, &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v, &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v, &ret); | 
|  |  | 
|  | ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v, &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v, &ret); | 
|  |  | 
|  | ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_clear_rx_errors(struct ub960_data *priv) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | for_each_rxport(priv, it) { | 
|  | ret = ub960_rxport_clear_errors(priv, it.nport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_get_strobe_pos(struct ub960_data *priv, | 
|  | unsigned int nport, s8 *strobe_pos) | 
|  | { | 
|  | u8 v; | 
|  | u8 clk_delay, data_delay; | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB960_IR_RX_ANA_STROBE_SET_CLK, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ? | 
|  | 0 : UB960_MANUAL_STROBE_EXTRA_DELAY; | 
|  |  | 
|  | ret = ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB960_IR_RX_ANA_STROBE_SET_DATA, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ? | 
|  | 0 : UB960_MANUAL_STROBE_EXTRA_DELAY; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK; | 
|  |  | 
|  | *strobe_pos = data_delay - clk_delay; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_set_strobe_pos(struct ub960_data *priv, | 
|  | unsigned int nport, s8 strobe_pos) | 
|  | { | 
|  | u8 clk_delay, data_delay; | 
|  | int ret = 0; | 
|  |  | 
|  | clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY; | 
|  | data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY; | 
|  |  | 
|  | if (strobe_pos < UB960_MIN_AEQ_STROBE_POS) | 
|  | clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY; | 
|  | else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS) | 
|  | data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY; | 
|  | else if (strobe_pos < 0) | 
|  | clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY; | 
|  | else if (strobe_pos > 0) | 
|  | data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY; | 
|  |  | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay, &ret); | 
|  |  | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_set_strobe_range(struct ub960_data *priv, s8 strobe_min, | 
|  | s8 strobe_max) | 
|  | { | 
|  | /* Convert the signed strobe pos to positive zero based value */ | 
|  | strobe_min -= UB960_MIN_AEQ_STROBE_POS; | 
|  | strobe_max -= UB960_MIN_AEQ_STROBE_POS; | 
|  |  | 
|  | return ub960_write(priv, UB960_XR_SFILTER_CFG, | 
|  | ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) | | 
|  | ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT), | 
|  | NULL); | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_get_eq_level(struct ub960_data *priv, | 
|  | unsigned int nport, u8 *eq_level) | 
|  | { | 
|  | int ret; | 
|  | u8 v; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | *eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) + | 
|  | (v & UB960_RR_AEQ_STATUS_STATUS_2); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_set_eq_level(struct ub960_data *priv, | 
|  | unsigned int nport, u8 eq_level) | 
|  | { | 
|  | u8 eq_stage_1_select_value, eq_stage_2_select_value; | 
|  | const unsigned int eq_stage_max = 7; | 
|  | int ret; | 
|  | u8 v; | 
|  |  | 
|  | if (eq_level <= eq_stage_max) { | 
|  | eq_stage_1_select_value = eq_level; | 
|  | eq_stage_2_select_value = 0; | 
|  | } else { | 
|  | eq_stage_1_select_value = eq_stage_max; | 
|  | eq_stage_2_select_value = eq_level - eq_stage_max; | 
|  | } | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK | | 
|  | UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK); | 
|  | v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT; | 
|  | v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT; | 
|  | v |= UB960_RR_AEQ_BYPASS_ENABLE; | 
|  |  | 
|  | ret = ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_set_eq_range(struct ub960_data *priv, | 
|  | unsigned int nport, u8 eq_min, u8 eq_max) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX, | 
|  | (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) | | 
|  | (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT), | 
|  | &ret); | 
|  |  | 
|  | /* Enable AEQ min setting */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2, | 
|  | UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR, | 
|  | UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport) | 
|  | { | 
|  | struct ub960_rxport *rxport = priv->rxports[nport]; | 
|  | int ret; | 
|  |  | 
|  | /* We also set common settings here. Should be moved elsewhere. */ | 
|  |  | 
|  | if (priv->strobe.manual) { | 
|  | /* Disable AEQ_SFILTER_EN */ | 
|  | ret = ub960_update_bits(priv, UB960_XR_AEQ_CTL1, | 
|  | UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } else { | 
|  | /* Enable SFILTER and error control */ | 
|  | ret = ub960_write(priv, UB960_XR_AEQ_CTL1, | 
|  | UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK | | 
|  | UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, | 
|  | NULL); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Set AEQ strobe range */ | 
|  | ret = ub960_rxport_set_strobe_range(priv, priv->strobe.min, | 
|  | priv->strobe.max); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* The rest are port specific */ | 
|  |  | 
|  | if (priv->strobe.manual) | 
|  | ret = ub960_rxport_set_strobe_pos(priv, nport, | 
|  | rxport->eq.strobe_pos); | 
|  | else | 
|  | ret = ub960_rxport_set_strobe_pos(priv, nport, 0); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (rxport->eq.manual_eq) { | 
|  | ret = ub960_rxport_set_eq_level(priv, nport, | 
|  | rxport->eq.manual.eq_level); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Enable AEQ Bypass */ | 
|  | ret = ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS, | 
|  | UB960_RR_AEQ_BYPASS_ENABLE, | 
|  | UB960_RR_AEQ_BYPASS_ENABLE, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } else { | 
|  | ret = ub960_rxport_set_eq_range(priv, nport, | 
|  | rxport->eq.aeq.eq_level_min, | 
|  | rxport->eq.aeq.eq_level_max); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Disable AEQ Bypass */ | 
|  | ret = ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS, | 
|  | UB960_RR_AEQ_BYPASS_ENABLE, 0, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport, | 
|  | bool *ok) | 
|  | { | 
|  | u8 rx_port_sts1, rx_port_sts2; | 
|  | u16 parity_errors; | 
|  | u8 csi_rx_sts; | 
|  | u8 csi_err_cnt; | 
|  | u8 bcc_sts; | 
|  | int ret; | 
|  | bool errors; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, | 
|  | &rx_port_sts1, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) { | 
|  | *ok = false; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, | 
|  | &rx_port_sts2, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, | 
|  | &csi_err_cnt, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, | 
|  | &parity_errors, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) || | 
|  | (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) || | 
|  | (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) || | 
|  | (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt || | 
|  | parity_errors; | 
|  |  | 
|  | *ok = !errors; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_lockup_wa_ub9702(struct ub960_data *priv) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | /* Toggle PI_MODE to avoid possible FPD RX lockup */ | 
|  |  | 
|  | ret = ub960_update_bits(priv, UB9702_RR_CHANNEL_MODE, GENMASK(4, 3), | 
|  | 2 << 3, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | usleep_range(1000, 5000); | 
|  |  | 
|  | return ub960_update_bits(priv, UB9702_RR_CHANNEL_MODE, GENMASK(4, 3), | 
|  | 0, NULL); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Wait for the RX ports to lock, have no errors and have stable strobe position | 
|  | * and EQ level. | 
|  | */ | 
|  | static int ub960_rxport_wait_locks(struct ub960_data *priv, | 
|  | unsigned long port_mask, | 
|  | unsigned int *lock_mask) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned long timeout; | 
|  | unsigned int link_ok_mask; | 
|  | unsigned int missing; | 
|  | unsigned int loops; | 
|  | u8 nport; | 
|  | int ret; | 
|  |  | 
|  | if (port_mask == 0) { | 
|  | if (lock_mask) | 
|  | *lock_mask = 0; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (port_mask >= BIT(priv->hw_data->num_rxports)) | 
|  | return -EINVAL; | 
|  |  | 
|  | timeout = jiffies + msecs_to_jiffies(1000); | 
|  | loops = 0; | 
|  | link_ok_mask = 0; | 
|  |  | 
|  | while (time_before(jiffies, timeout)) { | 
|  | bool fpd4_wa = false; | 
|  | missing = 0; | 
|  |  | 
|  | for_each_set_bit(nport, &port_mask, | 
|  | priv->hw_data->num_rxports) { | 
|  | struct ub960_rxport *rxport = priv->rxports[nport]; | 
|  | bool ok; | 
|  |  | 
|  | if (!rxport) | 
|  | continue; | 
|  |  | 
|  | ret = ub960_rxport_link_ok(priv, nport, &ok); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (!ok && rxport->cdr_mode == RXPORT_CDR_FPD4) | 
|  | fpd4_wa = true; | 
|  |  | 
|  | /* | 
|  | * We want the link to be ok for two consecutive loops, | 
|  | * as a link could get established just before our test | 
|  | * and drop soon after. | 
|  | */ | 
|  | if (!ok || !(link_ok_mask & BIT(nport))) | 
|  | missing++; | 
|  |  | 
|  | if (ok) | 
|  | link_ok_mask |= BIT(nport); | 
|  | else | 
|  | link_ok_mask &= ~BIT(nport); | 
|  | } | 
|  |  | 
|  | loops++; | 
|  |  | 
|  | if (missing == 0) | 
|  | break; | 
|  |  | 
|  | if (fpd4_wa) { | 
|  | ret = ub960_rxport_lockup_wa_ub9702(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * The sleep time of 10 ms was found by testing to give a lock | 
|  | * with a few iterations. It can be decreased if on some setups | 
|  | * the lock can be achieved much faster. | 
|  | */ | 
|  | fsleep(10 * USEC_PER_MSEC); | 
|  | } | 
|  |  | 
|  | if (lock_mask) | 
|  | *lock_mask = link_ok_mask; | 
|  |  | 
|  | dev_dbg(dev, "Wait locks done in %u loops\n", loops); | 
|  | for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) { | 
|  | struct ub960_rxport *rxport = priv->rxports[nport]; | 
|  | s8 strobe_pos, eq_level; | 
|  | u16 v; | 
|  |  | 
|  | if (!rxport) | 
|  | continue; | 
|  |  | 
|  | if (!(link_ok_mask & BIT(nport))) { | 
|  | dev_dbg(dev, "\trx%u: not locked\n", nport); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, | 
|  | &v, NULL); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (priv->hw_data->is_ub9702) { | 
|  | dev_dbg(dev, "\trx%u: locked, freq %llu Hz\n", | 
|  | nport, ((u64)v * HZ_PER_MHZ) >> 8); | 
|  | } else { | 
|  | ret = ub960_rxport_get_strobe_pos(priv, nport, | 
|  | &strobe_pos); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_rxport_get_eq_level(priv, nport, &eq_level); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, | 
|  | "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n", | 
|  | nport, strobe_pos, eq_level, | 
|  | ((u64)v * HZ_PER_MHZ) >> 8); | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv, | 
|  | struct ub960_rxport *rxport) | 
|  | { | 
|  | unsigned int mult; | 
|  | unsigned int div; | 
|  |  | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_RAW10: | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | mult = 1; | 
|  | div = 10; | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | mult = 2; | 
|  | div = 1; | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_NONSYNC: | 
|  | mult = 2; | 
|  | div = 5; | 
|  | break; | 
|  |  | 
|  | default: | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | return clk_get_rate(priv->refclk) * mult / div; | 
|  | } | 
|  |  | 
|  | static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv, | 
|  | struct ub960_rxport *rxport) | 
|  | { | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_RAW10: | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | return 2359400; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | return 47187500; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_NONSYNC: | 
|  | return 9437500; | 
|  |  | 
|  | default: | 
|  | return 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_serializer_write(struct ub960_rxport *rxport, u8 reg, | 
|  | u8 val, int *err) | 
|  | { | 
|  | struct ub960_data *priv = rxport->priv; | 
|  | struct device *dev = &priv->client->dev; | 
|  | union i2c_smbus_data data; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | data.byte = val; | 
|  |  | 
|  | ret = i2c_smbus_xfer(priv->client->adapter, rxport->ser.alias, 0, | 
|  | I2C_SMBUS_WRITE, reg, I2C_SMBUS_BYTE_DATA, &data); | 
|  | if (ret) | 
|  | dev_err(dev, | 
|  | "rx%u: cannot write serializer register 0x%02x (%d)!\n", | 
|  | rxport->nport, reg, ret); | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_serializer_read(struct ub960_rxport *rxport, u8 reg, | 
|  | u8 *val, int *err) | 
|  | { | 
|  | struct ub960_data *priv = rxport->priv; | 
|  | struct device *dev = &priv->client->dev; | 
|  | union i2c_smbus_data data = { 0 }; | 
|  | int ret; | 
|  |  | 
|  | if (err && *err) | 
|  | return *err; | 
|  |  | 
|  | ret = i2c_smbus_xfer(priv->client->adapter, rxport->ser.alias, | 
|  | priv->client->flags, I2C_SMBUS_READ, reg, | 
|  | I2C_SMBUS_BYTE_DATA, &data); | 
|  | if (ret) | 
|  | dev_err(dev, | 
|  | "rx%u: cannot read serializer register 0x%02x (%d)!\n", | 
|  | rxport->nport, reg, ret); | 
|  | else | 
|  | *val = data.byte; | 
|  |  | 
|  | if (ret && err) | 
|  | *err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_serializer_temp_ramp(struct ub960_rxport *rxport) | 
|  | { | 
|  | struct ub960_data *priv = rxport->priv; | 
|  | short temp_dynamic_offset[] = {-1, -1, 0, 0, 1, 1, 1, 3}; | 
|  | u8 temp_dynamic_cfg; | 
|  | u8 nport = rxport->nport; | 
|  | u8 ser_temp_code; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Configure temp ramp only on UB953 */ | 
|  | if (!fwnode_device_is_compatible(rxport->ser.fwnode, "ti,ds90ub953-q1")) | 
|  | return 0; | 
|  |  | 
|  | /* Read current serializer die temperature */ | 
|  | ub960_rxport_read(priv, nport, UB960_RR_SENSOR_STS_2, &ser_temp_code, | 
|  | &ret); | 
|  |  | 
|  | /* Enable I2C passthrough on back channel */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Select indirect page for analog regs on the serializer */ | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_IND_ACC_CTL, | 
|  | UB953_IND_TARGET_ANALOG << 2, &ret); | 
|  |  | 
|  | /* Set temperature ramp dynamic and static config */ | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_IND_ACC_ADDR, | 
|  | UB953_IND_ANA_TEMP_DYNAMIC_CFG, &ret); | 
|  | ub960_rxport_serializer_read(rxport, UB953_REG_IND_ACC_DATA, | 
|  | &temp_dynamic_cfg, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | temp_dynamic_cfg |= UB953_IND_ANA_TEMP_DYNAMIC_CFG_OV; | 
|  | temp_dynamic_cfg += temp_dynamic_offset[ser_temp_code]; | 
|  |  | 
|  | /* Update temp static config */ | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_IND_ACC_ADDR, | 
|  | UB953_IND_ANA_TEMP_STATIC_CFG, &ret); | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_IND_ACC_DATA, | 
|  | UB953_IND_ANA_TEMP_STATIC_CFG_MASK, &ret); | 
|  |  | 
|  | /* Update temperature ramp dynamic config */ | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_IND_ACC_ADDR, | 
|  | UB953_IND_ANA_TEMP_DYNAMIC_CFG, &ret); | 
|  |  | 
|  | /* Enable I2C auto ack on BC before we set dynamic cfg and reset */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_AUTO_ACK_ALL, | 
|  | UB960_RR_BCC_CONFIG_AUTO_ACK_ALL, &ret); | 
|  |  | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_IND_ACC_DATA, | 
|  | temp_dynamic_cfg, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Soft reset to apply PLL updates */ | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_RESET_CTL, | 
|  | UB953_REG_RESET_CTL_DIGITAL_RESET_0, | 
|  | &ret); | 
|  | msleep(20); | 
|  |  | 
|  | /* Disable I2C passthrough and auto-ack on BC */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH | | 
|  | UB960_RR_BCC_CONFIG_AUTO_ACK_ALL, | 
|  | 0x0, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_bc_ser_config(struct ub960_rxport *rxport) | 
|  | { | 
|  | struct ub960_data *priv = rxport->priv; | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 nport = rxport->nport; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Skip port if serializer's address is not known */ | 
|  | if (rxport->ser.addr < 0) { | 
|  | dev_dbg(dev, | 
|  | "rx%u: serializer address missing, skip configuration\n", | 
|  | nport); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Note: the code here probably only works for CSI-2 serializers in | 
|  | * sync mode. To support other serializers the BC related configuration | 
|  | * should be done before calling this function. | 
|  | */ | 
|  |  | 
|  | /* Enable I2C passthrough and auto-ack on BC */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH | | 
|  | UB960_RR_BCC_CONFIG_AUTO_ACK_ALL, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH | | 
|  | UB960_RR_BCC_CONFIG_AUTO_ACK_ALL, | 
|  | &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Disable BC alternate mode auto detect */ | 
|  | ub960_rxport_serializer_write(rxport, UB971_ENH_BC_CHK, 0x02, &ret); | 
|  | /* Decrease link detect timer */ | 
|  | ub960_rxport_serializer_write(rxport, UB953_REG_BC_CTRL, 0x06, &ret); | 
|  |  | 
|  | /* Disable I2C passthrough and auto-ack on BC */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH | | 
|  | UB960_RR_BCC_CONFIG_AUTO_ACK_ALL, | 
|  | 0x0, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport) | 
|  | { | 
|  | struct ub960_rxport *rxport = priv->rxports[nport]; | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata; | 
|  | struct i2c_board_info ser_info = { | 
|  | .fwnode = rxport->ser.fwnode, | 
|  | .platform_data = ser_pdata, | 
|  | }; | 
|  |  | 
|  | ser_pdata->port = nport; | 
|  | ser_pdata->atr = priv->atr; | 
|  | if (priv->hw_data->is_ub9702) | 
|  | ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport); | 
|  | else | 
|  | ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport); | 
|  |  | 
|  | /* | 
|  | * The serializer is added under the same i2c adapter as the | 
|  | * deserializer. This is not quite right, as the serializer is behind | 
|  | * the FPD-Link. | 
|  | */ | 
|  | ser_info.addr = rxport->ser.alias; | 
|  | rxport->ser.client = | 
|  | i2c_new_client_device(priv->client->adapter, &ser_info); | 
|  | if (IS_ERR(rxport->ser.client)) { | 
|  | dev_err(dev, "rx%u: cannot add %s i2c device", nport, | 
|  | ser_info.type); | 
|  | return PTR_ERR(rxport->ser.client); | 
|  | } | 
|  |  | 
|  | dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n", | 
|  | nport, rxport->ser.client->addr, | 
|  | rxport->ser.client->adapter->nr, rxport->ser.client->addr); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport) | 
|  | { | 
|  | struct ub960_rxport *rxport = priv->rxports[nport]; | 
|  |  | 
|  | i2c_unregister_device(rxport->ser.client); | 
|  | rxport->ser.client = NULL; | 
|  | } | 
|  |  | 
|  | /* Add serializer i2c devices for all initialized ports */ | 
|  | static int ub960_rxport_add_serializers(struct ub960_data *priv) | 
|  | { | 
|  | unsigned int failed_nport; | 
|  | int ret; | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | ret = ub960_rxport_add_serializer(priv, it.nport); | 
|  | if (ret) { | 
|  | failed_nport = it.nport; | 
|  | goto err_remove_sers; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_remove_sers: | 
|  | while (failed_nport--) { | 
|  | struct ub960_rxport *rxport = priv->rxports[failed_nport]; | 
|  |  | 
|  | if (!rxport) | 
|  | continue; | 
|  |  | 
|  | ub960_rxport_remove_serializer(priv, failed_nport); | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void ub960_rxport_remove_serializers(struct ub960_data *priv) | 
|  | { | 
|  | for_each_active_rxport(priv, it) | 
|  | ub960_rxport_remove_serializer(priv, it.nport); | 
|  | } | 
|  |  | 
|  | static int ub960_init_tx_port(struct ub960_data *priv, | 
|  | struct ub960_txport *txport) | 
|  | { | 
|  | unsigned int nport = txport->nport; | 
|  | u8 csi_ctl = 0; | 
|  |  | 
|  | /* | 
|  | * From the datasheet: "initial CSI Skew-Calibration | 
|  | * sequence [...] should be set when operating at 1.6 Gbps" | 
|  | */ | 
|  | if (priv->tx_data_rate == MHZ(1600)) | 
|  | csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN; | 
|  |  | 
|  | csi_ctl |= (4 - txport->num_data_lanes) << 4; | 
|  |  | 
|  | if (!txport->non_continous_clk) | 
|  | csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK; | 
|  |  | 
|  | return ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl, NULL); | 
|  | } | 
|  |  | 
|  | static int ub960_init_tx_ports_ub960(struct ub960_data *priv) | 
|  | { | 
|  | u8 speed_select; | 
|  |  | 
|  | switch (priv->tx_data_rate) { | 
|  | case MHZ(400): | 
|  | speed_select = 3; | 
|  | break; | 
|  | case MHZ(800): | 
|  | speed_select = 2; | 
|  | break; | 
|  | case MHZ(1200): | 
|  | speed_select = 1; | 
|  | break; | 
|  | case MHZ(1600): | 
|  | default: | 
|  | speed_select = 0; | 
|  | break; | 
|  | } | 
|  |  | 
|  | return ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select, NULL); | 
|  | } | 
|  |  | 
|  | static int ub960_init_tx_ports_ub9702(struct ub960_data *priv) | 
|  | { | 
|  | u8 speed_select; | 
|  | u8 ana_pll_div; | 
|  | u8 pll_div; | 
|  | int ret = 0; | 
|  |  | 
|  | switch (priv->tx_data_rate) { | 
|  | case MHZ(400): | 
|  | speed_select = 3; | 
|  | pll_div = 0x10; | 
|  | ana_pll_div = 0xa2; | 
|  | break; | 
|  | case MHZ(800): | 
|  | speed_select = 2; | 
|  | pll_div = 0x10; | 
|  | ana_pll_div = 0x92; | 
|  | break; | 
|  | case MHZ(1200): | 
|  | speed_select = 1; | 
|  | pll_div = 0x18; | 
|  | ana_pll_div = 0x90; | 
|  | break; | 
|  | case MHZ(1500): | 
|  | speed_select = 0; | 
|  | pll_div = 0x0f; | 
|  | ana_pll_div = 0x82; | 
|  | break; | 
|  | case MHZ(1600): | 
|  | default: | 
|  | speed_select = 0; | 
|  | pll_div = 0x10; | 
|  | ana_pll_div = 0x82; | 
|  | break; | 
|  | case MHZ(2500): | 
|  | speed_select = 0x10; | 
|  | pll_div = 0x19; | 
|  | ana_pll_div = 0x80; | 
|  | break; | 
|  | } | 
|  |  | 
|  | ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select, &ret); | 
|  | ub960_write(priv, UB9702_SR_CSI_PLL_DIV, pll_div, &ret); | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, | 
|  | UB9702_IR_CSI_ANA_CSIPLL_REG_1, ana_pll_div, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_init_tx_ports(struct ub960_data *priv) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | if (priv->hw_data->is_ub9702) | 
|  | ret = ub960_init_tx_ports_ub9702(priv); | 
|  | else | 
|  | ret = ub960_init_tx_ports_ub960(priv); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | for (unsigned int nport = 0; nport < priv->hw_data->num_txports; | 
|  | nport++) { | 
|  | struct ub960_txport *txport = priv->txports[nport]; | 
|  |  | 
|  | if (!txport) | 
|  | continue; | 
|  |  | 
|  | ret = ub960_init_tx_port(priv, txport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_init_rx_port_ub960(struct ub960_data *priv, | 
|  | struct ub960_rxport *rxport) | 
|  | { | 
|  | unsigned int nport = rxport->nport; | 
|  | u32 bc_freq_val; | 
|  | int ret = 0; | 
|  |  | 
|  | /* | 
|  | * Back channel frequency select. | 
|  | * Override FREQ_SELECT from the strap. | 
|  | * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1) | 
|  | * 2 - 10 Mbps | 
|  | * 6 - 50 Mbps (DS90UB953-Q1) | 
|  | * | 
|  | * Note that changing this setting will result in some errors on the back | 
|  | * channel for a short period of time. | 
|  | */ | 
|  |  | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_RAW10: | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | bc_freq_val = 0; | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_NONSYNC: | 
|  | bc_freq_val = 2; | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | bc_freq_val = 6; | 
|  | break; | 
|  |  | 
|  | default: | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK, | 
|  | bc_freq_val, &ret); | 
|  |  | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_RAW10: | 
|  | /* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, | 
|  | UB960_RR_PORT_CONFIG_FPD3_MODE_MASK, | 
|  | 0x3, &ret); | 
|  |  | 
|  | /* | 
|  | * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits | 
|  | */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, | 
|  | UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK, | 
|  | 0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT, | 
|  | &ret); | 
|  |  | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | /* Not implemented */ | 
|  | return -EINVAL; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | case RXPORT_MODE_CSI2_NONSYNC: | 
|  | /* CSI-2 Mode (DS90UB953-Q1 compatible) */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3, | 
|  | 0x0, &ret); | 
|  |  | 
|  | break; | 
|  | } | 
|  |  | 
|  | /* LV_POLARITY & FV_POLARITY */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3, | 
|  | rxport->lv_fv_pol, &ret); | 
|  |  | 
|  | /* Enable all interrupt sources from this port */ | 
|  | ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07, &ret); | 
|  | ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f, &ret); | 
|  |  | 
|  | /* Enable I2C_PASS_THROUGH */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, &ret); | 
|  |  | 
|  | /* Enable I2C communication to the serializer via the alias addr */ | 
|  | ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID, | 
|  | rxport->ser.alias << 1, &ret); | 
|  |  | 
|  | /* Configure EQ related settings */ | 
|  | ub960_rxport_config_eq(priv, nport); | 
|  |  | 
|  | /* Enable RX port */ | 
|  | ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport), | 
|  | &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_init_rx_ports_ub960(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int port_lock_mask; | 
|  | unsigned int port_mask; | 
|  | int ret; | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | ret = ub960_init_rx_port_ub960(priv, it.rxport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = ub960_reset(priv, false); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | port_mask = 0; | 
|  |  | 
|  | for_each_active_rxport(priv, it) | 
|  | port_mask |= BIT(it.nport); | 
|  |  | 
|  | ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (port_mask != port_lock_mask) { | 
|  | ret = -EIO; | 
|  | dev_err_probe(dev, ret, "Failed to lock all RX ports\n"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Set temperature ramp on serializer */ | 
|  | for_each_active_rxport(priv, it) { | 
|  | ret = ub960_serializer_temp_ramp(it.rxport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ub960_rxport_update_bits(priv, it.nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, | 
|  | &ret); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Clear any errors caused by switching the RX port settings while | 
|  | * probing. | 
|  | */ | 
|  | ret = ub960_clear_rx_errors(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * UB9702 specific initial RX port configuration | 
|  | */ | 
|  |  | 
|  | static int ub960_turn_off_rxport_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | /* Disable RX port */ | 
|  | ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), 0, &ret); | 
|  |  | 
|  | /* Disable FPD Rx and FPD BC CMR */ | 
|  | ub960_rxport_write(priv, nport, UB9702_RR_RX_CTL_2, 0x1b, &ret); | 
|  |  | 
|  | /* Disable FPD BC Tx */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, BIT(4), 0, | 
|  | &ret); | 
|  |  | 
|  | /* Disable internal RX blocks */ | 
|  | ub960_rxport_write(priv, nport, UB9702_RR_RX_CTL_1, 0x15, &ret); | 
|  |  | 
|  | /* Disable AEQ */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_CFG_2, 0x03, &ret); | 
|  |  | 
|  | /* PI disabled and oDAC disabled */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_CFG_4, 0x09, &ret); | 
|  |  | 
|  | /* AEQ configured for disabled link */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_CFG_1, 0x20, &ret); | 
|  |  | 
|  | /* disable AEQ clock and DFE */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_CFG_3, 0x45, &ret); | 
|  |  | 
|  | /* Powerdown FPD3 CDR */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_FPD3_CDR_CTRL_SEL_5, 0x82, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_set_bc_drv_config_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | u8 fpd_bc_ctl0; | 
|  | u8 fpd_bc_ctl1; | 
|  | u8 fpd_bc_ctl2; | 
|  | int ret = 0; | 
|  |  | 
|  | if (priv->rxports[nport]->cdr_mode == RXPORT_CDR_FPD4) { | 
|  | /* Set FPD PBC drv into FPD IV mode */ | 
|  |  | 
|  | fpd_bc_ctl0 = 0; | 
|  | fpd_bc_ctl1 = 0; | 
|  | fpd_bc_ctl2 = 0; | 
|  | } else { | 
|  | /* Set FPD PBC drv into FPD III mode */ | 
|  |  | 
|  | fpd_bc_ctl0 = 2; | 
|  | fpd_bc_ctl1 = 1; | 
|  | fpd_bc_ctl2 = 5; | 
|  | } | 
|  |  | 
|  | ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_FPD_BC_CTL0, GENMASK(7, 5), | 
|  | fpd_bc_ctl0 << 5, &ret); | 
|  |  | 
|  | ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_FPD_BC_CTL1, BIT(6), | 
|  | fpd_bc_ctl1 << 6, &ret); | 
|  |  | 
|  | ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_FPD_BC_CTL2, GENMASK(6, 3), | 
|  | fpd_bc_ctl2 << 3, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_set_fpd4_sync_mode_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | /* FPD4 Sync Mode */ | 
|  | ub960_rxport_write(priv, nport, UB9702_RR_CHANNEL_MODE, 0x0, &ret); | 
|  |  | 
|  | /* BC_FREQ_SELECT = (PLL_FREQ/3200) Mbps */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK, 6, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_set_bc_drv_config_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Set AEQ timer to 400us/step */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_SYSTEM_INIT_REG0, 0x2f, &ret); | 
|  |  | 
|  | /* Disable FPD4 Auto Recovery */ | 
|  | ub960_update_bits(priv, UB9702_SR_CSI_EXCLUSIVE_FWD2, GENMASK(5, 4), 0, | 
|  | &ret); | 
|  |  | 
|  | /* Enable RX port */ | 
|  | ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport), | 
|  | &ret); | 
|  |  | 
|  | /* Enable FPD4 Auto Recovery */ | 
|  | ub960_update_bits(priv, UB9702_SR_CSI_EXCLUSIVE_FWD2, GENMASK(5, 4), | 
|  | BIT(4), &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_set_fpd4_async_mode_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | /* FPD4 ASync Mode */ | 
|  | ub960_rxport_write(priv, nport, UB9702_RR_CHANNEL_MODE, 0x1, &ret); | 
|  |  | 
|  | /* 10Mbps w/ BC enabled */ | 
|  | /* BC_FREQ_SELECT=(PLL_FREQ/3200) Mbps */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK, 2, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_set_bc_drv_config_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Set AEQ timer to 400us/step */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_SYSTEM_INIT_REG0, 0x2f, &ret); | 
|  |  | 
|  | /* Disable FPD4 Auto Recover */ | 
|  | ub960_update_bits(priv, UB9702_SR_CSI_EXCLUSIVE_FWD2, GENMASK(5, 4), 0, | 
|  | &ret); | 
|  |  | 
|  | /* Enable RX port */ | 
|  | ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport), | 
|  | &ret); | 
|  |  | 
|  | /* Enable FPD4 Auto Recovery */ | 
|  | ub960_update_bits(priv, UB9702_SR_CSI_EXCLUSIVE_FWD2, GENMASK(5, 4), | 
|  | BIT(4), &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_set_fpd3_sync_mode_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | /* FPD3 Sync Mode */ | 
|  | ub960_rxport_write(priv, nport, UB9702_RR_CHANNEL_MODE, 0x2, &ret); | 
|  |  | 
|  | /* BC_FREQ_SELECT=(PLL_FREQ/3200) Mbps */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK, 6, &ret); | 
|  |  | 
|  | /* Set AEQ_LOCK_MODE = 1 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_FPD3_AEQ_CTRL_SEL_1, BIT(7), &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_set_bc_drv_config_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Enable RX port */ | 
|  | ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport), | 
|  | &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_set_raw10_dvp_mode_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | /* FPD3 RAW10 Mode */ | 
|  | ub960_rxport_write(priv, nport, UB9702_RR_CHANNEL_MODE, 0x5, &ret); | 
|  |  | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK, 0, &ret); | 
|  |  | 
|  | /* Set AEQ_LOCK_MODE = 1 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_FPD3_AEQ_CTRL_SEL_1, BIT(7), &ret); | 
|  |  | 
|  | /* | 
|  | * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits | 
|  | * 0b10 : 8-bit processing using upper 8 bits | 
|  | */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3 << 6, | 
|  | 0x2 << 6, &ret); | 
|  |  | 
|  | /* LV_POLARITY & FV_POLARITY */ | 
|  | ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3, | 
|  | priv->rxports[nport]->lv_fv_pol, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_set_bc_drv_config_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Enable RX port */ | 
|  | ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport), | 
|  | &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_configure_rx_port_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct ub960_rxport *rxport = priv->rxports[nport]; | 
|  | int ret; | 
|  |  | 
|  | if (!rxport) { | 
|  | ret = ub960_turn_off_rxport_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: disabled\n", nport); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | switch (rxport->cdr_mode) { | 
|  | case RXPORT_CDR_FPD4: | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | ret = ub960_set_fpd4_sync_mode_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: FPD-Link IV SYNC mode\n", nport); | 
|  | break; | 
|  | case RXPORT_MODE_CSI2_NONSYNC: | 
|  | ret = ub960_set_fpd4_async_mode_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: FPD-Link IV ASYNC mode\n", nport); | 
|  | break; | 
|  | default: | 
|  | dev_err(dev, "rx%u: unsupported FPD4 mode %u\n", nport, | 
|  | rxport->rx_mode); | 
|  | return -EINVAL; | 
|  | } | 
|  | break; | 
|  |  | 
|  | case RXPORT_CDR_FPD3: | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | ret = ub960_set_fpd3_sync_mode_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: FPD-Link III SYNC mode\n", nport); | 
|  | break; | 
|  | case RXPORT_MODE_RAW10: | 
|  | ret = ub960_set_raw10_dvp_mode_ub9702(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: FPD-Link III RAW10 DVP mode\n", | 
|  | nport); | 
|  | break; | 
|  | default: | 
|  | dev_err(&priv->client->dev, | 
|  | "rx%u: unsupported FPD3 mode %u\n", nport, | 
|  | rxport->rx_mode); | 
|  | return -EINVAL; | 
|  | } | 
|  | break; | 
|  |  | 
|  | default: | 
|  | dev_err(&priv->client->dev, "rx%u: unsupported CDR mode %u\n", | 
|  | nport, rxport->cdr_mode); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_lock_recovery_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | /* Assumption that max AEQ should be under 16 */ | 
|  | const u8 rx_aeq_limit = 16; | 
|  | u8 prev_aeq = 0xff; | 
|  | bool rx_lock; | 
|  |  | 
|  | for (unsigned int retry = 0; retry < 3; ++retry) { | 
|  | u8 port_sts1; | 
|  | u8 rx_aeq; | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, | 
|  | &port_sts1, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | rx_lock = port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS; | 
|  |  | 
|  | if (!rx_lock) { | 
|  | ret = ub960_rxport_lockup_wa_ub9702(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Restart AEQ by changing max to 0 --> 0x23 */ | 
|  | ret = ub960_write_ind(priv, | 
|  | UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, 0, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | msleep(20); | 
|  |  | 
|  | /* AEQ Restart */ | 
|  | ret = ub960_write_ind(priv, | 
|  | UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, | 
|  | 0x23, NULL); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | msleep(20); | 
|  | dev_dbg(dev, "rx%u: no lock, retry = %u\n", nport, | 
|  | retry); | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ret = ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL11, &rx_aeq, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (rx_aeq < rx_aeq_limit) { | 
|  | dev_dbg(dev, | 
|  | "rx%u: locked and AEQ normal before setting AEQ window\n", | 
|  | nport); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (rx_aeq != prev_aeq) { | 
|  | ret = ub960_rxport_lockup_wa_ub9702(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Restart AEQ by changing max to 0 --> 0x23 */ | 
|  | ret = ub960_write_ind(priv, | 
|  | UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, | 
|  | 0, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | msleep(20); | 
|  |  | 
|  | /* AEQ Restart */ | 
|  | ret = ub960_write_ind(priv, | 
|  | UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, | 
|  | 0x23, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | msleep(20); | 
|  |  | 
|  | dev_dbg(dev, | 
|  | "rx%u: high AEQ at initial check recovery loop, retry=%u\n", | 
|  | nport, retry); | 
|  |  | 
|  | prev_aeq = rx_aeq; | 
|  | } else { | 
|  | dev_dbg(dev, | 
|  | "rx%u: lossy cable detected, RX_AEQ %#x, RX_AEQ_LIMIT %#x, retry %u\n", | 
|  | nport, rx_aeq, rx_aeq_limit, retry); | 
|  | dev_dbg(dev, | 
|  | "rx%u: will continue with initiation sequence but high AEQ\n", | 
|  | nport); | 
|  | return 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | dev_err(dev, "rx%u: max number of retries: %s\n", nport, | 
|  | rx_lock ? "unstable AEQ" : "no lock"); | 
|  |  | 
|  | return -EIO; | 
|  | } | 
|  |  | 
|  | static int ub960_enable_aeq_lms_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 read_aeq_init; | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL11, &read_aeq_init, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: initial AEQ = %#x\n", nport, read_aeq_init); | 
|  |  | 
|  | /* Set AEQ Min */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL6, read_aeq_init, &ret); | 
|  | /* Set AEQ Max */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, read_aeq_init + 1, &ret); | 
|  | /* Set AEQ offset to 0 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL10, 0x0, &ret); | 
|  |  | 
|  | /* Enable AEQ tap2 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_EQ_CTRL_SEL_38, 0x00, &ret); | 
|  | /* Set VGA Gain 1 Gain 2 override to 0 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_VGA_CTRL_SEL_8, 0x00, &ret); | 
|  | /* Set VGA Initial Sweep Gain to 0 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_VGA_CTRL_SEL_6, 0x80, &ret); | 
|  | /* Set VGA_Adapt (VGA Gain) override to 0 (thermometer encoded) */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_VGA_CTRL_SEL_3, 0x00, &ret); | 
|  | /* Enable VGA_SWEEP */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_EQ_ADAPT_CTRL, 0x40, &ret); | 
|  | /* Disable VGA_SWEEP_GAIN_OV, disable VGA_TUNE_OV */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_EQ_OVERRIDE_CTRL, 0x00, &ret); | 
|  |  | 
|  | /* Set VGA HIGH Threshold to 43 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_VGA_CTRL_SEL_1, 0x2b, &ret); | 
|  | /* Set VGA LOW Threshold to 18 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_VGA_CTRL_SEL_2, 0x12, &ret); | 
|  | /* Set vga_sweep_th to 32 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_EQ_CTRL_SEL_15, 0x20, &ret); | 
|  | /* Set AEQ timer to 400us/step and parity threshold to 7 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_SYSTEM_INIT_REG0, 0xef, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: enable FPD-Link IV AEQ LMS\n", nport); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_enable_dfe_lms_ub9702(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Enable DFE LMS */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_EQ_CTRL_SEL_24, 0x40, &ret); | 
|  | /* Disable VGA Gain1 override */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_GAIN_CTRL_0, 0x20, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | usleep_range(1000, 5000); | 
|  |  | 
|  | /* Disable VGA Gain2 override */ | 
|  | ret = ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), | 
|  | UB9702_IR_RX_ANA_GAIN_CTRL_0, 0x00, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: enabled FPD-Link IV DFE LMS", nport); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_init_rx_ports_ub9702(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int port_lock_mask; | 
|  | unsigned int port_mask = 0; | 
|  | bool have_fpd4 = false; | 
|  | int ret; | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | ret = ub960_rxport_update_bits(priv, it.nport, | 
|  | UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_BC_ALWAYS_ON, | 
|  | UB960_RR_BCC_CONFIG_BC_ALWAYS_ON, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Disable FPD4 Auto Recovery */ | 
|  | ret = ub960_write(priv, UB9702_SR_CSI_EXCLUSIVE_FWD2, 0x0f, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | if (it.rxport->ser.addr >= 0) { | 
|  | /* | 
|  | * Set serializer's I2C address if set in the dts file, | 
|  | * and freeze it to prevent updates from the FC. | 
|  | */ | 
|  | ub960_rxport_write(priv, it.nport, UB960_RR_SER_ID, | 
|  | it.rxport->ser.addr << 1 | | 
|  | UB960_RR_SER_ID_FREEZE_DEVICE_ID, | 
|  | &ret); | 
|  | } | 
|  |  | 
|  | /* Set serializer I2C alias with auto-ack */ | 
|  | ub960_rxport_write(priv, it.nport, UB960_RR_SER_ALIAS_ID, | 
|  | it.rxport->ser.alias << 1 | | 
|  | UB960_RR_SER_ALIAS_ID_AUTO_ACK, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | if (fwnode_device_is_compatible(it.rxport->ser.fwnode, | 
|  | "ti,ds90ub971-q1")) { | 
|  | ret = ub960_rxport_bc_ser_config(it.rxport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  | } | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | /* Hold state machine in reset */ | 
|  | ub960_rxport_write(priv, it.nport, UB9702_RR_RX_SM_SEL_2, 0x10, | 
|  | &ret); | 
|  |  | 
|  | /* Set AEQ max to 0 */ | 
|  | ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(it.nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, 0, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, | 
|  | "rx%u: holding state machine and adjusting AEQ max to 0", | 
|  | it.nport); | 
|  | } | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | port_mask |= BIT(it.nport); | 
|  |  | 
|  | if (it.rxport->cdr_mode == RXPORT_CDR_FPD4) | 
|  | have_fpd4 = true; | 
|  | } | 
|  |  | 
|  | for_each_rxport(priv, it) { | 
|  | ret = ub960_configure_rx_port_ub9702(priv, it.nport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = ub960_reset(priv, false); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (have_fpd4) { | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | /* Release state machine */ | 
|  | ret = ub960_rxport_write(priv, it.nport, | 
|  | UB9702_RR_RX_SM_SEL_2, 0x0, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: state machine released\n", | 
|  | it.nport); | 
|  | } | 
|  |  | 
|  | /* Wait for SM to resume */ | 
|  | fsleep(5000); | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | ret = ub960_write_ind(priv, | 
|  | UB960_IND_TARGET_RX_ANA(it.nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL7, | 
|  | 0x23, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: AEQ restart\n", it.nport); | 
|  | } | 
|  |  | 
|  | /* Wait for lock */ | 
|  | fsleep(20000); | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | ret = ub960_lock_recovery_ub9702(priv, it.nport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | ret = ub960_enable_aeq_lms_ub9702(priv, it.nport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | /* Hold state machine in reset */ | 
|  | ret = ub960_rxport_write(priv, it.nport, | 
|  | UB9702_RR_RX_SM_SEL_2, 0x10, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = ub960_reset(priv, false); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | /* Release state machine */ | 
|  | ret = ub960_rxport_write(priv, it.nport, | 
|  | UB9702_RR_RX_SM_SEL_2, 0, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Wait time for stable lock */ | 
|  | fsleep(15000); | 
|  |  | 
|  | /* Set temperature ramp on serializer */ | 
|  | for_each_active_rxport(priv, it) { | 
|  | ret = ub960_serializer_temp_ramp(it.rxport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | ret = ub960_enable_dfe_lms_ub9702(priv, it.nport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Wait for DFE and LMS to adapt */ | 
|  | fsleep(5000); | 
|  |  | 
|  | ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (port_mask != port_lock_mask) { | 
|  | ret = -EIO; | 
|  | dev_err_probe(dev, ret, "Failed to lock all RX ports\n"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | /* Enable all interrupt sources from this port */ | 
|  | ub960_rxport_write(priv, it.nport, UB960_RR_PORT_ICR_HI, 0x07, | 
|  | &ret); | 
|  | ub960_rxport_write(priv, it.nport, UB960_RR_PORT_ICR_LO, 0x7f, | 
|  | &ret); | 
|  |  | 
|  | /* Clear serializer I2C alias auto-ack */ | 
|  | ub960_rxport_update_bits(priv, it.nport, UB960_RR_SER_ALIAS_ID, | 
|  | UB960_RR_SER_ALIAS_ID_AUTO_ACK, 0, | 
|  | &ret); | 
|  |  | 
|  | /* Enable I2C_PASS_THROUGH */ | 
|  | ub960_rxport_update_bits(priv, it.nport, UB960_RR_BCC_CONFIG, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, | 
|  | UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH, | 
|  | &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Enable FPD4 Auto Recovery, Recovery loop active */ | 
|  | ret = ub960_write(priv, UB9702_SR_CSI_EXCLUSIVE_FWD2, 0x18, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | for_each_active_rxport_fpd4(priv, it) { | 
|  | u8 final_aeq; | 
|  |  | 
|  | ret = ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(it.nport), | 
|  | UB9702_IR_RX_ANA_AEQ_ALP_SEL11, &final_aeq, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_dbg(dev, "rx%u: final AEQ = %#x\n", it.nport, final_aeq); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Clear any errors caused by switching the RX port settings while | 
|  | * probing. | 
|  | */ | 
|  |  | 
|  | ret = ub960_clear_rx_errors(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_rxport_handle_events(struct ub960_data *priv, u8 nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 rx_port_sts1; | 
|  | u8 rx_port_sts2; | 
|  | u8 csi_rx_sts; | 
|  | u8 bcc_sts; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Read interrupts (also clears most of them) */ | 
|  | ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &rx_port_sts1, | 
|  | &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &rx_port_sts2, | 
|  | &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts, &ret); | 
|  | ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts, &ret); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) { | 
|  | u16 v; | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, | 
|  | &v, NULL); | 
|  | if (!ret) | 
|  | dev_err(dev, "rx%u parity errors: %u\n", nport, v); | 
|  | } | 
|  |  | 
|  | if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR) | 
|  | dev_err(dev, "rx%u BCC CRC error\n", nport); | 
|  |  | 
|  | if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR) | 
|  | dev_err(dev, "rx%u BCC SEQ error\n", nport); | 
|  |  | 
|  | if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE) | 
|  | dev_err(dev, "rx%u line length unstable\n", nport); | 
|  |  | 
|  | if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR) | 
|  | dev_err(dev, "rx%u FPD3 encode error\n", nport); | 
|  |  | 
|  | if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR) | 
|  | dev_err(dev, "rx%u buffer error\n", nport); | 
|  |  | 
|  | if (csi_rx_sts) | 
|  | dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts); | 
|  |  | 
|  | if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR) | 
|  | dev_err(dev, "rx%u CSI ECC1 error\n", nport); | 
|  |  | 
|  | if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR) | 
|  | dev_err(dev, "rx%u CSI ECC2 error\n", nport); | 
|  |  | 
|  | if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR) | 
|  | dev_err(dev, "rx%u CSI checksum error\n", nport); | 
|  |  | 
|  | if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR) | 
|  | dev_err(dev, "rx%u CSI length error\n", nport); | 
|  |  | 
|  | if (bcc_sts) | 
|  | dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts); | 
|  |  | 
|  | if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR) | 
|  | dev_err(dev, "rx%u BCC response error", nport); | 
|  |  | 
|  | if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO) | 
|  | dev_err(dev, "rx%u BCC slave timeout", nport); | 
|  |  | 
|  | if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR) | 
|  | dev_err(dev, "rx%u BCC slave error", nport); | 
|  |  | 
|  | if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO) | 
|  | dev_err(dev, "rx%u BCC master timeout", nport); | 
|  |  | 
|  | if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR) | 
|  | dev_err(dev, "rx%u BCC master error", nport); | 
|  |  | 
|  | if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR) | 
|  | dev_err(dev, "rx%u BCC sequence error", nport); | 
|  |  | 
|  | if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) { | 
|  | u16 v; | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, | 
|  | &v, NULL); | 
|  | if (!ret) | 
|  | dev_dbg(dev, "rx%u line len changed: %u\n", nport, v); | 
|  | } | 
|  |  | 
|  | if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) { | 
|  | u16 v; | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, | 
|  | &v, NULL); | 
|  | if (!ret) | 
|  | dev_dbg(dev, "rx%u line count changed: %u\n", nport, v); | 
|  | } | 
|  |  | 
|  | if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) { | 
|  | dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport, | 
|  | (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ? | 
|  | "locked" : | 
|  | "unlocked", | 
|  | (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ? | 
|  | "passed" : | 
|  | "not passed", | 
|  | (rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ? | 
|  | "no clock" : | 
|  | "clock ok", | 
|  | (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ? | 
|  | "stable freq" : | 
|  | "unstable freq"); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* ----------------------------------------------------------------------------- | 
|  | * V4L2 | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * The current implementation only supports a simple VC mapping, where all VCs | 
|  | * from a one RX port will be mapped to the same VC. Also, the hardware | 
|  | * dictates that all streams from an RX port must go to a single TX port. | 
|  | * | 
|  | * This function decides the target VC numbers for each RX port with a simple | 
|  | * algorithm, so that for each TX port, we get VC numbers starting from 0, | 
|  | * and counting up. | 
|  | * | 
|  | * E.g. if all four RX ports are in use, of which the first two go to the | 
|  | * first TX port and the secont two go to the second TX port, we would get | 
|  | * the following VCs for the four RX ports: 0, 1, 0, 1. | 
|  | * | 
|  | * TODO: implement a more sophisticated VC mapping. As the driver cannot know | 
|  | * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this | 
|  | * probably needs to be somehow configurable. Device tree? | 
|  | */ | 
|  | static void ub960_get_vc_maps(struct ub960_data *priv, | 
|  | struct v4l2_subdev_state *state, u8 *vc) | 
|  | { | 
|  | u8 cur_vc[UB960_MAX_TX_NPORTS] = {}; | 
|  | struct v4l2_subdev_route *route; | 
|  | u8 handled_mask = 0; | 
|  |  | 
|  | for_each_active_route(&state->routing, route) { | 
|  | unsigned int rx, tx; | 
|  |  | 
|  | rx = ub960_pad_to_port(priv, route->sink_pad); | 
|  | if (BIT(rx) & handled_mask) | 
|  | continue; | 
|  |  | 
|  | tx = ub960_pad_to_port(priv, route->source_pad); | 
|  |  | 
|  | vc[rx] = cur_vc[tx]++; | 
|  | handled_mask |= BIT(rx); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  |  | 
|  | dev_dbg(dev, "enable TX port %u\n", nport); | 
|  |  | 
|  | return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL, | 
|  | UB960_TR_CSI_CTL_CSI_ENABLE, | 
|  | UB960_TR_CSI_CTL_CSI_ENABLE, NULL); | 
|  | } | 
|  |  | 
|  | static int ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  |  | 
|  | dev_dbg(dev, "disable TX port %u\n", nport); | 
|  |  | 
|  | return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL, | 
|  | UB960_TR_CSI_CTL_CSI_ENABLE, 0, NULL); | 
|  | } | 
|  |  | 
|  | static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  |  | 
|  | dev_dbg(dev, "enable RX port %u\n", nport); | 
|  |  | 
|  | /* Enable forwarding */ | 
|  | return ub960_update_bits(priv, UB960_SR_FWD_CTL1, | 
|  | UB960_SR_FWD_CTL1_PORT_DIS(nport), 0, NULL); | 
|  | } | 
|  |  | 
|  | static int ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  |  | 
|  | dev_dbg(dev, "disable RX port %u\n", nport); | 
|  |  | 
|  | /* Disable forwarding */ | 
|  | return ub960_update_bits(priv, UB960_SR_FWD_CTL1, | 
|  | UB960_SR_FWD_CTL1_PORT_DIS(nport), | 
|  | UB960_SR_FWD_CTL1_PORT_DIS(nport), NULL); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * The driver only supports using a single VC for each source. This function | 
|  | * checks that each source only provides streams using a single VC. | 
|  | */ | 
|  | static int ub960_validate_stream_vcs(struct ub960_data *priv) | 
|  | { | 
|  | for_each_active_rxport(priv, it) { | 
|  | struct v4l2_mbus_frame_desc desc; | 
|  | int ret; | 
|  | u8 vc; | 
|  |  | 
|  | ret = v4l2_subdev_call(it.rxport->source.sd, pad, | 
|  | get_frame_desc, it.rxport->source.pad, | 
|  | &desc); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2) | 
|  | continue; | 
|  |  | 
|  | if (desc.num_entries == 0) | 
|  | continue; | 
|  |  | 
|  | vc = desc.entry[0].bus.csi2.vc; | 
|  |  | 
|  | for (unsigned int i = 1; i < desc.num_entries; i++) { | 
|  | if (vc == desc.entry[i].bus.csi2.vc) | 
|  | continue; | 
|  |  | 
|  | dev_err(&priv->client->dev, | 
|  | "rx%u: source with multiple virtual-channels is not supported\n", | 
|  | it.nport); | 
|  | return -ENODEV; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_configure_ports_for_streaming(struct ub960_data *priv, | 
|  | struct v4l2_subdev_state *state) | 
|  | { | 
|  | u8 fwd_ctl; | 
|  | struct { | 
|  | u32 num_streams; | 
|  | u8 pixel_dt; | 
|  | u8 meta_dt; | 
|  | u32 meta_lines; | 
|  | u32 tx_port; | 
|  | } rx_data[UB960_MAX_RX_NPORTS] = {}; | 
|  | u8 vc_map[UB960_MAX_RX_NPORTS] = {}; | 
|  | struct v4l2_subdev_route *route; | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_validate_stream_vcs(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ub960_get_vc_maps(priv, state, vc_map); | 
|  |  | 
|  | for_each_active_route(&state->routing, route) { | 
|  | struct ub960_rxport *rxport; | 
|  | struct ub960_txport *txport; | 
|  | struct v4l2_mbus_framefmt *fmt; | 
|  | const struct ub960_format_info *ub960_fmt; | 
|  | unsigned int nport; | 
|  |  | 
|  | nport = ub960_pad_to_port(priv, route->sink_pad); | 
|  |  | 
|  | rxport = priv->rxports[nport]; | 
|  | if (!rxport) | 
|  | return -EINVAL; | 
|  |  | 
|  | txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)]; | 
|  | if (!txport) | 
|  | return -EINVAL; | 
|  |  | 
|  | rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad); | 
|  |  | 
|  | rx_data[nport].num_streams++; | 
|  |  | 
|  | /* For the rest, we are only interested in parallel busses */ | 
|  | if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC || | 
|  | rxport->rx_mode == RXPORT_MODE_CSI2_NONSYNC) | 
|  | continue; | 
|  |  | 
|  | if (rx_data[nport].num_streams > 2) | 
|  | return -EPIPE; | 
|  |  | 
|  | fmt = v4l2_subdev_state_get_format(state, route->sink_pad, | 
|  | route->sink_stream); | 
|  | if (!fmt) | 
|  | return -EPIPE; | 
|  |  | 
|  | ub960_fmt = ub960_find_format(fmt->code); | 
|  | if (!ub960_fmt) | 
|  | return -EPIPE; | 
|  |  | 
|  | if (ub960_fmt->meta) { | 
|  | if (fmt->height > 3) { | 
|  | dev_err(&priv->client->dev, | 
|  | "rx%u: unsupported metadata height %u\n", | 
|  | nport, fmt->height); | 
|  | return -EPIPE; | 
|  | } | 
|  |  | 
|  | rx_data[nport].meta_dt = ub960_fmt->datatype; | 
|  | rx_data[nport].meta_lines = fmt->height; | 
|  | } else { | 
|  | rx_data[nport].pixel_dt = ub960_fmt->datatype; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Configure RX ports */ | 
|  |  | 
|  | /* | 
|  | * Keep all port forwardings disabled by default. Forwarding will be | 
|  | * enabled in ub960_enable_rx_port. | 
|  | */ | 
|  | fwd_ctl = GENMASK(7, 4); | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | unsigned long nport = it.nport; | 
|  |  | 
|  | u8 vc = vc_map[nport]; | 
|  |  | 
|  | if (rx_data[nport].num_streams == 0) | 
|  | continue; | 
|  |  | 
|  | switch (it.rxport->rx_mode) { | 
|  | case RXPORT_MODE_RAW10: | 
|  | ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID, | 
|  | rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT), | 
|  | &ret); | 
|  |  | 
|  | ub960_rxport_write(priv, nport, | 
|  | UB960_RR_RAW_EMBED_DTYPE, | 
|  | (rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) | | 
|  | rx_data[nport].meta_dt, &ret); | 
|  |  | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | /* Not implemented */ | 
|  | break; | 
|  |  | 
|  | case RXPORT_MODE_CSI2_SYNC: | 
|  | case RXPORT_MODE_CSI2_NONSYNC: | 
|  | if (!priv->hw_data->is_ub9702) { | 
|  | /* Map all VCs from this port to the same VC */ | 
|  | ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP, | 
|  | (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) | | 
|  | (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) | | 
|  | (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) | | 
|  | (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)), | 
|  | &ret); | 
|  | } else { | 
|  | unsigned int i; | 
|  |  | 
|  | /* Map all VCs from this port to VC(nport) */ | 
|  | for (i = 0; i < 8; i++) | 
|  | ub960_rxport_write(priv, nport, | 
|  | UB9702_RR_VC_ID_MAP(i), | 
|  | (nport << 4) | nport, | 
|  | &ret); | 
|  | } | 
|  |  | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (rx_data[nport].tx_port == 1) | 
|  | fwd_ctl |= BIT(nport); /* forward to TX1 */ | 
|  | else | 
|  | fwd_ctl &= ~BIT(nport); /* forward to TX0 */ | 
|  | } | 
|  |  | 
|  | ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl, &ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void ub960_update_streaming_status(struct ub960_data *priv) | 
|  | { | 
|  | unsigned int i; | 
|  |  | 
|  | for (i = 0; i < UB960_MAX_NPORTS; i++) { | 
|  | if (priv->stream_enable_mask[i]) | 
|  | break; | 
|  | } | 
|  |  | 
|  | priv->streaming = i < UB960_MAX_NPORTS; | 
|  | } | 
|  |  | 
|  | static int ub960_enable_streams(struct v4l2_subdev *sd, | 
|  | struct v4l2_subdev_state *state, u32 source_pad, | 
|  | u64 source_streams_mask) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  | struct device *dev = &priv->client->dev; | 
|  | u64 sink_streams[UB960_MAX_RX_NPORTS] = {}; | 
|  | struct v4l2_subdev_route *route; | 
|  | unsigned int failed_port; | 
|  | int ret; | 
|  |  | 
|  | if (!priv->streaming) { | 
|  | dev_dbg(dev, "Prepare for streaming\n"); | 
|  | ret = ub960_configure_ports_for_streaming(priv, state); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Enable TX port if not yet enabled */ | 
|  | if (!priv->stream_enable_mask[source_pad]) { | 
|  | ret = ub960_enable_tx_port(priv, | 
|  | ub960_pad_to_port(priv, source_pad)); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | priv->stream_enable_mask[source_pad] |= source_streams_mask; | 
|  |  | 
|  | /* Collect sink streams per pad which we need to enable */ | 
|  | for_each_active_route(&state->routing, route) { | 
|  | unsigned int nport; | 
|  |  | 
|  | if (route->source_pad != source_pad) | 
|  | continue; | 
|  |  | 
|  | if (!(source_streams_mask & BIT_ULL(route->source_stream))) | 
|  | continue; | 
|  |  | 
|  | nport = ub960_pad_to_port(priv, route->sink_pad); | 
|  |  | 
|  | sink_streams[nport] |= BIT_ULL(route->sink_stream); | 
|  | } | 
|  |  | 
|  | for_each_rxport(priv, it) { | 
|  | unsigned int nport = it.nport; | 
|  |  | 
|  | if (!sink_streams[nport]) | 
|  | continue; | 
|  |  | 
|  | /* Enable the RX port if not yet enabled */ | 
|  | if (!priv->stream_enable_mask[nport]) { | 
|  | ret = ub960_enable_rx_port(priv, nport); | 
|  | if (ret) { | 
|  | failed_port = nport; | 
|  | goto err; | 
|  | } | 
|  | } | 
|  |  | 
|  | priv->stream_enable_mask[nport] |= sink_streams[nport]; | 
|  |  | 
|  | dev_dbg(dev, "enable RX port %u streams %#llx\n", nport, | 
|  | sink_streams[nport]); | 
|  |  | 
|  | ret = v4l2_subdev_enable_streams( | 
|  | priv->rxports[nport]->source.sd, | 
|  | priv->rxports[nport]->source.pad, | 
|  | sink_streams[nport]); | 
|  | if (ret) { | 
|  | priv->stream_enable_mask[nport] &= ~sink_streams[nport]; | 
|  |  | 
|  | if (!priv->stream_enable_mask[nport]) | 
|  | ub960_disable_rx_port(priv, nport); | 
|  |  | 
|  | failed_port = nport; | 
|  | goto err; | 
|  | } | 
|  | } | 
|  |  | 
|  | priv->streaming = true; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err: | 
|  | for (unsigned int nport = 0; nport < failed_port; nport++) { | 
|  | if (!sink_streams[nport]) | 
|  | continue; | 
|  |  | 
|  | dev_dbg(dev, "disable RX port %u streams %#llx\n", nport, | 
|  | sink_streams[nport]); | 
|  |  | 
|  | ret = v4l2_subdev_disable_streams( | 
|  | priv->rxports[nport]->source.sd, | 
|  | priv->rxports[nport]->source.pad, | 
|  | sink_streams[nport]); | 
|  | if (ret) | 
|  | dev_err(dev, "Failed to disable streams: %d\n", ret); | 
|  |  | 
|  | priv->stream_enable_mask[nport] &= ~sink_streams[nport]; | 
|  |  | 
|  | /* Disable RX port if no active streams */ | 
|  | if (!priv->stream_enable_mask[nport]) | 
|  | ub960_disable_rx_port(priv, nport); | 
|  | } | 
|  |  | 
|  | priv->stream_enable_mask[source_pad] &= ~source_streams_mask; | 
|  |  | 
|  | if (!priv->stream_enable_mask[source_pad]) | 
|  | ub960_disable_tx_port(priv, | 
|  | ub960_pad_to_port(priv, source_pad)); | 
|  |  | 
|  | ub960_update_streaming_status(priv); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_disable_streams(struct v4l2_subdev *sd, | 
|  | struct v4l2_subdev_state *state, | 
|  | u32 source_pad, u64 source_streams_mask) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  | struct device *dev = &priv->client->dev; | 
|  | u64 sink_streams[UB960_MAX_RX_NPORTS] = {}; | 
|  | struct v4l2_subdev_route *route; | 
|  | int ret; | 
|  |  | 
|  | /* Collect sink streams per pad which we need to disable */ | 
|  | for_each_active_route(&state->routing, route) { | 
|  | unsigned int nport; | 
|  |  | 
|  | if (route->source_pad != source_pad) | 
|  | continue; | 
|  |  | 
|  | if (!(source_streams_mask & BIT_ULL(route->source_stream))) | 
|  | continue; | 
|  |  | 
|  | nport = ub960_pad_to_port(priv, route->sink_pad); | 
|  |  | 
|  | sink_streams[nport] |= BIT_ULL(route->sink_stream); | 
|  | } | 
|  |  | 
|  | for_each_rxport(priv, it) { | 
|  | unsigned int nport = it.nport; | 
|  |  | 
|  | if (!sink_streams[nport]) | 
|  | continue; | 
|  |  | 
|  | dev_dbg(dev, "disable RX port %u streams %#llx\n", nport, | 
|  | sink_streams[nport]); | 
|  |  | 
|  | ret = v4l2_subdev_disable_streams( | 
|  | priv->rxports[nport]->source.sd, | 
|  | priv->rxports[nport]->source.pad, | 
|  | sink_streams[nport]); | 
|  | if (ret) | 
|  | dev_err(dev, "Failed to disable streams: %d\n", ret); | 
|  |  | 
|  | priv->stream_enable_mask[nport] &= ~sink_streams[nport]; | 
|  |  | 
|  | /* Disable RX port if no active streams */ | 
|  | if (!priv->stream_enable_mask[nport]) | 
|  | ub960_disable_rx_port(priv, nport); | 
|  | } | 
|  |  | 
|  | /* Disable TX port if no active streams */ | 
|  |  | 
|  | priv->stream_enable_mask[source_pad] &= ~source_streams_mask; | 
|  |  | 
|  | if (!priv->stream_enable_mask[source_pad]) | 
|  | ub960_disable_tx_port(priv, | 
|  | ub960_pad_to_port(priv, source_pad)); | 
|  |  | 
|  | ub960_update_streaming_status(priv); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int _ub960_set_routing(struct v4l2_subdev *sd, | 
|  | struct v4l2_subdev_state *state, | 
|  | struct v4l2_subdev_krouting *routing) | 
|  | { | 
|  | static const struct v4l2_mbus_framefmt format = { | 
|  | .width = 640, | 
|  | .height = 480, | 
|  | .code = MEDIA_BUS_FMT_UYVY8_1X16, | 
|  | .field = V4L2_FIELD_NONE, | 
|  | .colorspace = V4L2_COLORSPACE_SRGB, | 
|  | .ycbcr_enc = V4L2_YCBCR_ENC_601, | 
|  | .quantization = V4L2_QUANTIZATION_LIM_RANGE, | 
|  | .xfer_func = V4L2_XFER_FUNC_SRGB, | 
|  | }; | 
|  | int ret; | 
|  |  | 
|  | ret = v4l2_subdev_routing_validate(sd, routing, | 
|  | V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 | | 
|  | V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_set_routing(struct v4l2_subdev *sd, | 
|  | struct v4l2_subdev_state *state, | 
|  | enum v4l2_subdev_format_whence which, | 
|  | struct v4l2_subdev_krouting *routing) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  |  | 
|  | if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming) | 
|  | return -EBUSY; | 
|  |  | 
|  | return _ub960_set_routing(sd, state, routing); | 
|  | } | 
|  |  | 
|  | static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad, | 
|  | struct v4l2_mbus_frame_desc *fd) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  | struct v4l2_subdev_route *route; | 
|  | struct v4l2_subdev_state *state; | 
|  | int ret = 0; | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 vc_map[UB960_MAX_RX_NPORTS] = {}; | 
|  |  | 
|  | if (!ub960_pad_is_source(priv, pad)) | 
|  | return -EINVAL; | 
|  |  | 
|  | fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2; | 
|  |  | 
|  | state = v4l2_subdev_lock_and_get_active_state(&priv->sd); | 
|  |  | 
|  | ub960_get_vc_maps(priv, state, vc_map); | 
|  |  | 
|  | for_each_active_route(&state->routing, route) { | 
|  | struct v4l2_mbus_frame_desc_entry *source_entry = NULL; | 
|  | struct v4l2_mbus_frame_desc source_fd; | 
|  | unsigned int nport; | 
|  | unsigned int i; | 
|  |  | 
|  | if (route->source_pad != pad) | 
|  | continue; | 
|  |  | 
|  | nport = ub960_pad_to_port(priv, route->sink_pad); | 
|  |  | 
|  | ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad, | 
|  | get_frame_desc, | 
|  | priv->rxports[nport]->source.pad, | 
|  | &source_fd); | 
|  | if (ret) { | 
|  | dev_err(dev, | 
|  | "Failed to get source frame desc for pad %u\n", | 
|  | route->sink_pad); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < source_fd.num_entries; i++) { | 
|  | if (source_fd.entry[i].stream == route->sink_stream) { | 
|  | source_entry = &source_fd.entry[i]; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (!source_entry) { | 
|  | dev_err(dev, | 
|  | "Failed to find stream from source frame desc\n"); | 
|  | ret = -EPIPE; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | fd->entry[fd->num_entries].stream = route->source_stream; | 
|  | fd->entry[fd->num_entries].flags = source_entry->flags; | 
|  | fd->entry[fd->num_entries].length = source_entry->length; | 
|  | fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode; | 
|  |  | 
|  | fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport]; | 
|  |  | 
|  | if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) { | 
|  | fd->entry[fd->num_entries].bus.csi2.dt = | 
|  | source_entry->bus.csi2.dt; | 
|  | } else { | 
|  | const struct ub960_format_info *ub960_fmt; | 
|  | struct v4l2_mbus_framefmt *fmt; | 
|  |  | 
|  | fmt = v4l2_subdev_state_get_format(state, pad, | 
|  | route->source_stream); | 
|  |  | 
|  | if (!fmt) { | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | ub960_fmt = ub960_find_format(fmt->code); | 
|  | if (!ub960_fmt) { | 
|  | dev_err(dev, "Unable to find format\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | fd->entry[fd->num_entries].bus.csi2.dt = | 
|  | ub960_fmt->datatype; | 
|  | } | 
|  |  | 
|  | fd->num_entries++; | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | v4l2_subdev_unlock_state(state); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_set_fmt(struct v4l2_subdev *sd, | 
|  | struct v4l2_subdev_state *state, | 
|  | struct v4l2_subdev_format *format) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  | struct v4l2_mbus_framefmt *fmt; | 
|  |  | 
|  | if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming) | 
|  | return -EBUSY; | 
|  |  | 
|  | /* No transcoding, source and sink formats must match. */ | 
|  | if (ub960_pad_is_source(priv, format->pad)) | 
|  | return v4l2_subdev_get_fmt(sd, state, format); | 
|  |  | 
|  | /* | 
|  | * Default to the first format if the requested media bus code isn't | 
|  | * supported. | 
|  | */ | 
|  | if (!ub960_find_format(format->format.code)) | 
|  | format->format.code = ub960_formats[0].code; | 
|  |  | 
|  | fmt = v4l2_subdev_state_get_format(state, format->pad, format->stream); | 
|  | if (!fmt) | 
|  | return -EINVAL; | 
|  |  | 
|  | *fmt = format->format; | 
|  |  | 
|  | fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad, | 
|  | format->stream); | 
|  | if (!fmt) | 
|  | return -EINVAL; | 
|  |  | 
|  | *fmt = format->format; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_init_state(struct v4l2_subdev *sd, | 
|  | struct v4l2_subdev_state *state) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  |  | 
|  | struct v4l2_subdev_route routes[] = { | 
|  | { | 
|  | .sink_pad = 0, | 
|  | .sink_stream = 0, | 
|  | .source_pad = priv->hw_data->num_rxports, | 
|  | .source_stream = 0, | 
|  | .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE, | 
|  | }, | 
|  | }; | 
|  |  | 
|  | struct v4l2_subdev_krouting routing = { | 
|  | .num_routes = ARRAY_SIZE(routes), | 
|  | .routes = routes, | 
|  | }; | 
|  |  | 
|  | return _ub960_set_routing(sd, state, &routing); | 
|  | } | 
|  |  | 
|  | static const struct v4l2_subdev_pad_ops ub960_pad_ops = { | 
|  | .enable_streams = ub960_enable_streams, | 
|  | .disable_streams = ub960_disable_streams, | 
|  |  | 
|  | .set_routing = ub960_set_routing, | 
|  | .get_frame_desc = ub960_get_frame_desc, | 
|  |  | 
|  | .get_fmt = v4l2_subdev_get_fmt, | 
|  | .set_fmt = ub960_set_fmt, | 
|  | }; | 
|  |  | 
|  | static int ub960_log_status_ub960_sp_eq(struct ub960_data *priv, | 
|  | unsigned int nport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 eq_level; | 
|  | s8 strobe_pos; | 
|  | int ret; | 
|  | u8 v; | 
|  |  | 
|  | /* Strobe */ | 
|  |  | 
|  | ret = ub960_read(priv, UB960_XR_AEQ_CTL1, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\t%s strobe\n", | 
|  | (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" : | 
|  | "Manual"); | 
|  |  | 
|  | if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) { | 
|  | ret = ub960_read(priv, UB960_XR_SFILTER_CFG, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tStrobe range [%d, %d]\n", | 
|  | ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7, | 
|  | ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7); | 
|  | } | 
|  |  | 
|  | ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tStrobe pos %d\n", strobe_pos); | 
|  |  | 
|  | /* EQ */ | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\t%s EQ\n", | 
|  | (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" : | 
|  | "Adaptive"); | 
|  |  | 
|  | if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) { | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tEQ range [%u, %u]\n", | 
|  | (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf, | 
|  | (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf); | 
|  | } | 
|  |  | 
|  | ret = ub960_rxport_get_eq_level(priv, nport, &eq_level); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tEQ level %u\n", eq_level); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_log_status(struct v4l2_subdev *sd) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct v4l2_subdev_state *state; | 
|  | u16 v16 = 0; | 
|  | u8 v = 0; | 
|  | u8 id[UB960_SR_FPD3_RX_ID_LEN]; | 
|  | int ret = 0; | 
|  |  | 
|  | state = v4l2_subdev_lock_and_get_active_state(sd); | 
|  |  | 
|  | for (unsigned int i = 0; i < sizeof(id); i++) { | 
|  | ret = ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i], NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id); | 
|  |  | 
|  | for (unsigned int nport = 0; nport < priv->hw_data->num_txports; | 
|  | nport++) { | 
|  | struct ub960_txport *txport = priv->txports[nport]; | 
|  |  | 
|  | dev_info(dev, "TX %u\n", nport); | 
|  |  | 
|  | if (!txport) { | 
|  | dev_info(dev, "\tNot initialized\n"); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ret = ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1), | 
|  | v & (u8)BIT(0)); | 
|  |  | 
|  | ret = ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tframe counter %u\n", v16); | 
|  |  | 
|  | ret = ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tframe error counter %u\n", v16); | 
|  |  | 
|  | ret = ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tline counter %u\n", v16); | 
|  |  | 
|  | ret = ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tline error counter %u\n", v16); | 
|  | } | 
|  |  | 
|  | for_each_rxport(priv, it) { | 
|  | unsigned int nport = it.nport; | 
|  |  | 
|  | dev_info(dev, "RX %u\n", nport); | 
|  |  | 
|  | if (!it.rxport) { | 
|  | dev_info(dev, "\tNot initialized\n"); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (v & UB960_RR_RX_PORT_STS1_LOCK_STS) | 
|  | dev_info(dev, "\tLocked\n"); | 
|  | else | 
|  | dev_info(dev, "\tNot locked\n"); | 
|  |  | 
|  | dev_info(dev, "\trx_port_sts1 %#02x\n", v); | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v, | 
|  | NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\trx_port_sts2 %#02x\n", v); | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tlink freq %llu Hz\n", ((u64)v16 * HZ_PER_MHZ) >> 8); | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tparity errors %u\n", v16); | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tlines per frame %u\n", v16); | 
|  |  | 
|  | ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, | 
|  | &v16, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tbytes per line %u\n", v16); | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, | 
|  | &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tcsi_err_counter %u\n", v); | 
|  |  | 
|  | if (!priv->hw_data->is_ub9702) { | 
|  | ret = ub960_log_status_ub960_sp_eq(priv, nport); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* GPIOs */ | 
|  | for (unsigned int i = 0; i < UB960_NUM_BC_GPIOS; i++) { | 
|  | u8 ctl_reg; | 
|  | u8 ctl_shift; | 
|  |  | 
|  | ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2); | 
|  | ctl_shift = (i % 2) * 4; | 
|  |  | 
|  | ret = ub960_rxport_read(priv, nport, ctl_reg, &v, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | dev_info(dev, "\tGPIO%u: mode %u\n", i, | 
|  | (v >> ctl_shift) & 0xf); | 
|  | } | 
|  | } | 
|  |  | 
|  | v4l2_subdev_unlock_state(state); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = { | 
|  | .log_status = ub960_log_status, | 
|  | }; | 
|  |  | 
|  | static const struct v4l2_subdev_internal_ops ub960_internal_ops = { | 
|  | .init_state = ub960_init_state, | 
|  | }; | 
|  |  | 
|  | static const struct v4l2_subdev_ops ub960_subdev_ops = { | 
|  | .core = &ub960_subdev_core_ops, | 
|  | .pad = &ub960_pad_ops, | 
|  | }; | 
|  |  | 
|  | static const struct media_entity_operations ub960_entity_ops = { | 
|  | .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1, | 
|  | .link_validate = v4l2_subdev_link_validate, | 
|  | .has_pad_interdep = v4l2_subdev_has_pad_interdep, | 
|  | }; | 
|  |  | 
|  | /* ----------------------------------------------------------------------------- | 
|  | * Core | 
|  | */ | 
|  |  | 
|  | static irqreturn_t ub960_handle_events(int irq, void *arg) | 
|  | { | 
|  | struct ub960_data *priv = arg; | 
|  | u8 int_sts; | 
|  | u8 fwd_sts; | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts, NULL); | 
|  | if (ret || !int_sts) | 
|  | return IRQ_NONE; | 
|  |  | 
|  | dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts); | 
|  |  | 
|  | ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts, NULL); | 
|  | if (ret) | 
|  | return IRQ_NONE; | 
|  |  | 
|  | dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts); | 
|  |  | 
|  | for (unsigned int i = 0; i < priv->hw_data->num_txports; i++) { | 
|  | if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i)) { | 
|  | ret = ub960_csi_handle_events(priv, i); | 
|  | if (ret) | 
|  | return IRQ_NONE; | 
|  | } | 
|  | } | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(it.nport)) { | 
|  | ret = ub960_rxport_handle_events(priv, it.nport); | 
|  | if (ret) | 
|  | return IRQ_NONE; | 
|  | } | 
|  | } | 
|  |  | 
|  | return IRQ_HANDLED; | 
|  | } | 
|  |  | 
|  | static void ub960_handler_work(struct work_struct *work) | 
|  | { | 
|  | struct delayed_work *dwork = to_delayed_work(work); | 
|  | struct ub960_data *priv = | 
|  | container_of(dwork, struct ub960_data, poll_work); | 
|  |  | 
|  | ub960_handle_events(0, priv); | 
|  |  | 
|  | schedule_delayed_work(&priv->poll_work, | 
|  | msecs_to_jiffies(UB960_POLL_TIME_MS)); | 
|  | } | 
|  |  | 
|  | static void ub960_txport_free_ports(struct ub960_data *priv) | 
|  | { | 
|  | unsigned int nport; | 
|  |  | 
|  | for (nport = 0; nport < priv->hw_data->num_txports; nport++) { | 
|  | struct ub960_txport *txport = priv->txports[nport]; | 
|  |  | 
|  | if (!txport) | 
|  | continue; | 
|  |  | 
|  | kfree(txport); | 
|  | priv->txports[nport] = NULL; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void ub960_rxport_free_ports(struct ub960_data *priv) | 
|  | { | 
|  | for_each_active_rxport(priv, it) { | 
|  | fwnode_handle_put(it.rxport->source.ep_fwnode); | 
|  | fwnode_handle_put(it.rxport->ser.fwnode); | 
|  |  | 
|  | mutex_destroy(&it.rxport->aliased_addrs_lock); | 
|  |  | 
|  | kfree(it.rxport); | 
|  | priv->rxports[it.nport] = NULL; | 
|  | } | 
|  | } | 
|  |  | 
|  | static int | 
|  | ub960_parse_dt_rxport_link_properties(struct ub960_data *priv, | 
|  | struct fwnode_handle *link_fwnode, | 
|  | struct ub960_rxport *rxport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int nport = rxport->nport; | 
|  | u32 rx_mode; | 
|  | u32 cdr_mode; | 
|  | s32 strobe_pos; | 
|  | u32 eq_level; | 
|  | u32 ser_i2c_alias; | 
|  | u32 ser_i2c_addr; | 
|  | int ret; | 
|  |  | 
|  | cdr_mode = RXPORT_CDR_FPD3; | 
|  |  | 
|  | ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode); | 
|  | if (ret < 0 && ret != -EINVAL) { | 
|  | dev_err(dev, "rx%u: failed to read '%s': %d\n", nport, | 
|  | "ti,cdr-mode", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (cdr_mode > RXPORT_CDR_LAST) { | 
|  | dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) { | 
|  | dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | rxport->cdr_mode = cdr_mode; | 
|  |  | 
|  | ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode); | 
|  | if (ret < 0) { | 
|  | dev_err(dev, "rx%u: failed to read '%s': %d\n", nport, | 
|  | "ti,rx-mode", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (rx_mode > RXPORT_MODE_LAST) { | 
|  | dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | switch (rx_mode) { | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport, | 
|  | rx_mode); | 
|  | return -EINVAL; | 
|  | default: | 
|  | break; | 
|  | } | 
|  |  | 
|  | rxport->rx_mode = rx_mode; | 
|  |  | 
|  | /* EQ & Strobe related */ | 
|  |  | 
|  | /* Defaults */ | 
|  | rxport->eq.manual_eq = false; | 
|  | rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL; | 
|  | rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL; | 
|  |  | 
|  | ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos", | 
|  | &strobe_pos); | 
|  | if (ret) { | 
|  | if (ret != -EINVAL) { | 
|  | dev_err(dev, "rx%u: failed to read '%s': %d\n", nport, | 
|  | "ti,strobe-pos", ret); | 
|  | return ret; | 
|  | } | 
|  | } else { | 
|  | if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS || | 
|  | strobe_pos > UB960_MAX_MANUAL_STROBE_POS) { | 
|  | dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n", | 
|  | nport, strobe_pos); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* NOTE: ignored unless global manual strobe pos is also set */ | 
|  | rxport->eq.strobe_pos = strobe_pos; | 
|  | if (!priv->strobe.manual) | 
|  | dev_warn(dev, | 
|  | "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n", | 
|  | nport); | 
|  | } | 
|  |  | 
|  | ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level); | 
|  | if (ret) { | 
|  | if (ret != -EINVAL) { | 
|  | dev_err(dev, "rx%u: failed to read '%s': %d\n", nport, | 
|  | "ti,eq-level", ret); | 
|  | return ret; | 
|  | } | 
|  | } else { | 
|  | if (eq_level > UB960_MAX_EQ_LEVEL) { | 
|  | dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n", | 
|  | nport, eq_level); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | rxport->eq.manual_eq = true; | 
|  | rxport->eq.manual.eq_level = eq_level; | 
|  | } | 
|  |  | 
|  | ret = fwnode_property_read_u32(link_fwnode, "i2c-alias", | 
|  | &ser_i2c_alias); | 
|  | if (ret) { | 
|  | dev_err(dev, "rx%u: failed to read '%s': %d\n", nport, | 
|  | "i2c-alias", ret); | 
|  | return ret; | 
|  | } | 
|  | rxport->ser.alias = ser_i2c_alias; | 
|  |  | 
|  | rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer"); | 
|  | if (!rxport->ser.fwnode) { | 
|  | dev_err(dev, "rx%u: missing 'serializer' node\n", nport); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ret = fwnode_property_read_u32(rxport->ser.fwnode, "reg", | 
|  | &ser_i2c_addr); | 
|  | if (ret) | 
|  | rxport->ser.addr = -EINVAL; | 
|  | else | 
|  | rxport->ser.addr = ser_i2c_addr; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv, | 
|  | struct fwnode_handle *ep_fwnode, | 
|  | struct ub960_rxport *rxport) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct v4l2_fwnode_endpoint vep = {}; | 
|  | unsigned int nport = rxport->nport; | 
|  | bool hsync_hi; | 
|  | bool vsync_hi; | 
|  | int ret; | 
|  |  | 
|  | rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode); | 
|  | if (!rxport->source.ep_fwnode) { | 
|  | dev_err(dev, "rx%u: no remote endpoint\n", nport); | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | /* We currently have properties only for RAW modes */ | 
|  |  | 
|  | switch (rxport->rx_mode) { | 
|  | case RXPORT_MODE_RAW10: | 
|  | case RXPORT_MODE_RAW12_HF: | 
|  | case RXPORT_MODE_RAW12_LF: | 
|  | break; | 
|  | default: | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | vep.bus_type = V4L2_MBUS_PARALLEL; | 
|  | ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep); | 
|  | if (ret) { | 
|  | dev_err(dev, "rx%u: failed to parse endpoint data\n", nport); | 
|  | goto err_put_source_ep_fwnode; | 
|  | } | 
|  |  | 
|  | hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH); | 
|  | vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH); | 
|  |  | 
|  | /* LineValid and FrameValid are inverse to the h/vsync active */ | 
|  | rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) | | 
|  | (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_put_source_ep_fwnode: | 
|  | fwnode_handle_put(rxport->source.ep_fwnode); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport, | 
|  | struct fwnode_handle *link_fwnode, | 
|  | struct fwnode_handle *ep_fwnode) | 
|  | { | 
|  | static const char *vpoc_names[UB960_MAX_RX_NPORTS] = { | 
|  | "vpoc0", "vpoc1", "vpoc2", "vpoc3" | 
|  | }; | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct ub960_rxport *rxport; | 
|  | int ret; | 
|  |  | 
|  | rxport = kzalloc(sizeof(*rxport), GFP_KERNEL); | 
|  | if (!rxport) | 
|  | return -ENOMEM; | 
|  |  | 
|  | priv->rxports[nport] = rxport; | 
|  |  | 
|  | rxport->nport = nport; | 
|  | rxport->priv = priv; | 
|  |  | 
|  | ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport); | 
|  | if (ret) | 
|  | goto err_free_rxport; | 
|  |  | 
|  | rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]); | 
|  | if (IS_ERR(rxport->vpoc)) { | 
|  | ret = PTR_ERR(rxport->vpoc); | 
|  | if (ret == -ENODEV) { | 
|  | rxport->vpoc = NULL; | 
|  | } else { | 
|  | dev_err(dev, "rx%u: failed to get VPOC supply: %d\n", | 
|  | nport, ret); | 
|  | goto err_put_remote_fwnode; | 
|  | } | 
|  | } | 
|  |  | 
|  | ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport); | 
|  | if (ret) | 
|  | goto err_put_remote_fwnode; | 
|  |  | 
|  | mutex_init(&rxport->aliased_addrs_lock); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_put_remote_fwnode: | 
|  | fwnode_handle_put(rxport->ser.fwnode); | 
|  | err_free_rxport: | 
|  | priv->rxports[nport] = NULL; | 
|  | kfree(rxport); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static struct fwnode_handle * | 
|  | ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode, | 
|  | unsigned int nport) | 
|  | { | 
|  | struct fwnode_handle *link_fwnode; | 
|  | int ret; | 
|  |  | 
|  | fwnode_for_each_child_node(links_fwnode, link_fwnode) { | 
|  | u32 link_num; | 
|  |  | 
|  | if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@")) | 
|  | continue; | 
|  |  | 
|  | ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num); | 
|  | if (ret) { | 
|  | fwnode_handle_put(link_fwnode); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | if (nport == link_num) | 
|  | return link_fwnode; | 
|  | } | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static int ub960_parse_dt_rxports(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | struct fwnode_handle *links_fwnode; | 
|  | int ret; | 
|  |  | 
|  | links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links"); | 
|  | if (!links_fwnode) { | 
|  | dev_err(dev, "'links' node missing\n"); | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | /* Defaults, recommended by TI */ | 
|  | priv->strobe.min = 2; | 
|  | priv->strobe.max = 3; | 
|  |  | 
|  | priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe"); | 
|  |  | 
|  | for_each_rxport(priv, it) { | 
|  | struct fwnode_handle *link_fwnode; | 
|  | struct fwnode_handle *ep_fwnode; | 
|  | unsigned int nport = it.nport; | 
|  |  | 
|  | link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport); | 
|  | if (!link_fwnode) | 
|  | continue; | 
|  |  | 
|  | ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev), | 
|  | nport, 0, 0); | 
|  | if (!ep_fwnode) { | 
|  | fwnode_handle_put(link_fwnode); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ret = ub960_parse_dt_rxport(priv, nport, link_fwnode, | 
|  | ep_fwnode); | 
|  |  | 
|  | fwnode_handle_put(link_fwnode); | 
|  | fwnode_handle_put(ep_fwnode); | 
|  |  | 
|  | if (ret) { | 
|  | dev_err(dev, "rx%u: failed to parse RX port\n", nport); | 
|  | goto err_put_links; | 
|  | } | 
|  | } | 
|  |  | 
|  | fwnode_handle_put(links_fwnode); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_put_links: | 
|  | fwnode_handle_put(links_fwnode); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_parse_dt_txports(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | u32 nport; | 
|  | int ret; | 
|  |  | 
|  | for (nport = 0; nport < priv->hw_data->num_txports; nport++) { | 
|  | unsigned int port = nport + priv->hw_data->num_rxports; | 
|  | struct fwnode_handle *ep_fwnode; | 
|  |  | 
|  | ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev), | 
|  | port, 0, 0); | 
|  | if (!ep_fwnode) | 
|  | continue; | 
|  |  | 
|  | ret = ub960_parse_dt_txport(priv, ep_fwnode, nport); | 
|  |  | 
|  | fwnode_handle_put(ep_fwnode); | 
|  |  | 
|  | if (ret) | 
|  | break; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_parse_dt(struct ub960_data *priv) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | ret = ub960_parse_dt_rxports(priv); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = ub960_parse_dt_txports(priv); | 
|  | if (ret) | 
|  | goto err_free_rxports; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_free_rxports: | 
|  | ub960_rxport_free_ports(priv); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int ub960_notify_bound(struct v4l2_async_notifier *notifier, | 
|  | struct v4l2_subdev *subdev, | 
|  | struct v4l2_async_connection *asd) | 
|  | { | 
|  | struct ub960_data *priv = sd_to_ub960(notifier->sd); | 
|  | struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport; | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 nport = rxport->nport; | 
|  | int ret; | 
|  |  | 
|  | ret = media_entity_get_fwnode_pad(&subdev->entity, | 
|  | rxport->source.ep_fwnode, | 
|  | MEDIA_PAD_FL_SOURCE); | 
|  | if (ret < 0) { | 
|  | dev_err(dev, "Failed to find pad for %s\n", subdev->name); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | rxport->source.sd = subdev; | 
|  | rxport->source.pad = ret; | 
|  |  | 
|  | ret = media_create_pad_link(&rxport->source.sd->entity, | 
|  | rxport->source.pad, &priv->sd.entity, nport, | 
|  | MEDIA_LNK_FL_ENABLED | | 
|  | MEDIA_LNK_FL_IMMUTABLE); | 
|  | if (ret) { | 
|  | dev_err(dev, "Unable to link %s:%u -> %s:%u\n", | 
|  | rxport->source.sd->name, rxport->source.pad, | 
|  | priv->sd.name, nport); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | if (!it.rxport->source.sd) { | 
|  | dev_dbg(dev, "Waiting for more subdevs to be bound\n"); | 
|  | return 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ub960_notify_unbind(struct v4l2_async_notifier *notifier, | 
|  | struct v4l2_subdev *subdev, | 
|  | struct v4l2_async_connection *asd) | 
|  | { | 
|  | struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport; | 
|  |  | 
|  | rxport->source.sd = NULL; | 
|  | } | 
|  |  | 
|  | static const struct v4l2_async_notifier_operations ub960_notify_ops = { | 
|  | .bound = ub960_notify_bound, | 
|  | .unbind = ub960_notify_unbind, | 
|  | }; | 
|  |  | 
|  | static int ub960_v4l2_notifier_register(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | int ret; | 
|  |  | 
|  | v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd); | 
|  |  | 
|  | for_each_active_rxport(priv, it) { | 
|  | struct ub960_asd *asd; | 
|  |  | 
|  | asd = v4l2_async_nf_add_fwnode(&priv->notifier, | 
|  | it.rxport->source.ep_fwnode, | 
|  | struct ub960_asd); | 
|  | if (IS_ERR(asd)) { | 
|  | dev_err(dev, "Failed to add subdev for source %u: %pe", | 
|  | it.nport, asd); | 
|  | v4l2_async_nf_cleanup(&priv->notifier); | 
|  | return PTR_ERR(asd); | 
|  | } | 
|  |  | 
|  | asd->rxport = it.rxport; | 
|  | } | 
|  |  | 
|  | priv->notifier.ops = &ub960_notify_ops; | 
|  |  | 
|  | ret = v4l2_async_nf_register(&priv->notifier); | 
|  | if (ret) { | 
|  | dev_err(dev, "Failed to register subdev_notifier"); | 
|  | v4l2_async_nf_cleanup(&priv->notifier); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ub960_v4l2_notifier_unregister(struct ub960_data *priv) | 
|  | { | 
|  | v4l2_async_nf_unregister(&priv->notifier); | 
|  | v4l2_async_nf_cleanup(&priv->notifier); | 
|  | } | 
|  |  | 
|  | static int ub960_create_subdev(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | unsigned int i; | 
|  | int ret; | 
|  |  | 
|  | v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops); | 
|  | priv->sd.internal_ops = &ub960_internal_ops; | 
|  |  | 
|  | v4l2_ctrl_handler_init(&priv->ctrl_handler, 1); | 
|  | priv->sd.ctrl_handler = &priv->ctrl_handler; | 
|  |  | 
|  | v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ, | 
|  | ARRAY_SIZE(priv->tx_link_freq) - 1, 0, | 
|  | priv->tx_link_freq); | 
|  |  | 
|  | if (priv->ctrl_handler.error) { | 
|  | ret = priv->ctrl_handler.error; | 
|  | goto err_free_ctrl; | 
|  | } | 
|  |  | 
|  | priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE | | 
|  | V4L2_SUBDEV_FL_STREAMS; | 
|  | priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE; | 
|  | priv->sd.entity.ops = &ub960_entity_ops; | 
|  |  | 
|  | for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) { | 
|  | priv->pads[i].flags = ub960_pad_is_sink(priv, i) ? | 
|  | MEDIA_PAD_FL_SINK : | 
|  | MEDIA_PAD_FL_SOURCE; | 
|  | } | 
|  |  | 
|  | ret = media_entity_pads_init(&priv->sd.entity, | 
|  | priv->hw_data->num_rxports + | 
|  | priv->hw_data->num_txports, | 
|  | priv->pads); | 
|  | if (ret) | 
|  | goto err_free_ctrl; | 
|  |  | 
|  | priv->sd.state_lock = priv->sd.ctrl_handler->lock; | 
|  |  | 
|  | ret = v4l2_subdev_init_finalize(&priv->sd); | 
|  | if (ret) | 
|  | goto err_entity_cleanup; | 
|  |  | 
|  | ret = ub960_v4l2_notifier_register(priv); | 
|  | if (ret) { | 
|  | dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret); | 
|  | goto err_subdev_cleanup; | 
|  | } | 
|  |  | 
|  | ret = v4l2_async_register_subdev(&priv->sd); | 
|  | if (ret) { | 
|  | dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret); | 
|  | goto err_unreg_notif; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_unreg_notif: | 
|  | ub960_v4l2_notifier_unregister(priv); | 
|  | err_subdev_cleanup: | 
|  | v4l2_subdev_cleanup(&priv->sd); | 
|  | err_entity_cleanup: | 
|  | media_entity_cleanup(&priv->sd.entity); | 
|  | err_free_ctrl: | 
|  | v4l2_ctrl_handler_free(&priv->ctrl_handler); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void ub960_destroy_subdev(struct ub960_data *priv) | 
|  | { | 
|  | ub960_v4l2_notifier_unregister(priv); | 
|  | v4l2_async_unregister_subdev(&priv->sd); | 
|  |  | 
|  | v4l2_subdev_cleanup(&priv->sd); | 
|  |  | 
|  | media_entity_cleanup(&priv->sd.entity); | 
|  | v4l2_ctrl_handler_free(&priv->ctrl_handler); | 
|  | } | 
|  |  | 
|  | static const struct regmap_config ub960_regmap_config = { | 
|  | .name = "ds90ub960", | 
|  |  | 
|  | .reg_bits = 8, | 
|  | .val_bits = 8, | 
|  |  | 
|  | .max_register = 0xff, | 
|  |  | 
|  | /* | 
|  | * We do locking in the driver to cover the TX/RX port selection and the | 
|  | * indirect register access. | 
|  | */ | 
|  | .disable_locking = true, | 
|  | }; | 
|  |  | 
|  | static int ub960_get_hw_resources(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  |  | 
|  | priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config); | 
|  | if (IS_ERR(priv->regmap)) | 
|  | return PTR_ERR(priv->regmap); | 
|  |  | 
|  | priv->vddio = devm_regulator_get(dev, "vddio"); | 
|  | if (IS_ERR(priv->vddio)) | 
|  | return dev_err_probe(dev, PTR_ERR(priv->vddio), | 
|  | "cannot get VDDIO regulator\n"); | 
|  |  | 
|  | /* get power-down pin from DT */ | 
|  | priv->pd_gpio = | 
|  | devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH); | 
|  | if (IS_ERR(priv->pd_gpio)) | 
|  | return dev_err_probe(dev, PTR_ERR(priv->pd_gpio), | 
|  | "Cannot get powerdown GPIO\n"); | 
|  |  | 
|  | priv->refclk = devm_clk_get(dev, "refclk"); | 
|  | if (IS_ERR(priv->refclk)) | 
|  | return dev_err_probe(dev, PTR_ERR(priv->refclk), | 
|  | "Cannot get REFCLK\n"); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ub960_enable_core_hw(struct ub960_data *priv) | 
|  | { | 
|  | struct device *dev = &priv->client->dev; | 
|  | u8 rev_mask; | 
|  | int ret; | 
|  | u8 dev_sts; | 
|  | u8 refclk_freq; | 
|  |  | 
|  | ret = regulator_enable(priv->vddio); | 
|  | if (ret) | 
|  | return dev_err_probe(dev, ret, | 
|  | "failed to enable VDDIO regulator\n"); | 
|  |  | 
|  | ret = clk_prepare_enable(priv->refclk); | 
|  | if (ret) { | 
|  | dev_err_probe(dev, ret, "Failed to enable refclk\n"); | 
|  | goto err_disable_vddio; | 
|  | } | 
|  |  | 
|  | if (priv->pd_gpio) { | 
|  | gpiod_set_value_cansleep(priv->pd_gpio, 1); | 
|  | /* wait min 2 ms for reset to complete */ | 
|  | fsleep(2000); | 
|  | gpiod_set_value_cansleep(priv->pd_gpio, 0); | 
|  | /* wait min 2 ms for power up to finish */ | 
|  | fsleep(2000); | 
|  | } | 
|  |  | 
|  | ret = ub960_reset(priv, true); | 
|  | if (ret) | 
|  | goto err_pd_gpio; | 
|  |  | 
|  | /* Runtime check register accessibility */ | 
|  | ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask, NULL); | 
|  | if (ret) { | 
|  | dev_err_probe(dev, ret, "Cannot read first register, abort\n"); | 
|  | goto err_pd_gpio; | 
|  | } | 
|  |  | 
|  | dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model, | 
|  | rev_mask); | 
|  |  | 
|  | ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts, NULL); | 
|  | if (ret) | 
|  | goto err_pd_gpio; | 
|  |  | 
|  | if (priv->hw_data->is_ub9702) | 
|  | ret = ub960_read(priv, UB9702_SR_REFCLK_FREQ, &refclk_freq, | 
|  | NULL); | 
|  | else | 
|  | ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq, | 
|  | NULL); | 
|  | if (ret) | 
|  | goto err_pd_gpio; | 
|  |  | 
|  | dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n", | 
|  | !!(dev_sts & BIT(4)), refclk_freq, | 
|  | clk_get_rate(priv->refclk) / HZ_PER_MHZ); | 
|  |  | 
|  | /* Disable all RX ports by default */ | 
|  | ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0, NULL); | 
|  | if (ret) | 
|  | goto err_pd_gpio; | 
|  |  | 
|  | /* release GPIO lock */ | 
|  | if (priv->hw_data->is_ub9702) { | 
|  | ret = ub960_update_bits(priv, UB960_SR_RESET, | 
|  | UB960_SR_RESET_GPIO_LOCK_RELEASE, | 
|  | UB960_SR_RESET_GPIO_LOCK_RELEASE, | 
|  | NULL); | 
|  | if (ret) | 
|  | goto err_pd_gpio; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_pd_gpio: | 
|  | gpiod_set_value_cansleep(priv->pd_gpio, 1); | 
|  | clk_disable_unprepare(priv->refclk); | 
|  | err_disable_vddio: | 
|  | regulator_disable(priv->vddio); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void ub960_disable_core_hw(struct ub960_data *priv) | 
|  | { | 
|  | gpiod_set_value_cansleep(priv->pd_gpio, 1); | 
|  | clk_disable_unprepare(priv->refclk); | 
|  | regulator_disable(priv->vddio); | 
|  | } | 
|  |  | 
|  | static int ub960_probe(struct i2c_client *client) | 
|  | { | 
|  | struct device *dev = &client->dev; | 
|  | struct ub960_data *priv; | 
|  | int ret; | 
|  |  | 
|  | priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL); | 
|  | if (!priv) | 
|  | return -ENOMEM; | 
|  |  | 
|  | priv->client = client; | 
|  |  | 
|  | priv->hw_data = device_get_match_data(dev); | 
|  |  | 
|  | mutex_init(&priv->reg_lock); | 
|  |  | 
|  | INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work); | 
|  |  | 
|  | /* | 
|  | * Initialize these to invalid values so that the first reg writes will | 
|  | * configure the target. | 
|  | */ | 
|  | priv->reg_current.indirect_target = 0xff; | 
|  | priv->reg_current.rxport = 0xff; | 
|  | priv->reg_current.txport = 0xff; | 
|  |  | 
|  | ret = ub960_get_hw_resources(priv); | 
|  | if (ret) | 
|  | goto err_mutex_destroy; | 
|  |  | 
|  | ret = ub960_enable_core_hw(priv); | 
|  | if (ret) | 
|  | goto err_mutex_destroy; | 
|  |  | 
|  | ret = ub960_parse_dt(priv); | 
|  | if (ret) | 
|  | goto err_disable_core_hw; | 
|  |  | 
|  | ret = ub960_init_tx_ports(priv); | 
|  | if (ret) | 
|  | goto err_free_ports; | 
|  |  | 
|  | ret = ub960_rxport_enable_vpocs(priv); | 
|  | if (ret) | 
|  | goto err_free_ports; | 
|  |  | 
|  | if (priv->hw_data->is_ub9702) | 
|  | ret = ub960_init_rx_ports_ub9702(priv); | 
|  | else | 
|  | ret = ub960_init_rx_ports_ub960(priv); | 
|  |  | 
|  | if (ret) | 
|  | goto err_disable_vpocs; | 
|  |  | 
|  | ret = ub960_init_atr(priv); | 
|  | if (ret) | 
|  | goto err_disable_vpocs; | 
|  |  | 
|  | ret = ub960_rxport_add_serializers(priv); | 
|  | if (ret) | 
|  | goto err_uninit_atr; | 
|  |  | 
|  | ret = ub960_create_subdev(priv); | 
|  | if (ret) | 
|  | goto err_free_sers; | 
|  |  | 
|  | if (client->irq) | 
|  | dev_warn(dev, "irq support not implemented, using polling\n"); | 
|  |  | 
|  | schedule_delayed_work(&priv->poll_work, | 
|  | msecs_to_jiffies(UB960_POLL_TIME_MS)); | 
|  |  | 
|  | #ifdef UB960_DEBUG_I2C_RX_ID | 
|  | for_each_rxport(priv, it) | 
|  | ub960_write(priv, UB960_SR_I2C_RX_ID(it.nport), | 
|  | (UB960_DEBUG_I2C_RX_ID + it.nport) << 1, NULL); | 
|  | #endif | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_free_sers: | 
|  | ub960_rxport_remove_serializers(priv); | 
|  | err_uninit_atr: | 
|  | ub960_uninit_atr(priv); | 
|  | err_disable_vpocs: | 
|  | ub960_rxport_disable_vpocs(priv); | 
|  | err_free_ports: | 
|  | ub960_rxport_free_ports(priv); | 
|  | ub960_txport_free_ports(priv); | 
|  | err_disable_core_hw: | 
|  | ub960_disable_core_hw(priv); | 
|  | err_mutex_destroy: | 
|  | mutex_destroy(&priv->reg_lock); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void ub960_remove(struct i2c_client *client) | 
|  | { | 
|  | struct v4l2_subdev *sd = i2c_get_clientdata(client); | 
|  | struct ub960_data *priv = sd_to_ub960(sd); | 
|  |  | 
|  | cancel_delayed_work_sync(&priv->poll_work); | 
|  |  | 
|  | ub960_destroy_subdev(priv); | 
|  | ub960_rxport_remove_serializers(priv); | 
|  | ub960_uninit_atr(priv); | 
|  | ub960_rxport_disable_vpocs(priv); | 
|  | ub960_rxport_free_ports(priv); | 
|  | ub960_txport_free_ports(priv); | 
|  | ub960_disable_core_hw(priv); | 
|  | mutex_destroy(&priv->reg_lock); | 
|  | } | 
|  |  | 
|  | static const struct ub960_hw_data ds90ub960_hw = { | 
|  | .model = "ub960", | 
|  | .num_rxports = 4, | 
|  | .num_txports = 2, | 
|  | }; | 
|  |  | 
|  | static const struct ub960_hw_data ds90ub9702_hw = { | 
|  | .model = "ub9702", | 
|  | .num_rxports = 4, | 
|  | .num_txports = 2, | 
|  | .is_ub9702 = true, | 
|  | .is_fpdlink4 = true, | 
|  | }; | 
|  |  | 
|  | static const struct i2c_device_id ub960_id[] = { | 
|  | { "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw }, | 
|  | { "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw }, | 
|  | {} | 
|  | }; | 
|  | MODULE_DEVICE_TABLE(i2c, ub960_id); | 
|  |  | 
|  | static const struct of_device_id ub960_dt_ids[] = { | 
|  | { .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw }, | 
|  | { .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw }, | 
|  | {} | 
|  | }; | 
|  | MODULE_DEVICE_TABLE(of, ub960_dt_ids); | 
|  |  | 
|  | static struct i2c_driver ds90ub960_driver = { | 
|  | .probe		= ub960_probe, | 
|  | .remove		= ub960_remove, | 
|  | .id_table	= ub960_id, | 
|  | .driver = { | 
|  | .name	= "ds90ub960", | 
|  | .of_match_table = ub960_dt_ids, | 
|  | }, | 
|  | }; | 
|  | module_i2c_driver(ds90ub960_driver); | 
|  |  | 
|  | MODULE_LICENSE("GPL"); | 
|  | MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver"); | 
|  | MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>"); | 
|  | MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>"); | 
|  | MODULE_IMPORT_NS("I2C_ATR"); |