// SPDX-License-Identifier: GPL-2.0 /* Copyright (c) 2024 Rockchip Electronics Co., Ltd. */ #include #include #include #include #include #include #include "dev.h" #include "regs.h" #include "common.h" #include "isp_stats.h" #include "isp_stats_v33.h" #include "isp_params_v33.h" #define ISP33_3A_MEAS_DONE BIT(31) static void isp3_module_done(struct rkisp_isp_stats_vdev *stats_vdev, u32 reg, u32 value) { void __iomem *base = stats_vdev->dev->hw_dev->base_addr; writel(value, base + reg); } static u32 isp3_stats_read(struct rkisp_isp_stats_vdev *stats_vdev, u32 addr) { return rkisp_read(stats_vdev->dev, addr, true); } static void isp3_stats_write(struct rkisp_isp_stats_vdev *stats_vdev, u32 addr, u32 value) { rkisp_write(stats_vdev->dev, addr, value, true); } static int rkisp_stats_get_sharp_stats(struct rkisp_isp_stats_vdev *stats_vdev, struct rkisp33_stat_buffer *pbuf) { struct isp33_sharp_stat *sharp; u32 i, val; if (!pbuf) return 0; val = isp3_stats_read(stats_vdev, ISP3X_SHARP_EN); if (val & 0x1) { sharp = &pbuf->stat.sharp; for (i = 0; i < ISP33_SHARP_NOISE_CURVE_NUM / 2; i++) { val = isp3_stats_read(stats_vdev, ISP33_SHARP_NOISE_CURVE0 + i * 4); sharp->noise_curve[i * 2] = val & 0x7ff; sharp->noise_curve[i * 2 + 1] = (val >> 16) & 0x7ff; } val = isp3_stats_read(stats_vdev, ISP33_SHARP_NOISE_CURVE8); sharp->noise_curve[i * 2] = val & 0x7ff; pbuf->meas_type |= ISP33_STAT_SHARP; } return 0; } static int rkisp_stats_get_bay3d_stats(struct rkisp_isp_stats_vdev *stats_vdev, struct rkisp33_stat_buffer *pbuf) { struct isp33_bay3d_stat *bay3d; u32 i, val; if (!pbuf) return 0; val = isp3_stats_read(stats_vdev, ISP33_BAY3D_CTRL0); if (val & 0x1) { bay3d = &pbuf->stat.bay3d; val = isp3_stats_read(stats_vdev, ISP33_BAY3D_TNRSUM); bay3d->sigma_num = val; for (i = 0; i < ISP33_BAY3D_TNRSIG_NUM / 2; i++) { val = isp3_stats_read(stats_vdev, ISP33_BAY3D_TNRYO0 + i * 4); bay3d->sigma_y[i * 2] = val & 0xfff; bay3d->sigma_y[i * 2 + 1] = (val >> 16) & 0xfff; } pbuf->meas_type |= ISP33_STAT_BAY3D; } return 0; } static int rkisp_stats_get_hist_stats(struct rkisp_isp_stats_vdev *stats_vdev, struct rkisp33_stat_buffer *pbuf) { struct rkisp_device *dev = stats_vdev->dev; struct rkisp_isp_params_vdev *params = &dev->params_vdev; struct rkisp_isp_params_val_v33 *priv_val = params->priv_val; struct isp33_isp_params_cfg *params_rec = params->isp33_params + dev->unite_index; struct isp33_hist_cfg *arg_rec = ¶ms_rec->others.hist_cfg; struct isp33_hist_stat *hist; int val, i, j, timeout; val = isp3_stats_read(stats_vdev, ISP33_HIST_CTRL); if (val & 0x1) { val = isp3_stats_read(stats_vdev, ISP33_HIST_STAB); arg_rec->stab_frame_cnt0 = val & 0xf; arg_rec->stab_frame_cnt1 = (val & 0xf0) >> 4; for (i = 0; i < priv_val->hist_blk_num; i++) { val = ISP33_IIR_RD_ID(i) | ISP33_IIR_RD_P; isp3_stats_write(stats_vdev, ISP33_HIST_RW, val); timeout = 5; while (timeout--) { val = isp3_stats_read(stats_vdev, ISP33_HIST_RW); if (val & ISP33_IIR_RDATA_VAL) break; udelay(2); } if (timeout < 0) { v4l2_warn(&dev->v4l2_dev, "%s hist read:%d timeout\n", __func__, i); return 0; } for (j = 0; j < ISP33_HIST_IIR_NUM / 2; j++) { val = isp3_stats_read(stats_vdev, ISP33_HIST_IIR0 + 4 * j); arg_rec->iir[i][2 * j] = val & 0x3FF; arg_rec->iir[i][2 * j + 1] = val >> 16; } } if (dev->is_frm_rd) arg_rec->iir_wr = true; if (pbuf) { hist = &pbuf->stat.hist; memcpy(hist->iir, arg_rec->iir, sizeof(hist->iir)); pbuf->meas_type |= ISP33_STAT_HIST; } } return 0; } static int rkisp_stats_get_enh_stats(struct rkisp_isp_stats_vdev *stats_vdev, struct rkisp33_stat_buffer *pbuf) { struct rkisp_device *dev = stats_vdev->dev; struct rkisp_isp_params_vdev *params = &dev->params_vdev; struct rkisp_isp_params_val_v33 *priv_val = params->priv_val; struct isp33_isp_params_cfg *params_rec = params->isp33_params + dev->unite_index; struct isp33_enh_cfg *arg_rec = ¶ms_rec->others.enh_cfg; struct isp33_enh_stat *enh; int val, i, j, timeout; val = isp3_stats_read(stats_vdev, ISP33_ENH_CTRL); if (val & 0x1) { enh = &pbuf->stat.enh; val = isp3_stats_read(stats_vdev, ISP33_ENH_PRE_FRAME); arg_rec->pre_wet_frame_cnt0 = val & 0xf; arg_rec->pre_wet_frame_cnt1 = (val & 0xf0) >> 4; for (i = 0; i < priv_val->enh_row; i++) { val = ISP33_IIR_RD_ID(i) | ISP33_IIR_RD_P; isp3_stats_write(stats_vdev, ISP33_ENH_IIR_RW, val); timeout = 5; while (timeout--) { val = isp3_stats_read(stats_vdev, ISP33_ENH_IIR_RW); if (val & ISP33_IIR_RDATA_VAL) break; udelay(2); } if (timeout < 0) { v4l2_warn(&dev->v4l2_dev, "%s enh read:%d timeout\n", __func__, i); return 0; } for (j = 0; j < priv_val->enh_col / 4; j++) { val = isp3_stats_read(stats_vdev, ISP33_ENH_IIR0 + 4 * j); arg_rec->iir[i][4 * j] = val & 0xFF; arg_rec->iir[i][4 * j + 1] = (val & 0xff00) >> 8; arg_rec->iir[i][4 * j + 2] = (val & 0xff0000) >> 16; arg_rec->iir[i][4 * j + 3] = (val & 0xff000000) >> 24; } } if (dev->is_frm_rd) arg_rec->iir_wr = true; if (pbuf) { enh = &pbuf->stat.enh; memcpy(enh->iir, arg_rec->iir, sizeof(enh->iir)); pbuf->meas_type |= ISP33_STAT_ENH; } } return 0; } static int rkisp_stats_update_buf(struct rkisp_isp_stats_vdev *stats_vdev) { struct rkisp_device *dev = stats_vdev->dev; struct rkisp_buffer *buf; unsigned long flags; u32 size = stats_vdev->vdev_fmt.fmt.meta.buffersize / dev->unite_div; u32 val, addr = 0, offset = 0; int i, ret = 0; spin_lock_irqsave(&stats_vdev->rd_lock, flags); if (!stats_vdev->nxt_buf && !list_empty(&stats_vdev->stat)) { buf = list_first_entry(&stats_vdev->stat, struct rkisp_buffer, queue); list_del(&buf->queue); stats_vdev->nxt_buf = buf; } spin_unlock_irqrestore(&stats_vdev->rd_lock, flags); if (stats_vdev->nxt_buf) { addr = stats_vdev->nxt_buf->buff_addr[0]; if (!dev->hw_dev->is_single) { stats_vdev->cur_buf = stats_vdev->nxt_buf; stats_vdev->nxt_buf = NULL; } } else if (stats_vdev->stats_buf[0].mem_priv) { addr = stats_vdev->stats_buf[0].dma_addr; } else { ret = -EINVAL; } if (addr) { for (i = 0; i < dev->unite_div; i++) { val = addr + i * size; rkisp_idx_write(dev, ISP39_W3A_AEBIG_ADDR, val, i, false); offset = sizeof(struct isp33_rawae_stat) + sizeof(struct isp33_rawhist_stat); val += offset; rkisp_idx_write(dev, ISP39_W3A_AE0_ADDR, val, i, false); val += offset; rkisp_idx_write(dev, ISP39_W3A_AWB_ADDR, val, i, false); } v4l2_dbg(4, rkisp_debug, &dev->v4l2_dev, "%s BASE:0x%x SHD AEBIG:0x%x AE0:0x%x AWB:0x%x\n", __func__, addr, isp3_stats_read(stats_vdev, ISP39_W3A_AEBIG_ADDR_SHD), isp3_stats_read(stats_vdev, ISP39_W3A_AE0_ADDR_SHD), isp3_stats_read(stats_vdev, ISP39_W3A_AWB_ADDR_SHD)); } return ret; } static void rkisp_stats_info2ddr(struct rkisp_isp_stats_vdev *stats_vdev, struct rkisp33_stat_buffer *pbuf) { struct rkisp_device *dev = stats_vdev->dev; struct rkisp_isp_params_val_v33 *priv_val; struct rkisp_dummy_buffer *buf; int idx, buf_fd = -1; u32 reg = 0, ctrl, mask; if (dev->is_aiisp_en) return; priv_val = dev->params_vdev.priv_val; if (!priv_val->buf_info_owner && priv_val->buf_info_idx >= 0) { priv_val->buf_info_idx = -1; rkisp_clear_bits(dev, ISP3X_GAIN_CTRL, ISP3X_GAIN_2DDR_EN, false); rkisp_clear_bits(dev, ISP3X_RAWAWB_CTRL, ISP32_RAWAWB_2DDR_PATH_EN, false); return; } if (priv_val->buf_info_owner == RKISP_INFO2DRR_OWNER_GAIN) { reg = ISP3X_GAIN_CTRL; ctrl = ISP3X_GAIN_2DDR_EN; mask = ISP3X_GAIN_2DDR_EN; } else { reg = ISP3X_RAWAWB_CTRL; ctrl = ISP32_RAWAWB_2DDR_PATH_EN; mask = ISP32_RAWAWB_2DDR_PATH_EN | ISP32_RAWAWB_2DDR_PATH_DS; } idx = priv_val->buf_info_idx; if (idx >= 0) { buf = &priv_val->buf_info[idx]; rkisp_finish_buffer(dev, buf); v4l2_dbg(4, rkisp_debug, &dev->v4l2_dev, "%s data:0x%x 0x%x:0x%x\n", __func__, *(u32 *)buf->vaddr, reg, rkisp_read(dev, reg, true)); if (*(u32 *)buf->vaddr != RKISP_INFO2DDR_BUF_INIT && pbuf && (reg != ISP3X_RAWAWB_CTRL || !(rkisp_read(dev, reg, true) & ISP32_RAWAWB_2DDR_PATH_ERR))) { pbuf->stat.info2ddr.buf_fd = buf->dma_fd; pbuf->stat.info2ddr.owner = priv_val->buf_info_owner; pbuf->meas_type |= ISP33_STAT_INFO2DDR; buf_fd = buf->dma_fd; } else if (reg == ISP3X_RAWAWB_CTRL && rkisp_read(dev, reg, true) & ISP32_RAWAWB_2DDR_PATH_ERR) { v4l2_warn(&dev->v4l2_dev, "rawawb2ddr path error idx:%d\n", idx); } else { u32 v0 = rkisp_read(dev, reg, false); u32 v1 = rkisp_read_reg_cache(dev, reg); if ((v0 & mask) != (v1 & mask)) rkisp_write(dev, reg, v0 | (v1 & mask), false); } if (buf_fd == -1) return; } /* get next unused buf to hw */ for (idx = 0; idx < priv_val->buf_info_cnt; idx++) { buf = &priv_val->buf_info[idx]; if (*(u32 *)buf->vaddr == RKISP_INFO2DDR_BUF_INIT) break; } if (idx == priv_val->buf_info_cnt) { rkisp_clear_bits(dev, reg, ctrl, false); priv_val->buf_info_idx = -1; } else { buf = &priv_val->buf_info[idx]; rkisp_write(dev, ISP3X_MI_GAIN_WR_BASE, buf->dma_addr, false); if (dev->hw_dev->is_single) rkisp_write(dev, ISP3X_MI_WR_CTRL2, ISP3X_GAINSELF_UPD, true); if (priv_val->buf_info_idx < 0) rkisp_set_bits(dev, reg, 0, ctrl, false); priv_val->buf_info_idx = idx; } } static void rkisp_stats_send_meas_v33(struct rkisp_isp_stats_vdev *stats_vdev, struct rkisp_isp_readout_work *meas_work) { struct rkisp_isp_params_vdev *params_vdev = &stats_vdev->dev->params_vdev; struct rkisp_device *dev = stats_vdev->dev; struct rkisp_buffer *cur_buf = stats_vdev->cur_buf; struct rkisp33_stat_buffer *cur_stat_buf = NULL; u32 size = stats_vdev->vdev_fmt.fmt.meta.buffersize; u32 val, w3a_int, cur_frame_id = meas_work->frame_id; bool is_dummy = false; unsigned long flags; w3a_int = isp3_stats_read(stats_vdev, ISP39_W3A_INT_STAT); if (w3a_int) isp3_stats_write(stats_vdev, ISP39_W3A_INT_STAT, w3a_int); if (!stats_vdev->rdbk_drop) { if (!cur_buf && stats_vdev->stats_buf[0].mem_priv) { rkisp_finish_buffer(stats_vdev->dev, &stats_vdev->stats_buf[0]); cur_stat_buf = stats_vdev->stats_buf[0].vaddr; cur_stat_buf->frame_id = cur_frame_id; cur_stat_buf->params_id = params_vdev->cur_frame_id; is_dummy = true; } else if (cur_buf) { cur_stat_buf = cur_buf->vaddr[0]; cur_stat_buf->frame_id = cur_frame_id; cur_stat_buf->params_id = params_vdev->cur_frame_id; } /* buffer done when frame of right handle */ if (dev->unite_div > ISP_UNITE_DIV1) { if (dev->unite_index == ISP_UNITE_LEFT) { cur_buf = NULL; is_dummy = false; } else if (cur_stat_buf) { cur_stat_buf = (void *)cur_stat_buf + size / 2; cur_stat_buf->frame_id = cur_frame_id; cur_stat_buf->params_id = params_vdev->cur_frame_id; } } if (dev->unite_div < ISP_UNITE_DIV2 || dev->unite_index == ISP_UNITE_RIGHT) { /* config buf for next frame */ stats_vdev->cur_buf = NULL; if (stats_vdev->nxt_buf) { stats_vdev->cur_buf = stats_vdev->nxt_buf; stats_vdev->nxt_buf = NULL; } rkisp_stats_update_buf(stats_vdev); } } else { cur_buf = NULL; } if (w3a_int & ISP39_W3A_INT_ERR_MASK) { val = isp3_stats_read(stats_vdev, ISP3X_RAWAE_BIG1_BASE); if (val & ISP33_3A_MEAS_DONE) isp3_module_done(stats_vdev, ISP3X_RAWAE_BIG1_BASE, val); val = isp3_stats_read(stats_vdev, ISP3X_RAWAE_LITE_BASE); if (val & ISP33_3A_MEAS_DONE) isp3_module_done(stats_vdev, ISP3X_RAWAE_LITE_BASE, val); val = isp3_stats_read(stats_vdev, ISP3X_RAWHIST_BIG1_BASE); if (val & ISP33_3A_MEAS_DONE) isp3_module_done(stats_vdev, ISP3X_RAWHIST_BIG1_BASE, val); val = isp3_stats_read(stats_vdev, ISP3X_RAWHIST_LITE_BASE); if (val & ISP33_3A_MEAS_DONE) isp3_module_done(stats_vdev, ISP3X_RAWHIST_LITE_BASE, val); val = isp3_stats_read(stats_vdev, ISP3X_RAWAWB_BASE); if (val & ISP33_3A_MEAS_DONE) isp3_module_done(stats_vdev, ISP3X_RAWAWB_BASE, val); v4l2_warn(&dev->v4l2_dev, "id:%d stats seq:%d error:0x%x overflow(aebig:%d ae0:%d awb:%d wcfifo(wr:%d rd:%d))\n", dev->unite_index, cur_frame_id, w3a_int, !!(w3a_int & ISP39_W3A_INT_AEBIG_OVF), !!(w3a_int & ISP39_W3A_INT_AE0_OVF), !!(w3a_int & ISP39_W3A_INT_AWB_OVF), !!(w3a_int & ISP39_W3A_INT_WCFIFO_WR_ERR), !!(w3a_int & ISP39_W3A_INT_WCFIFO_RD_ERR)); } else { if (meas_work->isp3a_ris & ISP3X_3A_RAWAWB && cur_stat_buf) cur_stat_buf->meas_type |= ISP33_STAT_RAWAWB; if (meas_work->isp3a_ris & ISP3X_3A_RAWAE_BIG && cur_stat_buf) cur_stat_buf->meas_type |= ISP33_STAT_RAWAE3; if (meas_work->isp3a_ris & ISP3X_3A_RAWHIST_BIG && cur_stat_buf) cur_stat_buf->meas_type |= ISP33_STAT_RAWHST3; if (meas_work->isp3a_ris & ISP3X_3A_RAWAE_CH0 && cur_stat_buf) cur_stat_buf->meas_type |= ISP33_STAT_RAWAE0; if (meas_work->isp3a_ris & ISP3X_3A_RAWHIST_CH0 && cur_stat_buf) cur_stat_buf->meas_type |= ISP33_STAT_RAWHST0; } if (meas_work->isp_ris & ISP3X_FRAME) { rkisp_stats_get_bay3d_stats(stats_vdev, cur_stat_buf); rkisp_stats_get_sharp_stats(stats_vdev, cur_stat_buf); rkisp_stats_get_enh_stats(stats_vdev, cur_stat_buf); rkisp_stats_get_hist_stats(stats_vdev, cur_stat_buf); } if (cur_stat_buf && (dev->is_first_double || dev->is_wait_aiq)) { cur_stat_buf->meas_type |= ISP33_STAT_RTT_FST; dev_info(dev->dev, "stats seq:%d meas_type:0x%x for fast\n", cur_frame_id, cur_stat_buf->meas_type); } if (is_dummy) { spin_lock_irqsave(&stats_vdev->rd_lock, flags); if (!list_empty(&stats_vdev->stat)) { cur_buf = list_first_entry(&stats_vdev->stat, struct rkisp_buffer, queue); list_del(&cur_buf->queue); } spin_unlock_irqrestore(&stats_vdev->rd_lock, flags); if (cur_buf) { memcpy(cur_buf->vaddr[0], stats_vdev->stats_buf[0].vaddr, size); cur_stat_buf = cur_buf->vaddr[0]; } } if (cur_buf && cur_stat_buf) { cur_stat_buf->frame_id = cur_frame_id; cur_stat_buf->params_id = params_vdev->cur_frame_id; cur_stat_buf->stat.info2ddr.buf_fd = -1; cur_stat_buf->stat.info2ddr.owner = 0; rkisp_stats_info2ddr(stats_vdev, cur_stat_buf); vb2_set_plane_payload(&cur_buf->vb.vb2_buf, 0, size); cur_buf->vb.sequence = cur_frame_id; cur_buf->vb.vb2_buf.timestamp = meas_work->timestamp; vb2_buffer_done(&cur_buf->vb.vb2_buf, VB2_BUF_STATE_DONE); } v4l2_dbg(4, rkisp_debug, &stats_vdev->dev->v4l2_dev, "%s seq:%d params_id:%d ris:0x%x buf:%p meas_type:0x%x\n", __func__, cur_frame_id, params_vdev->cur_frame_id, meas_work->isp3a_ris, cur_buf, !cur_stat_buf ? 0 : cur_stat_buf->meas_type); } static void rkisp_stats_isr_v33(struct rkisp_isp_stats_vdev *stats_vdev, u32 isp_ris, u32 isp3a_ris) { struct rkisp_isp_readout_work work; u32 cur_frame_id, isp_mis_tmp = 0; u32 temp_isp3a_ris; rkisp_dmarx_get_frame(stats_vdev->dev, &cur_frame_id, NULL, NULL, true); temp_isp3a_ris = isp3_stats_read(stats_vdev, ISP3X_ISP_3A_RIS); isp_mis_tmp = temp_isp3a_ris; if (isp_mis_tmp) { isp3_stats_write(stats_vdev, ISP3X_ISP_3A_ICR, isp_mis_tmp); isp_mis_tmp &= isp3_stats_read(stats_vdev, ISP3X_ISP_3A_MIS); if (isp_mis_tmp) v4l2_err(stats_vdev->vnode.vdev.v4l2_dev, "isp3A icr 3A info err: 0x%x 0x%x\n", isp_mis_tmp, isp3a_ris); } if (isp_ris & ISP3X_FRAME) { work.readout = RKISP_ISP_READOUT_MEAS; work.frame_id = cur_frame_id; work.isp_ris = isp_ris; work.isp3a_ris = temp_isp3a_ris; work.timestamp = ktime_get_ns(); rkisp_stats_send_meas_v33(stats_vdev, &work); } } static void rkisp_get_stat_size_v33(struct rkisp_isp_stats_vdev *stats_vdev, unsigned int sizes[]) { int mult = stats_vdev->dev->unite_div; sizes[0] = ALIGN(sizeof(struct rkisp33_stat_buffer), 16); sizes[0] *= mult; stats_vdev->vdev_fmt.fmt.meta.buffersize = sizes[0]; } static struct rkisp_isp_stats_ops rkisp_isp_stats_ops_tbl = { .isr_hdl = rkisp_stats_isr_v33, .send_meas = rkisp_stats_send_meas_v33, .get_stat_size = rkisp_get_stat_size_v33, }; void rkisp_stats_first_ddr_config_v33(struct rkisp_isp_stats_vdev *stats_vdev) { struct rkisp_device *dev = stats_vdev->dev; u32 val, size = 0, div = dev->unite_div; if (dev->isp_sdev.in_fmt.fmt_type == FMT_YUV) return; rkisp_get_stat_size_v33(stats_vdev, &size); stats_vdev->stats_buf[0].is_need_vaddr = true; stats_vdev->stats_buf[0].size = size; if (rkisp_alloc_buffer(dev, &stats_vdev->stats_buf[0])) v4l2_warn(&dev->v4l2_dev, "stats alloc buf fail\n"); else memset(stats_vdev->stats_buf[0].vaddr, 0, size); if (rkisp_stats_update_buf(stats_vdev) < 0) { v4l2_err(&dev->v4l2_dev, "no stats buf to enable w3a\n"); return; } rkisp_unite_set_bits(dev, ISP3X_SWS_CFG, 0, ISP3X_3A_DDR_WRITE_EN, false); val = ISP39_W3A_EN | ISP39_W3A_AUTO_CLR_EN | ISP39_W3A_FORCE_UPD; rkisp_unite_write(dev, ISP39_W3A_CTRL0, val, false); rkisp_unite_write(dev, ISP39_W3A_WR_SIZE, size / div, false); if (stats_vdev->nxt_buf) { stats_vdev->cur_buf = stats_vdev->nxt_buf; stats_vdev->nxt_buf = NULL; } } void rkisp_stats_next_ddr_config_v33(struct rkisp_isp_stats_vdev *stats_vdev) { struct rkisp_device *dev = stats_vdev->dev; struct rkisp_hw_dev *hw = dev->hw_dev; if (!stats_vdev->streamon || dev->isp_sdev.in_fmt.fmt_type == FMT_YUV) return; /* pingpong buf */ if (hw->is_single) rkisp_stats_update_buf(stats_vdev); } void rkisp_init_stats_vdev_v33(struct rkisp_isp_stats_vdev *stats_vdev) { stats_vdev->ops = &rkisp_isp_stats_ops_tbl; } void rkisp_uninit_stats_vdev_v33(struct rkisp_isp_stats_vdev *stats_vdev) { }