OSDN Git Service

media: rkvdec: Add the VP9 backend
authorAndrzej Pietrasiewicz <andrzej.p@collabora.com>
Wed, 17 Nov 2021 11:52:56 +0000 (11:52 +0000)
committerMauro Carvalho Chehab <mchehab+huawei@kernel.org>
Mon, 22 Nov 2021 07:49:23 +0000 (07:49 +0000)
The Rockchip VDEC supports VP9 profile 0 up to 4096x2304@30fps. Add
a backend for this new format.

Signed-off-by: Boris Brezillon <boris.brezillon@collabora.com>
Signed-off-by: Ezequiel Garcia <ezequiel@collabora.com>
Signed-off-by: Adrian Ratiu <adrian.ratiu@collabora.com>
Co-developed-by: Andrzej Pietrasiewicz <andrzej.p@collabora.com>
Signed-off-by: Andrzej Pietrasiewicz <andrzej.p@collabora.com>
Signed-off-by: Hans Verkuil <hverkuil-cisco@xs4all.nl>
Signed-off-by: Mauro Carvalho Chehab <mchehab+huawei@kernel.org>
drivers/staging/media/rkvdec/Kconfig
drivers/staging/media/rkvdec/Makefile
drivers/staging/media/rkvdec/rkvdec-vp9.c [new file with mode: 0644]
drivers/staging/media/rkvdec/rkvdec.c
drivers/staging/media/rkvdec/rkvdec.h

index c02199b..dc7292f 100644 (file)
@@ -9,6 +9,7 @@ config VIDEO_ROCKCHIP_VDEC
        select VIDEOBUF2_VMALLOC
        select V4L2_MEM2MEM_DEV
        select V4L2_H264
+       select V4L2_VP9
        help
          Support for the Rockchip Video Decoder IP present on Rockchip SoCs,
          which accelerates video decoding.
index c08fed0..cb86b42 100644 (file)
@@ -1,3 +1,3 @@
 obj-$(CONFIG_VIDEO_ROCKCHIP_VDEC) += rockchip-vdec.o
 
-rockchip-vdec-y += rkvdec.o rkvdec-h264.o
+rockchip-vdec-y += rkvdec.o rkvdec-h264.o rkvdec-vp9.o
diff --git a/drivers/staging/media/rkvdec/rkvdec-vp9.c b/drivers/staging/media/rkvdec/rkvdec-vp9.c
new file mode 100644 (file)
index 0000000..311a126
--- /dev/null
@@ -0,0 +1,1072 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Rockchip Video Decoder VP9 backend
+ *
+ * Copyright (C) 2019 Collabora, Ltd.
+ *     Boris Brezillon <boris.brezillon@collabora.com>
+ * Copyright (C) 2021 Collabora, Ltd.
+ *     Andrzej Pietrasiewicz <andrzej.p@collabora.com>
+ *
+ * Copyright (C) 2016 Rockchip Electronics Co., Ltd.
+ *     Alpha Lin <Alpha.Lin@rock-chips.com>
+ */
+
+/*
+ * For following the vp9 spec please start reading this driver
+ * code from rkvdec_vp9_run() followed by rkvdec_vp9_done().
+ */
+
+#include <linux/kernel.h>
+#include <linux/vmalloc.h>
+#include <media/v4l2-mem2mem.h>
+#include <media/v4l2-vp9.h>
+
+#include "rkvdec.h"
+#include "rkvdec-regs.h"
+
+#define RKVDEC_VP9_PROBE_SIZE          4864
+#define RKVDEC_VP9_COUNT_SIZE          13232
+#define RKVDEC_VP9_MAX_SEGMAP_SIZE     73728
+
+struct rkvdec_vp9_intra_mode_probs {
+       u8 y_mode[105];
+       u8 uv_mode[23];
+};
+
+struct rkvdec_vp9_intra_only_frame_probs {
+       u8 coef_intra[4][2][128];
+       struct rkvdec_vp9_intra_mode_probs intra_mode[10];
+};
+
+struct rkvdec_vp9_inter_frame_probs {
+       u8 y_mode[4][9];
+       u8 comp_mode[5];
+       u8 comp_ref[5];
+       u8 single_ref[5][2];
+       u8 inter_mode[7][3];
+       u8 interp_filter[4][2];
+       u8 padding0[11];
+       u8 coef[2][4][2][128];
+       u8 uv_mode_0_2[3][9];
+       u8 padding1[5];
+       u8 uv_mode_3_5[3][9];
+       u8 padding2[5];
+       u8 uv_mode_6_8[3][9];
+       u8 padding3[5];
+       u8 uv_mode_9[9];
+       u8 padding4[7];
+       u8 padding5[16];
+       struct {
+               u8 joint[3];
+               u8 sign[2];
+               u8 classes[2][10];
+               u8 class0_bit[2];
+               u8 bits[2][10];
+               u8 class0_fr[2][2][3];
+               u8 fr[2][3];
+               u8 class0_hp[2];
+               u8 hp[2];
+       } mv;
+};
+
+struct rkvdec_vp9_probs {
+       u8 partition[16][3];
+       u8 pred[3];
+       u8 tree[7];
+       u8 skip[3];
+       u8 tx32[2][3];
+       u8 tx16[2][2];
+       u8 tx8[2][1];
+       u8 is_inter[4];
+       /* 128 bit alignment */
+       u8 padding0[3];
+       union {
+               struct rkvdec_vp9_inter_frame_probs inter;
+               struct rkvdec_vp9_intra_only_frame_probs intra_only;
+       };
+};
+
+/* Data structure describing auxiliary buffer format. */
+struct rkvdec_vp9_priv_tbl {
+       struct rkvdec_vp9_probs probs;
+       u8 segmap[2][RKVDEC_VP9_MAX_SEGMAP_SIZE];
+};
+
+struct rkvdec_vp9_refs_counts {
+       u32 eob[2];
+       u32 coeff[3];
+};
+
+struct rkvdec_vp9_inter_frame_symbol_counts {
+       u32 partition[16][4];
+       u32 skip[3][2];
+       u32 inter[4][2];
+       u32 tx32p[2][4];
+       u32 tx16p[2][4];
+       u32 tx8p[2][2];
+       u32 y_mode[4][10];
+       u32 uv_mode[10][10];
+       u32 comp[5][2];
+       u32 comp_ref[5][2];
+       u32 single_ref[5][2][2];
+       u32 mv_mode[7][4];
+       u32 filter[4][3];
+       u32 mv_joint[4];
+       u32 sign[2][2];
+       /* add 1 element for align */
+       u32 classes[2][11 + 1];
+       u32 class0[2][2];
+       u32 bits[2][10][2];
+       u32 class0_fp[2][2][4];
+       u32 fp[2][4];
+       u32 class0_hp[2][2];
+       u32 hp[2][2];
+       struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6];
+};
+
+struct rkvdec_vp9_intra_frame_symbol_counts {
+       u32 partition[4][4][4];
+       u32 skip[3][2];
+       u32 intra[4][2];
+       u32 tx32p[2][4];
+       u32 tx16p[2][4];
+       u32 tx8p[2][2];
+       struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6];
+};
+
+struct rkvdec_vp9_run {
+       struct rkvdec_run base;
+       const struct v4l2_ctrl_vp9_frame *decode_params;
+};
+
+struct rkvdec_vp9_frame_info {
+       u32 valid : 1;
+       u32 segmapid : 1;
+       u32 frame_context_idx : 2;
+       u32 reference_mode : 2;
+       u32 tx_mode : 3;
+       u32 interpolation_filter : 3;
+       u32 flags;
+       u64 timestamp;
+       struct v4l2_vp9_segmentation seg;
+       struct v4l2_vp9_loop_filter lf;
+};
+
+struct rkvdec_vp9_ctx {
+       struct rkvdec_aux_buf priv_tbl;
+       struct rkvdec_aux_buf count_tbl;
+       struct v4l2_vp9_frame_symbol_counts inter_cnts;
+       struct v4l2_vp9_frame_symbol_counts intra_cnts;
+       struct v4l2_vp9_frame_context probability_tables;
+       struct v4l2_vp9_frame_context frame_context[4];
+       struct rkvdec_vp9_frame_info cur;
+       struct rkvdec_vp9_frame_info last;
+};
+
+static void write_coeff_plane(const u8 coef[6][6][3], u8 *coeff_plane)
+{
+       unsigned int idx = 0, byte_count = 0;
+       int k, m, n;
+       u8 p;
+
+       for (k = 0; k < 6; k++) {
+               for (m = 0; m < 6; m++) {
+                       for (n = 0; n < 3; n++) {
+                               p = coef[k][m][n];
+                               coeff_plane[idx++] = p;
+                               byte_count++;
+                               if (byte_count == 27) {
+                                       idx += 5;
+                                       byte_count = 0;
+                               }
+                       }
+               }
+       }
+}
+
+static void init_intra_only_probs(struct rkvdec_ctx *ctx,
+                                 const struct rkvdec_vp9_run *run)
+{
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu;
+       struct rkvdec_vp9_intra_only_frame_probs *rkprobs;
+       const struct v4l2_vp9_frame_context *probs;
+       unsigned int i, j, k;
+
+       rkprobs = &tbl->probs.intra_only;
+       probs = &vp9_ctx->probability_tables;
+
+       /*
+        * intra only 149 x 128 bits ,aligned to 152 x 128 bits coeff related
+        * prob 64 x 128 bits
+        */
+       for (i = 0; i < ARRAY_SIZE(probs->coef); i++) {
+               for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++)
+                       write_coeff_plane(probs->coef[i][j][0],
+                                         rkprobs->coef_intra[i][j]);
+       }
+
+       /* intra mode prob  80 x 128 bits */
+       for (i = 0; i < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob); i++) {
+               unsigned int byte_count = 0;
+               int idx = 0;
+
+               /* vp9_kf_y_mode_prob */
+               for (j = 0; j < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0]); j++) {
+                       for (k = 0; k < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0][0]);
+                            k++) {
+                               u8 val = v4l2_vp9_kf_y_mode_prob[i][j][k];
+
+                               rkprobs->intra_mode[i].y_mode[idx++] = val;
+                               byte_count++;
+                               if (byte_count == 27) {
+                                       byte_count = 0;
+                                       idx += 5;
+                               }
+                       }
+               }
+
+       }
+
+       for (i = 0; i < sizeof(v4l2_vp9_kf_uv_mode_prob); ++i) {
+               const u8 *ptr = (const u8 *)v4l2_vp9_kf_uv_mode_prob;
+
+               rkprobs->intra_mode[i / 23].uv_mode[i % 23] = ptr[i];
+       }
+}
+
+static void init_inter_probs(struct rkvdec_ctx *ctx,
+                            const struct rkvdec_vp9_run *run)
+{
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu;
+       struct rkvdec_vp9_inter_frame_probs *rkprobs;
+       const struct v4l2_vp9_frame_context *probs;
+       unsigned int i, j, k;
+
+       rkprobs = &tbl->probs.inter;
+       probs = &vp9_ctx->probability_tables;
+
+       /*
+        * inter probs
+        * 151 x 128 bits, aligned to 152 x 128 bits
+        * inter only
+        * intra_y_mode & inter_block info 6 x 128 bits
+        */
+
+       memcpy(rkprobs->y_mode, probs->y_mode, sizeof(rkprobs->y_mode));
+       memcpy(rkprobs->comp_mode, probs->comp_mode,
+              sizeof(rkprobs->comp_mode));
+       memcpy(rkprobs->comp_ref, probs->comp_ref,
+              sizeof(rkprobs->comp_ref));
+       memcpy(rkprobs->single_ref, probs->single_ref,
+              sizeof(rkprobs->single_ref));
+       memcpy(rkprobs->inter_mode, probs->inter_mode,
+              sizeof(rkprobs->inter_mode));
+       memcpy(rkprobs->interp_filter, probs->interp_filter,
+              sizeof(rkprobs->interp_filter));
+
+       /* 128 x 128 bits coeff related */
+       for (i = 0; i < ARRAY_SIZE(probs->coef); i++) {
+               for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++) {
+                       for (k = 0; k < ARRAY_SIZE(probs->coef[0][0]); k++)
+                               write_coeff_plane(probs->coef[i][j][k],
+                                                 rkprobs->coef[k][i][j]);
+               }
+       }
+
+       /* intra uv mode 6 x 128 */
+       memcpy(rkprobs->uv_mode_0_2, &probs->uv_mode[0],
+              sizeof(rkprobs->uv_mode_0_2));
+       memcpy(rkprobs->uv_mode_3_5, &probs->uv_mode[3],
+              sizeof(rkprobs->uv_mode_3_5));
+       memcpy(rkprobs->uv_mode_6_8, &probs->uv_mode[6],
+              sizeof(rkprobs->uv_mode_6_8));
+       memcpy(rkprobs->uv_mode_9, &probs->uv_mode[9],
+              sizeof(rkprobs->uv_mode_9));
+
+       /* mv related 6 x 128 */
+       memcpy(rkprobs->mv.joint, probs->mv.joint,
+              sizeof(rkprobs->mv.joint));
+       memcpy(rkprobs->mv.sign, probs->mv.sign,
+              sizeof(rkprobs->mv.sign));
+       memcpy(rkprobs->mv.classes, probs->mv.classes,
+              sizeof(rkprobs->mv.classes));
+       memcpy(rkprobs->mv.class0_bit, probs->mv.class0_bit,
+              sizeof(rkprobs->mv.class0_bit));
+       memcpy(rkprobs->mv.bits, probs->mv.bits,
+              sizeof(rkprobs->mv.bits));
+       memcpy(rkprobs->mv.class0_fr, probs->mv.class0_fr,
+              sizeof(rkprobs->mv.class0_fr));
+       memcpy(rkprobs->mv.fr, probs->mv.fr,
+              sizeof(rkprobs->mv.fr));
+       memcpy(rkprobs->mv.class0_hp, probs->mv.class0_hp,
+              sizeof(rkprobs->mv.class0_hp));
+       memcpy(rkprobs->mv.hp, probs->mv.hp,
+              sizeof(rkprobs->mv.hp));
+}
+
+static void init_probs(struct rkvdec_ctx *ctx,
+                      const struct rkvdec_vp9_run *run)
+{
+       const struct v4l2_ctrl_vp9_frame *dec_params;
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu;
+       struct rkvdec_vp9_probs *rkprobs = &tbl->probs;
+       const struct v4l2_vp9_segmentation *seg;
+       const struct v4l2_vp9_frame_context *probs;
+       bool intra_only;
+
+       dec_params = run->decode_params;
+       probs = &vp9_ctx->probability_tables;
+       seg = &dec_params->seg;
+
+       memset(rkprobs, 0, sizeof(*rkprobs));
+
+       intra_only = !!(dec_params->flags &
+                       (V4L2_VP9_FRAME_FLAG_KEY_FRAME |
+                        V4L2_VP9_FRAME_FLAG_INTRA_ONLY));
+
+       /* sb info  5 x 128 bit */
+       memcpy(rkprobs->partition,
+              intra_only ? v4l2_vp9_kf_partition_probs : probs->partition,
+              sizeof(rkprobs->partition));
+
+       memcpy(rkprobs->pred, seg->pred_probs, sizeof(rkprobs->pred));
+       memcpy(rkprobs->tree, seg->tree_probs, sizeof(rkprobs->tree));
+       memcpy(rkprobs->skip, probs->skip, sizeof(rkprobs->skip));
+       memcpy(rkprobs->tx32, probs->tx32, sizeof(rkprobs->tx32));
+       memcpy(rkprobs->tx16, probs->tx16, sizeof(rkprobs->tx16));
+       memcpy(rkprobs->tx8, probs->tx8, sizeof(rkprobs->tx8));
+       memcpy(rkprobs->is_inter, probs->is_inter, sizeof(rkprobs->is_inter));
+
+       if (intra_only)
+               init_intra_only_probs(ctx, run);
+       else
+               init_inter_probs(ctx, run);
+}
+
+struct rkvdec_vp9_ref_reg {
+       u32 reg_frm_size;
+       u32 reg_hor_stride;
+       u32 reg_y_stride;
+       u32 reg_yuv_stride;
+       u32 reg_ref_base;
+};
+
+static struct rkvdec_vp9_ref_reg ref_regs[] = {
+       {
+               .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(0),
+               .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(0),
+               .reg_y_stride = RKVDEC_VP9_LAST_FRAME_YSTRIDE,
+               .reg_yuv_stride = RKVDEC_VP9_LAST_FRAME_YUVSTRIDE,
+               .reg_ref_base = RKVDEC_REG_VP9_LAST_FRAME_BASE,
+       },
+       {
+               .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(1),
+               .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(1),
+               .reg_y_stride = RKVDEC_VP9_GOLDEN_FRAME_YSTRIDE,
+               .reg_yuv_stride = 0,
+               .reg_ref_base = RKVDEC_REG_VP9_GOLDEN_FRAME_BASE,
+       },
+       {
+               .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(2),
+               .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(2),
+               .reg_y_stride = RKVDEC_VP9_ALTREF_FRAME_YSTRIDE,
+               .reg_yuv_stride = 0,
+               .reg_ref_base = RKVDEC_REG_VP9_ALTREF_FRAME_BASE,
+       }
+};
+
+static struct rkvdec_decoded_buffer *
+get_ref_buf(struct rkvdec_ctx *ctx, struct vb2_v4l2_buffer *dst, u64 timestamp)
+{
+       struct v4l2_m2m_ctx *m2m_ctx = ctx->fh.m2m_ctx;
+       struct vb2_queue *cap_q = &m2m_ctx->cap_q_ctx.q;
+       int buf_idx;
+
+       /*
+        * If a ref is unused or invalid, address of current destination
+        * buffer is returned.
+        */
+       buf_idx = vb2_find_timestamp(cap_q, timestamp, 0);
+       if (buf_idx < 0)
+               return vb2_to_rkvdec_decoded_buf(&dst->vb2_buf);
+
+       return vb2_to_rkvdec_decoded_buf(vb2_get_buffer(cap_q, buf_idx));
+}
+
+static dma_addr_t get_mv_base_addr(struct rkvdec_decoded_buffer *buf)
+{
+       unsigned int aligned_pitch, aligned_height, yuv_len;
+
+       aligned_height = round_up(buf->vp9.height, 64);
+       aligned_pitch = round_up(buf->vp9.width * buf->vp9.bit_depth, 512) / 8;
+       yuv_len = (aligned_height * aligned_pitch * 3) / 2;
+
+       return vb2_dma_contig_plane_dma_addr(&buf->base.vb.vb2_buf, 0) +
+              yuv_len;
+}
+
+static void config_ref_registers(struct rkvdec_ctx *ctx,
+                                const struct rkvdec_vp9_run *run,
+                                struct rkvdec_decoded_buffer *ref_buf,
+                                struct rkvdec_vp9_ref_reg *ref_reg)
+{
+       unsigned int aligned_pitch, aligned_height, y_len, yuv_len;
+       struct rkvdec_dev *rkvdec = ctx->dev;
+
+       aligned_height = round_up(ref_buf->vp9.height, 64);
+       writel_relaxed(RKVDEC_VP9_FRAMEWIDTH(ref_buf->vp9.width) |
+                      RKVDEC_VP9_FRAMEHEIGHT(ref_buf->vp9.height),
+                      rkvdec->regs + ref_reg->reg_frm_size);
+
+       writel_relaxed(vb2_dma_contig_plane_dma_addr(&ref_buf->base.vb.vb2_buf, 0),
+                      rkvdec->regs + ref_reg->reg_ref_base);
+
+       if (&ref_buf->base.vb == run->base.bufs.dst)
+               return;
+
+       aligned_pitch = round_up(ref_buf->vp9.width * ref_buf->vp9.bit_depth, 512) / 8;
+       y_len = aligned_height * aligned_pitch;
+       yuv_len = (y_len * 3) / 2;
+
+       writel_relaxed(RKVDEC_HOR_Y_VIRSTRIDE(aligned_pitch / 16) |
+                      RKVDEC_HOR_UV_VIRSTRIDE(aligned_pitch / 16),
+                      rkvdec->regs + ref_reg->reg_hor_stride);
+       writel_relaxed(RKVDEC_VP9_REF_YSTRIDE(y_len / 16),
+                      rkvdec->regs + ref_reg->reg_y_stride);
+
+       if (!ref_reg->reg_yuv_stride)
+               return;
+
+       writel_relaxed(RKVDEC_VP9_REF_YUVSTRIDE(yuv_len / 16),
+                      rkvdec->regs + ref_reg->reg_yuv_stride);
+}
+
+static void config_seg_registers(struct rkvdec_ctx *ctx, unsigned int segid)
+{
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       const struct v4l2_vp9_segmentation *seg;
+       struct rkvdec_dev *rkvdec = ctx->dev;
+       s16 feature_val;
+       int feature_id;
+       u32 val = 0;
+
+       seg = vp9_ctx->last.valid ? &vp9_ctx->last.seg : &vp9_ctx->cur.seg;
+       feature_id = V4L2_VP9_SEG_LVL_ALT_Q;
+       if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
+               feature_val = seg->feature_data[segid][feature_id];
+               val |= RKVDEC_SEGID_FRAME_QP_DELTA_EN(1) |
+                      RKVDEC_SEGID_FRAME_QP_DELTA(feature_val);
+       }
+
+       feature_id = V4L2_VP9_SEG_LVL_ALT_L;
+       if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
+               feature_val = seg->feature_data[segid][feature_id];
+               val |= RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE_EN(1) |
+                      RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE(feature_val);
+       }
+
+       feature_id = V4L2_VP9_SEG_LVL_REF_FRAME;
+       if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
+               feature_val = seg->feature_data[segid][feature_id];
+               val |= RKVDEC_SEGID_REFERINFO_EN(1) |
+                      RKVDEC_SEGID_REFERINFO(feature_val);
+       }
+
+       feature_id = V4L2_VP9_SEG_LVL_SKIP;
+       if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid))
+               val |= RKVDEC_SEGID_FRAME_SKIP_EN(1);
+
+       if (!segid &&
+           (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE))
+               val |= RKVDEC_SEGID_ABS_DELTA(1);
+
+       writel_relaxed(val, rkvdec->regs + RKVDEC_VP9_SEGID_GRP(segid));
+}
+
+static void update_dec_buf_info(struct rkvdec_decoded_buffer *buf,
+                               const struct v4l2_ctrl_vp9_frame *dec_params)
+{
+       buf->vp9.width = dec_params->frame_width_minus_1 + 1;
+       buf->vp9.height = dec_params->frame_height_minus_1 + 1;
+       buf->vp9.bit_depth = dec_params->bit_depth;
+}
+
+static void update_ctx_cur_info(struct rkvdec_vp9_ctx *vp9_ctx,
+                               struct rkvdec_decoded_buffer *buf,
+                               const struct v4l2_ctrl_vp9_frame *dec_params)
+{
+       vp9_ctx->cur.valid = true;
+       vp9_ctx->cur.reference_mode = dec_params->reference_mode;
+       vp9_ctx->cur.interpolation_filter = dec_params->interpolation_filter;
+       vp9_ctx->cur.flags = dec_params->flags;
+       vp9_ctx->cur.timestamp = buf->base.vb.vb2_buf.timestamp;
+       vp9_ctx->cur.seg = dec_params->seg;
+       vp9_ctx->cur.lf = dec_params->lf;
+}
+
+static void update_ctx_last_info(struct rkvdec_vp9_ctx *vp9_ctx)
+{
+       vp9_ctx->last = vp9_ctx->cur;
+}
+
+static void config_registers(struct rkvdec_ctx *ctx,
+                            const struct rkvdec_vp9_run *run)
+{
+       unsigned int y_len, uv_len, yuv_len, bit_depth, aligned_height, aligned_pitch, stream_len;
+       const struct v4l2_ctrl_vp9_frame *dec_params;
+       struct rkvdec_decoded_buffer *ref_bufs[3];
+       struct rkvdec_decoded_buffer *dst, *last, *mv_ref;
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       u32 val, last_frame_info = 0;
+       const struct v4l2_vp9_segmentation *seg;
+       struct rkvdec_dev *rkvdec = ctx->dev;
+       dma_addr_t addr;
+       bool intra_only;
+       unsigned int i;
+
+       dec_params = run->decode_params;
+       dst = vb2_to_rkvdec_decoded_buf(&run->base.bufs.dst->vb2_buf);
+       ref_bufs[0] = get_ref_buf(ctx, &dst->base.vb, dec_params->last_frame_ts);
+       ref_bufs[1] = get_ref_buf(ctx, &dst->base.vb, dec_params->golden_frame_ts);
+       ref_bufs[2] = get_ref_buf(ctx, &dst->base.vb, dec_params->alt_frame_ts);
+
+       if (vp9_ctx->last.valid)
+               last = get_ref_buf(ctx, &dst->base.vb, vp9_ctx->last.timestamp);
+       else
+               last = dst;
+
+       update_dec_buf_info(dst, dec_params);
+       update_ctx_cur_info(vp9_ctx, dst, dec_params);
+       seg = &dec_params->seg;
+
+       intra_only = !!(dec_params->flags &
+                       (V4L2_VP9_FRAME_FLAG_KEY_FRAME |
+                        V4L2_VP9_FRAME_FLAG_INTRA_ONLY));
+
+       writel_relaxed(RKVDEC_MODE(RKVDEC_MODE_VP9),
+                      rkvdec->regs + RKVDEC_REG_SYSCTRL);
+
+       bit_depth = dec_params->bit_depth;
+       aligned_height = round_up(ctx->decoded_fmt.fmt.pix_mp.height, 64);
+
+       aligned_pitch = round_up(ctx->decoded_fmt.fmt.pix_mp.width *
+                                bit_depth,
+                                512) / 8;
+       y_len = aligned_height * aligned_pitch;
+       uv_len = y_len / 2;
+       yuv_len = y_len + uv_len;
+
+       writel_relaxed(RKVDEC_Y_HOR_VIRSTRIDE(aligned_pitch / 16) |
+                      RKVDEC_UV_HOR_VIRSTRIDE(aligned_pitch / 16),
+                      rkvdec->regs + RKVDEC_REG_PICPAR);
+       writel_relaxed(RKVDEC_Y_VIRSTRIDE(y_len / 16),
+                      rkvdec->regs + RKVDEC_REG_Y_VIRSTRIDE);
+       writel_relaxed(RKVDEC_YUV_VIRSTRIDE(yuv_len / 16),
+                      rkvdec->regs + RKVDEC_REG_YUV_VIRSTRIDE);
+
+       stream_len = vb2_get_plane_payload(&run->base.bufs.src->vb2_buf, 0);
+       writel_relaxed(RKVDEC_STRM_LEN(stream_len),
+                      rkvdec->regs + RKVDEC_REG_STRM_LEN);
+
+       /*
+        * Reset count buffer, because decoder only output intra related syntax
+        * counts when decoding intra frame, but update entropy need to update
+        * all the probabilities.
+        */
+       if (intra_only)
+               memset(vp9_ctx->count_tbl.cpu, 0, vp9_ctx->count_tbl.size);
+
+       vp9_ctx->cur.segmapid = vp9_ctx->last.segmapid;
+       if (!intra_only &&
+           !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) &&
+           (!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) ||
+            (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP)))
+               vp9_ctx->cur.segmapid++;
+
+       for (i = 0; i < ARRAY_SIZE(ref_bufs); i++)
+               config_ref_registers(ctx, run, ref_bufs[i], &ref_regs[i]);
+
+       for (i = 0; i < 8; i++)
+               config_seg_registers(ctx, i);
+
+       writel_relaxed(RKVDEC_VP9_TX_MODE(vp9_ctx->cur.tx_mode) |
+                      RKVDEC_VP9_FRAME_REF_MODE(dec_params->reference_mode),
+                      rkvdec->regs + RKVDEC_VP9_CPRHEADER_CONFIG);
+
+       if (!intra_only) {
+               const struct v4l2_vp9_loop_filter *lf;
+               s8 delta;
+
+               if (vp9_ctx->last.valid)
+                       lf = &vp9_ctx->last.lf;
+               else
+                       lf = &vp9_ctx->cur.lf;
+
+               val = 0;
+               for (i = 0; i < ARRAY_SIZE(lf->ref_deltas); i++) {
+                       delta = lf->ref_deltas[i];
+                       val |= RKVDEC_REF_DELTAS_LASTFRAME(i, delta);
+               }
+
+               writel_relaxed(val,
+                              rkvdec->regs + RKVDEC_VP9_REF_DELTAS_LASTFRAME);
+
+               for (i = 0; i < ARRAY_SIZE(lf->mode_deltas); i++) {
+                       delta = lf->mode_deltas[i];
+                       last_frame_info |= RKVDEC_MODE_DELTAS_LASTFRAME(i,
+                                                                       delta);
+               }
+       }
+
+       if (vp9_ctx->last.valid && !intra_only &&
+           vp9_ctx->last.seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED)
+               last_frame_info |= RKVDEC_SEG_EN_LASTFRAME;
+
+       if (vp9_ctx->last.valid &&
+           vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_SHOW_FRAME)
+               last_frame_info |= RKVDEC_LAST_SHOW_FRAME;
+
+       if (vp9_ctx->last.valid &&
+           vp9_ctx->last.flags &
+           (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY))
+               last_frame_info |= RKVDEC_LAST_INTRA_ONLY;
+
+       if (vp9_ctx->last.valid &&
+           last->vp9.width == dst->vp9.width &&
+           last->vp9.height == dst->vp9.height)
+               last_frame_info |= RKVDEC_LAST_WIDHHEIGHT_EQCUR;
+
+       writel_relaxed(last_frame_info,
+                      rkvdec->regs + RKVDEC_VP9_INFO_LASTFRAME);
+
+       writel_relaxed(stream_len - dec_params->compressed_header_size -
+                      dec_params->uncompressed_header_size,
+                      rkvdec->regs + RKVDEC_VP9_LASTTILE_SIZE);
+
+       for (i = 0; !intra_only && i < ARRAY_SIZE(ref_bufs); i++) {
+               unsigned int refw = ref_bufs[i]->vp9.width;
+               unsigned int refh = ref_bufs[i]->vp9.height;
+               u32 hscale, vscale;
+
+               hscale = (refw << 14) / dst->vp9.width;
+               vscale = (refh << 14) / dst->vp9.height;
+               writel_relaxed(RKVDEC_VP9_REF_HOR_SCALE(hscale) |
+                              RKVDEC_VP9_REF_VER_SCALE(vscale),
+                              rkvdec->regs + RKVDEC_VP9_REF_SCALE(i));
+       }
+
+       addr = vb2_dma_contig_plane_dma_addr(&dst->base.vb.vb2_buf, 0);
+       writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_DECOUT_BASE);
+       addr = vb2_dma_contig_plane_dma_addr(&run->base.bufs.src->vb2_buf, 0);
+       writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_STRM_RLC_BASE);
+       writel_relaxed(vp9_ctx->priv_tbl.dma +
+                      offsetof(struct rkvdec_vp9_priv_tbl, probs),
+                      rkvdec->regs + RKVDEC_REG_CABACTBL_PROB_BASE);
+       writel_relaxed(vp9_ctx->count_tbl.dma,
+                      rkvdec->regs + RKVDEC_REG_VP9COUNT_BASE);
+
+       writel_relaxed(vp9_ctx->priv_tbl.dma +
+                      offsetof(struct rkvdec_vp9_priv_tbl, segmap) +
+                      (RKVDEC_VP9_MAX_SEGMAP_SIZE * vp9_ctx->cur.segmapid),
+                      rkvdec->regs + RKVDEC_REG_VP9_SEGIDCUR_BASE);
+       writel_relaxed(vp9_ctx->priv_tbl.dma +
+                      offsetof(struct rkvdec_vp9_priv_tbl, segmap) +
+                      (RKVDEC_VP9_MAX_SEGMAP_SIZE * (!vp9_ctx->cur.segmapid)),
+                      rkvdec->regs + RKVDEC_REG_VP9_SEGIDLAST_BASE);
+
+       if (!intra_only &&
+           !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) &&
+           vp9_ctx->last.valid)
+               mv_ref = last;
+       else
+               mv_ref = dst;
+
+       writel_relaxed(get_mv_base_addr(mv_ref),
+                      rkvdec->regs + RKVDEC_VP9_REF_COLMV_BASE);
+
+       writel_relaxed(ctx->decoded_fmt.fmt.pix_mp.width |
+                      (ctx->decoded_fmt.fmt.pix_mp.height << 16),
+                      rkvdec->regs + RKVDEC_REG_PERFORMANCE_CYCLE);
+}
+
+static int validate_dec_params(struct rkvdec_ctx *ctx,
+                              const struct v4l2_ctrl_vp9_frame *dec_params)
+{
+       unsigned int aligned_width, aligned_height;
+
+       /* We only support profile 0. */
+       if (dec_params->profile != 0) {
+               dev_err(ctx->dev->dev, "unsupported profile %d\n",
+                       dec_params->profile);
+               return -EINVAL;
+       }
+
+       aligned_width = round_up(dec_params->frame_width_minus_1 + 1, 64);
+       aligned_height = round_up(dec_params->frame_height_minus_1 + 1, 64);
+
+       /*
+        * Userspace should update the capture/decoded format when the
+        * resolution changes.
+        */
+       if (aligned_width != ctx->decoded_fmt.fmt.pix_mp.width ||
+           aligned_height != ctx->decoded_fmt.fmt.pix_mp.height) {
+               dev_err(ctx->dev->dev,
+                       "unexpected bitstream resolution %dx%d\n",
+                       dec_params->frame_width_minus_1 + 1,
+                       dec_params->frame_height_minus_1 + 1);
+               return -EINVAL;
+       }
+
+       return 0;
+}
+
+static int rkvdec_vp9_run_preamble(struct rkvdec_ctx *ctx,
+                                  struct rkvdec_vp9_run *run)
+{
+       const struct v4l2_ctrl_vp9_frame *dec_params;
+       const struct v4l2_ctrl_vp9_compressed_hdr *prob_updates;
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       struct v4l2_ctrl *ctrl;
+       unsigned int fctx_idx;
+       int ret;
+
+       /* v4l2-specific stuff */
+       rkvdec_run_preamble(ctx, &run->base);
+
+       ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl,
+                             V4L2_CID_STATELESS_VP9_FRAME);
+       if (WARN_ON(!ctrl))
+               return -EINVAL;
+       dec_params = ctrl->p_cur.p;
+
+       ret = validate_dec_params(ctx, dec_params);
+       if (ret)
+               return ret;
+
+       run->decode_params = dec_params;
+
+       ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl, V4L2_CID_STATELESS_VP9_COMPRESSED_HDR);
+       if (WARN_ON(!ctrl))
+               return -EINVAL;
+       prob_updates = ctrl->p_cur.p;
+       vp9_ctx->cur.tx_mode = prob_updates->tx_mode;
+
+       /*
+        * vp9 stuff
+        *
+        * by this point the userspace has done all parts of 6.2 uncompressed_header()
+        * except this fragment:
+        * if ( FrameIsIntra || error_resilient_mode ) {
+        *      setup_past_independence ( )
+        *      if ( frame_type == KEY_FRAME || error_resilient_mode == 1 ||
+        *           reset_frame_context == 3 ) {
+        *              for ( i = 0; i < 4; i ++ ) {
+        *                      save_probs( i )
+        *              }
+        *      } else if ( reset_frame_context == 2 ) {
+        *              save_probs( frame_context_idx )
+        *      }
+        *      frame_context_idx = 0
+        * }
+        */
+       fctx_idx = v4l2_vp9_reset_frame_ctx(dec_params, vp9_ctx->frame_context);
+       vp9_ctx->cur.frame_context_idx = fctx_idx;
+
+       /* 6.1 frame(sz): load_probs() and load_probs2() */
+       vp9_ctx->probability_tables = vp9_ctx->frame_context[fctx_idx];
+
+       /*
+        * The userspace has also performed 6.3 compressed_header(), but handling the
+        * probs in a special way. All probs which need updating, except MV-related,
+        * have been read from the bitstream and translated through inv_map_table[],
+        * but no 6.3.6 inv_recenter_nonneg(v, m) has been performed. The values passed
+        * by userspace are either translated values (there are no 0 values in
+        * inv_map_table[]), or zero to indicate no update. All MV-related probs which need
+        * updating have been read from the bitstream and (mv_prob << 1) | 1 has been
+        * performed. The values passed by userspace are either new values
+        * to replace old ones (the above mentioned shift and bitwise or never result in
+        * a zero) or zero to indicate no update.
+        * fw_update_probs() performs actual probs updates or leaves probs as-is
+        * for values for which a zero was passed from userspace.
+        */
+       v4l2_vp9_fw_update_probs(&vp9_ctx->probability_tables, prob_updates, dec_params);
+
+       return 0;
+}
+
+static int rkvdec_vp9_run(struct rkvdec_ctx *ctx)
+{
+       struct rkvdec_dev *rkvdec = ctx->dev;
+       struct rkvdec_vp9_run run = { };
+       int ret;
+
+       ret = rkvdec_vp9_run_preamble(ctx, &run);
+       if (ret) {
+               rkvdec_run_postamble(ctx, &run.base);
+               return ret;
+       }
+
+       /* Prepare probs. */
+       init_probs(ctx, &run);
+
+       /* Configure hardware registers. */
+       config_registers(ctx, &run);
+
+       rkvdec_run_postamble(ctx, &run.base);
+
+       schedule_delayed_work(&rkvdec->watchdog_work, msecs_to_jiffies(2000));
+
+       writel(1, rkvdec->regs + RKVDEC_REG_PREF_LUMA_CACHE_COMMAND);
+       writel(1, rkvdec->regs + RKVDEC_REG_PREF_CHR_CACHE_COMMAND);
+
+       writel(0xe, rkvdec->regs + RKVDEC_REG_STRMD_ERR_EN);
+       /* Start decoding! */
+       writel(RKVDEC_INTERRUPT_DEC_E | RKVDEC_CONFIG_DEC_CLK_GATE_E |
+              RKVDEC_TIMEOUT_E | RKVDEC_BUF_EMPTY_E,
+              rkvdec->regs + RKVDEC_REG_INTERRUPT);
+
+       return 0;
+}
+
+#define copy_tx_and_skip(p1, p2)                               \
+do {                                                           \
+       memcpy((p1)->tx8, (p2)->tx8, sizeof((p1)->tx8));        \
+       memcpy((p1)->tx16, (p2)->tx16, sizeof((p1)->tx16));     \
+       memcpy((p1)->tx32, (p2)->tx32, sizeof((p1)->tx32));     \
+       memcpy((p1)->skip, (p2)->skip, sizeof((p1)->skip));     \
+} while (0)
+
+static void rkvdec_vp9_done(struct rkvdec_ctx *ctx,
+                           struct vb2_v4l2_buffer *src_buf,
+                           struct vb2_v4l2_buffer *dst_buf,
+                           enum vb2_buffer_state result)
+{
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       unsigned int fctx_idx;
+
+       /* v4l2-specific stuff */
+       if (result == VB2_BUF_STATE_ERROR)
+               goto out_update_last;
+
+       /*
+        * vp9 stuff
+        *
+        * 6.1.2 refresh_probs()
+        *
+        * In the spec a complementary condition goes last in 6.1.2 refresh_probs(),
+        * but it makes no sense to perform all the activities from the first "if"
+        * there if we actually are not refreshing the frame context. On top of that,
+        * because of 6.2 uncompressed_header() whenever error_resilient_mode == 1,
+        * refresh_frame_context == 0. Consequently, if we don't jump to out_update_last
+        * it means error_resilient_mode must be 0.
+        */
+       if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_REFRESH_FRAME_CTX))
+               goto out_update_last;
+
+       fctx_idx = vp9_ctx->cur.frame_context_idx;
+
+       if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_PARALLEL_DEC_MODE)) {
+               /* error_resilient_mode == 0 && frame_parallel_decoding_mode == 0 */
+               struct v4l2_vp9_frame_context *probs = &vp9_ctx->probability_tables;
+               bool frame_is_intra = vp9_ctx->cur.flags &
+                   (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY);
+               struct tx_and_skip {
+                       u8 tx8[2][1];
+                       u8 tx16[2][2];
+                       u8 tx32[2][3];
+                       u8 skip[3];
+               } _tx_skip, *tx_skip = &_tx_skip;
+               struct v4l2_vp9_frame_symbol_counts *counts;
+
+               /* buffer the forward-updated TX and skip probs */
+               if (frame_is_intra)
+                       copy_tx_and_skip(tx_skip, probs);
+
+               /* 6.1.2 refresh_probs(): load_probs() and load_probs2() */
+               *probs = vp9_ctx->frame_context[fctx_idx];
+
+               /* if FrameIsIntra then undo the effect of load_probs2() */
+               if (frame_is_intra)
+                       copy_tx_and_skip(probs, tx_skip);
+
+               counts = frame_is_intra ? &vp9_ctx->intra_cnts : &vp9_ctx->inter_cnts;
+               v4l2_vp9_adapt_coef_probs(probs, counts,
+                                         !vp9_ctx->last.valid ||
+                                         vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME,
+                                         frame_is_intra);
+               if (!frame_is_intra) {
+                       const struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts;
+                       u32 classes[2][11];
+                       int i;
+
+                       inter_cnts = vp9_ctx->count_tbl.cpu;
+                       for (i = 0; i < ARRAY_SIZE(classes); ++i)
+                               memcpy(classes[i], inter_cnts->classes[i], sizeof(classes[0]));
+                       counts->classes = &classes;
+
+                       /* load_probs2() already done */
+                       v4l2_vp9_adapt_noncoef_probs(&vp9_ctx->probability_tables, counts,
+                                                    vp9_ctx->cur.reference_mode,
+                                                    vp9_ctx->cur.interpolation_filter,
+                                                    vp9_ctx->cur.tx_mode, vp9_ctx->cur.flags);
+               }
+       }
+
+       /* 6.1.2 refresh_probs(): save_probs(fctx_idx) */
+       vp9_ctx->frame_context[fctx_idx] = vp9_ctx->probability_tables;
+
+out_update_last:
+       update_ctx_last_info(vp9_ctx);
+}
+
+static void rkvdec_init_v4l2_vp9_count_tbl(struct rkvdec_ctx *ctx)
+{
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       struct rkvdec_vp9_intra_frame_symbol_counts *intra_cnts = vp9_ctx->count_tbl.cpu;
+       struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts = vp9_ctx->count_tbl.cpu;
+       int i, j, k, l, m;
+
+       vp9_ctx->inter_cnts.partition = &inter_cnts->partition;
+       vp9_ctx->inter_cnts.skip = &inter_cnts->skip;
+       vp9_ctx->inter_cnts.intra_inter = &inter_cnts->inter;
+       vp9_ctx->inter_cnts.tx32p = &inter_cnts->tx32p;
+       vp9_ctx->inter_cnts.tx16p = &inter_cnts->tx16p;
+       vp9_ctx->inter_cnts.tx8p = &inter_cnts->tx8p;
+
+       vp9_ctx->intra_cnts.partition = (u32 (*)[16][4])(&intra_cnts->partition);
+       vp9_ctx->intra_cnts.skip = &intra_cnts->skip;
+       vp9_ctx->intra_cnts.intra_inter = &intra_cnts->intra;
+       vp9_ctx->intra_cnts.tx32p = &intra_cnts->tx32p;
+       vp9_ctx->intra_cnts.tx16p = &intra_cnts->tx16p;
+       vp9_ctx->intra_cnts.tx8p = &intra_cnts->tx8p;
+
+       vp9_ctx->inter_cnts.y_mode = &inter_cnts->y_mode;
+       vp9_ctx->inter_cnts.uv_mode = &inter_cnts->uv_mode;
+       vp9_ctx->inter_cnts.comp = &inter_cnts->comp;
+       vp9_ctx->inter_cnts.comp_ref = &inter_cnts->comp_ref;
+       vp9_ctx->inter_cnts.single_ref = &inter_cnts->single_ref;
+       vp9_ctx->inter_cnts.mv_mode = &inter_cnts->mv_mode;
+       vp9_ctx->inter_cnts.filter = &inter_cnts->filter;
+       vp9_ctx->inter_cnts.mv_joint = &inter_cnts->mv_joint;
+       vp9_ctx->inter_cnts.sign = &inter_cnts->sign;
+       /*
+        * rk hardware actually uses "u32 classes[2][11 + 1];"
+        * instead of "u32 classes[2][11];", so this must be explicitly
+        * copied into vp9_ctx->classes when passing the data to the
+        * vp9 library function
+        */
+       vp9_ctx->inter_cnts.class0 = &inter_cnts->class0;
+       vp9_ctx->inter_cnts.bits = &inter_cnts->bits;
+       vp9_ctx->inter_cnts.class0_fp = &inter_cnts->class0_fp;
+       vp9_ctx->inter_cnts.fp = &inter_cnts->fp;
+       vp9_ctx->inter_cnts.class0_hp = &inter_cnts->class0_hp;
+       vp9_ctx->inter_cnts.hp = &inter_cnts->hp;
+
+#define INNERMOST_LOOP \
+       do {                                                                            \
+               for (m = 0; m < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0][0]); ++m) {\
+                       vp9_ctx->inter_cnts.coeff[i][j][k][l][m] =                      \
+                               &inter_cnts->ref_cnt[k][i][j][l][m].coeff;              \
+                       vp9_ctx->inter_cnts.eob[i][j][k][l][m][0] =                     \
+                               &inter_cnts->ref_cnt[k][i][j][l][m].eob[0];             \
+                       vp9_ctx->inter_cnts.eob[i][j][k][l][m][1] =                     \
+                               &inter_cnts->ref_cnt[k][i][j][l][m].eob[1];             \
+                                                                                       \
+                       vp9_ctx->intra_cnts.coeff[i][j][k][l][m] =                      \
+                               &intra_cnts->ref_cnt[k][i][j][l][m].coeff;              \
+                       vp9_ctx->intra_cnts.eob[i][j][k][l][m][0] =                     \
+                               &intra_cnts->ref_cnt[k][i][j][l][m].eob[0];             \
+                       vp9_ctx->intra_cnts.eob[i][j][k][l][m][1] =                     \
+                               &intra_cnts->ref_cnt[k][i][j][l][m].eob[1];             \
+               }                                                                       \
+       } while (0)
+
+       for (i = 0; i < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff); ++i)
+               for (j = 0; j < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0]); ++j)
+                       for (k = 0; k < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0]); ++k)
+                               for (l = 0; l < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0]); ++l)
+                                       INNERMOST_LOOP;
+#undef INNERMOST_LOOP
+}
+
+static int rkvdec_vp9_start(struct rkvdec_ctx *ctx)
+{
+       struct rkvdec_dev *rkvdec = ctx->dev;
+       struct rkvdec_vp9_priv_tbl *priv_tbl;
+       struct rkvdec_vp9_ctx *vp9_ctx;
+       unsigned char *count_tbl;
+       int ret;
+
+       vp9_ctx = kzalloc(sizeof(*vp9_ctx), GFP_KERNEL);
+       if (!vp9_ctx)
+               return -ENOMEM;
+
+       ctx->priv = vp9_ctx;
+
+       priv_tbl = dma_alloc_coherent(rkvdec->dev, sizeof(*priv_tbl),
+                                     &vp9_ctx->priv_tbl.dma, GFP_KERNEL);
+       if (!priv_tbl) {
+               ret = -ENOMEM;
+               goto err_free_ctx;
+       }
+
+       vp9_ctx->priv_tbl.size = sizeof(*priv_tbl);
+       vp9_ctx->priv_tbl.cpu = priv_tbl;
+       memset(priv_tbl, 0, sizeof(*priv_tbl));
+
+       count_tbl = dma_alloc_coherent(rkvdec->dev, RKVDEC_VP9_COUNT_SIZE,
+                                      &vp9_ctx->count_tbl.dma, GFP_KERNEL);
+       if (!count_tbl) {
+               ret = -ENOMEM;
+               goto err_free_priv_tbl;
+       }
+
+       vp9_ctx->count_tbl.size = RKVDEC_VP9_COUNT_SIZE;
+       vp9_ctx->count_tbl.cpu = count_tbl;
+       memset(count_tbl, 0, sizeof(*count_tbl));
+       rkvdec_init_v4l2_vp9_count_tbl(ctx);
+
+       return 0;
+
+err_free_priv_tbl:
+       dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size,
+                         vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma);
+
+err_free_ctx:
+       kfree(vp9_ctx);
+       return ret;
+}
+
+static void rkvdec_vp9_stop(struct rkvdec_ctx *ctx)
+{
+       struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
+       struct rkvdec_dev *rkvdec = ctx->dev;
+
+       dma_free_coherent(rkvdec->dev, vp9_ctx->count_tbl.size,
+                         vp9_ctx->count_tbl.cpu, vp9_ctx->count_tbl.dma);
+       dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size,
+                         vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma);
+       kfree(vp9_ctx);
+}
+
+static int rkvdec_vp9_adjust_fmt(struct rkvdec_ctx *ctx,
+                                struct v4l2_format *f)
+{
+       struct v4l2_pix_format_mplane *fmt = &f->fmt.pix_mp;
+
+       fmt->num_planes = 1;
+       if (!fmt->plane_fmt[0].sizeimage)
+               fmt->plane_fmt[0].sizeimage = fmt->width * fmt->height * 2;
+       return 0;
+}
+
+const struct rkvdec_coded_fmt_ops rkvdec_vp9_fmt_ops = {
+       .adjust_fmt = rkvdec_vp9_adjust_fmt,
+       .start = rkvdec_vp9_start,
+       .stop = rkvdec_vp9_stop,
+       .run = rkvdec_vp9_run,
+       .done = rkvdec_vp9_done,
+};
index 4fd4a29..ad2624c 100644 (file)
@@ -99,10 +99,30 @@ static const struct rkvdec_ctrls rkvdec_h264_ctrls = {
        .num_ctrls = ARRAY_SIZE(rkvdec_h264_ctrl_descs),
 };
 
-static const u32 rkvdec_h264_decoded_fmts[] = {
+static const u32 rkvdec_h264_vp9_decoded_fmts[] = {
        V4L2_PIX_FMT_NV12,
 };
 
+static const struct rkvdec_ctrl_desc rkvdec_vp9_ctrl_descs[] = {
+       {
+               .cfg.id = V4L2_CID_STATELESS_VP9_FRAME,
+       },
+       {
+               .cfg.id = V4L2_CID_STATELESS_VP9_COMPRESSED_HDR,
+       },
+       {
+               .cfg.id = V4L2_CID_MPEG_VIDEO_VP9_PROFILE,
+               .cfg.min = V4L2_MPEG_VIDEO_VP9_PROFILE_0,
+               .cfg.max = V4L2_MPEG_VIDEO_VP9_PROFILE_0,
+               .cfg.def = V4L2_MPEG_VIDEO_VP9_PROFILE_0,
+       },
+};
+
+static const struct rkvdec_ctrls rkvdec_vp9_ctrls = {
+       .ctrls = rkvdec_vp9_ctrl_descs,
+       .num_ctrls = ARRAY_SIZE(rkvdec_vp9_ctrl_descs),
+};
+
 static const struct rkvdec_coded_fmt_desc rkvdec_coded_fmts[] = {
        {
                .fourcc = V4L2_PIX_FMT_H264_SLICE,
@@ -116,8 +136,23 @@ static const struct rkvdec_coded_fmt_desc rkvdec_coded_fmts[] = {
                },
                .ctrls = &rkvdec_h264_ctrls,
                .ops = &rkvdec_h264_fmt_ops,
-               .num_decoded_fmts = ARRAY_SIZE(rkvdec_h264_decoded_fmts),
-               .decoded_fmts = rkvdec_h264_decoded_fmts,
+               .num_decoded_fmts = ARRAY_SIZE(rkvdec_h264_vp9_decoded_fmts),
+               .decoded_fmts = rkvdec_h264_vp9_decoded_fmts,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_VP9_FRAME,
+               .frmsize = {
+                       .min_width = 64,
+                       .max_width = 4096,
+                       .step_width = 64,
+                       .min_height = 64,
+                       .max_height = 2304,
+                       .step_height = 64,
+               },
+               .ctrls = &rkvdec_vp9_ctrls,
+               .ops = &rkvdec_vp9_fmt_ops,
+               .num_decoded_fmts = ARRAY_SIZE(rkvdec_h264_vp9_decoded_fmts),
+               .decoded_fmts = rkvdec_h264_vp9_decoded_fmts,
        }
 };
 
index 52ac387..2f4ea17 100644 (file)
@@ -42,14 +42,18 @@ struct rkvdec_run {
 
 struct rkvdec_vp9_decoded_buffer_info {
        /* Info needed when the decoded frame serves as a reference frame. */
-       u16 width;
-       u16 height;
-       u32 bit_depth : 4;
+       unsigned short width;
+       unsigned short height;
+       unsigned int bit_depth : 4;
 };
 
 struct rkvdec_decoded_buffer {
        /* Must be the first field in this struct. */
        struct v4l2_m2m_buffer base;
+
+       union {
+               struct rkvdec_vp9_decoded_buffer_info vp9;
+       };
 };
 
 static inline struct rkvdec_decoded_buffer *
@@ -116,4 +120,6 @@ void rkvdec_run_preamble(struct rkvdec_ctx *ctx, struct rkvdec_run *run);
 void rkvdec_run_postamble(struct rkvdec_ctx *ctx, struct rkvdec_run *run);
 
 extern const struct rkvdec_coded_fmt_ops rkvdec_h264_fmt_ops;
+extern const struct rkvdec_coded_fmt_ops rkvdec_vp9_fmt_ops;
+
 #endif /* RKVDEC_H_ */