2 * Copyright © 2011 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the
6 * "Software"), to deal in the Software without restriction, including
7 * without limitation the rights to use, copy, modify, merge, publish,
8 * distribute, sub license, and/or sell copies of the Software, and to
9 * permit persons to whom the Software is furnished to do so, subject to
10 * the following conditions:
12 * The above copyright notice and this permission notice (including the
13 * next paragraph) shall be included in all copies or substantial portions
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
17 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
18 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
19 * IN NO EVENT SHALL PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR
20 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
21 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
22 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
30 #include <va/va_enc_h264.h>
31 #include <va/va_enc_mpeg2.h>
32 #include <va/va_enc_vp8.h>
33 #include <va/va_enc_hevc.h>
36 #include "i965_encoder_utils.h"
38 #define BITSTREAM_ALLOCATE_STEPPING 4096
40 #define NAL_REF_IDC_NONE 0
41 #define NAL_REF_IDC_LOW 1
42 #define NAL_REF_IDC_MEDIUM 2
43 #define NAL_REF_IDC_HIGH 3
51 #define SLICE_TYPE_P 0
52 #define SLICE_TYPE_B 1
53 #define SLICE_TYPE_I 2
55 #define IS_I_SLICE(type) (SLICE_TYPE_I == (type) || SLICE_TYPE_I == (type - 5))
56 #define IS_P_SLICE(type) (SLICE_TYPE_P == (type) || SLICE_TYPE_P == (type - 5))
57 #define IS_B_SLICE(type) (SLICE_TYPE_B == (type) || SLICE_TYPE_B == (type - 5))
59 #define ENTROPY_MODE_CAVLC 0
60 #define ENTROPY_MODE_CABAC 1
62 #define PROFILE_IDC_BASELINE 66
63 #define PROFILE_IDC_MAIN 77
64 #define PROFILE_IDC_HIGH 100
70 #define IDR_WRADL_NUT 19
71 #define IDR_NLP_NUT 20
72 #define SLICE_TRAIL_N_NUT 0
73 #define SLICE_TRAIL_R_NUT 1
74 #define PREFIX_SEI_NUT 39
75 #define SUFFIX_SEI_NUT 40
77 struct __avc_bitstream {
80 int max_size_in_dword;
83 typedef struct __avc_bitstream avc_bitstream;
86 swap32(unsigned int val)
88 unsigned char *pval = (unsigned char *)&val;
90 return ((pval[0] << 24) |
97 avc_bitstream_start(avc_bitstream *bs)
99 bs->max_size_in_dword = BITSTREAM_ALLOCATE_STEPPING;
100 bs->buffer = calloc(bs->max_size_in_dword * sizeof(int), 1);
105 avc_bitstream_end(avc_bitstream *bs)
107 int pos = (bs->bit_offset >> 5);
108 int bit_offset = (bs->bit_offset & 0x1f);
109 int bit_left = 32 - bit_offset;
112 bs->buffer[pos] = swap32((bs->buffer[pos] << bit_left));
119 avc_bitstream_put_ui(avc_bitstream *bs, unsigned int val, int size_in_bits)
121 int pos = (bs->bit_offset >> 5);
122 int bit_offset = (bs->bit_offset & 0x1f);
123 int bit_left = 32 - bit_offset;
128 if (size_in_bits < 32)
129 val &= ((1 << size_in_bits) - 1);
131 bs->bit_offset += size_in_bits;
133 if (bit_left > size_in_bits) {
134 bs->buffer[pos] = (bs->buffer[pos] << size_in_bits | val);
136 size_in_bits -= bit_left;
137 if (bit_left == 32) {
138 bs->buffer[pos] = val;
140 bs->buffer[pos] = (bs->buffer[pos] << bit_left) | (val >> size_in_bits);
142 bs->buffer[pos] = swap32(bs->buffer[pos]);
144 if (pos + 1 == bs->max_size_in_dword) {
145 bs->max_size_in_dword += BITSTREAM_ALLOCATE_STEPPING;
146 bs->buffer = realloc(bs->buffer, bs->max_size_in_dword * sizeof(unsigned int));
152 bs->buffer[pos + 1] = val;
157 avc_bitstream_put_ue(avc_bitstream *bs, unsigned int val)
159 int size_in_bits = 0;
167 avc_bitstream_put_ui(bs, 0, size_in_bits - 1); // leading zero
168 avc_bitstream_put_ui(bs, val, size_in_bits);
172 avc_bitstream_put_se(avc_bitstream *bs, int val)
174 unsigned int new_val;
179 new_val = 2 * val - 1;
181 avc_bitstream_put_ue(bs, new_val);
185 avc_bitstream_byte_aligning(avc_bitstream *bs, int bit)
187 int bit_offset = (bs->bit_offset & 0x7);
188 int bit_left = 8 - bit_offset;
194 assert(bit == 0 || bit == 1);
197 new_val = (1 << bit_left) - 1;
201 avc_bitstream_put_ui(bs, new_val, bit_left);
203 static void avc_rbsp_trailing_bits(avc_bitstream *bs)
205 avc_bitstream_put_ui(bs, 1, 1);
206 avc_bitstream_byte_aligning(bs, 0);
208 static void nal_start_code_prefix(avc_bitstream *bs)
210 avc_bitstream_put_ui(bs, 0x00000001, 32);
213 static void nal_header(avc_bitstream *bs, int nal_ref_idc, int nal_unit_type)
215 avc_bitstream_put_ui(bs, 0, 1); /* forbidden_zero_bit: 0 */
216 avc_bitstream_put_ui(bs, nal_ref_idc, 2);
217 avc_bitstream_put_ui(bs, nal_unit_type, 5);
221 slice_header(avc_bitstream *bs,
222 VAEncSequenceParameterBufferH264 *sps_param,
223 VAEncPictureParameterBufferH264 *pic_param,
224 VAEncSliceParameterBufferH264 *slice_param)
226 int first_mb_in_slice = slice_param->macroblock_address;
228 avc_bitstream_put_ue(bs, first_mb_in_slice); /* first_mb_in_slice: 0 */
229 avc_bitstream_put_ue(bs, slice_param->slice_type); /* slice_type */
230 avc_bitstream_put_ue(bs, slice_param->pic_parameter_set_id); /* pic_parameter_set_id: 0 */
231 avc_bitstream_put_ui(bs, pic_param->frame_num, sps_param->seq_fields.bits.log2_max_frame_num_minus4 + 4); /* frame_num */
233 /* frame_mbs_only_flag == 1 */
234 if (!sps_param->seq_fields.bits.frame_mbs_only_flag) {
239 if (pic_param->pic_fields.bits.idr_pic_flag)
240 avc_bitstream_put_ue(bs, slice_param->idr_pic_id); /* idr_pic_id: 0 */
242 if (sps_param->seq_fields.bits.pic_order_cnt_type == 0) {
243 avc_bitstream_put_ui(bs, pic_param->CurrPic.TopFieldOrderCnt, sps_param->seq_fields.bits.log2_max_pic_order_cnt_lsb_minus4 + 4);
244 /* pic_order_present_flag == 0 */
250 /* redundant_pic_cnt_present_flag == 0 */
253 if (IS_P_SLICE(slice_param->slice_type)) {
254 avc_bitstream_put_ui(bs, slice_param->num_ref_idx_active_override_flag, 1); /* num_ref_idx_active_override_flag: */
256 if (slice_param->num_ref_idx_active_override_flag)
257 avc_bitstream_put_ue(bs, slice_param->num_ref_idx_l0_active_minus1);
259 /* ref_pic_list_reordering */
260 avc_bitstream_put_ui(bs, 0, 1); /* ref_pic_list_reordering_flag_l0: 0 */
261 } else if (IS_B_SLICE(slice_param->slice_type)) {
262 avc_bitstream_put_ui(bs, slice_param->direct_spatial_mv_pred_flag, 1); /* direct_spatial_mv_pred: 1 */
264 avc_bitstream_put_ui(bs, slice_param->num_ref_idx_active_override_flag, 1); /* num_ref_idx_active_override_flag: */
266 if (slice_param->num_ref_idx_active_override_flag) {
267 avc_bitstream_put_ue(bs, slice_param->num_ref_idx_l0_active_minus1);
268 avc_bitstream_put_ue(bs, slice_param->num_ref_idx_l1_active_minus1);
271 /* ref_pic_list_reordering */
272 avc_bitstream_put_ui(bs, 0, 1); /* ref_pic_list_reordering_flag_l0: 0 */
273 avc_bitstream_put_ui(bs, 0, 1); /* ref_pic_list_reordering_flag_l1: 0 */
276 if ((pic_param->pic_fields.bits.weighted_pred_flag &&
277 IS_P_SLICE(slice_param->slice_type)) ||
278 ((pic_param->pic_fields.bits.weighted_bipred_idc == 1) &&
279 IS_B_SLICE(slice_param->slice_type))) {
280 /* FIXME: fill weight/offset table */
284 /* dec_ref_pic_marking */
285 if (pic_param->pic_fields.bits.reference_pic_flag) { /* nal_ref_idc != 0 */
286 unsigned char no_output_of_prior_pics_flag = 0;
287 unsigned char long_term_reference_flag = 0;
288 unsigned char adaptive_ref_pic_marking_mode_flag = 0;
290 if (pic_param->pic_fields.bits.idr_pic_flag) {
291 avc_bitstream_put_ui(bs, no_output_of_prior_pics_flag, 1); /* no_output_of_prior_pics_flag: 0 */
292 avc_bitstream_put_ui(bs, long_term_reference_flag, 1); /* long_term_reference_flag: 0 */
294 avc_bitstream_put_ui(bs, adaptive_ref_pic_marking_mode_flag, 1); /* adaptive_ref_pic_marking_mode_flag: 0 */
298 if (pic_param->pic_fields.bits.entropy_coding_mode_flag &&
299 !IS_I_SLICE(slice_param->slice_type))
300 avc_bitstream_put_ue(bs, slice_param->cabac_init_idc); /* cabac_init_idc: 0 */
302 avc_bitstream_put_se(bs, slice_param->slice_qp_delta); /* slice_qp_delta: 0 */
304 /* ignore for SP/SI */
306 if (pic_param->pic_fields.bits.deblocking_filter_control_present_flag) {
307 avc_bitstream_put_ue(bs, slice_param->disable_deblocking_filter_idc); /* disable_deblocking_filter_idc: 0 */
309 if (slice_param->disable_deblocking_filter_idc != 1) {
310 avc_bitstream_put_se(bs, slice_param->slice_alpha_c0_offset_div2); /* slice_alpha_c0_offset_div2: 2 */
311 avc_bitstream_put_se(bs, slice_param->slice_beta_offset_div2); /* slice_beta_offset_div2: 2 */
315 if (pic_param->pic_fields.bits.entropy_coding_mode_flag) {
316 avc_bitstream_byte_aligning(bs, 1);
321 build_avc_slice_header(VAEncSequenceParameterBufferH264 *sps_param,
322 VAEncPictureParameterBufferH264 *pic_param,
323 VAEncSliceParameterBufferH264 *slice_param,
324 unsigned char **slice_header_buffer)
327 int is_idr = !!pic_param->pic_fields.bits.idr_pic_flag;
328 int is_ref = !!pic_param->pic_fields.bits.reference_pic_flag;
330 avc_bitstream_start(&bs);
331 nal_start_code_prefix(&bs);
333 if (IS_I_SLICE(slice_param->slice_type)) {
334 nal_header(&bs, NAL_REF_IDC_HIGH, is_idr ? NAL_IDR : NAL_NON_IDR);
335 } else if (IS_P_SLICE(slice_param->slice_type)) {
337 nal_header(&bs, NAL_REF_IDC_MEDIUM, NAL_NON_IDR);
339 assert(IS_B_SLICE(slice_param->slice_type));
341 nal_header(&bs, is_ref ? NAL_REF_IDC_LOW : NAL_REF_IDC_NONE, NAL_NON_IDR);
344 slice_header(&bs, sps_param, pic_param, slice_param);
346 avc_bitstream_end(&bs);
347 *slice_header_buffer = (unsigned char *)bs.buffer;
349 return bs.bit_offset;
353 build_avc_sei_buffering_period(int cpb_removal_length,
354 unsigned int init_cpb_removal_delay,
355 unsigned int init_cpb_removal_delay_offset,
356 unsigned char **sei_buffer)
358 unsigned char *byte_buf;
361 avc_bitstream nal_bs;
362 avc_bitstream sei_bs;
364 avc_bitstream_start(&sei_bs);
365 avc_bitstream_put_ue(&sei_bs, 0); /*seq_parameter_set_id*/
366 avc_bitstream_put_ui(&sei_bs, init_cpb_removal_delay, cpb_removal_length);
367 avc_bitstream_put_ui(&sei_bs, init_cpb_removal_delay_offset, cpb_removal_length);
368 if (sei_bs.bit_offset & 0x7) {
369 avc_bitstream_put_ui(&sei_bs, 1, 1);
371 avc_bitstream_end(&sei_bs);
372 byte_size = (sei_bs.bit_offset + 7) / 8;
374 avc_bitstream_start(&nal_bs);
375 nal_start_code_prefix(&nal_bs);
376 nal_header(&nal_bs, NAL_REF_IDC_NONE, NAL_SEI);
378 avc_bitstream_put_ui(&nal_bs, 0, 8);
379 avc_bitstream_put_ui(&nal_bs, byte_size, 8);
381 byte_buf = (unsigned char *)sei_bs.buffer;
382 for (i = 0; i < byte_size; i++) {
383 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
387 avc_rbsp_trailing_bits(&nal_bs);
388 avc_bitstream_end(&nal_bs);
390 *sei_buffer = (unsigned char *)nal_bs.buffer;
392 return nal_bs.bit_offset;
396 build_avc_sei_pic_timing(unsigned int cpb_removal_length, unsigned int cpb_removal_delay,
397 unsigned int dpb_output_length, unsigned int dpb_output_delay,
398 unsigned char **sei_buffer)
400 unsigned char *byte_buf;
403 avc_bitstream nal_bs;
404 avc_bitstream sei_bs;
406 avc_bitstream_start(&sei_bs);
407 avc_bitstream_put_ui(&sei_bs, cpb_removal_delay, cpb_removal_length);
408 avc_bitstream_put_ui(&sei_bs, dpb_output_delay, dpb_output_length);
409 if (sei_bs.bit_offset & 0x7) {
410 avc_bitstream_put_ui(&sei_bs, 1, 1);
412 avc_bitstream_end(&sei_bs);
413 byte_size = (sei_bs.bit_offset + 7) / 8;
415 avc_bitstream_start(&nal_bs);
416 nal_start_code_prefix(&nal_bs);
417 nal_header(&nal_bs, NAL_REF_IDC_NONE, NAL_SEI);
419 avc_bitstream_put_ui(&nal_bs, 0x01, 8);
420 avc_bitstream_put_ui(&nal_bs, byte_size, 8);
422 byte_buf = (unsigned char *)sei_bs.buffer;
423 for (i = 0; i < byte_size; i++) {
424 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
428 avc_rbsp_trailing_bits(&nal_bs);
429 avc_bitstream_end(&nal_bs);
431 *sei_buffer = (unsigned char *)nal_bs.buffer;
433 return nal_bs.bit_offset;
438 build_avc_sei_buffer_timing(unsigned int init_cpb_removal_length,
439 unsigned int init_cpb_removal_delay,
440 unsigned int init_cpb_removal_delay_offset,
441 unsigned int cpb_removal_length,
442 unsigned int cpb_removal_delay,
443 unsigned int dpb_output_length,
444 unsigned int dpb_output_delay,
445 unsigned char **sei_buffer)
447 unsigned char *byte_buf;
448 int bp_byte_size, i, pic_byte_size;
450 avc_bitstream nal_bs;
451 avc_bitstream sei_bp_bs, sei_pic_bs;
453 avc_bitstream_start(&sei_bp_bs);
454 avc_bitstream_put_ue(&sei_bp_bs, 0); /*seq_parameter_set_id*/
455 avc_bitstream_put_ui(&sei_bp_bs, init_cpb_removal_delay, cpb_removal_length);
456 avc_bitstream_put_ui(&sei_bp_bs, init_cpb_removal_delay_offset, cpb_removal_length);
457 if (sei_bp_bs.bit_offset & 0x7) {
458 avc_bitstream_put_ui(&sei_bp_bs, 1, 1);
460 avc_bitstream_end(&sei_bp_bs);
461 bp_byte_size = (sei_bp_bs.bit_offset + 7) / 8;
463 avc_bitstream_start(&sei_pic_bs);
464 avc_bitstream_put_ui(&sei_pic_bs, cpb_removal_delay, cpb_removal_length);
465 avc_bitstream_put_ui(&sei_pic_bs, dpb_output_delay, dpb_output_length);
466 if (sei_pic_bs.bit_offset & 0x7) {
467 avc_bitstream_put_ui(&sei_pic_bs, 1, 1);
469 avc_bitstream_end(&sei_pic_bs);
470 pic_byte_size = (sei_pic_bs.bit_offset + 7) / 8;
472 avc_bitstream_start(&nal_bs);
473 nal_start_code_prefix(&nal_bs);
474 nal_header(&nal_bs, NAL_REF_IDC_NONE, NAL_SEI);
476 /* Write the SEI buffer period data */
477 avc_bitstream_put_ui(&nal_bs, 0, 8);
478 avc_bitstream_put_ui(&nal_bs, bp_byte_size, 8);
480 byte_buf = (unsigned char *)sei_bp_bs.buffer;
481 for (i = 0; i < bp_byte_size; i++) {
482 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
485 /* write the SEI timing data */
486 avc_bitstream_put_ui(&nal_bs, 0x01, 8);
487 avc_bitstream_put_ui(&nal_bs, pic_byte_size, 8);
489 byte_buf = (unsigned char *)sei_pic_bs.buffer;
490 for (i = 0; i < pic_byte_size; i++) {
491 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
495 avc_rbsp_trailing_bits(&nal_bs);
496 avc_bitstream_end(&nal_bs);
498 *sei_buffer = (unsigned char *)nal_bs.buffer;
500 return nal_bs.bit_offset;
504 build_mpeg2_slice_header(VAEncSequenceParameterBufferMPEG2 *sps_param,
505 VAEncPictureParameterBufferMPEG2 *pic_param,
506 VAEncSliceParameterBufferMPEG2 *slice_param,
507 unsigned char **slice_header_buffer)
511 avc_bitstream_start(&bs);
512 avc_bitstream_end(&bs);
513 *slice_header_buffer = (unsigned char *)bs.buffer;
515 return bs.bit_offset;
518 static void binarize_qindex_delta(avc_bitstream *bs, int qindex_delta)
520 if (qindex_delta == 0)
521 avc_bitstream_put_ui(bs, 0, 1);
523 avc_bitstream_put_ui(bs, 1, 1);
524 avc_bitstream_put_ui(bs, abs(qindex_delta), 4);
526 if (qindex_delta < 0)
527 avc_bitstream_put_ui(bs, 1, 1);
529 avc_bitstream_put_ui(bs, 0, 1);
533 void binarize_vp8_frame_header(VAEncSequenceParameterBufferVP8 *seq_param,
534 VAEncPictureParameterBufferVP8 *pic_param,
535 VAQMatrixBufferVP8 *q_matrix,
536 struct gen6_mfc_context *mfc_context,
537 struct intel_encoder_context *encoder_context)
541 int is_intra_frame = !pic_param->pic_flags.bits.frame_type;
542 int log2num = pic_param->pic_flags.bits.num_token_partitions;
544 /* modify picture paramters */
545 pic_param->pic_flags.bits.loop_filter_adj_enable = 1;
546 pic_param->pic_flags.bits.mb_no_coeff_skip = 1;
547 pic_param->pic_flags.bits.forced_lf_adjustment = 1;
548 pic_param->pic_flags.bits.refresh_entropy_probs = 1;
549 pic_param->pic_flags.bits.segmentation_enabled = 0;
551 pic_param->pic_flags.bits.loop_filter_type = pic_param->pic_flags.bits.version / 2;
552 if (pic_param->pic_flags.bits.version > 1)
553 pic_param->loop_filter_level[0] = 0;
555 avc_bitstream_start(&bs);
557 if (is_intra_frame) {
558 avc_bitstream_put_ui(&bs, 0, 1);
559 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.clamping_type , 1);
562 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.segmentation_enabled, 1);
564 if (pic_param->pic_flags.bits.segmentation_enabled) {
565 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.update_mb_segmentation_map, 1);
566 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.update_segment_feature_data, 1);
567 if (pic_param->pic_flags.bits.update_segment_feature_data) {
571 if (pic_param->pic_flags.bits.update_mb_segmentation_map) {
572 for (i = 0; i < 3; i++) {
573 if (mfc_context->vp8_state.mb_segment_tree_probs[i] == 255)
574 avc_bitstream_put_ui(&bs, 0, 1);
576 avc_bitstream_put_ui(&bs, 1, 1);
577 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.mb_segment_tree_probs[i], 8);
583 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.loop_filter_type, 1);
584 avc_bitstream_put_ui(&bs, pic_param->loop_filter_level[0], 6);
585 avc_bitstream_put_ui(&bs, pic_param->sharpness_level, 3);
587 mfc_context->vp8_state.frame_header_lf_update_pos = bs.bit_offset;
589 if (pic_param->pic_flags.bits.forced_lf_adjustment) {
590 avc_bitstream_put_ui(&bs, 1, 1);//mode_ref_lf_delta_enable = 1
591 avc_bitstream_put_ui(&bs, 1, 1);//mode_ref_lf_delta_update = 1
593 for (i = 0; i < 4; i++) {
594 avc_bitstream_put_ui(&bs, 1, 1);
595 if (pic_param->ref_lf_delta[i] > 0) {
596 avc_bitstream_put_ui(&bs, (abs(pic_param->ref_lf_delta[i]) & 0x3F), 6);
597 avc_bitstream_put_ui(&bs, 0, 1);
599 avc_bitstream_put_ui(&bs, (abs(pic_param->ref_lf_delta[i]) & 0x3F), 6);
600 avc_bitstream_put_ui(&bs, 1, 1);
604 for (i = 0; i < 4; i++) {
605 avc_bitstream_put_ui(&bs, 1, 1);
606 if (pic_param->mode_lf_delta[i] > 0) {
607 avc_bitstream_put_ui(&bs, (abs(pic_param->mode_lf_delta[i]) & 0x3F), 6);
608 avc_bitstream_put_ui(&bs, 0, 1);
610 avc_bitstream_put_ui(&bs, (abs(pic_param->mode_lf_delta[i]) & 0x3F), 6);
611 avc_bitstream_put_ui(&bs, 1, 1);
616 avc_bitstream_put_ui(&bs, 0, 1);//mode_ref_lf_delta_enable = 0
619 avc_bitstream_put_ui(&bs, log2num, 2);
621 mfc_context->vp8_state.frame_header_qindex_update_pos = bs.bit_offset;
623 avc_bitstream_put_ui(&bs, q_matrix->quantization_index[0], 7);
625 for (i = 0; i < 5; i++)
626 binarize_qindex_delta(&bs, q_matrix->quantization_index_delta[i]);
628 if (!is_intra_frame) {
629 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.refresh_golden_frame, 1);
630 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.refresh_alternate_frame, 1);
632 if (!pic_param->pic_flags.bits.refresh_golden_frame)
633 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.copy_buffer_to_golden, 2);
635 if (!pic_param->pic_flags.bits.refresh_alternate_frame)
636 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.copy_buffer_to_alternate, 2);
638 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.sign_bias_golden, 1);
639 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.sign_bias_alternate, 1);
642 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.refresh_entropy_probs, 1);
645 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.refresh_last, 1);
647 mfc_context->vp8_state.frame_header_token_update_pos = bs.bit_offset;
649 for (i = 0; i < 4 * 8 * 3 * 11; i++)
650 avc_bitstream_put_ui(&bs, 0, 1); //don't update coeff_probs
652 avc_bitstream_put_ui(&bs, pic_param->pic_flags.bits.mb_no_coeff_skip, 1);
653 if (pic_param->pic_flags.bits.mb_no_coeff_skip)
654 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.prob_skip_false, 8);
656 if (!is_intra_frame) {
657 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.prob_intra, 8);
658 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.prob_last, 8);
659 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.prob_gf, 8);
661 avc_bitstream_put_ui(&bs, 1, 1); //y_mode_update_flag = 1
662 for (i = 0; i < 4; i++) {
663 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.y_mode_probs[i], 8);
666 avc_bitstream_put_ui(&bs, 1, 1); //uv_mode_update_flag = 1
667 for (i = 0; i < 3; i++) {
668 avc_bitstream_put_ui(&bs, mfc_context->vp8_state.uv_mode_probs[i], 8);
671 mfc_context->vp8_state.frame_header_bin_mv_upate_pos = bs.bit_offset;
673 for (i = 0; i < 2 ; i++) {
674 for (j = 0; j < 19; j++) {
675 avc_bitstream_put_ui(&bs, 0, 1);
676 //avc_bitstream_put_ui(&bs, mfc_context->vp8_state.mv_probs[i][j], 7);
681 avc_bitstream_end(&bs);
683 mfc_context->vp8_state.vp8_frame_header = (unsigned char *)bs.buffer;
684 mfc_context->vp8_state.frame_header_bit_count = bs.bit_offset;
687 /* HEVC to do for internal header generated*/
689 void nal_header_hevc(avc_bitstream *bs, int nal_unit_type, int temporalid)
691 /* forbidden_zero_bit: 0 */
692 avc_bitstream_put_ui(bs, 0, 1);
694 avc_bitstream_put_ui(bs, nal_unit_type, 6);
695 /* layer_id. currently it is zero */
696 avc_bitstream_put_ui(bs, 0, 6);
698 avc_bitstream_put_ui(bs, temporalid + 1, 3);
701 int build_hevc_sei_buffering_period(int init_cpb_removal_delay_length,
702 unsigned int init_cpb_removal_delay,
703 unsigned int init_cpb_removal_delay_offset,
704 unsigned char **sei_buffer)
706 unsigned char *byte_buf;
708 //unsigned int cpb_removal_delay;
710 avc_bitstream nal_bs;
711 avc_bitstream sei_bp_bs;
713 avc_bitstream_start(&sei_bp_bs);
714 avc_bitstream_put_ue(&sei_bp_bs, 0); /*seq_parameter_set_id*/
715 /* SEI buffer period info */
716 /* NALHrdBpPresentFlag == 1 */
717 avc_bitstream_put_ui(&sei_bp_bs, init_cpb_removal_delay, init_cpb_removal_delay_length);
718 avc_bitstream_put_ui(&sei_bp_bs, init_cpb_removal_delay_offset, init_cpb_removal_delay_length);
719 if (sei_bp_bs.bit_offset & 0x7) {
720 avc_bitstream_put_ui(&sei_bp_bs, 1, 1);
722 avc_bitstream_end(&sei_bp_bs);
723 bp_byte_size = (sei_bp_bs.bit_offset + 7) / 8;
725 avc_bitstream_start(&nal_bs);
726 nal_start_code_prefix(&nal_bs);
727 nal_header_hevc(&nal_bs, PREFIX_SEI_NUT , 0);
729 /* Write the SEI buffer period data */
730 avc_bitstream_put_ui(&nal_bs, 0, 8);
731 avc_bitstream_put_ui(&nal_bs, bp_byte_size, 8);
733 byte_buf = (unsigned char *)sei_bp_bs.buffer;
734 for (i = 0; i < bp_byte_size; i++) {
735 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
739 avc_rbsp_trailing_bits(&nal_bs);
740 avc_bitstream_end(&nal_bs);
742 *sei_buffer = (unsigned char *)nal_bs.buffer;
744 return nal_bs.bit_offset;
747 int build_hevc_idr_sei_buffer_timing(unsigned int init_cpb_removal_delay_length,
748 unsigned int init_cpb_removal_delay,
749 unsigned int init_cpb_removal_delay_offset,
750 unsigned int cpb_removal_length,
751 unsigned int cpb_removal_delay,
752 unsigned int dpb_output_length,
753 unsigned int dpb_output_delay,
754 unsigned char **sei_buffer)
756 unsigned char *byte_buf;
757 int bp_byte_size, i, pic_byte_size;
758 //unsigned int cpb_removal_delay;
760 avc_bitstream nal_bs;
761 avc_bitstream sei_bp_bs, sei_pic_bs;
763 avc_bitstream_start(&sei_bp_bs);
764 avc_bitstream_put_ue(&sei_bp_bs, 0); /*seq_parameter_set_id*/
765 /* SEI buffer period info */
766 /* NALHrdBpPresentFlag == 1 */
767 avc_bitstream_put_ui(&sei_bp_bs, init_cpb_removal_delay, init_cpb_removal_delay_length);
768 avc_bitstream_put_ui(&sei_bp_bs, init_cpb_removal_delay_offset, init_cpb_removal_delay_length);
769 if (sei_bp_bs.bit_offset & 0x7) {
770 avc_bitstream_put_ui(&sei_bp_bs, 1, 1);
772 avc_bitstream_end(&sei_bp_bs);
773 bp_byte_size = (sei_bp_bs.bit_offset + 7) / 8;
775 /* SEI pic timing info */
776 avc_bitstream_start(&sei_pic_bs);
777 /* The info of CPB and DPB delay is controlled by CpbDpbDelaysPresentFlag,
778 * which is derived as 1 if one of the following conditions is true:
779 * nal_hrd_parameters_present_flag is present in the avc_bitstream and is equal to 1,
780 * vcl_hrd_parameters_present_flag is present in the avc_bitstream and is equal to 1,
782 //cpb_removal_delay = (hevc_context.current_cpb_removal - hevc_context.prev_idr_cpb_removal);
783 avc_bitstream_put_ui(&sei_pic_bs, cpb_removal_delay, cpb_removal_length);
784 avc_bitstream_put_ui(&sei_pic_bs, dpb_output_delay, dpb_output_length);
785 if (sei_pic_bs.bit_offset & 0x7) {
786 avc_bitstream_put_ui(&sei_pic_bs, 1, 1);
788 /* The pic_structure_present_flag determines whether the pic_structure
789 * info is written into the SEI pic timing info.
790 * Currently it is set to zero.
792 avc_bitstream_end(&sei_pic_bs);
793 pic_byte_size = (sei_pic_bs.bit_offset + 7) / 8;
795 avc_bitstream_start(&nal_bs);
796 nal_start_code_prefix(&nal_bs);
797 nal_header_hevc(&nal_bs, PREFIX_SEI_NUT , 0);
799 /* Write the SEI buffer period data */
800 avc_bitstream_put_ui(&nal_bs, 0, 8);
801 avc_bitstream_put_ui(&nal_bs, bp_byte_size, 8);
803 byte_buf = (unsigned char *)sei_bp_bs.buffer;
804 for (i = 0; i < bp_byte_size; i++) {
805 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
808 /* write the SEI pic timing data */
809 avc_bitstream_put_ui(&nal_bs, 0x01, 8);
810 avc_bitstream_put_ui(&nal_bs, pic_byte_size, 8);
812 byte_buf = (unsigned char *)sei_pic_bs.buffer;
813 for (i = 0; i < pic_byte_size; i++) {
814 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
818 avc_rbsp_trailing_bits(&nal_bs);
819 avc_bitstream_end(&nal_bs);
821 *sei_buffer = (unsigned char *)nal_bs.buffer;
823 return nal_bs.bit_offset;
826 int build_hevc_sei_pic_timing(unsigned int cpb_removal_length, unsigned int cpb_removal_delay,
827 unsigned int dpb_output_length, unsigned int dpb_output_delay,
828 unsigned char **sei_buffer)
830 unsigned char *byte_buf;
831 int i, pic_byte_size;
832 //unsigned int cpb_removal_delay;
834 avc_bitstream nal_bs;
835 avc_bitstream sei_pic_bs;
837 avc_bitstream_start(&sei_pic_bs);
838 /* The info of CPB and DPB delay is controlled by CpbDpbDelaysPresentFlag,
839 * which is derived as 1 if one of the following conditions is true:
840 * nal_hrd_parameters_present_flag is present in the avc_bitstream and is equal to 1,
841 * vcl_hrd_parameters_present_flag is present in the avc_bitstream and is equal to 1,
843 //cpb_removal_delay = (hevc_context.current_cpb_removal - hevc_context.current_idr_cpb_removal);
844 avc_bitstream_put_ui(&sei_pic_bs, cpb_removal_delay, cpb_removal_length);
845 avc_bitstream_put_ui(&sei_pic_bs, dpb_output_delay, dpb_output_length);
846 if (sei_pic_bs.bit_offset & 0x7) {
847 avc_bitstream_put_ui(&sei_pic_bs, 1, 1);
850 /* The pic_structure_present_flag determines whether the pic_structure
851 * info is written into the SEI pic timing info.
852 * Currently it is set to zero.
854 avc_bitstream_end(&sei_pic_bs);
855 pic_byte_size = (sei_pic_bs.bit_offset + 7) / 8;
857 avc_bitstream_start(&nal_bs);
858 nal_start_code_prefix(&nal_bs);
859 nal_header_hevc(&nal_bs, PREFIX_SEI_NUT , 0);
861 /* write the SEI Pic timing data */
862 avc_bitstream_put_ui(&nal_bs, 0x01, 8);
863 avc_bitstream_put_ui(&nal_bs, pic_byte_size, 8);
865 byte_buf = (unsigned char *)sei_pic_bs.buffer;
866 for (i = 0; i < pic_byte_size; i++) {
867 avc_bitstream_put_ui(&nal_bs, byte_buf[i], 8);
871 avc_rbsp_trailing_bits(&nal_bs);
872 avc_bitstream_end(&nal_bs);
874 *sei_buffer = (unsigned char *)nal_bs.buffer;
876 return nal_bs.bit_offset;
879 typedef struct _RefPicSet {
880 unsigned char num_negative_pics;
881 unsigned char num_positive_pics;
882 unsigned char delta_poc_s0_minus1[8];
883 unsigned char used_by_curr_pic_s0_flag[8];
884 unsigned char delta_poc_s1_minus1[8];
885 unsigned char used_by_curr_pic_s1_flag[8];
886 unsigned int inter_ref_pic_set_prediction_flag;
889 void hevc_short_term_ref_pic_set(avc_bitstream *bs, VAEncSliceParameterBufferHEVC *slice_param, int curPicOrderCnt)
891 hevcRefPicSet hevc_rps;
892 int rps_idx = 1, ref_idx = 0;
895 hevc_rps.inter_ref_pic_set_prediction_flag = 0;
896 /* s0: between I and P/B; s1 : between P and B */
897 hevc_rps.num_negative_pics = (slice_param->slice_type != HEVC_SLICE_I) ? 1 : 0;
898 hevc_rps.num_positive_pics = (slice_param->slice_type == HEVC_SLICE_B) ? 1 : 0;
899 hevc_rps.delta_poc_s0_minus1[0] = 0;
900 hevc_rps.used_by_curr_pic_s0_flag[0] = 0;
901 hevc_rps.delta_poc_s1_minus1[0] = 0;
902 hevc_rps.used_by_curr_pic_s1_flag[0] = 0;
903 if (slice_param->num_ref_idx_l0_active_minus1 == 0) {
904 hevc_rps.delta_poc_s0_minus1[0] = (slice_param->slice_type == HEVC_SLICE_I) ? 0 : (curPicOrderCnt - slice_param->ref_pic_list0[0].pic_order_cnt - 1); //0;
905 hevc_rps.used_by_curr_pic_s0_flag[0] = 1;
907 if (slice_param->num_ref_idx_l1_active_minus1 == 0) {
908 hevc_rps.delta_poc_s1_minus1[0] = (slice_param->slice_type == HEVC_SLICE_I) ? 0 : (slice_param->ref_pic_list1[0].pic_order_cnt - curPicOrderCnt - 1);
909 hevc_rps.used_by_curr_pic_s1_flag[0] = 1;
913 avc_bitstream_put_ui(bs, hevc_rps.inter_ref_pic_set_prediction_flag, 1);
915 if (hevc_rps.inter_ref_pic_set_prediction_flag) {
919 avc_bitstream_put_ue(bs, hevc_rps.num_negative_pics);
920 avc_bitstream_put_ue(bs, hevc_rps.num_positive_pics);
922 for (i = 0; i < hevc_rps.num_negative_pics; i++) {
923 avc_bitstream_put_ue(bs, hevc_rps.delta_poc_s0_minus1[ref_idx]);
924 avc_bitstream_put_ui(bs, hevc_rps.used_by_curr_pic_s0_flag[ref_idx], 1);
926 for (i = 0; i < hevc_rps.num_positive_pics; i++) {
927 avc_bitstream_put_ue(bs, hevc_rps.delta_poc_s1_minus1[ref_idx]);
928 avc_bitstream_put_ui(bs, hevc_rps.used_by_curr_pic_s1_flag[ref_idx], 1);
935 static void slice_rbsp(avc_bitstream *bs,
937 VAEncSequenceParameterBufferHEVC *seq_param,
938 VAEncPictureParameterBufferHEVC *pic_param,
939 VAEncSliceParameterBufferHEVC *slice_param)
941 int log2_cu_size = seq_param->log2_min_luma_coding_block_size_minus3 + 3;
942 int log2_ctb_size = seq_param->log2_diff_max_min_luma_coding_block_size + log2_cu_size;
943 int ctb_size = 1 << log2_ctb_size;
945 int picture_width_in_ctb = (seq_param->pic_width_in_luma_samples + ctb_size - 1) / ctb_size;
946 int picture_height_in_ctb = (seq_param->pic_height_in_luma_samples + ctb_size - 1) / ctb_size;
948 /* first_slice_segment_in_pic_flag */
949 if (slice_index == 0) {
950 avc_bitstream_put_ui(bs, 1, 1);
952 avc_bitstream_put_ui(bs, 0, 1);
955 /* no_output_of_prior_pics_flag */
956 if (pic_param->pic_fields.bits.idr_pic_flag)
957 avc_bitstream_put_ui(bs, 1, 1);
959 /* slice_pic_parameter_set_id */
960 avc_bitstream_put_ue(bs, 0);
962 /* not the first slice */
969 num_ctus = picture_width_in_ctb * picture_height_in_ctb;
970 bit_size = ceilf(log2f(num_ctus));
972 if (pic_param->pic_fields.bits.dependent_slice_segments_enabled_flag) {
973 avc_bitstream_put_ui(bs,
974 slice_param->slice_fields.bits.dependent_slice_segment_flag, 1);
976 /* slice_segment_address is based on Ceil(log2(PictureSizeinCtbs)) */
977 avc_bitstream_put_ui(bs, slice_param->slice_segment_address, bit_size);
979 if (!slice_param->slice_fields.bits.dependent_slice_segment_flag) {
980 /* slice_reserved_flag */
983 avc_bitstream_put_ue(bs, slice_param->slice_type);
984 /* use the inferred the value of pic_output_flag */
986 /* colour_plane_id */
987 if (seq_param->seq_fields.bits.separate_colour_plane_flag) {
988 avc_bitstream_put_ui(bs, slice_param->slice_fields.bits.colour_plane_id, 1);
991 if (!pic_param->pic_fields.bits.idr_pic_flag) {
992 int Log2MaxPicOrderCntLsb = 8;
993 avc_bitstream_put_ui(bs, pic_param->decoded_curr_pic.pic_order_cnt, Log2MaxPicOrderCntLsb);
995 //if (!slice_param->short_term_ref_pic_set_sps_flag)
997 /* short_term_ref_pic_set_sps_flag.
998 * Use zero and then pass the RPS from slice_header
1000 avc_bitstream_put_ui(bs, 0, 1);
1002 * Add the short_term reference picture set
1004 hevc_short_term_ref_pic_set(bs, slice_param, pic_param->decoded_curr_pic.pic_order_cnt);
1006 /* long term reference present flag. unpresent */
1009 /* sps temporal MVP*/
1010 if (seq_param->seq_fields.bits.sps_temporal_mvp_enabled_flag) {
1011 avc_bitstream_put_ui(bs,
1012 slice_param->slice_fields.bits.slice_temporal_mvp_enabled_flag, 1);
1016 /* long term reference present flag. unpresent */
1018 /* sample adaptive offset enabled flag */
1019 if (seq_param->seq_fields.bits.sample_adaptive_offset_enabled_flag) {
1020 avc_bitstream_put_ui(bs, slice_param->slice_fields.bits.slice_sao_luma_flag, 1);
1021 avc_bitstream_put_ui(bs, slice_param->slice_fields.bits.slice_sao_chroma_flag, 1);
1024 if (slice_param->slice_type != HEVC_SLICE_I) {
1025 /* num_ref_idx_active_override_flag. 0 */
1026 avc_bitstream_put_ui(bs, 0, 1);
1027 /* lists_modification_flag is unpresent NumPocTotalCurr > 1 ,here it is 1*/
1029 /* No reference picture set modification */
1031 /* MVD_l1_zero_flag */
1032 if (slice_param->slice_type == HEVC_SLICE_B)
1033 avc_bitstream_put_ui(bs, slice_param->slice_fields.bits.mvd_l1_zero_flag, 1);
1035 /* cabac_init_present_flag. 0 */
1037 /* slice_temporal_mvp_enabled_flag. */
1038 if (slice_param->slice_fields.bits.slice_temporal_mvp_enabled_flag) {
1039 if (slice_param->slice_type == HEVC_SLICE_B)
1040 avc_bitstream_put_ui(bs, slice_param->slice_fields.bits.collocated_from_l0_flag, 1);
1042 * TBD: Add the collocated_ref_idx.
1045 if (((pic_param->pic_fields.bits.weighted_pred_flag) &&
1046 (slice_param->slice_type == HEVC_SLICE_P)) ||
1047 ((pic_param->pic_fields.bits.weighted_bipred_flag) &&
1048 (slice_param->slice_type == HEVC_SLICE_B))) {
1050 * add the weighted table
1053 avc_bitstream_put_ue(bs, 5 - slice_param->max_num_merge_cand);
1055 /* slice_qp_delta */
1056 avc_bitstream_put_ue(bs, slice_param->slice_qp_delta);
1058 /* slice_cb/cr_qp_offset is controlled by pps_slice_chroma_qp_offsets_present_flag
1059 * The present flag is set to 1.
1061 avc_bitstream_put_ue(bs, slice_param->slice_cb_qp_offset);
1062 avc_bitstream_put_ue(bs, slice_param->slice_cr_qp_offset);
1065 * deblocking_filter_override_flag is controlled by
1066 * deblocking_filter_override_enabled_flag.
1067 * The override_enabled_flag is zero.
1068 * deblocking_filter_override_flag is zero. then
1069 * slice_deblocking_filter_disabled_flag is also zero
1070 * (It is inferred to be equal to pps_deblocking_filter_disabled_flag.
1073 /* slice_loop_filter_across_slices_enabled_flag is controlled
1074 * by pps_loop_filter_across_slices_enabled_flag &&
1075 * (slice_sao_luma_flag | | slice_sao_chroma_flag | |
1076 * !slice_deblocking_filter_disabled_flag ))
1081 if (pic_param->pic_fields.bits.tiles_enabled_flag ||
1082 pic_param->pic_fields.bits.entropy_coding_sync_enabled_flag) {
1084 * Add the Entry-points && tile definition.
1088 /* slice_segment_header_extension_present_flag. Not present */
1090 /* byte_alignment */
1091 avc_rbsp_trailing_bits(bs);
1094 int get_hevc_slice_nalu_type(VAEncPictureParameterBufferHEVC *pic_param)
1096 if (pic_param->pic_fields.bits.idr_pic_flag)
1097 return IDR_WRADL_NUT;
1098 else if (pic_param->pic_fields.bits.reference_pic_flag)
1099 return SLICE_TRAIL_R_NUT;
1101 return SLICE_TRAIL_N_NUT;
1104 int build_hevc_slice_header(VAEncSequenceParameterBufferHEVC *seq_param,
1105 VAEncPictureParameterBufferHEVC *pic_param,
1106 VAEncSliceParameterBufferHEVC *slice_param,
1107 unsigned char **header_buffer,
1112 avc_bitstream_start(&bs);
1113 nal_start_code_prefix(&bs);
1114 nal_header_hevc(&bs, get_hevc_slice_nalu_type(pic_param), 0);
1115 slice_rbsp(&bs, slice_index, seq_param, pic_param, slice_param);
1116 avc_bitstream_end(&bs);
1118 *header_buffer = (unsigned char *)bs.buffer;
1119 return bs.bit_offset;
1123 intel_avc_find_skipemulcnt(unsigned char *buf, int bits_length)
1126 int leading_zero_cnt, byte_length, zero_byte;
1130 #define NAL_UNIT_TYPE_MASK 0x1f
1131 #define HW_MAX_SKIP_LENGTH 15
1133 byte_length = ALIGN(bits_length, 32) >> 3;
1136 leading_zero_cnt = 0;
1138 for (i = 0; i < byte_length - 4; i++) {
1139 if (((buf[i] == 0) && (buf[i + 1] == 0) && (buf[i + 2] == 1)) ||
1140 ((buf[i] == 0) && (buf[i + 1] == 0) && (buf[i + 2] == 0) && (buf[i + 3] == 1))) {
1147 /* warning message is complained. But anyway it will be inserted. */
1148 WARN_ONCE("Invalid packed header data. "
1149 "Can't find the 000001 start_prefix code\n");
1152 i = leading_zero_cnt;
1155 if (!((buf[i] == 0) && (buf[i + 1] == 0) && (buf[i + 2] == 1)))
1158 skip_cnt = leading_zero_cnt + zero_byte + 3;
1160 /* the unit header byte is accounted */
1161 nal_unit_type = (buf[skip_cnt]) & NAL_UNIT_TYPE_MASK;
1164 if (nal_unit_type == 14 || nal_unit_type == 20 || nal_unit_type == 21) {
1165 /* more unit header bytes are accounted for MVC/SVC */
1168 if (skip_cnt > HW_MAX_SKIP_LENGTH) {
1169 WARN_ONCE("Too many leading zeros are padded for packed data. "
1170 "It is beyond the HW range.!!!\n");