1 /*
2 * Copyright 2021 Rockchip Electronics Co. LTD
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16 #define MODULE_TAG "av1d_cbs"
17
18 #include <string.h>
19
20 #include "mpp_mem.h"
21 #include "mpp_debug.h"
22 #include "mpp_bitread.h"
23 #include "mpp_bitwrite.h"
24 #include "rk_hdr_meta_com.h"
25
26 #include "av1d_parser.h"
27
28 #ifndef UINT32_MAX
29 #define UINT32_MAX 0xFFFFFFFF
30 #endif
31
32 #ifndef INT_MAX
33 #define INT_MAX 2147483647 /* maximum (signed) int value */
34 #endif
35
36 #define BUFFER_PADDING_SIZE 64
37 #define MAX_UINT_BITS(length) ((UINT64_C(1) << (length)) - 1)
38 #define MAX_INT_BITS(length) ((INT64_C(1) << ((length) - 1)) - 1)
39 #define MIN_INT_BITS(length) (-(INT64_C(1) << ((length) - 1)))
40
41 /**
42 * Clip a signed integer into the -(2^p),(2^p-1) range.
43 * @param a value to clip
44 * @param p bit position to clip at
45 * @return clipped value
46 */
mpp_clip_uintp2(RK_S32 a,RK_S32 p)47 static RK_U32 mpp_clip_uintp2(RK_S32 a, RK_S32 p)
48 {
49 if (a & ~((1 << p) - 1)) return -a >> 31 & ((1 << p) - 1);
50 else return a;
51 }
52
mpp_av1_read_uvlc(BitReadCtx_t * gbc,const char * name,RK_U32 * write_to,RK_U32 range_min,RK_U32 range_max)53 static RK_S32 mpp_av1_read_uvlc(BitReadCtx_t *gbc, const char *name, RK_U32 *write_to,
54 RK_U32 range_min, RK_U32 range_max)
55 {
56 RK_U32 value;
57
58 mpp_read_ue(gbc, &value);
59
60 if (value < range_min || value > range_max) {
61 mpp_err_f("%s out of range: "
62 "%d, but must be in [%d,%d].\n",
63 name, value, range_min, range_max);
64 return MPP_NOK;
65 }
66 *write_to = value;
67 return MPP_OK;
68 }
69
70
mpp_av1_read_leb128(BitReadCtx_t * gbc,RK_U64 * write_to)71 static RK_S32 mpp_av1_read_leb128(BitReadCtx_t *gbc, RK_U64 *write_to)
72 {
73 RK_U64 value;
74 RK_S32 err = 0, i;
75
76 value = 0;
77 for (i = 0; i < 8; i++) {
78 RK_U32 byte;
79
80 READ_BITS(gbc, 8, &byte);
81
82 if (err < 0)
83 return err;
84
85 value |= (RK_U64)(byte & 0x7f) << (i * 7);
86 if (!(byte & 0x80))
87 break;
88 }
89
90 if (value > UINT32_MAX)
91 return MPP_NOK;
92
93
94 *write_to = value;
95 return MPP_OK;
96
97 __bitread_error:
98 return MPP_NOK;
99
100 }
101
mpp_av1_read_ns(BitReadCtx_t * gbc,const char * name,RK_U32 n,RK_U32 * write_to)102 static RK_S32 mpp_av1_read_ns(BitReadCtx_t *gbc, const char *name,
103 RK_U32 n, RK_U32 *write_to)
104 {
105 RK_U32 m, v, extra_bit, value;
106 RK_S32 w;
107
108 w = mpp_log2(n) + 1;
109 m = (1 << w) - n;
110
111 if (mpp_get_bits_left(gbc) < w) {
112 mpp_err_f("Invalid non-symmetric value at "
113 "%s: bitstream ended.\n", name);
114 return MPP_NOK;
115 }
116 if (w - 1 > 0)
117 READ_BITS(gbc, w - 1, &v);
118 else
119 v = 0;
120
121 if (v < m) {
122 value = v;
123 } else {
124 READ_ONEBIT(gbc, &extra_bit);
125 value = (v << 1) - m + extra_bit;
126 }
127
128 *write_to = value;
129 return MPP_OK;
130
131 __bitread_error:
132 return MPP_NOK;
133
134 }
135
mpp_av1_read_increment(BitReadCtx_t * gbc,RK_U32 range_min,RK_U32 range_max,const char * name,RK_U32 * write_to)136 static RK_S32 mpp_av1_read_increment(BitReadCtx_t *gbc, RK_U32 range_min,
137 RK_U32 range_max, const char *name,
138 RK_U32 *write_to)
139 {
140 RK_U32 value;
141 RK_S32 i;
142 RK_S8 bits[33];
143
144 mpp_assert(range_min <= range_max && range_max - range_min < sizeof(bits) - 1);
145
146 for (i = 0, value = range_min; value < range_max;) {
147 RK_U8 tmp = 0;
148 if (mpp_get_bits_left(gbc) < 1) {
149 mpp_err_f("Invalid increment value at "
150 "%s: bitstream ended.\n", name);
151 return MPP_NOK;
152 }
153 READ_ONEBIT(gbc, &tmp);
154 if (tmp) {
155 bits[i++] = '1';
156 ++value;
157 } else {
158 bits[i++] = '0';
159 break;
160 }
161 }
162 *write_to = value;
163 return MPP_OK;
164
165 __bitread_error:
166 return MPP_NOK;
167 }
168
mpp_av1_read_unsigned(BitReadCtx_t * gbc,RK_S32 width,const char * name,RK_U32 * write_to,RK_U32 range_min,RK_U32 range_max)169 RK_S32 mpp_av1_read_unsigned(BitReadCtx_t *gbc,
170 RK_S32 width, const char *name,
171 RK_U32 *write_to, RK_U32 range_min,
172 RK_U32 range_max)
173 {
174 RK_U32 value;
175
176 mpp_assert(width > 0 && width <= 32);
177
178 if (mpp_get_bits_left(gbc) < width) {
179 mpp_err_f("Invalid value at "
180 "%s: bitstream ended.\n", name);
181 return MPP_NOK;
182 }
183
184 READ_BITS_LONG(gbc, width, &value);
185
186 if (value < range_min || value > range_max) {
187 mpp_err_f("%s out of range: "
188 "%d, but must be in [%d,%d].\n",
189 name, value, range_min, range_max);
190 return MPP_NOK;
191 }
192
193 *write_to = value;
194 return 0;
195
196 __bitread_error:
197 return MPP_NOK;
198
199 }
200
sign_extend(RK_S32 val,RK_U8 bits)201 static RK_S32 sign_extend(RK_S32 val, RK_U8 bits)
202 {
203 RK_U8 shift = 8 * sizeof(RK_S32) - bits;
204 RK_S32 v = { (RK_U8) val << shift };
205 return v >> shift;
206 }
207
mpp_av1_read_signed(BitReadCtx_t * gbc,RK_S32 width,const char * name,RK_S32 * write_to,RK_S32 range_min,RK_S32 range_max)208 RK_S32 mpp_av1_read_signed(BitReadCtx_t *gbc,
209 RK_S32 width, const char *name,
210 RK_S32 *write_to, RK_S32 range_min,
211 RK_S32 range_max)
212 {
213 RK_S32 value;
214
215 mpp_assert(width > 0 && width <= 32);
216
217 if (mpp_get_bits_left(gbc) < width) {
218 mpp_err_f("Invalid value at "
219 "%s: bitstream ended.\n", name);
220 return MPP_NOK;
221 }
222
223 READ_BITS_LONG(gbc, width, &value);
224 value = sign_extend(value, width);
225 if (value < range_min || value > range_max) {
226 mpp_err_f("%s out of range: "
227 "%d, but must be in [%d,%d].\n",
228 name, value, range_min, range_max);
229 return MPP_NOK;
230 }
231
232 *write_to = value;
233 return 0;
234
235 __bitread_error:
236 return MPP_NOK;
237
238 }
239
mpp_av1_read_subexp(BitReadCtx_t * gbc,RK_U32 range_max,RK_U32 * write_to)240 static RK_S32 mpp_av1_read_subexp(BitReadCtx_t *gbc,
241 RK_U32 range_max, RK_U32 *write_to)
242 {
243 RK_U32 value;
244 RK_S32 err;
245 RK_U32 max_len, len, range_offset, range_bits;
246
247 max_len = mpp_log2(range_max - 1) - 3;
248
249 err = mpp_av1_read_increment(gbc, 0, max_len, "subexp_more_bits", &len);
250 if (err < 0)
251 return err;
252
253 if (len) {
254 range_bits = 2 + len;
255 range_offset = 1 << range_bits;
256 } else {
257 range_bits = 3;
258 range_offset = 0;
259 }
260
261 if (len < max_len) {
262 err = mpp_av1_read_unsigned(gbc, range_bits,
263 "subexp_bits", &value,
264 0, MAX_UINT_BITS(range_bits));
265 if (err < 0)
266 return err;
267
268 } else {
269 err = mpp_av1_read_ns(gbc, "subexp_final_bits", range_max - range_offset,
270 &value);
271 if (err < 0)
272 return err;
273 }
274 value += range_offset;
275
276 *write_to = value;
277 return err;
278 }
279
280
mpp_av1_tile_log2(RK_S32 blksize,RK_S32 target)281 static RK_S32 mpp_av1_tile_log2(RK_S32 blksize, RK_S32 target)
282 {
283 RK_S32 k;
284 for (k = 0; (blksize << k) < target; k++);
285 return k;
286 }
287
mpp_av1_get_relative_dist(const AV1RawSequenceHeader * seq,RK_U32 a,RK_U32 b)288 static RK_S32 mpp_av1_get_relative_dist(const AV1RawSequenceHeader *seq,
289 RK_U32 a, RK_U32 b)
290 {
291 RK_U32 diff, m;
292 if (!seq->enable_order_hint)
293 return 0;
294 diff = a - b;
295 m = 1 << seq->order_hint_bits_minus_1;
296 diff = (diff & (m - 1)) - (diff & m);
297 return diff;
298 }
299
mpp_av1_get_payload_bytes_left(BitReadCtx_t * gbc)300 static size_t mpp_av1_get_payload_bytes_left(BitReadCtx_t *gbc)
301 {
302 size_t size = 0;
303 RK_U8 value = 0;
304 RK_S32 i = 0;
305
306 for (i = 0; mpp_get_bits_left(gbc) >= 8; i++) {
307 READ_BITS(gbc, 8, &value);
308 if (value)
309 size = i;
310 }
311 return size;
312
313 __bitread_error:
314 return MPP_NOK;
315
316 }
317
318 #define CHECK(call) do { \
319 err = (call); \
320 if (err < 0) \
321 return err; \
322 } while (0)
323
324
325 #define SUBSCRIPTS(subs, ...) (subs > 0 ? ((RK_S32[subs + 1]){ subs, __VA_ARGS__ }) : NULL)
326 #define fb(width, name) \
327 xf(width, name, current->name, 0, MAX_UINT_BITS(width), 0, )
328 #define fc(width, name, range_min, range_max) \
329 xf(width, name, current->name, range_min, range_max, 0, )
330 #define flag(name) fb(1, name)
331 #define su(width, name) \
332 xsu(width, name, current->name, 0, )
333
334 #define fbs(width, name, subs, ...) \
335 xf(width, name, current->name, 0, MAX_UINT_BITS(width), subs, __VA_ARGS__)
336 #define fcs(width, name, range_min, range_max, subs, ...) \
337 xf(width, name, current->name, range_min, range_max, subs, __VA_ARGS__)
338 #define flags(name, subs, ...) \
339 xf(1, name, current->name, 0, 1, subs, __VA_ARGS__)
340 #define sus(width, name, subs, ...) \
341 xsu(width, name, current->name, subs, __VA_ARGS__)
342
343 #define xf(width, name, var, range_min, range_max, subs, ...) do { \
344 RK_U32 value; \
345 CHECK(mpp_av1_read_unsigned(gb, width, #name, \
346 &value, range_min, range_max)); \
347 var = value; \
348 } while (0)
349
350 #define xsu(width, name, var, subs, ...) do { \
351 RK_S32 value; \
352 CHECK(mpp_av1_read_signed(gb, width, #name, \
353 &value, \
354 MIN_INT_BITS(width), \
355 MAX_INT_BITS(width))); \
356 var = value; \
357 } while (0)
358
359 #define uvlc(name, range_min, range_max) do { \
360 RK_U32 value; \
361 CHECK(mpp_av1_read_uvlc(gb, #name, \
362 &value, range_min, range_max)); \
363 current->name = value; \
364 } while (0)
365
366 #define ns(max_value, name) do { \
367 RK_U32 value; \
368 CHECK(mpp_av1_read_ns(gb, #name, max_value, \
369 &value)); \
370 current->name = value; \
371 } while (0)
372
373 #define increment(name, min, max) do { \
374 RK_U32 value; \
375 CHECK(mpp_av1_read_increment(gb, min, max, #name, &value)); \
376 current->name = value; \
377 } while (0)
378
379 #define subexp(name, max) do { \
380 RK_U32 value = 0; \
381 CHECK(mpp_av1_read_subexp(gb, max, \
382 &value)); \
383 current->name = value; \
384 } while (0)
385
386 #define delta_q(name) do { \
387 RK_U8 delta_coded; \
388 RK_S8 delta_q; \
389 xf(1, name.delta_coded, delta_coded, 0, 1, 0, ); \
390 if (delta_coded) \
391 xsu(1 + 6, name.delta_q, delta_q, 0, ); \
392 else \
393 delta_q = 0; \
394 current->name = delta_q; \
395 } while (0)
396
397 #define leb128(name) do { \
398 RK_U64 value; \
399 CHECK(mpp_av1_read_leb128(gb, &value)); \
400 current->name = value; \
401 } while (0)
402
403 #define infer(name, value) do { \
404 current->name = value; \
405 } while (0)
406
407 #define byte_alignment(gb) (mpp_get_bits_count(gb) % 8)
408
mpp_av1_read_obu_header(AV1Context * ctx,BitReadCtx_t * gb,AV1RawOBUHeader * current)409 static RK_S32 mpp_av1_read_obu_header(AV1Context *ctx, BitReadCtx_t *gb,
410 AV1RawOBUHeader *current)
411 {
412 RK_S32 err;
413
414 fc(1, obu_forbidden_bit, 0, 0);
415
416 fc(4, obu_type, 0, AV1_OBU_PADDING);
417 flag(obu_extension_flag);
418 flag(obu_has_size_field);
419
420 fc(1, obu_reserved_1bit, 0, 0);
421
422 if (current->obu_extension_flag) {
423 fb(3, temporal_id);
424 fb(2, spatial_id);
425 fc(3, extension_header_reserved_3bits, 0, 0);
426 } else {
427 infer(temporal_id, 0);
428 infer(spatial_id, 0);
429 }
430
431 ctx->temporal_id = current->temporal_id;
432 ctx->spatial_id = current->spatial_id;
433
434 return 0;
435 }
436
mpp_av1_trailing_bits(AV1Context * ctx,BitReadCtx_t * gb,RK_S32 nb_bits)437 static RK_S32 mpp_av1_trailing_bits(AV1Context *ctx, BitReadCtx_t *gb, RK_S32 nb_bits)
438 {
439 (void)ctx;
440 mpp_assert(nb_bits > 0);
441
442 // fixed(1, trailing_one_bit, 1);
443 mpp_skip_bits(gb, 1);
444
445 --nb_bits;
446
447 while (nb_bits > 0) {
448 // fixed(1, trailing_zero_bit, 0);
449 mpp_skip_bits(gb, 1);
450 --nb_bits;
451 }
452
453 return 0;
454 }
455
mpp_av1_byte_alignment(AV1Context * ctx,BitReadCtx_t * gb)456 static RK_S32 mpp_av1_byte_alignment(AV1Context *ctx, BitReadCtx_t *gb)
457 {
458
459 (void)ctx;
460
461 while (byte_alignment(gb) != 0)
462 mpp_skip_bits(gb, 1);
463 //fixed(1, zero_bit, 0);
464
465 return 0;
466 }
467
mpp_av1_color_config(AV1Context * ctx,BitReadCtx_t * gb,AV1RawColorConfig * current,RK_S32 seq_profile)468 static RK_S32 mpp_av1_color_config(AV1Context *ctx, BitReadCtx_t *gb,
469 AV1RawColorConfig *current, RK_S32 seq_profile)
470 {
471 RK_S32 err;
472
473 flag(high_bitdepth);
474
475 if (seq_profile == PROFILE_AV1_PROFESSIONAL &&
476 current->high_bitdepth) {
477 flag(twelve_bit);
478 ctx->bit_depth = current->twelve_bit ? 12 : 10;
479 } else {
480 ctx->bit_depth = current->high_bitdepth ? 10 : 8;
481 }
482
483 if (seq_profile == PROFILE_AV1_HIGH)
484 infer(mono_chrome, 0);
485 else
486 flag(mono_chrome);
487 ctx->num_planes = current->mono_chrome ? 1 : 3;
488
489 flag(color_description_present_flag);
490 if (current->color_description_present_flag) {
491 fb(8, color_primaries);
492 fb(8, transfer_characteristics);
493 fb(8, matrix_coefficients);
494 if (current->transfer_characteristics == MPP_FRAME_TRC_BT2020_10 ||
495 current->transfer_characteristics == MPP_FRAME_TRC_SMPTEST2084)
496 ctx->is_hdr = 1;
497 } else {
498 infer(color_primaries, MPP_FRAME_PRI_UNSPECIFIED);
499 infer(transfer_characteristics, MPP_FRAME_TRC_UNSPECIFIED);
500 infer(matrix_coefficients, MPP_FRAME_SPC_UNSPECIFIED);
501 }
502
503 if (current->mono_chrome) {
504 flag(color_range);
505
506 infer(subsampling_x, 1);
507 infer(subsampling_y, 1);
508 infer(chroma_sample_position, AV1_CSP_UNKNOWN);
509 infer(separate_uv_delta_q, 0);
510
511 } else if (current->color_primaries == MPP_FRAME_PRI_BT709 &&
512 current->transfer_characteristics == MPP_FRAME_TRC_IEC61966_2_1 &&
513 current->matrix_coefficients == MPP_FRAME_SPC_RGB) {
514 infer(color_range, 1);
515 infer(subsampling_x, 0);
516 infer(subsampling_y, 0);
517 flag(separate_uv_delta_q);
518
519 } else {
520 flag(color_range);
521
522 if (seq_profile == PROFILE_AV1_MAIN) {
523 infer(subsampling_x, 1);
524 infer(subsampling_y, 1);
525 } else if (seq_profile == PROFILE_AV1_HIGH) {
526 infer(subsampling_x, 0);
527 infer(subsampling_y, 0);
528 } else {
529 if (ctx->bit_depth == 12) {
530 fb(1, subsampling_x);
531 if (current->subsampling_x)
532 fb(1, subsampling_y);
533 else
534 infer(subsampling_y, 0);
535 } else {
536 infer(subsampling_x, 1);
537 infer(subsampling_y, 0);
538 }
539 }
540 if (current->subsampling_x && current->subsampling_y) {
541 fc(2, chroma_sample_position, AV1_CSP_UNKNOWN,
542 AV1_CSP_COLOCATED);
543 }
544
545 flag(separate_uv_delta_q);
546 }
547
548 return 0;
549 }
550
mpp_av1_timing_info(AV1Context * ctx,BitReadCtx_t * gb,AV1RawTimingInfo * current)551 static RK_S32 mpp_av1_timing_info(AV1Context *ctx, BitReadCtx_t *gb,
552 AV1RawTimingInfo *current)
553 {
554 (void)ctx;
555 RK_S32 err;
556
557 fc(32, num_units_in_display_tick, 1, MAX_UINT_BITS(32));
558 fc(32, time_scale, 1, MAX_UINT_BITS(32));
559
560 flag(equal_picture_interval);
561 if (current->equal_picture_interval)
562 uvlc(num_ticks_per_picture_minus_1, 0, MAX_UINT_BITS(32) - 1);
563
564 return 0;
565 }
566
mpp_av1_decoder_model_info(AV1Context * ctx,BitReadCtx_t * gb,AV1RawDecoderModelInfo * current)567 static RK_S32 mpp_av1_decoder_model_info(AV1Context *ctx, BitReadCtx_t *gb,
568 AV1RawDecoderModelInfo *current)
569 {
570 RK_S32 err;
571 (void)ctx;
572 fb(5, buffer_delay_length_minus_1);
573 fb(32, num_units_in_decoding_tick);
574 fb(5, buffer_removal_time_length_minus_1);
575 fb(5, frame_presentation_time_length_minus_1);
576
577 return 0;
578 }
579
mpp_av1_sequence_header_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawSequenceHeader * current)580 static RK_S32 mpp_av1_sequence_header_obu(AV1Context *ctx, BitReadCtx_t *gb,
581 AV1RawSequenceHeader *current)
582 {
583 RK_S32 i, err;
584
585 fc(3, seq_profile, PROFILE_AV1_MAIN,
586 PROFILE_AV1_PROFESSIONAL);
587 flag(still_picture);
588 flag(reduced_still_picture_header);
589
590 if (current->reduced_still_picture_header) {
591 infer(timing_info_present_flag, 0);
592 infer(decoder_model_info_present_flag, 0);
593 infer(initial_display_delay_present_flag, 0);
594 infer(operating_points_cnt_minus_1, 0);
595 infer(operating_point_idc[0], 0);
596
597 fb(5, seq_level_idx[0]);
598
599 infer(seq_tier[0], 0);
600 infer(decoder_model_present_for_this_op[0], 0);
601 infer(initial_display_delay_present_for_this_op[0], 0);
602
603 } else {
604 flag(timing_info_present_flag);
605 if (current->timing_info_present_flag) {
606 CHECK(mpp_av1_timing_info(ctx, gb, ¤t->timing_info));
607
608 flag(decoder_model_info_present_flag);
609 if (current->decoder_model_info_present_flag) {
610 CHECK(mpp_av1_decoder_model_info
611 (ctx, gb, ¤t->decoder_model_info));
612 }
613 } else {
614 infer(decoder_model_info_present_flag, 0);
615 }
616
617 flag(initial_display_delay_present_flag);
618
619 fb(5, operating_points_cnt_minus_1);
620 for (i = 0; i <= current->operating_points_cnt_minus_1; i++) {
621 fbs(12, operating_point_idc[i], 1, i);
622 fbs(5, seq_level_idx[i], 1, i);
623
624 if (current->seq_level_idx[i] > 7)
625 flags(seq_tier[i], 1, i);
626 else
627 infer(seq_tier[i], 0);
628
629 if (current->decoder_model_info_present_flag) {
630 flags(decoder_model_present_for_this_op[i], 1, i);
631 if (current->decoder_model_present_for_this_op[i]) {
632 RK_S32 n = current->decoder_model_info.buffer_delay_length_minus_1 + 1;
633 fbs(n, decoder_buffer_delay[i], 1, i);
634 fbs(n, encoder_buffer_delay[i], 1, i);
635 flags(low_delay_mode_flag[i], 1, i);
636 }
637 } else {
638 infer(decoder_model_present_for_this_op[i], 0);
639 }
640
641 if (current->initial_display_delay_present_flag) {
642 flags(initial_display_delay_present_for_this_op[i], 1, i);
643 if (current->initial_display_delay_present_for_this_op[i])
644 fbs(4, initial_display_delay_minus_1[i], 1, i);
645 }
646 }
647 }
648
649 fb(4, frame_width_bits_minus_1);
650 fb(4, frame_height_bits_minus_1);
651
652 fb(current->frame_width_bits_minus_1 + 1, max_frame_width_minus_1);
653 fb(current->frame_height_bits_minus_1 + 1, max_frame_height_minus_1);
654
655 if (current->reduced_still_picture_header)
656 infer(frame_id_numbers_present_flag, 0);
657 else
658 flag(frame_id_numbers_present_flag);
659 if (current->frame_id_numbers_present_flag) {
660 fb(4, delta_frame_id_length_minus_2);
661 fb(3, additional_frame_id_length_minus_1);
662 }
663
664 flag(use_128x128_superblock);
665 flag(enable_filter_intra);
666 flag(enable_intra_edge_filter);
667
668 if (current->reduced_still_picture_header) {
669 infer(enable_interintra_compound, 0);
670 infer(enable_masked_compound, 0);
671 infer(enable_warped_motion, 0);
672 infer(enable_dual_filter, 0);
673 infer(enable_order_hint, 0);
674 infer(enable_jnt_comp, 0);
675 infer(enable_ref_frame_mvs, 0);
676
677 infer(seq_force_screen_content_tools,
678 AV1_SELECT_SCREEN_CONTENT_TOOLS);
679 infer(seq_force_integer_mv,
680 AV1_SELECT_INTEGER_MV);
681 } else {
682 flag(enable_interintra_compound);
683 flag(enable_masked_compound);
684 flag(enable_warped_motion);
685 flag(enable_dual_filter);
686
687 flag(enable_order_hint);
688 if (current->enable_order_hint) {
689 flag(enable_jnt_comp);
690 flag(enable_ref_frame_mvs);
691 } else {
692 infer(enable_jnt_comp, 0);
693 infer(enable_ref_frame_mvs, 0);
694 }
695
696 flag(seq_choose_screen_content_tools);
697 if (current->seq_choose_screen_content_tools)
698 infer(seq_force_screen_content_tools,
699 AV1_SELECT_SCREEN_CONTENT_TOOLS);
700 else
701 fb(1, seq_force_screen_content_tools);
702 if (current->seq_force_screen_content_tools > 0) {
703 flag(seq_choose_integer_mv);
704 if (current->seq_choose_integer_mv)
705 infer(seq_force_integer_mv,
706 AV1_SELECT_INTEGER_MV);
707 else
708 fb(1, seq_force_integer_mv);
709 } else {
710 infer(seq_force_integer_mv, AV1_SELECT_INTEGER_MV);
711 }
712
713 if (current->enable_order_hint)
714 fb(3, order_hint_bits_minus_1);
715 }
716
717 flag(enable_superres);
718 flag(enable_cdef);
719 flag(enable_restoration);
720
721 CHECK(mpp_av1_color_config(ctx, gb, ¤t->color_config,
722 current->seq_profile));
723
724 flag(film_grain_params_present);
725
726 return 0;
727 }
728
mpp_av1_temporal_delimiter_obu(AV1Context * ctx,BitReadCtx_t * gb)729 static RK_S32 mpp_av1_temporal_delimiter_obu(AV1Context *ctx, BitReadCtx_t *gb)
730 {
731 (void)gb;
732 ctx->seen_frame_header = 0;
733
734 return 0;
735 }
736
737 /* spec 7.8 */
mpp_av1_set_frame_refs(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)738 static RK_S32 mpp_av1_set_frame_refs(AV1Context *ctx, BitReadCtx_t *gb,
739 AV1RawFrameHeader *current)
740 {
741 (void)gb;
742 const AV1RawSequenceHeader *seq = ctx->sequence_header;
743 static const RK_U8 ref_frame_list[AV1_NUM_REF_FRAMES - 2] = {
744 AV1_REF_FRAME_LAST2, AV1_REF_FRAME_LAST3, AV1_REF_FRAME_BWDREF,
745 AV1_REF_FRAME_ALTREF2, AV1_REF_FRAME_ALTREF
746 };
747 RK_S8 ref_frame_idx[AV1_REFS_PER_FRAME], used_frame[AV1_NUM_REF_FRAMES];
748 RK_S8 shifted_order_hints[AV1_NUM_REF_FRAMES];
749 RK_S32 cur_frame_hint, latest_order_hint, earliest_order_hint, ref;
750 RK_S32 i, j;
751
752 for (i = 0; i < AV1_REFS_PER_FRAME; i++)
753 ref_frame_idx[i] = -1;
754 ref_frame_idx[AV1_REF_FRAME_LAST - AV1_REF_FRAME_LAST] = current->last_frame_idx;
755 ref_frame_idx[AV1_REF_FRAME_GOLDEN - AV1_REF_FRAME_LAST] = current->golden_frame_idx;
756
757 /*
758 * An array usedFrame marking which reference frames
759 * have been used is prepared as follows:
760 */
761 for (i = 0; i < AV1_NUM_REF_FRAMES; i++)
762 used_frame[i] = 0;
763 used_frame[current->last_frame_idx] = 1;
764 used_frame[current->golden_frame_idx] = 1;
765
766 /*
767 * An array shiftedOrderHints (containing the expected output order shifted
768 * such that the current frame has hint equal to curFrameHint) is prepared as follows:
769 */
770 cur_frame_hint = 1 << (seq->order_hint_bits_minus_1);
771 for (i = 0; i < AV1_NUM_REF_FRAMES; i++)
772 shifted_order_hints[i] = cur_frame_hint +
773 mpp_av1_get_relative_dist(seq, ctx->ref_s[i].order_hint,
774 ctx->order_hint);
775
776 latest_order_hint = shifted_order_hints[current->last_frame_idx];
777 earliest_order_hint = shifted_order_hints[current->golden_frame_idx];
778
779 /* find_latest_backward */
780 ref = -1;
781 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
782 RK_S32 hint = shifted_order_hints[i];
783 if (!used_frame[i] && hint >= cur_frame_hint &&
784 (ref < 0 || hint >= latest_order_hint)) {
785 ref = i;
786 latest_order_hint = hint;
787 }
788 }
789 /*
790 * The ALTREF_FRAME reference is set to be a backward reference to the frame
791 * with highest output order as follows:
792 */
793 if (ref >= 0) {
794 ref_frame_idx[AV1_REF_FRAME_ALTREF - AV1_REF_FRAME_LAST] = ref;
795 used_frame[ref] = 1;
796 }
797
798 /* find_earliest_backward */
799 ref = -1;
800 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
801 RK_S32 hint = shifted_order_hints[i];
802 if (!used_frame[i] && hint >= cur_frame_hint &&
803 (ref < 0 || hint < earliest_order_hint)) {
804 ref = i;
805 earliest_order_hint = hint;
806 }
807 }
808 /*
809 * The BWDREF_FRAME reference is set to be a backward reference to
810 * the closest frame as follows:
811 */
812 if (ref >= 0) {
813 ref_frame_idx[AV1_REF_FRAME_BWDREF - AV1_REF_FRAME_LAST] = ref;
814 used_frame[ref] = 1;
815 }
816
817 ref = -1;
818 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
819 RK_S32 hint = shifted_order_hints[i];
820 if (!used_frame[i] && hint >= cur_frame_hint &&
821 (ref < 0 || hint < earliest_order_hint)) {
822 ref = i;
823 earliest_order_hint = hint;
824 }
825 }
826
827 /*
828 * The ALTREF2_FRAME reference is set to the next closest
829 * backward reference as follows:
830 */
831 if (ref >= 0) {
832 ref_frame_idx[AV1_REF_FRAME_ALTREF2 - AV1_REF_FRAME_LAST] = ref;
833 used_frame[ref] = 1;
834 }
835
836 /*
837 * The remaining references are set to be forward references
838 * in anti-chronological order as follows:
839 */
840 for (i = 0; i < AV1_REFS_PER_FRAME - 2; i++) {
841 RK_S32 ref_frame = ref_frame_list[i];
842 if (ref_frame_idx[ref_frame - AV1_REF_FRAME_LAST] < 0 ) {
843 /* find_latest_forward */
844 ref = -1;
845 for (j = 0; j < AV1_NUM_REF_FRAMES; j++) {
846 RK_S32 hint = shifted_order_hints[j];
847 if (!used_frame[j] && hint < cur_frame_hint &&
848 (ref < 0 || hint >= latest_order_hint)) {
849 ref = j;
850 latest_order_hint = hint;
851 }
852 }
853 if (ref >= 0) {
854 ref_frame_idx[ref_frame - AV1_REF_FRAME_LAST] = ref;
855 used_frame[ref] = 1;
856 }
857 }
858 }
859
860 /*
861 * Finally, any remaining references are set to the reference
862 * frame with smallest output order as follows:
863 */
864 ref = -1;
865 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
866 RK_S32 hint = shifted_order_hints[i];
867 if (ref < 0 || hint < earliest_order_hint) {
868 ref = i;
869 earliest_order_hint = hint;
870 }
871 }
872 for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
873 if (ref_frame_idx[i] < 0)
874 ref_frame_idx[i] = ref;
875 infer(ref_frame_idx[i], ref_frame_idx[i]);
876 }
877
878 return 0;
879 }
880
mpp_av1_superres_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)881 static RK_S32 mpp_av1_superres_params(AV1Context *ctx, BitReadCtx_t *gb,
882 AV1RawFrameHeader *current)
883 {
884 const AV1RawSequenceHeader *seq = ctx->sequence_header;
885 RK_S32 denom, err;
886
887 if (seq->enable_superres)
888 flag(use_superres);
889 else
890 infer(use_superres, 0);
891
892 if (current->use_superres) {
893 fb(3, coded_denom);
894 denom = current->coded_denom + AV1_SUPERRES_DENOM_MIN;
895 } else {
896 denom = AV1_SUPERRES_NUM;
897 }
898
899 ctx->upscaled_width = ctx->frame_width;
900 ctx->frame_width = (ctx->upscaled_width * AV1_SUPERRES_NUM +
901 denom / 2) / denom;
902 return 0;
903 }
904
mpp_av1_frame_size(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)905 static RK_S32 mpp_av1_frame_size(AV1Context *ctx, BitReadCtx_t *gb,
906 AV1RawFrameHeader *current)
907 {
908 const AV1RawSequenceHeader *seq = ctx->sequence_header;
909 RK_S32 err;
910
911 if (current->frame_size_override_flag) {
912 fb(seq->frame_width_bits_minus_1 + 1, frame_width_minus_1);
913 fb(seq->frame_height_bits_minus_1 + 1, frame_height_minus_1);
914 } else {
915 infer(frame_width_minus_1, seq->max_frame_width_minus_1);
916 infer(frame_height_minus_1, seq->max_frame_height_minus_1);
917 }
918
919 ctx->frame_width = current->frame_width_minus_1 + 1;
920 ctx->frame_height = current->frame_height_minus_1 + 1;
921
922 CHECK(mpp_av1_superres_params(ctx, gb, current));
923
924 return 0;
925 }
926
mpp_av1_render_size(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)927 static RK_S32 mpp_av1_render_size(AV1Context *ctx, BitReadCtx_t *gb,
928 AV1RawFrameHeader *current)
929 {
930 RK_S32 err;
931
932 flag(render_and_frame_size_different);
933
934 if (current->render_and_frame_size_different) {
935 fb(16, render_width_minus_1);
936 fb(16, render_height_minus_1);
937 } else {
938 infer(render_width_minus_1, current->frame_width_minus_1);
939 infer(render_height_minus_1, current->frame_height_minus_1);
940 }
941
942 ctx->render_width = current->render_width_minus_1 + 1;
943 ctx->render_height = current->render_height_minus_1 + 1;
944
945 return 0;
946 }
947
mpp_av1_frame_size_with_refs(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)948 static RK_S32 mpp_av1_frame_size_with_refs(AV1Context *ctx, BitReadCtx_t *gb,
949 AV1RawFrameHeader *current)
950 {
951 RK_S32 i, err;
952
953 for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
954 flags(found_ref[i], 1, i);
955 if (current->found_ref[i]) {
956 AV1ReferenceFrameState *ref =
957 &ctx->ref_s[current->ref_frame_idx[i]];
958
959 if (!ref->valid) {
960 mpp_err_f("Missing reference frame needed for frame size "
961 "(ref = %d, ref_frame_idx = %d).\n",
962 i, current->ref_frame_idx[i]);
963 return MPP_ERR_PROTOL;
964 }
965
966 infer(frame_width_minus_1, ref->upscaled_width - 1);
967 infer(frame_height_minus_1, ref->frame_height - 1);
968 infer(render_width_minus_1, ref->render_width - 1);
969 infer(render_height_minus_1, ref->render_height - 1);
970
971 ctx->upscaled_width = ref->upscaled_width;
972 ctx->frame_width = ctx->upscaled_width;
973 ctx->frame_height = ref->frame_height;
974 ctx->render_width = ref->render_width;
975 ctx->render_height = ref->render_height;
976 break;
977 }
978 }
979
980 if (i >= AV1_REFS_PER_FRAME) {
981 CHECK(mpp_av1_frame_size(ctx, gb, current));
982 CHECK(mpp_av1_render_size(ctx, gb, current));
983 } else {
984 CHECK(mpp_av1_superres_params(ctx, gb, current));
985 }
986
987 return 0;
988 }
989
mpp_av1_interpolation_filter(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)990 static RK_S32 mpp_av1_interpolation_filter(AV1Context *ctx, BitReadCtx_t *gb,
991 AV1RawFrameHeader *current)
992 {
993 RK_S32 err;
994 (void)ctx;
995 flag(is_filter_switchable);
996 if (current->is_filter_switchable)
997 infer(interpolation_filter,
998 AV1_INTERPOLATION_FILTER_SWITCHABLE);
999 else
1000 fb(2, interpolation_filter);
1001
1002 return 0;
1003 }
1004
mpp_av1_tile_info(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1005 static RK_S32 mpp_av1_tile_info(AV1Context *ctx, BitReadCtx_t *gb,
1006 AV1RawFrameHeader *current)
1007 {
1008 const AV1RawSequenceHeader *seq = ctx->sequence_header;
1009 RK_S32 mi_cols, mi_rows, sb_cols, sb_rows, sb_shift, sb_size;
1010 RK_S32 max_tile_width_sb, max_tile_height_sb, max_tile_area_sb;
1011 RK_S32 min_log2_tile_cols, max_log2_tile_cols, max_log2_tile_rows;
1012 RK_S32 min_log2_tiles, min_log2_tile_rows;
1013 RK_S32 i, err;
1014
1015 mi_cols = 2 * ((ctx->frame_width + 7) >> 3);
1016 mi_rows = 2 * ((ctx->frame_height + 7) >> 3);
1017
1018 sb_cols = seq->use_128x128_superblock ? ((mi_cols + 31) >> 5)
1019 : ((mi_cols + 15) >> 4);
1020 sb_rows = seq->use_128x128_superblock ? ((mi_rows + 31) >> 5)
1021 : ((mi_rows + 15) >> 4);
1022
1023 sb_shift = seq->use_128x128_superblock ? 5 : 4;
1024 sb_size = sb_shift + 2;
1025
1026 max_tile_width_sb = AV1_MAX_TILE_WIDTH >> sb_size;
1027 max_tile_area_sb = AV1_MAX_TILE_AREA >> (2 * sb_size);
1028
1029 min_log2_tile_cols = mpp_av1_tile_log2(max_tile_width_sb, sb_cols);
1030 max_log2_tile_cols = mpp_av1_tile_log2(1, MPP_MIN(sb_cols, AV1_MAX_TILE_COLS));
1031 max_log2_tile_rows = mpp_av1_tile_log2(1, MPP_MIN(sb_rows, AV1_MAX_TILE_ROWS));
1032 min_log2_tiles = MPP_MAX(min_log2_tile_cols,
1033 mpp_av1_tile_log2(max_tile_area_sb, sb_rows * sb_cols));
1034
1035 flag(uniform_tile_spacing_flag);
1036
1037 if (current->uniform_tile_spacing_flag) {
1038 RK_S32 tile_width_sb, tile_height_sb;
1039
1040 increment(tile_cols_log2, min_log2_tile_cols, max_log2_tile_cols);
1041
1042 tile_width_sb = (sb_cols + (1 << current->tile_cols_log2) - 1) >>
1043 current->tile_cols_log2;
1044 current->tile_cols = (sb_cols + tile_width_sb - 1) / tile_width_sb;
1045
1046 min_log2_tile_rows = MPP_MAX(min_log2_tiles - current->tile_cols_log2, 0);
1047
1048 increment(tile_rows_log2, min_log2_tile_rows, max_log2_tile_rows);
1049
1050 tile_height_sb = (sb_rows + (1 << current->tile_rows_log2) - 1) >>
1051 current->tile_rows_log2;
1052 current->tile_rows = (sb_rows + tile_height_sb - 1) / tile_height_sb;
1053
1054 for (i = 0; i < current->tile_cols - 1; i++)
1055 infer(width_in_sbs_minus_1[i], tile_width_sb - 1);
1056 infer(width_in_sbs_minus_1[i],
1057 sb_cols - (current->tile_cols - 1) * tile_width_sb - 1);
1058 for (i = 0; i < current->tile_rows - 1; i++)
1059 infer(height_in_sbs_minus_1[i], tile_height_sb - 1);
1060 infer(height_in_sbs_minus_1[i],
1061 sb_rows - (current->tile_rows - 1) * tile_height_sb - 1);
1062
1063 } else {
1064 RK_S32 widest_tile_sb, start_sb, size_sb, max_width, max_height;
1065
1066 widest_tile_sb = 0;
1067
1068 start_sb = 0;
1069 for (i = 0; start_sb < sb_cols && i < AV1_MAX_TILE_COLS; i++) {
1070 max_width = MPP_MIN(sb_cols - start_sb, max_tile_width_sb);
1071 ns(max_width, width_in_sbs_minus_1[i]);
1072 //ns(max_width, width_in_sbs_minus_1[i]);
1073 size_sb = current->width_in_sbs_minus_1[i] + 1;
1074 widest_tile_sb = MPP_MAX(size_sb, widest_tile_sb);
1075 start_sb += size_sb;
1076 }
1077 current->tile_cols_log2 = mpp_av1_tile_log2(1, i);
1078 current->tile_cols = i;
1079
1080 if (min_log2_tiles > 0)
1081 max_tile_area_sb = (sb_rows * sb_cols) >> (min_log2_tiles + 1);
1082 else
1083 max_tile_area_sb = sb_rows * sb_cols;
1084 max_tile_height_sb = MPP_MAX(max_tile_area_sb / widest_tile_sb, 1);
1085
1086 start_sb = 0;
1087 for (i = 0; start_sb < sb_rows && i < AV1_MAX_TILE_ROWS; i++) {
1088 max_height = MPP_MIN(sb_rows - start_sb, max_tile_height_sb);
1089 ns(max_height, height_in_sbs_minus_1[i]);
1090 size_sb = current->height_in_sbs_minus_1[i] + 1;
1091 start_sb += size_sb;
1092 }
1093 current->tile_rows_log2 = mpp_av1_tile_log2(1, i);
1094 current->tile_rows = i;
1095 }
1096
1097 if (current->tile_cols_log2 > 0 ||
1098 current->tile_rows_log2 > 0) {
1099 fb(current->tile_cols_log2 + current->tile_rows_log2,
1100 context_update_tile_id);
1101 fb(2, tile_size_bytes_minus1);
1102 } else {
1103 infer(context_update_tile_id, 0);
1104 current->tile_size_bytes_minus1 = 3;
1105 }
1106
1107 ctx->tile_cols = current->tile_cols;
1108 ctx->tile_rows = current->tile_rows;
1109
1110 return 0;
1111 }
1112
mpp_av1_quantization_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1113 static RK_S32 mpp_av1_quantization_params(AV1Context *ctx, BitReadCtx_t *gb,
1114 AV1RawFrameHeader *current)
1115 {
1116 const AV1RawSequenceHeader *seq = ctx->sequence_header;
1117 RK_S32 err;
1118
1119 fb(8, base_q_idx);
1120
1121 delta_q(delta_q_y_dc);
1122
1123 if (ctx->num_planes > 1) {
1124 if (seq->color_config.separate_uv_delta_q)
1125 flag(diff_uv_delta);
1126 else
1127 infer(diff_uv_delta, 0);
1128
1129 delta_q(delta_q_u_dc);
1130 delta_q(delta_q_u_ac);
1131
1132 if (current->diff_uv_delta) {
1133 delta_q(delta_q_v_dc);
1134 delta_q(delta_q_v_ac);
1135 } else {
1136 infer(delta_q_v_dc, current->delta_q_u_dc);
1137 infer(delta_q_v_ac, current->delta_q_u_ac);
1138 }
1139 } else {
1140 infer(delta_q_u_dc, 0);
1141 infer(delta_q_u_ac, 0);
1142 infer(delta_q_v_dc, 0);
1143 infer(delta_q_v_ac, 0);
1144 }
1145
1146 flag(using_qmatrix);
1147 if (current->using_qmatrix) {
1148 fb(4, qm_y);
1149 fb(4, qm_u);
1150 if (seq->color_config.separate_uv_delta_q)
1151 fb(4, qm_v);
1152 else
1153 infer(qm_v, current->qm_u);
1154 }
1155
1156 return 0;
1157 }
1158
mpp_av1_segmentation_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1159 static RK_S32 mpp_av1_segmentation_params(AV1Context *ctx, BitReadCtx_t *gb,
1160 AV1RawFrameHeader *current)
1161 {
1162 static const RK_U8 bits[AV1_SEG_LVL_MAX] = { 8, 6, 6, 6, 6, 3, 0, 0 };
1163 static const RK_U8 sign[AV1_SEG_LVL_MAX] = { 1, 1, 1, 1, 1, 0, 0, 0 };
1164 static const RK_U8 default_feature_enabled[AV1_SEG_LVL_MAX] = { 0 };
1165 static const RK_S16 default_feature_value[AV1_SEG_LVL_MAX] = { 0 };
1166 RK_S32 i, j, err;
1167
1168 flag(segmentation_enabled);
1169
1170 if (current->segmentation_enabled) {
1171 if (current->primary_ref_frame == AV1_PRIMARY_REF_NONE) {
1172 infer(segmentation_update_map, 1);
1173 infer(segmentation_temporal_update, 0);
1174 infer(segmentation_update_data, 1);
1175 } else {
1176 flag(segmentation_update_map);
1177 if (current->segmentation_update_map)
1178 flag(segmentation_temporal_update);
1179 else
1180 infer(segmentation_temporal_update, 0);
1181 flag(segmentation_update_data);
1182 }
1183
1184 for (i = 0; i < AV1_MAX_SEGMENTS; i++) {
1185 const RK_U8 *ref_feature_enabled;
1186 const RK_S16 *ref_feature_value;
1187
1188 if (current->primary_ref_frame == AV1_PRIMARY_REF_NONE) {
1189 ref_feature_enabled = default_feature_enabled;
1190 ref_feature_value = default_feature_value;
1191 } else {
1192 ref_feature_enabled =
1193 ctx->ref_s[current->ref_frame_idx[current->primary_ref_frame]].feature_enabled[i];
1194 ref_feature_value =
1195 ctx->ref_s[current->ref_frame_idx[current->primary_ref_frame]].feature_value[i];
1196 }
1197
1198 for (j = 0; j < AV1_SEG_LVL_MAX; j++) {
1199 if (current->segmentation_update_data) {
1200 flags(feature_enabled[i][j], 2, i, j);
1201 if (current->feature_enabled[i][j] && bits[j] > 0) {
1202 if (sign[j]) {
1203 RK_S32 sign_, data;
1204
1205 READ_ONEBIT(gb, &sign_);
1206 READ_BITS(gb, bits[j], &data);
1207 if (sign_) data -= (1 << bits[j]);
1208 current->feature_value[i][j] = data;
1209 } else
1210 fbs(bits[j], feature_value[i][j], 2, i, j);
1211 } else {
1212 infer(feature_value[i][j], 0);
1213 }
1214 } else {
1215 infer(feature_enabled[i][j], ref_feature_enabled[j]);
1216 infer(feature_value[i][j], ref_feature_value[j]);
1217 }
1218 }
1219 }
1220 } else {
1221 for (i = 0; i < AV1_MAX_SEGMENTS; i++) {
1222 for (j = 0; j < AV1_SEG_LVL_MAX; j++) {
1223 infer(feature_enabled[i][j], 0);
1224 infer(feature_value[i][j], 0);
1225 }
1226 }
1227 }
1228
1229 infer(segmentation_id_last_active, 0);
1230 infer(segmentation_id_preskip, 0);
1231 for (i = 0; i < AV1_MAX_SEGMENTS; i++) {
1232 for (j = 0; j < AV1_SEG_LVL_MAX; j++) {
1233 if (current->feature_enabled[i][j]) {
1234 infer(segmentation_id_last_active, i);
1235 if ( j > AV1_SEG_LVL_REF_FRAME)
1236 infer(segmentation_id_preskip, 1);
1237 }
1238 }
1239 }
1240
1241 return 0;
1242 __BITREAD_ERR:
1243 return MPP_ERR_STREAM;
1244 }
1245
mpp_av1_delta_q_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1246 static RK_S32 mpp_av1_delta_q_params(AV1Context *ctx, BitReadCtx_t *gb,
1247 AV1RawFrameHeader *current)
1248 {
1249 RK_S32 err;
1250 (void)ctx;
1251 if (current->base_q_idx > 0)
1252 flag(delta_q_present);
1253 else
1254 infer(delta_q_present, 0);
1255
1256 if (current->delta_q_present)
1257 fb(2, delta_q_res);
1258
1259 return 0;
1260 }
1261
mpp_av1_delta_lf_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1262 static RK_S32 mpp_av1_delta_lf_params(AV1Context *ctx, BitReadCtx_t *gb,
1263 AV1RawFrameHeader *current)
1264 {
1265 RK_S32 err;
1266 (void)ctx;
1267 if (current->delta_q_present) {
1268 if (!current->allow_intrabc)
1269 flag(delta_lf_present);
1270 else
1271 infer(delta_lf_present, 0);
1272 if (current->delta_lf_present) {
1273 fb(2, delta_lf_res);
1274 flag(delta_lf_multi);
1275 } else {
1276 infer(delta_lf_res, 0);
1277 infer(delta_lf_multi, 0);
1278 }
1279 } else {
1280 infer(delta_lf_present, 0);
1281 infer(delta_lf_res, 0);
1282 infer(delta_lf_multi, 0);
1283 }
1284
1285 return 0;
1286 }
1287
mpp_av1_loop_filter_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1288 static RK_S32 mpp_av1_loop_filter_params(AV1Context *ctx, BitReadCtx_t *gb,
1289 AV1RawFrameHeader *current)
1290 {
1291 static const RK_S8 default_loop_filter_ref_deltas[AV1_TOTAL_REFS_PER_FRAME] =
1292 { 1, 0, 0, 0, -1, 0, -1, -1 };
1293 static const RK_S8 default_loop_filter_mode_deltas[2] = { 0, 0 };
1294 RK_S32 i, err;
1295
1296 if (ctx->coded_lossless || current->allow_intrabc) {
1297 infer(loop_filter_level[0], 0);
1298 infer(loop_filter_level[1], 0);
1299 infer(loop_filter_ref_deltas[AV1_REF_FRAME_INTRA], 1);
1300 infer(loop_filter_ref_deltas[AV1_REF_FRAME_LAST], 0);
1301 infer(loop_filter_ref_deltas[AV1_REF_FRAME_LAST2], 0);
1302 infer(loop_filter_ref_deltas[AV1_REF_FRAME_LAST3], 0);
1303 infer(loop_filter_ref_deltas[AV1_REF_FRAME_BWDREF], 0);
1304 infer(loop_filter_ref_deltas[AV1_REF_FRAME_GOLDEN], -1);
1305 infer(loop_filter_ref_deltas[AV1_REF_FRAME_ALTREF], -1);
1306 infer(loop_filter_ref_deltas[AV1_REF_FRAME_ALTREF2], -1);
1307 for (i = 0; i < 2; i++)
1308 infer(loop_filter_mode_deltas[i], 0);
1309 return 0;
1310 }
1311
1312 fb(6, loop_filter_level[0]);
1313 fb(6, loop_filter_level[1]);
1314
1315 if (ctx->num_planes > 1) {
1316 if (current->loop_filter_level[0] ||
1317 current->loop_filter_level[1]) {
1318 fb(6, loop_filter_level[2]);
1319 fb(6, loop_filter_level[3]);
1320 }
1321 }
1322
1323 av1d_dbg(AV1D_DBG_HEADER, "orderhint %d loop_filter_level %d %d %d %d\n",
1324 current->order_hint,
1325 current->loop_filter_level[0], current->loop_filter_level[1],
1326 current->loop_filter_level[2], current->loop_filter_level[3]);
1327 fb(3, loop_filter_sharpness);
1328
1329 flag(loop_filter_delta_enabled);
1330 if (current->loop_filter_delta_enabled) {
1331 const RK_S8 *ref_loop_filter_ref_deltas, *ref_loop_filter_mode_deltas;
1332
1333 if (current->primary_ref_frame == AV1_PRIMARY_REF_NONE) {
1334 ref_loop_filter_ref_deltas = default_loop_filter_ref_deltas;
1335 ref_loop_filter_mode_deltas = default_loop_filter_mode_deltas;
1336 } else {
1337 ref_loop_filter_ref_deltas =
1338 ctx->ref_s[current->ref_frame_idx[current->primary_ref_frame]].loop_filter_ref_deltas;
1339 ref_loop_filter_mode_deltas =
1340 ctx->ref_s[current->ref_frame_idx[current->primary_ref_frame]].loop_filter_mode_deltas;
1341 }
1342
1343 flag(loop_filter_delta_update);
1344 for (i = 0; i < AV1_TOTAL_REFS_PER_FRAME; i++) {
1345 if (current->loop_filter_delta_update)
1346 flags(update_ref_delta[i], 1, i);
1347 else
1348 infer(update_ref_delta[i], 0);
1349 if (current->update_ref_delta[i])
1350 sus(1 + 6, loop_filter_ref_deltas[i], 1, i);
1351 else
1352 infer(loop_filter_ref_deltas[i], ref_loop_filter_ref_deltas[i]);
1353 }
1354 for (i = 0; i < 2; i++) {
1355 if (current->loop_filter_delta_update)
1356 flags(update_mode_delta[i], 1, i);
1357 else
1358 infer(update_mode_delta[i], 0);
1359 if (current->update_mode_delta[i])
1360 sus(1 + 6, loop_filter_mode_deltas[i], 1, i);
1361 else
1362 infer(loop_filter_mode_deltas[i], ref_loop_filter_mode_deltas[i]);
1363 }
1364 } else {
1365 for (i = 0; i < AV1_TOTAL_REFS_PER_FRAME; i++)
1366 infer(loop_filter_ref_deltas[i], default_loop_filter_ref_deltas[i]);
1367 for (i = 0; i < 2; i++)
1368 infer(loop_filter_mode_deltas[i], default_loop_filter_mode_deltas[i]);
1369 }
1370
1371 return 0;
1372 }
1373
mpp_av1_cdef_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1374 static RK_S32 mpp_av1_cdef_params(AV1Context *ctx, BitReadCtx_t *gb,
1375 AV1RawFrameHeader *current)
1376 {
1377 const AV1RawSequenceHeader *seq = ctx->sequence_header;
1378 RK_S32 i, err;
1379 if (ctx->coded_lossless || current->allow_intrabc ||
1380 !seq->enable_cdef) {
1381 infer(cdef_damping_minus_3, 0);
1382 infer(cdef_bits, 0);
1383 infer(cdef_y_pri_strength[0], 0);
1384 infer(cdef_y_sec_strength[0], 0);
1385 infer(cdef_uv_pri_strength[0], 0);
1386 infer(cdef_uv_sec_strength[0], 0);
1387
1388 return 0;
1389 }
1390
1391 fb(2, cdef_damping_minus_3);
1392 fb(2, cdef_bits);
1393
1394 for (i = 0; i < (1 << current->cdef_bits); i++) {
1395 fbs(4, cdef_y_pri_strength[i], 1, i);
1396 fbs(2, cdef_y_sec_strength[i], 1, i);
1397
1398 if (ctx->num_planes > 1) {
1399 fbs(4, cdef_uv_pri_strength[i], 1, i);
1400 fbs(2, cdef_uv_sec_strength[i], 1, i);
1401 }
1402 }
1403
1404 return 0;
1405 }
1406
mpp_av1_lr_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1407 static RK_S32 mpp_av1_lr_params(AV1Context *ctx, BitReadCtx_t *gb,
1408 AV1RawFrameHeader *current)
1409 {
1410 const AV1RawSequenceHeader *seq = ctx->sequence_header;
1411 RK_S32 uses_lr, uses_chroma_lr;
1412 RK_S32 i, err;
1413
1414 if (ctx->all_lossless || current->allow_intrabc ||
1415 !seq->enable_restoration) {
1416 return 0;
1417 }
1418
1419 uses_lr = uses_chroma_lr = 0;
1420 for (i = 0; i < ctx->num_planes; i++) {
1421 fbs(2, lr_type[i], 1, i);
1422
1423 if (current->lr_type[i] != AV1_RESTORE_NONE) {
1424 uses_lr = 1;
1425 if (i > 0)
1426 uses_chroma_lr = 1;
1427 }
1428 }
1429
1430 if (uses_lr) {
1431 if (seq->use_128x128_superblock)
1432 increment(lr_unit_shift, 1, 2);
1433 else
1434 increment(lr_unit_shift, 0, 2);
1435
1436 if (seq->color_config.subsampling_x &&
1437 seq->color_config.subsampling_y && uses_chroma_lr) {
1438 fb(1, lr_uv_shift);
1439 } else {
1440 infer(lr_uv_shift, 0);
1441 }
1442 }
1443
1444 return 0;
1445 }
1446
mpp_av1_read_tx_mode(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1447 static RK_S32 mpp_av1_read_tx_mode(AV1Context *ctx, BitReadCtx_t *gb,
1448 AV1RawFrameHeader *current)
1449 {
1450 RK_S32 err;
1451
1452 if (ctx->coded_lossless)
1453 infer(tx_mode, 0);
1454 else {
1455 flag(tx_mode);
1456 current->tx_mode = current->tx_mode ? TX_MODE_SELECT : TX_MODE_LARGEST;
1457 }
1458
1459 return 0;
1460 }
1461
mpp_av1_frame_reference_mode(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1462 static RK_S32 mpp_av1_frame_reference_mode(AV1Context *ctx, BitReadCtx_t *gb,
1463 AV1RawFrameHeader *current)
1464 {
1465 RK_S32 err;
1466 (void)ctx;
1467 if (current->frame_type == AV1_FRAME_INTRA_ONLY ||
1468 current->frame_type == AV1_FRAME_KEY)
1469 infer(reference_select, 0);
1470 else
1471 flag(reference_select);
1472
1473 return 0;
1474 }
1475
mpp_av1_skip_mode_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1476 static RK_S32 mpp_av1_skip_mode_params(AV1Context *ctx, BitReadCtx_t *gb,
1477 AV1RawFrameHeader *current)
1478 {
1479 const AV1RawSequenceHeader *seq = ctx->sequence_header;
1480 RK_S32 skip_mode_allowed;
1481 RK_S32 err;
1482
1483 ctx->skip_ref0 = 0;
1484 ctx->skip_ref1 = 0;
1485
1486 if (current->frame_type == AV1_FRAME_KEY ||
1487 current->frame_type == AV1_FRAME_INTRA_ONLY ||
1488 !current->reference_select || !seq->enable_order_hint) {
1489 skip_mode_allowed = 0;
1490 } else {
1491 RK_S32 forward_idx, backward_idx;
1492 RK_S32 forward_hint, backward_hint;
1493 RK_S32 ref_hint, dist, i;
1494
1495 forward_idx = -1;
1496 backward_idx = -1;
1497 forward_hint = -1;
1498 backward_hint = -1;
1499 for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
1500 ref_hint = ctx->ref_s[current->ref_frame_idx[i]].order_hint;
1501 dist = mpp_av1_get_relative_dist(seq, ref_hint,
1502 ctx->order_hint);
1503 if (dist < 0) {
1504 if (forward_idx < 0 ||
1505 mpp_av1_get_relative_dist(seq, ref_hint,
1506 forward_hint) > 0) {
1507 forward_idx = i;
1508 forward_hint = ref_hint;
1509 }
1510 } else if (dist > 0) {
1511 if (backward_idx < 0 ||
1512 mpp_av1_get_relative_dist(seq, ref_hint,
1513 backward_hint) < 0) {
1514 backward_idx = i;
1515 backward_hint = ref_hint;
1516 }
1517 }
1518 }
1519
1520 if (forward_idx < 0) {
1521 skip_mode_allowed = 0;
1522 } else if (backward_idx >= 0) {
1523 skip_mode_allowed = 1;
1524 ctx->skip_ref0 = MPP_MIN(forward_idx, backward_idx) + 1;
1525 ctx->skip_ref1 = MPP_MAX(forward_idx, backward_idx) + 1;
1526 // Frames for skip mode are forward_idx and backward_idx.
1527 } else {
1528 RK_S32 second_forward_idx;
1529 RK_S32 second_forward_hint;
1530 second_forward_idx = -1;
1531 for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
1532 ref_hint = ctx->ref_s[current->ref_frame_idx[i]].order_hint;
1533 if (mpp_av1_get_relative_dist(seq, ref_hint,
1534 forward_hint) < 0) {
1535 if (second_forward_idx < 0 ||
1536 mpp_av1_get_relative_dist(seq, ref_hint,
1537 second_forward_hint) > 0) {
1538 second_forward_idx = i;
1539 second_forward_hint = ref_hint;
1540 }
1541 }
1542 }
1543
1544 if (second_forward_idx < 0) {
1545 skip_mode_allowed = 0;
1546 } else {
1547 ctx->skip_ref0 = MPP_MIN(forward_idx, second_forward_idx) + 1;
1548 ctx->skip_ref1 = MPP_MAX(forward_idx, second_forward_idx) + 1;
1549 skip_mode_allowed = 1;
1550 // Frames for skip mode are forward_idx and second_forward_idx.
1551 }
1552 }
1553 }
1554
1555 if (skip_mode_allowed)
1556 flag(skip_mode_present);
1557 else
1558 infer(skip_mode_present, 0);
1559
1560 return 0;
1561 }
1562
mpp_av1_global_motion_param(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current,RK_S32 type,RK_S32 ref,RK_S32 idx)1563 static RK_S32 mpp_av1_global_motion_param(AV1Context *ctx, BitReadCtx_t *gb,
1564 AV1RawFrameHeader *current,
1565 RK_S32 type, RK_S32 ref, RK_S32 idx)
1566 {
1567 RK_U32 abs_bits, prec_bits, num_syms;
1568 RK_S32 err;
1569 (void)ctx;
1570 if (idx < 2) {
1571 if (type == AV1_WARP_MODEL_TRANSLATION) {
1572 abs_bits = AV1_GM_ABS_TRANS_ONLY_BITS - !current->allow_high_precision_mv;
1573 prec_bits = AV1_GM_TRANS_ONLY_PREC_BITS - !current->allow_high_precision_mv;
1574 } else {
1575 abs_bits = AV1_GM_ABS_TRANS_BITS;
1576 prec_bits = AV1_GM_TRANS_PREC_BITS;
1577 }
1578 } else {
1579 abs_bits = AV1_GM_ABS_ALPHA_BITS;
1580 prec_bits = AV1_GM_ALPHA_PREC_BITS;
1581 }
1582
1583 num_syms = 2 * (1 << abs_bits) + 1;
1584 subexp(gm_params[ref][idx], num_syms);// 2, ref, idx);
1585
1586 // Actual gm_params value is not reconstructed here.
1587 (void)prec_bits;
1588
1589 return 0;
1590 }
1591
1592 /*
1593 * Actual gm_params value is not reconstructed here.
1594 * Real gm_params update in av1d_parser.c->global_motion_params()
1595 */
mpp_av1_global_motion_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1596 static RK_S32 mpp_av1_global_motion_params(AV1Context *ctx, BitReadCtx_t *gb,
1597 AV1RawFrameHeader *current)
1598 {
1599 RK_S32 ref, type;
1600 RK_S32 err;
1601
1602 if (current->frame_type == AV1_FRAME_KEY ||
1603 current->frame_type == AV1_FRAME_INTRA_ONLY)
1604 return 0;
1605
1606 for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
1607 flags(is_global[ref], 1, ref);
1608 if (current->is_global[ref]) {
1609 flags(is_rot_zoom[ref], 1, ref);
1610 if (current->is_rot_zoom[ref]) {
1611 type = AV1_WARP_MODEL_ROTZOOM;
1612 } else {
1613 flags(is_translation[ref], 1, ref);
1614 type = current->is_translation[ref] ? AV1_WARP_MODEL_TRANSLATION
1615 : AV1_WARP_MODEL_AFFINE;
1616 }
1617 } else {
1618 type = AV1_WARP_MODEL_IDENTITY;
1619 }
1620
1621 if (type >= AV1_WARP_MODEL_ROTZOOM) {
1622 CHECK(mpp_av1_global_motion_param(ctx, gb, current, type, ref, 2));
1623 CHECK(mpp_av1_global_motion_param(ctx, gb, current, type, ref, 3));
1624 if (type == AV1_WARP_MODEL_AFFINE) {
1625 CHECK(mpp_av1_global_motion_param(ctx, gb, current, type, ref, 4));
1626 CHECK(mpp_av1_global_motion_param(ctx, gb, current, type, ref, 5));
1627 } else {
1628 current->gm_params[ref][4] = -current->gm_params[ref][3];
1629 current->gm_params[ref][5] = current->gm_params[ref][2];
1630 }
1631 }
1632 if (type >= AV1_WARP_MODEL_TRANSLATION) {
1633 CHECK(mpp_av1_global_motion_param(ctx, gb, current, type, ref, 0));
1634 CHECK(mpp_av1_global_motion_param(ctx, gb, current, type, ref, 1));
1635 }
1636 }
1637
1638 return 0;
1639 }
1640
mpp_av1_film_grain_params(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFilmGrainParams * current,AV1RawFrameHeader * frame_header)1641 static RK_S32 mpp_av1_film_grain_params(AV1Context *ctx, BitReadCtx_t *gb,
1642 AV1RawFilmGrainParams *current,
1643 AV1RawFrameHeader *frame_header)
1644 {
1645 const AV1RawSequenceHeader *seq = ctx->sequence_header;
1646 RK_S32 num_pos_luma, num_pos_chroma;
1647 RK_S32 i, err;
1648
1649 if (!seq->film_grain_params_present ||
1650 (!frame_header->show_frame && !frame_header->showable_frame))
1651 return 0;
1652
1653 flag(apply_grain);
1654
1655 if (!current->apply_grain)
1656 return 0;
1657
1658 fb(16, grain_seed);
1659
1660 if (frame_header->frame_type == AV1_FRAME_INTER)
1661 flag(update_grain);
1662 else
1663 infer(update_grain, 1);
1664
1665 if (!current->update_grain) {
1666 fb(3, film_grain_params_ref_idx);
1667 return 0;
1668 }
1669
1670 fc(4, num_y_points, 0, 14);
1671 for (i = 0; i < current->num_y_points; i++) {
1672 fcs(8, point_y_value[i],
1673 i ? current->point_y_value[i - 1] + 1 : 0,
1674 MAX_UINT_BITS(8) - (current->num_y_points - i - 1),
1675 1, i);
1676 fbs(8, point_y_scaling[i], 1, i);
1677 }
1678
1679 if (seq->color_config.mono_chrome)
1680 infer(chroma_scaling_from_luma, 0);
1681 else
1682 flag(chroma_scaling_from_luma);
1683
1684 if (seq->color_config.mono_chrome ||
1685 current->chroma_scaling_from_luma ||
1686 (seq->color_config.subsampling_x == 1 &&
1687 seq->color_config.subsampling_y == 1 &&
1688 current->num_y_points == 0)) {
1689 infer(num_cb_points, 0);
1690 infer(num_cr_points, 0);
1691 } else {
1692 fc(4, num_cb_points, 0, 10);
1693 for (i = 0; i < current->num_cb_points; i++) {
1694 fcs(8, point_cb_value[i],
1695 i ? current->point_cb_value[i - 1] + 1 : 0,
1696 MAX_UINT_BITS(8) - (current->num_cb_points - i - 1),
1697 1, i);
1698 fbs(8, point_cb_scaling[i], 1, i);
1699 }
1700 fc(4, num_cr_points, 0, 10);
1701 for (i = 0; i < current->num_cr_points; i++) {
1702 fcs(8, point_cr_value[i],
1703 i ? current->point_cr_value[i - 1] + 1 : 0,
1704 MAX_UINT_BITS(8) - (current->num_cr_points - i - 1),
1705 1, i);
1706 fbs(8, point_cr_scaling[i], 1, i);
1707 }
1708 }
1709
1710 fb(2, grain_scaling_minus_8);
1711 fb(2, ar_coeff_lag);
1712 num_pos_luma = 2 * current->ar_coeff_lag * (current->ar_coeff_lag + 1);
1713 if (current->num_y_points) {
1714 num_pos_chroma = num_pos_luma + 1;
1715 for (i = 0; i < num_pos_luma; i++)
1716 fbs(8, ar_coeffs_y_plus_128[i], 1, i);
1717 } else {
1718 num_pos_chroma = num_pos_luma;
1719 }
1720 if (current->chroma_scaling_from_luma || current->num_cb_points) {
1721 for (i = 0; i < num_pos_chroma; i++)
1722 fbs(8, ar_coeffs_cb_plus_128[i], 1, i);
1723 }
1724 if (current->chroma_scaling_from_luma || current->num_cr_points) {
1725 for (i = 0; i < num_pos_chroma; i++)
1726 fbs(8, ar_coeffs_cr_plus_128[i], 1, i);
1727 }
1728 fb(2, ar_coeff_shift_minus_6);
1729 fb(2, grain_scale_shift);
1730 if (current->num_cb_points) {
1731 fb(8, cb_mult);
1732 fb(8, cb_luma_mult);
1733 fb(9, cb_offset);
1734 }
1735 if (current->num_cr_points) {
1736 fb(8, cr_mult);
1737 fb(8, cr_luma_mult);
1738 fb(9, cr_offset);
1739 }
1740
1741 flag(overlap_flag);
1742 flag(clip_to_restricted_range);
1743
1744 return 0;
1745 }
1746
mpp_av1_uncompressed_header(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current)1747 static RK_S32 mpp_av1_uncompressed_header(AV1Context *ctx, BitReadCtx_t *gb,
1748 AV1RawFrameHeader *current)
1749 {
1750 const AV1RawSequenceHeader *seq;
1751 RK_S32 id_len, diff_len, all_frames, frame_is_intra, order_hint_bits;
1752 RK_S32 i, err;
1753
1754 if (!ctx->sequence_header) {
1755 mpp_err_f("No sequence header available: "
1756 "unable to decode frame header.\n");
1757 return MPP_ERR_UNKNOW;
1758 }
1759 seq = ctx->sequence_header;
1760
1761 id_len = seq->additional_frame_id_length_minus_1 +
1762 seq->delta_frame_id_length_minus_2 + 3;
1763 all_frames = (1 << AV1_NUM_REF_FRAMES) - 1;
1764
1765 if (seq->reduced_still_picture_header) {
1766 infer(show_existing_frame, 0);
1767 infer(frame_type, AV1_FRAME_KEY);
1768 infer(show_frame, 1);
1769 infer(showable_frame, 0);
1770 frame_is_intra = 1;
1771
1772 } else {
1773 flag(show_existing_frame);
1774
1775 if (current->show_existing_frame) {
1776 AV1ReferenceFrameState *ref;
1777
1778 fb(3, frame_to_show_map_idx);
1779 ref = &ctx->ref_s[current->frame_to_show_map_idx];
1780
1781 if (!ref->valid) {
1782 mpp_err_f("Missing reference frame needed for "
1783 "show_existing_frame (frame_to_show_map_idx = %d).\n",
1784 current->frame_to_show_map_idx);
1785 return MPP_ERR_UNKNOW;
1786 }
1787
1788 if (seq->decoder_model_info_present_flag &&
1789 !seq->timing_info.equal_picture_interval) {
1790 fb(seq->decoder_model_info.frame_presentation_time_length_minus_1 + 1,
1791 frame_presentation_time);
1792 }
1793
1794 if (seq->frame_id_numbers_present_flag)
1795 fb(id_len, display_frame_id);
1796
1797 infer(frame_type, ref->frame_type);
1798 if (current->frame_type == AV1_FRAME_KEY) {
1799 infer(refresh_frame_flags, all_frames);
1800
1801 // Section 7.21
1802 infer(current_frame_id, ref->frame_id);
1803 ctx->upscaled_width = ref->upscaled_width;
1804 ctx->frame_width = ref->frame_width;
1805 ctx->frame_height = ref->frame_height;
1806 ctx->render_width = ref->render_width;
1807 ctx->render_height = ref->render_height;
1808 ctx->bit_depth = ref->bit_depth;
1809 ctx->order_hint = ref->order_hint;
1810 } else
1811 infer(refresh_frame_flags, 0);
1812
1813 infer(frame_width_minus_1, ref->upscaled_width - 1);
1814 infer(frame_height_minus_1, ref->frame_height - 1);
1815 infer(render_width_minus_1, ref->render_width - 1);
1816 infer(render_height_minus_1, ref->render_height - 1);
1817
1818 return 0;
1819 }
1820
1821 fb(2, frame_type);
1822 frame_is_intra = (current->frame_type == AV1_FRAME_INTRA_ONLY ||
1823 current->frame_type == AV1_FRAME_KEY);
1824
1825 ctx->frame_is_intra = frame_is_intra;
1826 if (current->frame_type == AV1_FRAME_KEY) {
1827 RK_U32 refresh_frame_flags = (1 << NUM_REF_FRAMES) - 1;
1828
1829 Av1GetCDFs(ctx, current->frame_to_show_map_idx);
1830 Av1StoreCDFs(ctx, refresh_frame_flags);
1831 }
1832
1833 flag(show_frame);
1834 if (current->show_frame &&
1835 seq->decoder_model_info_present_flag &&
1836 !seq->timing_info.equal_picture_interval) {
1837 fb(seq->decoder_model_info.frame_presentation_time_length_minus_1 + 1,
1838 frame_presentation_time);
1839 }
1840 if (current->show_frame)
1841 infer(showable_frame, current->frame_type != AV1_FRAME_KEY);
1842 else
1843 flag(showable_frame);
1844
1845 if (current->frame_type == AV1_FRAME_SWITCH ||
1846 (current->frame_type == AV1_FRAME_KEY && current->show_frame))
1847 infer(error_resilient_mode, 1);
1848 else
1849 flag(error_resilient_mode);
1850 }
1851
1852 if (current->frame_type == AV1_FRAME_KEY && current->show_frame) {
1853 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
1854 ctx->ref_s[i].valid = 0;
1855 ctx->ref_s[i].order_hint = 0;
1856 }
1857 }
1858
1859 flag(disable_cdf_update);
1860
1861 if (seq->seq_force_screen_content_tools ==
1862 AV1_SELECT_SCREEN_CONTENT_TOOLS) {
1863 flag(allow_screen_content_tools);
1864 } else {
1865 infer(allow_screen_content_tools,
1866 seq->seq_force_screen_content_tools);
1867 }
1868 if (current->allow_screen_content_tools) {
1869 if (seq->seq_force_integer_mv == AV1_SELECT_INTEGER_MV)
1870 flag(force_integer_mv);
1871 else
1872 infer(force_integer_mv, seq->seq_force_integer_mv);
1873 } else {
1874 infer(force_integer_mv, 0);
1875 }
1876
1877 if (seq->frame_id_numbers_present_flag) {
1878 fb(id_len, current_frame_id);
1879
1880 diff_len = seq->delta_frame_id_length_minus_2 + 2;
1881 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
1882 if (current->current_frame_id > (RK_S32)(1 << diff_len)) {
1883 if (ctx->ref_s[i].frame_id > current->current_frame_id ||
1884 ctx->ref_s[i].frame_id < (current->current_frame_id -
1885 (RK_S32)(1 << diff_len)))
1886 ctx->ref_s[i].valid = 0;
1887 } else {
1888 if (ctx->ref_s[i].frame_id > current->current_frame_id &&
1889 ctx->ref_s[i].frame_id < ((RK_S32)(1 << id_len) +
1890 current->current_frame_id -
1891 (RK_S32)(1 << diff_len)))
1892 ctx->ref_s[i].valid = 0;
1893 }
1894 }
1895 } else {
1896 infer(current_frame_id, 0);
1897 }
1898
1899 if (current->frame_type == AV1_FRAME_SWITCH)
1900 infer(frame_size_override_flag, 1);
1901 else if (seq->reduced_still_picture_header)
1902 infer(frame_size_override_flag, 0);
1903 else
1904 flag(frame_size_override_flag);
1905
1906 order_hint_bits =
1907 seq->enable_order_hint ? seq->order_hint_bits_minus_1 + 1 : 0;
1908 if (order_hint_bits > 0)
1909 fb(order_hint_bits, order_hint);
1910 else
1911 infer(order_hint, 0);
1912 ctx->order_hint = current->order_hint;
1913
1914 if (frame_is_intra || current->error_resilient_mode)
1915 infer(primary_ref_frame, AV1_PRIMARY_REF_NONE);
1916 else
1917 fb(3, primary_ref_frame);
1918
1919 if (seq->decoder_model_info_present_flag) {
1920 flag(buffer_removal_time_present_flag);
1921 if (current->buffer_removal_time_present_flag) {
1922 for (i = 0; i <= seq->operating_points_cnt_minus_1; i++) {
1923 if (seq->decoder_model_present_for_this_op[i]) {
1924 RK_S32 op_pt_idc = seq->operating_point_idc[i];
1925 RK_S32 in_temporal_layer = (op_pt_idc >> ctx->temporal_id ) & 1;
1926 RK_S32 in_spatial_layer = (op_pt_idc >> (ctx->spatial_id + 8)) & 1;
1927 if (seq->operating_point_idc[i] == 0 ||
1928 (in_temporal_layer && in_spatial_layer)) {
1929 fbs(seq->decoder_model_info.buffer_removal_time_length_minus_1 + 1,
1930 buffer_removal_time[i], 1, i);
1931 }
1932 }
1933 }
1934 }
1935 }
1936
1937 if (current->frame_type == AV1_FRAME_SWITCH ||
1938 (current->frame_type == AV1_FRAME_KEY && current->show_frame))
1939 infer(refresh_frame_flags, all_frames);
1940 else
1941 fb(8, refresh_frame_flags);
1942
1943 ctx->refresh_frame_flags = current->refresh_frame_flags;
1944 if (!frame_is_intra || current->refresh_frame_flags != all_frames) {
1945 if (seq->enable_order_hint) {
1946 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) {
1947 if (current->error_resilient_mode)
1948 fbs(order_hint_bits, ref_order_hint[i], 1, i);
1949 else
1950 infer(ref_order_hint[i], ctx->ref_s[i].order_hint);
1951 if (current->ref_order_hint[i] != ctx->ref_s[i].order_hint)
1952 ctx->ref_s[i].valid = 0;
1953 }
1954 }
1955 }
1956
1957 current->ref_frame_valued = 1;
1958 if (current->frame_type == AV1_FRAME_KEY ||
1959 current->frame_type == AV1_FRAME_INTRA_ONLY) {
1960 CHECK(mpp_av1_frame_size(ctx, gb, current));
1961 CHECK(mpp_av1_render_size(ctx, gb, current));
1962
1963 if (current->allow_screen_content_tools &&
1964 ctx->upscaled_width == ctx->frame_width)
1965 flag(allow_intrabc);
1966 else
1967 infer(allow_intrabc, 0);
1968
1969 current->ref_frame_valued = 0;
1970 } else {
1971 if (!seq->enable_order_hint) {
1972 infer(frame_refs_short_signaling, 0);
1973 } else {
1974 flag(frame_refs_short_signaling);
1975 if (current->frame_refs_short_signaling) {
1976 fb(3, last_frame_idx);
1977 fb(3, golden_frame_idx);
1978 CHECK(mpp_av1_set_frame_refs(ctx, gb, current));
1979 }
1980 }
1981
1982 for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
1983 if (!current->frame_refs_short_signaling)
1984 fbs(3, ref_frame_idx[i], 1, i);
1985 if (seq->frame_id_numbers_present_flag) {
1986 fbs(seq->delta_frame_id_length_minus_2 + 2,
1987 delta_frame_id_minus1[i], 1, i);
1988 }
1989 }
1990
1991 if (current->frame_size_override_flag &&
1992 !current->error_resilient_mode) {
1993 CHECK(mpp_av1_frame_size_with_refs(ctx, gb, current));
1994 } else {
1995 CHECK(mpp_av1_frame_size(ctx, gb, current));
1996 CHECK(mpp_av1_render_size(ctx, gb, current));
1997 }
1998
1999 if (current->force_integer_mv)
2000 infer(allow_high_precision_mv, 0);
2001 else
2002 flag(allow_high_precision_mv);
2003
2004 CHECK(mpp_av1_interpolation_filter(ctx, gb, current));
2005
2006 flag(is_motion_mode_switchable);
2007
2008 if (current->error_resilient_mode ||
2009 !seq->enable_ref_frame_mvs)
2010 infer(use_ref_frame_mvs, 0);
2011 else
2012 flag(use_ref_frame_mvs);
2013
2014 infer(allow_intrabc, 0);
2015 }
2016
2017 if (!frame_is_intra) {
2018 // Derive reference frame sign biases.
2019 }
2020
2021 if (seq->reduced_still_picture_header || current->disable_cdf_update)
2022 infer(disable_frame_end_update_cdf, 1);
2023 else
2024 flag(disable_frame_end_update_cdf);
2025
2026 ctx->disable_frame_end_update_cdf = current->disable_frame_end_update_cdf;
2027
2028 if (current->use_ref_frame_mvs) {
2029 // Perform motion field estimation process.
2030 }
2031 av1d_dbg(AV1D_DBG_HEADER, "ptile_info in %d", mpp_get_bits_count(gb));
2032 CHECK(mpp_av1_tile_info(ctx, gb, current));
2033 av1d_dbg(AV1D_DBG_HEADER, "ptile_info out %d", mpp_get_bits_count(gb));
2034
2035 CHECK(mpp_av1_quantization_params(ctx, gb, current));
2036 av1d_dbg(AV1D_DBG_HEADER, "quantization out %d", mpp_get_bits_count(gb));
2037
2038 CHECK(mpp_av1_segmentation_params(ctx, gb, current));
2039 av1d_dbg(AV1D_DBG_HEADER, "segmentation out %d", mpp_get_bits_count(gb));
2040
2041 CHECK(mpp_av1_delta_q_params(ctx, gb, current));
2042 av1d_dbg(AV1D_DBG_HEADER, "delta_q out %d", mpp_get_bits_count(gb));
2043
2044 CHECK(mpp_av1_delta_lf_params(ctx, gb, current));
2045 av1d_dbg(AV1D_DBG_HEADER, "lf out %d", mpp_get_bits_count(gb));
2046
2047 // Init coeff CDFs / load previous segments.
2048 if (current->error_resilient_mode || frame_is_intra || current->primary_ref_frame == AV1_PRIMARY_REF_NONE) {
2049 // Init non-coeff CDFs.
2050 // Setup past independence.
2051 ctx->cdfs = &ctx->default_cdfs;
2052 ctx->cdfs_ndvc = &ctx->default_cdfs_ndvc;
2053 Av1DefaultCoeffProbs(current->base_q_idx, ctx->cdfs);
2054 } else {
2055 // Load CDF tables from previous frame.
2056 // Load params from previous frame.
2057 RK_U32 idx = current->ref_frame_idx[current->primary_ref_frame];
2058
2059 Av1GetCDFs(ctx, idx);
2060 }
2061 av1d_dbg(AV1D_DBG_HEADER, "show_existing_frame_index %d primary_ref_frame %d %d (%d) refresh_frame_flags %d base_q_idx %d\n",
2062 current->frame_to_show_map_idx,
2063 current->ref_frame_idx[current->primary_ref_frame],
2064 ctx->ref[current->ref_frame_idx[current->primary_ref_frame]].slot_index,
2065 current->primary_ref_frame,
2066 current->refresh_frame_flags,
2067 current->base_q_idx);
2068 Av1StoreCDFs(ctx, current->refresh_frame_flags);
2069
2070 ctx->coded_lossless = 1;
2071 for (i = 0; i < AV1_MAX_SEGMENTS; i++) {
2072 RK_S32 qindex;
2073 if (current->feature_enabled[i][AV1_SEG_LVL_ALT_Q]) {
2074 qindex = (current->base_q_idx +
2075 current->feature_value[i][AV1_SEG_LVL_ALT_Q]);
2076 } else {
2077 qindex = current->base_q_idx;
2078 }
2079 qindex = mpp_clip_uintp2(qindex, 8);
2080
2081 if (qindex || current->delta_q_y_dc ||
2082 current->delta_q_u_ac || current->delta_q_u_dc ||
2083 current->delta_q_v_ac || current->delta_q_v_dc) {
2084 ctx->coded_lossless = 0;
2085 }
2086 }
2087 ctx->all_lossless = ctx->coded_lossless &&
2088 ctx->frame_width == ctx->upscaled_width;
2089 av1d_dbg(AV1D_DBG_HEADER, "filter in %d", mpp_get_bits_count(gb));
2090
2091 CHECK(mpp_av1_loop_filter_params(ctx, gb, current));
2092 av1d_dbg(AV1D_DBG_HEADER, "cdef in %d", mpp_get_bits_count(gb));
2093
2094 CHECK(mpp_av1_cdef_params(ctx, gb, current));
2095 av1d_dbg(AV1D_DBG_HEADER, "lr in %d", mpp_get_bits_count(gb));
2096
2097 CHECK(mpp_av1_lr_params(ctx, gb, current));
2098 av1d_dbg(AV1D_DBG_HEADER, "read_tx in %d", mpp_get_bits_count(gb));
2099
2100 CHECK(mpp_av1_read_tx_mode(ctx, gb, current));
2101 av1d_dbg(AV1D_DBG_HEADER, "reference in%d", mpp_get_bits_count(gb));
2102
2103 CHECK(mpp_av1_frame_reference_mode(ctx, gb, current));
2104 av1d_dbg(AV1D_DBG_HEADER, "kip_mode in %d", mpp_get_bits_count(gb));
2105
2106 CHECK(mpp_av1_skip_mode_params(ctx, gb, current));
2107
2108 if (frame_is_intra || current->error_resilient_mode ||
2109 !seq->enable_warped_motion)
2110 infer(allow_warped_motion, 0);
2111 else
2112 flag(allow_warped_motion);
2113
2114 flag(reduced_tx_set);
2115 av1d_dbg(AV1D_DBG_HEADER, "motion in%d", mpp_get_bits_count(gb));
2116
2117 CHECK(mpp_av1_global_motion_params(ctx, gb, current));
2118 av1d_dbg(AV1D_DBG_HEADER, "grain in %d", mpp_get_bits_count(gb));
2119 CHECK(mpp_av1_film_grain_params(ctx, gb, ¤t->film_grain, current));
2120 av1d_dbg(AV1D_DBG_HEADER, "film_grain out %d", mpp_get_bits_count(gb));
2121
2122 av1d_dbg(AV1D_DBG_REF, "Frame %d: size %dx%d "
2123 "upscaled %d render %dx%d subsample %dx%d "
2124 "bitdepth %d tiles %dx%d.\n", ctx->order_hint,
2125 ctx->frame_width, ctx->frame_height, ctx->upscaled_width,
2126 ctx->render_width, ctx->render_height,
2127 seq->color_config.subsampling_x + 1,
2128 seq->color_config.subsampling_y + 1, ctx->bit_depth,
2129 ctx->tile_rows, ctx->tile_cols);
2130
2131 return 0;
2132 }
2133
mpp_av1_frame_header_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrameHeader * current,RK_S32 redundant,void * rw_buffer_ref)2134 static RK_S32 mpp_av1_frame_header_obu(AV1Context *ctx, BitReadCtx_t *gb,
2135 AV1RawFrameHeader *current, RK_S32 redundant,
2136 void *rw_buffer_ref)
2137 {
2138 RK_S32 start_pos, fh_bits, fh_bytes, err;
2139 RK_U8 *fh_start;
2140 (void)rw_buffer_ref;
2141 if (ctx->seen_frame_header) {
2142 if (!redundant) {
2143 mpp_err_f("Invalid repeated "
2144 "frame header OBU.\n");
2145 return MPP_ERR_UNKNOW;
2146 } else {
2147 BitReadCtx_t fh;
2148 size_t i, b;
2149 RK_U32 val;
2150
2151 // mpp_assert(ctx->frame_header_ref && ctx->frame_header);
2152
2153 mpp_set_bitread_ctx(&fh, ctx->frame_header,
2154 ctx->frame_header_size);
2155
2156 for (i = 0; i < ctx->frame_header_size; i += 8) {
2157 b = MPP_MIN(ctx->frame_header_size - i, 8);
2158 mpp_read_bits(&fh, b, (RK_S32*)&val);
2159 xf(b, frame_header_copy[i],
2160 val, val, val, 1, i / 8);
2161 }
2162 }
2163 } else {
2164
2165 start_pos = mpp_get_bits_count(gb);
2166
2167 CHECK(mpp_av1_uncompressed_header(ctx, gb, current));
2168
2169 ctx->tile_num = 0;
2170
2171 if (current->show_existing_frame) {
2172 ctx->seen_frame_header = 0;
2173 } else {
2174 ctx->seen_frame_header = 1;
2175
2176 fh_bits = mpp_get_bits_count(gb) - start_pos;
2177 fh_start = (RK_U8*)gb->buf + start_pos / 8;
2178
2179 fh_bytes = (fh_bits + 7) / 8;
2180 ctx->frame_header_size = fh_bits;
2181 MPP_FREE(ctx->frame_header);
2182 ctx->frame_header =
2183 mpp_malloc(RK_U8, fh_bytes + BUFFER_PADDING_SIZE);
2184 if (!ctx->frame_header) {
2185 mpp_err_f("frame header malloc failed\n");
2186 return MPP_ERR_NOMEM;
2187 }
2188 memcpy(ctx->frame_header, fh_start, fh_bytes);
2189 }
2190 }
2191
2192 return 0;
2193 }
2194
mpp_av1_tile_group_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawTileGroup * current)2195 static RK_S32 mpp_av1_tile_group_obu(AV1Context *ctx, BitReadCtx_t *gb,
2196 AV1RawTileGroup *current)
2197 {
2198 RK_S32 num_tiles, tile_bits;
2199 RK_S32 err;
2200
2201 num_tiles = ctx->tile_cols * ctx->tile_rows;
2202 if (num_tiles > 1)
2203 flag(tile_start_and_end_present_flag);
2204 else
2205 infer(tile_start_and_end_present_flag, 0);
2206
2207 if (num_tiles == 1 || !current->tile_start_and_end_present_flag) {
2208 infer(tg_start, 0);
2209 infer(tg_end, num_tiles - 1);
2210 } else {
2211 tile_bits = mpp_av1_tile_log2(1, ctx->tile_cols) +
2212 mpp_av1_tile_log2(1, ctx->tile_rows);
2213 fc(tile_bits, tg_start, ctx->tile_num, num_tiles - 1);
2214 fc(tile_bits, tg_end, current->tg_start, num_tiles - 1);
2215 }
2216
2217 ctx->tile_num = current->tg_end + 1;
2218
2219 CHECK(mpp_av1_byte_alignment(ctx, gb));
2220
2221 // Reset header for next frame.
2222 if (current->tg_end == num_tiles - 1)
2223 ctx->seen_frame_header = 0;
2224 // Tile data follows.
2225
2226 return 0;
2227 }
2228
mpp_av1_frame_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawFrame * current,void * rw_buffer_ref)2229 static RK_S32 mpp_av1_frame_obu(AV1Context *ctx, BitReadCtx_t *gb,
2230 AV1RawFrame *current,
2231 void *rw_buffer_ref)
2232 {
2233 RK_S32 err;
2234 RK_U32 start_pos = mpp_get_bits_count(gb);
2235
2236 CHECK(mpp_av1_frame_header_obu(ctx, gb, ¤t->header,
2237 0, rw_buffer_ref));
2238
2239 CHECK(mpp_av1_byte_alignment(ctx, gb));
2240
2241 CHECK(mpp_av1_tile_group_obu(ctx, gb, ¤t->tile_group));
2242 ctx->frame_tag_size += (mpp_get_bits_count(gb) - start_pos + 7) >> 3;
2243
2244 return 0;
2245 }
2246
mpp_av1_tile_list_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawTileList * current)2247 static RK_S32 mpp_av1_tile_list_obu(AV1Context *ctx, BitReadCtx_t *gb,
2248 AV1RawTileList *current)
2249 {
2250 RK_S32 err;
2251 (void)ctx;
2252 fb(8, output_frame_width_in_tiles_minus_1);
2253 fb(8, output_frame_height_in_tiles_minus_1);
2254
2255 fb(16, tile_count_minus_1);
2256
2257 // Tile data follows.
2258
2259 return 0;
2260 }
2261
mpp_av1_metadata_hdr_cll(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadataHDRCLL * current)2262 static RK_S32 mpp_av1_metadata_hdr_cll(AV1Context *ctx, BitReadCtx_t *gb,
2263 AV1RawMetadataHDRCLL *current)
2264 {
2265 RK_S32 err;
2266 (void)ctx;
2267 fb(16, max_cll);
2268 fb(16, max_fall);
2269
2270 ctx->content_light.MaxCLL = current->max_cll;
2271 ctx->content_light.MaxFALL = current->max_fall;
2272
2273 return 0;
2274 }
2275
mpp_av1_metadata_hdr_mdcv(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadataHDRMDCV * current)2276 static RK_S32 mpp_av1_metadata_hdr_mdcv(AV1Context *ctx, BitReadCtx_t *gb,
2277 AV1RawMetadataHDRMDCV *current)
2278 {
2279 RK_S32 err, i;
2280 (void)ctx;
2281 for (i = 0; i < 3; i++) {
2282 fbs(16, primary_chromaticity_x[i], 1, i);
2283 fbs(16, primary_chromaticity_y[i], 1, i);
2284 }
2285
2286 fb(16, white_point_chromaticity_x);
2287 fb(16, white_point_chromaticity_y);
2288
2289 fc(32, luminance_max, 1, MAX_UINT_BITS(32));
2290 // luminance_min must be lower than luminance_max. Convert luminance_max from
2291 // 24.8 fixed point to 18.14 fixed point in order to compare them.
2292 fc(32, luminance_min, 0, MPP_MIN(((RK_U64)current->luminance_max << 6) - 1,
2293 MAX_UINT_BITS(32)));
2294
2295 for (i = 0; i < 3; i++) {
2296 ctx->mastering_display.display_primaries[i][0] = current->primary_chromaticity_x[i];
2297 ctx->mastering_display.display_primaries[i][1] = current->primary_chromaticity_y[i];
2298 }
2299 ctx->mastering_display.white_point[0] = current->white_point_chromaticity_x;
2300 ctx->mastering_display.white_point[1] = current->white_point_chromaticity_y;
2301 ctx->mastering_display.max_luminance = current->luminance_max;
2302 ctx->mastering_display.min_luminance = current->luminance_min;
2303
2304 return 0;
2305 }
2306
mpp_av1_scalability_structure(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadataScalability * current)2307 static RK_S32 mpp_av1_scalability_structure(AV1Context *ctx, BitReadCtx_t *gb,
2308 AV1RawMetadataScalability *current)
2309 {
2310 const AV1RawSequenceHeader *seq;
2311 RK_S32 err, i, j;
2312
2313 if (!ctx->sequence_header) {
2314 mpp_err_f("No sequence header available: "
2315 "unable to parse scalability metadata.\n");
2316 return MPP_ERR_UNKNOW;
2317 }
2318 seq = ctx->sequence_header;
2319
2320 fb(2, spatial_layers_cnt_minus_1);
2321 flag(spatial_layer_dimensions_present_flag);
2322 flag(spatial_layer_description_present_flag);
2323 flag(temporal_group_description_present_flag);
2324 fc(3, scalability_structure_reserved_3bits, 0, 0);
2325 if (current->spatial_layer_dimensions_present_flag) {
2326 for (i = 0; i <= current->spatial_layers_cnt_minus_1; i++) {
2327 fcs(16, spatial_layer_max_width[i],
2328 0, seq->max_frame_width_minus_1 + 1, 1, i);
2329 fcs(16, spatial_layer_max_height[i],
2330 0, seq->max_frame_height_minus_1 + 1, 1, i);
2331 }
2332 }
2333 if (current->spatial_layer_description_present_flag) {
2334 for (i = 0; i <= current->spatial_layers_cnt_minus_1; i++)
2335 fbs(8, spatial_layer_ref_id[i], 1, i);
2336 }
2337 if (current->temporal_group_description_present_flag) {
2338 fb(8, temporal_group_size);
2339 for (i = 0; i < current->temporal_group_size; i++) {
2340 fbs(3, temporal_group_temporal_id[i], 1, i);
2341 flags(temporal_group_temporal_switching_up_point_flag[i], 1, i);
2342 flags(temporal_group_spatial_switching_up_point_flag[i], 1, i);
2343 fbs(3, temporal_group_ref_cnt[i], 1, i);
2344 for (j = 0; j < current->temporal_group_ref_cnt[i]; j++) {
2345 fbs(8, temporal_group_ref_pic_diff[i][j], 2, i, j);
2346 }
2347 }
2348 }
2349
2350 return 0;
2351 }
2352
mpp_av1_metadata_scalability(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadataScalability * current)2353 static RK_S32 mpp_av1_metadata_scalability(AV1Context *ctx, BitReadCtx_t *gb,
2354 AV1RawMetadataScalability *current)
2355 {
2356 RK_S32 err;
2357
2358 fb(8, scalability_mode_idc);
2359
2360 if (current->scalability_mode_idc == AV1_SCALABILITY_SS)
2361 CHECK(mpp_av1_scalability_structure(ctx, gb, current));
2362
2363 return 0;
2364 }
2365
mpp_av1_get_dlby_rpu(AV1Context * ctx,BitReadCtx_t * gb)2366 static RK_S32 mpp_av1_get_dlby_rpu(AV1Context *ctx, BitReadCtx_t *gb)
2367 {
2368 MppFrameHdrDynamicMeta *hdr_dynamic_meta = ctx->hdr_dynamic_meta;
2369 RK_U32 emdf_payload_size = 0;
2370
2371 /* skip emdf_container{} */
2372 SKIP_BITS(gb, 3);
2373 SKIP_BITS(gb, 2);
2374 SKIP_BITS(gb, 5);
2375 SKIP_BITS(gb, 5);
2376 SKIP_BITS(gb, 1);
2377 SKIP_BITS(gb, 5);
2378 SKIP_BITS(gb, 1);
2379 /* skip emdf_payload_config{} */
2380 SKIP_BITS(gb, 5);
2381
2382 /* get payload size */
2383 #define VARIABLE_BITS8(gb, value) \
2384 for (;;) { \
2385 RK_U32 tmp, flag; \
2386 \
2387 READ_BITS(gb, 8, &tmp); \
2388 value += tmp; \
2389 READ_ONEBIT(gb, &flag); \
2390 if (!flag) break; \
2391 value <<= 8; \
2392 value += (1 << 8); \
2393 }
2394
2395 VARIABLE_BITS8(gb, emdf_payload_size);
2396 if (!hdr_dynamic_meta) {
2397 hdr_dynamic_meta = mpp_calloc_size(MppFrameHdrDynamicMeta,
2398 sizeof(MppFrameHdrDynamicMeta) + SZ_1K);
2399 if (!hdr_dynamic_meta) {
2400 mpp_err_f("malloc hdr dynamic data failed!\n");
2401 return MPP_ERR_NOMEM;
2402 }
2403 }
2404
2405 RK_U32 i;
2406 MppWriteCtx bit_ctx;
2407
2408 mpp_writer_init(&bit_ctx, hdr_dynamic_meta->data, SZ_1K);
2409
2410 mpp_writer_put_raw_bits(&bit_ctx, 0, 24);
2411 mpp_writer_put_raw_bits(&bit_ctx, 1, 8);
2412 mpp_writer_put_raw_bits(&bit_ctx, 0x19, 8);
2413 for (i = 0; i < emdf_payload_size; i++) {
2414 RK_U8 data;
2415
2416 READ_BITS(gb, 8, &data);
2417 mpp_writer_put_bits(&bit_ctx, data, 8);
2418 }
2419
2420 hdr_dynamic_meta->size = mpp_writer_bytes(&bit_ctx);
2421 hdr_dynamic_meta->hdr_fmt = DLBY;
2422 av1d_dbg(AV1D_DBG_STRMIN, "dlby rpu size %d -> %d\n",
2423 emdf_payload_size, hdr_dynamic_meta->size);
2424
2425 ctx->hdr_dynamic_meta = hdr_dynamic_meta;
2426 ctx->hdr_dynamic = 1;
2427 ctx->is_hdr = 1;
2428
2429 if (av1d_debug & AV1D_DBG_DUMP_RPU) {
2430 RK_U8 *p = hdr_dynamic_meta->data;
2431 char fname[128];
2432 FILE *fp_in = NULL;
2433 static RK_U32 g_frame_no = 0;
2434
2435 sprintf(fname, "/data/video/meta_%d.txt", g_frame_no++);
2436 fp_in = fopen(fname, "wb");
2437 mpp_err("open %s %p\n", fname, fp_in);
2438 if (fp_in)
2439 fwrite(p, 1, hdr_dynamic_meta->size, fp_in);
2440 fflush(fp_in);
2441 fclose(fp_in);
2442 }
2443
2444 return 0;
2445
2446 __BITREAD_ERR:
2447 return MPP_ERR_STREAM;
2448 }
2449
mpp_av1_fill_dynamic_meta(AV1Context * ctx,const RK_U8 * data,RK_U32 size,RK_U32 hdr_fmt)2450 static void mpp_av1_fill_dynamic_meta(AV1Context *ctx, const RK_U8 *data, RK_U32 size, RK_U32 hdr_fmt)
2451 {
2452 MppFrameHdrDynamicMeta *hdr_dynamic_meta = ctx->hdr_dynamic_meta;
2453
2454 if (hdr_dynamic_meta && (hdr_dynamic_meta->size < size)) {
2455 mpp_free(hdr_dynamic_meta);
2456 hdr_dynamic_meta = NULL;
2457 }
2458
2459 if (!hdr_dynamic_meta) {
2460 hdr_dynamic_meta = mpp_calloc_size(MppFrameHdrDynamicMeta,
2461 sizeof(MppFrameHdrDynamicMeta) + size);
2462 if (!hdr_dynamic_meta) {
2463 mpp_err_f("malloc hdr dynamic data failed!\n");
2464 return;
2465 }
2466 }
2467 if (size && data) {
2468 switch (hdr_fmt) {
2469 case HDR10PLUS: {
2470 memcpy((RK_U8*)hdr_dynamic_meta->data, (RK_U8*)data, size);
2471 } break;
2472 default: break;
2473 }
2474 hdr_dynamic_meta->size = size;
2475 hdr_dynamic_meta->hdr_fmt = hdr_fmt;
2476
2477 ctx->hdr_dynamic_meta = hdr_dynamic_meta;
2478 ctx->hdr_dynamic = 1;
2479 ctx->is_hdr = 1;
2480 }
2481 }
2482
mpp_av1_metadata_itut_t35(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadataITUTT35 * current)2483 static RK_S32 mpp_av1_metadata_itut_t35(AV1Context *ctx, BitReadCtx_t *gb,
2484 AV1RawMetadataITUTT35 *current)
2485 {
2486 RK_S32 err;
2487
2488 fb(8, itu_t_t35_country_code);
2489 if (current->itu_t_t35_country_code == 0xff)
2490 fb(8, itu_t_t35_country_code_extension_byte);
2491
2492 current->payload_size = mpp_get_bits_left(gb) / 8 - 1;
2493
2494 av1d_dbg(AV1D_DBG_STRMIN, "%s itu_t_t35_country_code %d payload_size %d\n",
2495 __func__, current->itu_t_t35_country_code, current->payload_size);
2496
2497 fb(16, itu_t_t35_terminal_provider_code);
2498
2499 av1d_dbg(AV1D_DBG_STRMIN, "itu_t_t35_country_code 0x%x\n",
2500 current->itu_t_t35_country_code);
2501 av1d_dbg(AV1D_DBG_STRMIN, "itu_t_t35_terminal_provider_code 0x%x\n",
2502 current->itu_t_t35_terminal_provider_code);
2503
2504 switch (current->itu_t_t35_terminal_provider_code) {
2505 case 0x3B: {/* dlby provider_code is 0x3b*/
2506 READ_BITS_LONG(gb, 32, ¤t->itu_t_t35_terminal_provider_oriented_code);
2507 av1d_dbg(AV1D_DBG_STRMIN, "itu_t_t35_terminal_provider_oriented_code 0x%x\n",
2508 current->itu_t_t35_terminal_provider_oriented_code);
2509 if (current->itu_t_t35_terminal_provider_oriented_code == 0x800)
2510 mpp_av1_get_dlby_rpu(ctx, gb);
2511 } break;
2512 case 0x3C: {/* smpte2094_40 provider_code is 0x3c*/
2513 const RK_U16 smpte2094_40_provider_oriented_code = 0x0001;
2514 const RK_U8 smpte2094_40_application_identifier = 0x04;
2515 RK_U8 application_identifier;
2516
2517 fb(16, itu_t_t35_terminal_provider_oriented_code);
2518 av1d_dbg(AV1D_DBG_STRMIN, "itu_t_t35_terminal_provider_oriented_code 0x%x\n",
2519 current->itu_t_t35_terminal_provider_oriented_code);
2520 READ_BITS(gb, 8, &application_identifier);
2521 /* hdr10plus priverder_oriented_code is 0x0001, application_identifier is 0x04 */
2522 if (current->itu_t_t35_terminal_provider_oriented_code == smpte2094_40_provider_oriented_code &&
2523 application_identifier == smpte2094_40_application_identifier)
2524 mpp_av1_fill_dynamic_meta(ctx, gb->data_, mpp_get_bits_left(gb) >> 3, HDR10PLUS);
2525 } break;
2526 default:
2527 break;
2528 }
2529
2530 return 0;
2531 __BITREAD_ERR:
2532 return 0;
2533 }
2534
mpp_av1_metadata_timecode(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadataTimecode * current)2535 static RK_S32 mpp_av1_metadata_timecode(AV1Context *ctx, BitReadCtx_t *gb,
2536 AV1RawMetadataTimecode *current)
2537 {
2538 RK_S32 err;
2539 (void)ctx;
2540
2541 fb(5, counting_type);
2542 flag(full_timestamp_flag);
2543 flag(discontinuity_flag);
2544 flag(cnt_dropped_flag);
2545 fb(9, n_frames);
2546
2547 if (current->full_timestamp_flag) {
2548 fc(6, seconds_value, 0, 59);
2549 fc(6, minutes_value, 0, 59);
2550 fc(5, hours_value, 0, 23);
2551 } else {
2552 flag(seconds_flag);
2553 if (current->seconds_flag) {
2554 fc(6, seconds_value, 0, 59);
2555 flag(minutes_flag);
2556 if (current->minutes_flag) {
2557 fc(6, minutes_value, 0, 59);
2558 flag(hours_flag);
2559 if (current->hours_flag)
2560 fc(5, hours_value, 0, 23);
2561 }
2562 }
2563 }
2564
2565 fb(5, time_offset_length);
2566 if (current->time_offset_length > 0)
2567 fb(current->time_offset_length, time_offset_value);
2568 else
2569 infer(time_offset_length, 0);
2570
2571 return 0;
2572 }
2573
mpp_av1_metadata_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawMetadata * current)2574 static RK_S32 mpp_av1_metadata_obu(AV1Context *ctx, BitReadCtx_t *gb,
2575 AV1RawMetadata *current)
2576 {
2577 RK_S32 err;
2578
2579 leb128(metadata_type);
2580 av1d_dbg(AV1D_DBG_STRMIN, "%s meta type %lld\n", __func__, current->metadata_type);
2581 switch (current->metadata_type) {
2582 case AV1_METADATA_TYPE_HDR_CLL:
2583 CHECK(mpp_av1_metadata_hdr_cll(ctx, gb, ¤t->metadata.hdr_cll));
2584 break;
2585 case AV1_METADATA_TYPE_HDR_MDCV:
2586 CHECK(mpp_av1_metadata_hdr_mdcv(ctx, gb, ¤t->metadata.hdr_mdcv));
2587 break;
2588 case AV1_METADATA_TYPE_SCALABILITY:
2589 CHECK(mpp_av1_metadata_scalability(ctx, gb, ¤t->metadata.scalability));
2590 break;
2591 case AV1_METADATA_TYPE_ITUT_T35:
2592 CHECK(mpp_av1_metadata_itut_t35(ctx, gb, ¤t->metadata.itut_t35));
2593 break;
2594 case AV1_METADATA_TYPE_TIMECODE:
2595 CHECK(mpp_av1_metadata_timecode(ctx, gb, ¤t->metadata.timecode));
2596 break;
2597 default:
2598 mpp_err_f("unknown metadata type %lld\n", current->metadata_type);
2599 break;
2600 }
2601
2602 return 0;
2603 }
2604
mpp_av1_padding_obu(AV1Context * ctx,BitReadCtx_t * gb,AV1RawPadding * current)2605 static RK_S32 mpp_av1_padding_obu(AV1Context *ctx, BitReadCtx_t *gb,
2606 AV1RawPadding *current)
2607 {
2608 RK_S32 err;
2609 RK_U32 i;
2610 (void)ctx;
2611 current->payload_size = mpp_av1_get_payload_bytes_left(gb);
2612
2613 current->payload = mpp_malloc(RK_U8, current->payload_size);
2614 if (!current->payload )
2615 return MPP_ERR_NOMEM;
2616
2617 for (i = 0; i < current->payload_size; i++)
2618 xf(8, obu_padding_byte[i], current->payload[i], 0x00, 0xff, 1, i);
2619
2620 return 0;
2621 }
2622
2623
2624
mpp_insert_unit(Av1UnitFragment * frag,RK_S32 position)2625 static MPP_RET mpp_insert_unit(Av1UnitFragment *frag, RK_S32 position)
2626 {
2627 Av1ObuUnit *units;
2628
2629 if (frag->nb_units < frag->nb_units_allocated) {
2630 units = frag->units;
2631
2632 if (position < frag->nb_units)
2633 memmove(units + position + 1, units + position,
2634 (frag->nb_units - position) * sizeof(*units));
2635 } else {
2636 units = mpp_malloc(Av1ObuUnit, frag->nb_units * 2 + 1);
2637 if (!units)
2638 return MPP_ERR_NOMEM;
2639
2640 frag->nb_units_allocated = 2 * frag->nb_units_allocated + 1;
2641
2642 if (position > 0)
2643 memcpy(units, frag->units, position * sizeof(*units));
2644
2645 if (position < frag->nb_units)
2646 memcpy(units + position + 1, frag->units + position,
2647 (frag->nb_units - position) * sizeof(*units));
2648 }
2649
2650 memset(units + position, 0, sizeof(*units));
2651
2652 if (units != frag->units) {
2653 mpp_free(frag->units);
2654 frag->units = units;
2655 }
2656
2657 ++frag->nb_units;
2658
2659 return MPP_OK;
2660 }
2661
mpp_insert_unit_data(Av1UnitFragment * frag,RK_S32 position,Av1UnitType type,RK_U8 * data,size_t data_size)2662 static MPP_RET mpp_insert_unit_data(Av1UnitFragment *frag,
2663 RK_S32 position,
2664 Av1UnitType type,
2665 RK_U8 *data, size_t data_size)
2666 {
2667 Av1ObuUnit *unit;
2668 MPP_RET ret;
2669
2670 if (position == -1)
2671 position = frag->nb_units;
2672
2673 mpp_assert(position >= 0 && position <= frag->nb_units);
2674 ret = mpp_insert_unit(frag, position);
2675 if (ret < 0) {
2676 return ret;
2677 }
2678
2679 unit = &frag->units[position];
2680 unit->type = type;
2681 unit->data = data;
2682 unit->data_size = data_size;
2683
2684 return MPP_OK;
2685 }
2686
mpp_av1_split_fragment(AV1Context * ctx,Av1UnitFragment * frag,RK_S32 header_flag)2687 RK_S32 mpp_av1_split_fragment(AV1Context *ctx, Av1UnitFragment *frag, RK_S32 header_flag)
2688 {
2689 BitReadCtx_t gbc;
2690 RK_U8 *data;
2691 size_t size;
2692 RK_U64 obu_length;
2693 RK_S32 pos, err;
2694
2695 data = frag->data;
2696 size = frag->data_size;
2697
2698 if (INT_MAX / 8 < size) {
2699 mpp_err( "Invalid fragment: "
2700 "too large (%d bytes).\n", size);
2701 err = MPP_NOK;
2702 goto fail;
2703 }
2704
2705 if (header_flag && size && data[0] & 0x80) {
2706 // first bit is nonzero, the extradata does not consist purely of
2707 // OBUs. Expect MP4/Matroska AV1CodecConfigurationRecord
2708 RK_S32 config_record_version = data[0] & 0x7f;
2709
2710 if (config_record_version != 1) {
2711 mpp_err(
2712 "Unknown version %d of AV1CodecConfigurationRecord "
2713 "found!\n",
2714 config_record_version);
2715 err = MPP_NOK;
2716 goto fail;
2717 }
2718
2719 if (size <= 4) {
2720 if (size < 4) {
2721 av1d_dbg(AV1D_DBG_STRMIN,
2722 "Undersized AV1CodecConfigurationRecord v%d found!\n",
2723 config_record_version);
2724 err = MPP_NOK;
2725 goto fail;
2726 }
2727
2728 goto success;
2729 }
2730
2731 // In AV1CodecConfigurationRecord v1, actual OBUs start after
2732 // four bytes. Thus set the offset as required for properly
2733 // parsing them.
2734 data += 4;
2735 size -= 4;
2736 }
2737
2738 while (size > 0) {
2739 AV1RawOBUHeader header;
2740 RK_U64 obu_size = 0;
2741
2742 mpp_set_bitread_ctx(&gbc, data, size);
2743
2744 err = mpp_av1_read_obu_header(ctx, &gbc, &header);
2745 if (err < 0)
2746 goto fail;
2747
2748 if (header.obu_has_size_field) {
2749 if (mpp_get_bits_left(&gbc) < 8) {
2750 mpp_err( "Invalid OBU: fragment "
2751 "too short (%d bytes).\n", size);
2752 err = MPP_NOK;
2753 goto fail;
2754 }
2755 err = mpp_av1_read_leb128(&gbc, &obu_size);
2756 if (err < 0)
2757 goto fail;
2758 } else
2759 obu_size = size - 1 - header.obu_extension_flag;
2760
2761 pos = mpp_get_bits_count(&gbc);
2762
2763 mpp_assert(pos % 8 == 0 && pos / 8 <= (RK_S32)size);
2764 obu_length = pos / 8 + obu_size;
2765
2766 if (size < obu_length) {
2767 mpp_err( "Invalid OBU length: "
2768 "%lld, but only %d bytes remaining in fragment.\n",
2769 obu_length, size);
2770 err = MPP_NOK;
2771 goto fail;
2772 }
2773 err = mpp_insert_unit_data(frag, -1, header.obu_type,
2774 data, obu_length);
2775 if (err < 0)
2776 goto fail;
2777
2778 data += obu_length;
2779 size -= obu_length;
2780 }
2781
2782 success:
2783 err = 0;
2784 fail:
2785 return err;
2786 }
2787
mpp_av1_ref_tile_data(Av1ObuUnit * unit,BitReadCtx_t * gbc,AV1RawTileData * td)2788 static RK_S32 mpp_av1_ref_tile_data(Av1ObuUnit *unit,
2789 BitReadCtx_t *gbc,
2790 AV1RawTileData *td)
2791 {
2792 RK_S32 pos;
2793
2794 pos = mpp_get_bits_count(gbc);
2795 if (pos >= (RK_S32)(8 * unit->data_size)) {
2796 mpp_err( "Bitstream ended before "
2797 "any data in tile group (%d bits read).\n", pos);
2798 return MPP_NOK;
2799 }
2800 // Must be byte-aligned at this point.
2801 mpp_assert(pos % 8 == 0);
2802
2803 td->offset = pos / 8;
2804 td->data = unit->data + pos / 8;
2805 td->data_size = unit->data_size - pos / 8;
2806
2807 return 0;
2808 }
2809
mpp_av1_alloc_unit_content(Av1ObuUnit * unit)2810 static MPP_RET mpp_av1_alloc_unit_content(Av1ObuUnit *unit)
2811 {
2812 (void)unit;
2813 MPP_FREE(unit->content);
2814 unit->content = mpp_calloc(AV1RawOBU, 1);
2815 if (!unit->content) {
2816 return MPP_ERR_NOMEM; // drop_obu()
2817 }
2818 return MPP_OK;
2819 }
2820
mpp_av1_read_unit(AV1Context * ctx,Av1ObuUnit * unit)2821 MPP_RET mpp_av1_read_unit(AV1Context *ctx, Av1ObuUnit *unit)
2822 {
2823 AV1RawOBU *obu;
2824 BitReadCtx_t gbc;
2825 RK_S32 err = 0, start_pos, end_pos, hdr_start_pos;
2826
2827 err = mpp_av1_alloc_unit_content(unit);
2828
2829 if (err < 0)
2830 return err;
2831
2832 obu = unit->content;
2833
2834 mpp_set_bitread_ctx(&gbc, unit->data, unit->data_size);
2835
2836 hdr_start_pos = mpp_get_bits_count(&gbc);
2837
2838 err = mpp_av1_read_obu_header(ctx, &gbc, &obu->header);
2839 if (err < 0)
2840 return err;
2841 mpp_assert(obu->header.obu_type == unit->type);
2842
2843 if (obu->header.obu_has_size_field) {
2844 RK_U64 obu_size = 0;
2845 err = mpp_av1_read_leb128(&gbc, &obu_size);
2846 if (err < 0)
2847 return err;
2848 obu->obu_size = obu_size;
2849 } else {
2850 if (unit->data_size < (RK_U32)(1 + obu->header.obu_extension_flag)) {
2851 mpp_err( "Invalid OBU length: "
2852 "unit too short (%d).\n", unit->data_size);
2853 return MPP_NOK;
2854 }
2855 obu->obu_size = unit->data_size - 1 - obu->header.obu_extension_flag;
2856 }
2857
2858 start_pos = mpp_get_bits_count(&gbc);
2859 if (!ctx->fist_tile_group)
2860 ctx->frame_tag_size += ((start_pos - hdr_start_pos + 7) >> 3);
2861 if (obu->header.obu_extension_flag) {
2862 if (obu->header.obu_type != AV1_OBU_SEQUENCE_HEADER &&
2863 obu->header.obu_type != AV1_OBU_TEMPORAL_DELIMITER &&
2864 ctx->operating_point_idc) {
2865 RK_S32 in_temporal_layer =
2866 (ctx->operating_point_idc >> ctx->temporal_id ) & 1;
2867 RK_S32 in_spatial_layer =
2868 (ctx->operating_point_idc >> (ctx->spatial_id + 8)) & 1;
2869 if (!in_temporal_layer || !in_spatial_layer) {
2870 return MPP_ERR_PROTOL; // drop_obu()
2871 }
2872 }
2873 }
2874 av1d_dbg(AV1D_DBG_HEADER, "obu type %d size %d\n",
2875 obu->header.obu_type, obu->obu_size);
2876 switch (obu->header.obu_type) {
2877 case AV1_OBU_SEQUENCE_HEADER: {
2878 err = mpp_av1_sequence_header_obu(ctx, &gbc,
2879 &obu->obu.sequence_header);
2880 if (err < 0)
2881 return err;
2882 ctx->frame_tag_size += obu->obu_size;
2883 if (ctx->operating_point >= 0) {
2884 AV1RawSequenceHeader *sequence_header = &obu->obu.sequence_header;
2885
2886 if (ctx->operating_point > sequence_header->operating_points_cnt_minus_1) {
2887 mpp_err("Invalid Operating Point %d requested. "
2888 "Must not be higher than %u.\n",
2889 ctx->operating_point, sequence_header->operating_points_cnt_minus_1);
2890 return MPP_ERR_PROTOL;
2891 }
2892 ctx->operating_point_idc = sequence_header->operating_point_idc[ctx->operating_point];
2893 }
2894
2895 ctx->sequence_header = NULL;
2896 ctx->sequence_header = &obu->obu.sequence_header;
2897 } break;
2898 case AV1_OBU_TEMPORAL_DELIMITER: {
2899 err = mpp_av1_temporal_delimiter_obu(ctx, &gbc);
2900 if (err < 0)
2901 return err;
2902 } break;
2903 case AV1_OBU_FRAME_HEADER:
2904 case AV1_OBU_REDUNDANT_FRAME_HEADER: {
2905 err = mpp_av1_frame_header_obu(ctx, &gbc,
2906 &obu->obu.frame_header,
2907 obu->header.obu_type ==
2908 AV1_OBU_REDUNDANT_FRAME_HEADER,
2909 NULL);
2910 if (err < 0)
2911 return err;
2912 ctx->frame_tag_size += obu->obu_size;
2913 } break;
2914 case AV1_OBU_TILE_GROUP: {
2915 RK_U32 cur_pos = mpp_get_bits_count(&gbc);
2916
2917 err = mpp_av1_tile_group_obu(ctx, &gbc, &obu->obu.tile_group);
2918 if (err < 0)
2919 return err;
2920 if (!ctx->fist_tile_group)
2921 ctx->frame_tag_size += MPP_ALIGN(mpp_get_bits_count(&gbc) - cur_pos, 8) / 8;
2922 ctx->fist_tile_group = 1;
2923 err = mpp_av1_ref_tile_data(unit, &gbc,
2924 &obu->obu.tile_group.tile_data);
2925 if (err < 0)
2926 return err;
2927 } break;
2928 case AV1_OBU_FRAME: {
2929 err = mpp_av1_frame_obu(ctx, &gbc, &obu->obu.frame,
2930 NULL);
2931 if (err < 0)
2932 return err;
2933
2934 err = mpp_av1_ref_tile_data(unit, &gbc,
2935 &obu->obu.frame.tile_group.tile_data);
2936 if (err < 0)
2937 return err;
2938 } break;
2939 case AV1_OBU_TILE_LIST: {
2940 err = mpp_av1_tile_list_obu(ctx, &gbc, &obu->obu.tile_list);
2941 if (err < 0)
2942 return err;
2943
2944 err = mpp_av1_ref_tile_data(unit, &gbc,
2945 &obu->obu.tile_list.tile_data);
2946 if (err < 0)
2947 return err;
2948 } break;
2949 case AV1_OBU_METADATA: {
2950 ctx->frame_tag_size += obu->obu_size;
2951 err = mpp_av1_metadata_obu(ctx, &gbc, &obu->obu.metadata);
2952 if (err < 0)
2953 return err;
2954 } break;
2955 case AV1_OBU_PADDING: {
2956 err = mpp_av1_padding_obu(ctx, &gbc, &obu->obu.padding);
2957 if (err < 0)
2958 return err;
2959 } break;
2960 default:
2961 return MPP_ERR_VALUE;
2962 }
2963
2964 end_pos = mpp_get_bits_count(&gbc);
2965 mpp_assert(end_pos <= (RK_S32)(unit->data_size * 8));
2966
2967 if (obu->obu_size > 0 &&
2968 obu->header.obu_type != AV1_OBU_TILE_GROUP &&
2969 obu->header.obu_type != AV1_OBU_TILE_LIST &&
2970 obu->header.obu_type != AV1_OBU_FRAME) {
2971 RK_S32 nb_bits = obu->obu_size * 8 + start_pos - end_pos;
2972
2973 if (nb_bits <= 0)
2974 return MPP_NOK;
2975
2976 err = mpp_av1_trailing_bits(ctx, &gbc, nb_bits);
2977 if (err < 0)
2978 return err;
2979 }
2980
2981 return 0;
2982 }
2983
mpp_av1_read_fragment_content(AV1Context * ctx,Av1UnitFragment * frag)2984 RK_S32 mpp_av1_read_fragment_content(AV1Context *ctx, Av1UnitFragment *frag)
2985 {
2986 int err, i, j;
2987 AV1RawOBU *obu;
2988
2989 ctx->frame_tag_size = 0;
2990 ctx->fist_tile_group = 0;
2991 for (i = 0; i < frag->nb_units; i++) {
2992 Av1ObuUnit *unit = &frag->units[i];
2993 if (ctx->unit_types) {
2994 for (j = 0; j < ctx->nb_unit_types; j++) {
2995 if (ctx->unit_types[j] == unit->type)
2996 break;
2997 }
2998 if (j >= ctx->nb_unit_types)
2999 continue;
3000 }
3001 MPP_FREE(unit->content);
3002 mpp_assert(unit->data);
3003 err = mpp_av1_read_unit(ctx, unit);
3004
3005 if (err == MPP_ERR_VALUE) {
3006 mpp_err_f("Decomposition unimplemented for unit %d "
3007 "(type %d).\n", i, unit->type);
3008 } else if (err == MPP_ERR_PROTOL) {
3009 mpp_err_f("Skipping decomposition of"
3010 "unit %d (type %d).\n", i, unit->type);
3011 MPP_FREE(unit->content);
3012 unit->content = NULL;
3013 } else if (err < 0) {
3014 mpp_err_f("Failed to read unit %d (type %d).\n", i, unit->type);
3015 return err;
3016 }
3017 obu = unit->content;
3018 av1d_dbg(AV1D_DBG_HEADER, "obu->header.obu_type %d, obu->obu_size = %d ctx->frame_tag_size %d",
3019 obu->header.obu_type, obu->obu_size, ctx->frame_tag_size);
3020 }
3021 return 0;
3022 }
3023
mpp_av1_set_context_with_sequence(Av1CodecContext * ctx,const AV1RawSequenceHeader * seq)3024 int mpp_av1_set_context_with_sequence(Av1CodecContext *ctx,
3025 const AV1RawSequenceHeader *seq)
3026 {
3027 int width = seq->max_frame_width_minus_1 + 1;
3028 int height = seq->max_frame_height_minus_1 + 1;
3029
3030 ctx->profile = seq->seq_profile;
3031 ctx->level = seq->seq_level_idx[0];
3032
3033 ctx->color_range =
3034 seq->color_config.color_range ? MPP_FRAME_RANGE_JPEG : MPP_FRAME_RANGE_MPEG;
3035 ctx->color_primaries = seq->color_config.color_primaries;
3036 ctx->colorspace = seq->color_config.matrix_coefficients;
3037 ctx->color_trc = seq->color_config.transfer_characteristics;
3038
3039 switch (seq->color_config.chroma_sample_position) {
3040 case AV1_CSP_VERTICAL:
3041 ctx->chroma_sample_location = MPP_CHROMA_LOC_LEFT;
3042 break;
3043 case AV1_CSP_COLOCATED:
3044 ctx->chroma_sample_location = MPP_CHROMA_LOC_TOPLEFT;
3045 break;
3046 }
3047
3048 if (ctx->width != width || ctx->height != height) {
3049 ctx->width = width;
3050 ctx->height = height;
3051 }
3052 return 0;
3053 }
3054
mpp_av1_fragment_reset(Av1UnitFragment * frag)3055 void mpp_av1_fragment_reset(Av1UnitFragment *frag)
3056 {
3057 int i;
3058
3059 for (i = 0; i < frag->nb_units; i++) {
3060 Av1ObuUnit *unit = &frag->units[i];
3061 MPP_FREE(unit->content);
3062 unit->data = NULL;
3063 unit->data_size = 0;
3064 }
3065 frag->nb_units = 0;
3066 frag->data = NULL;
3067 frag->data_size = 0;
3068 }
3069
mpp_av1_assemble_fragment(AV1Context * ctx,Av1UnitFragment * frag)3070 RK_S32 mpp_av1_assemble_fragment(AV1Context *ctx, Av1UnitFragment *frag)
3071 {
3072 size_t size, pos;
3073 RK_S32 i;
3074 (void)ctx;
3075 size = 0;
3076 for (i = 0; i < frag->nb_units; i++)
3077 size += frag->units[i].data_size;
3078
3079 frag->data = mpp_malloc(RK_U8, size + BUFFER_PADDING_SIZE);
3080 if (!frag->data)
3081 return MPP_ERR_NOMEM;
3082
3083 memset(frag->data + size, 0, BUFFER_PADDING_SIZE);
3084
3085 pos = 0;
3086 for (i = 0; i < frag->nb_units; i++) {
3087 memcpy(frag->data + pos, frag->units[i].data,
3088 frag->units[i].data_size);
3089 pos += frag->units[i].data_size;
3090 }
3091 mpp_assert(pos == size);
3092 frag->data_size = size;
3093
3094 return 0;
3095 }
3096
mpp_av1_flush(AV1Context * ctx)3097 void mpp_av1_flush(AV1Context *ctx)
3098 {
3099 // ctx->sequencframe_headere_header = NULL;
3100 // ctx-> = NULL;
3101
3102 memset(ctx->ref_s, 0, sizeof(ctx->ref_s));
3103 ctx->operating_point_idc = 0;
3104 ctx->seen_frame_header = 0;
3105 ctx->tile_num = 0;
3106 }
3107
mpp_av1_close(AV1Context * ctx)3108 void mpp_av1_close(AV1Context *ctx)
3109 {
3110 MPP_FREE(ctx->frame_header);
3111 MPP_FREE(ctx->sequence_header);
3112 MPP_FREE(ctx->raw_frame_header);
3113 }
3114
mpp_av1_free_metadata(void * unit,RK_U8 * content)3115 void mpp_av1_free_metadata(void *unit, RK_U8 *content)
3116 {
3117 AV1RawOBU *obu = (AV1RawOBU*)content;
3118 (void)unit;
3119 mpp_assert(obu->header.obu_type == AV1_OBU_METADATA);
3120 MPP_FREE(content);
3121 }
3122