]> git.sesse.net Git - ffmpeg/blob - libavcodec/vdpau.c
Merge commit '16c7a8a142277aa3c98b4802ee57daaebe6999e5'
[ffmpeg] / libavcodec / vdpau.c
1 /*
2  * Video Decode and Presentation API for UNIX (VDPAU) is used for
3  * HW decode acceleration for MPEG-1/2, MPEG-4 ASP, H.264 and VC-1.
4  *
5  * Copyright (c) 2008 NVIDIA
6  *
7  * This file is part of FFmpeg.
8  *
9  * FFmpeg is free software; you can redistribute it and/or
10  * modify it under the terms of the GNU Lesser General Public
11  * License as published by the Free Software Foundation; either
12  * version 2.1 of the License, or (at your option) any later version.
13  *
14  * FFmpeg is distributed in the hope that it will be useful,
15  * but WITHOUT ANY WARRANTY; without even the implied warranty of
16  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
17  * Lesser General Public License for more details.
18  *
19  * You should have received a copy of the GNU Lesser General Public
20  * License along with FFmpeg; if not, write to the Free Software
21  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
22  */
23
24 #include <limits.h>
25 #include "libavutil/avassert.h"
26 #include "avcodec.h"
27 #include "internal.h"
28 #include "h264.h"
29 #include "vc1.h"
30
31 #undef NDEBUG
32 #include <assert.h>
33
34 #include "vdpau.h"
35 #include "vdpau_internal.h"
36
37 /**
38  * @addtogroup VDPAU_Decoding
39  *
40  * @{
41  */
42
43 static int vdpau_error(VdpStatus status)
44 {
45     switch (status) {
46     case VDP_STATUS_OK:
47         return 0;
48     case VDP_STATUS_NO_IMPLEMENTATION:
49         return AVERROR(ENOSYS);
50     case VDP_STATUS_DISPLAY_PREEMPTED:
51         return AVERROR(EIO);
52     case VDP_STATUS_INVALID_HANDLE:
53         return AVERROR(EBADF);
54     case VDP_STATUS_INVALID_POINTER:
55         return AVERROR(EFAULT);
56     case VDP_STATUS_RESOURCES:
57         return AVERROR(ENOBUFS);
58     case VDP_STATUS_HANDLE_DEVICE_MISMATCH:
59         return AVERROR(EXDEV);
60     case VDP_STATUS_ERROR:
61         return AVERROR(EIO);
62     default:
63         return AVERROR(EINVAL);
64     }
65 }
66
67 AVVDPAUContext *av_alloc_vdpaucontext(void)
68 {
69     return av_vdpau_alloc_context();
70 }
71
72 MAKE_ACCESSORS(AVVDPAUContext, vdpau_hwaccel, AVVDPAU_Render2, render2)
73
74 int ff_vdpau_common_init(AVCodecContext *avctx, VdpDecoderProfile profile,
75                          int level)
76 {
77     VDPAUHWContext *hwctx = avctx->hwaccel_context;
78     VDPAUContext *vdctx = avctx->internal->hwaccel_priv_data;
79     VdpVideoSurfaceQueryCapabilities *surface_query_caps;
80     VdpDecoderQueryCapabilities *decoder_query_caps;
81     VdpDecoderCreate *create;
82     void *func;
83     VdpStatus status;
84     VdpBool supported;
85     uint32_t max_level, max_mb, max_width, max_height;
86     /* See vdpau/vdpau.h for alignment constraints. */
87     uint32_t width  = (avctx->coded_width + 1) & ~1;
88     uint32_t height = (avctx->coded_height + 3) & ~3;
89
90     vdctx->width            = UINT32_MAX;
91     vdctx->height           = UINT32_MAX;
92
93     if (!hwctx) {
94         vdctx->device  = VDP_INVALID_HANDLE;
95         av_log(avctx, AV_LOG_WARNING, "hwaccel_context has not been setup by the user application, cannot initialize\n");
96         return 0;
97     }
98
99     if (hwctx->context.decoder != VDP_INVALID_HANDLE) {
100         vdctx->decoder = hwctx->context.decoder;
101         vdctx->render  = hwctx->context.render;
102         vdctx->device  = VDP_INVALID_HANDLE;
103         return 0; /* Decoder created by user */
104     }
105     hwctx->reset            = 0;
106
107     vdctx->device           = hwctx->device;
108     vdctx->get_proc_address = hwctx->get_proc_address;
109
110     if (hwctx->flags & AV_HWACCEL_FLAG_IGNORE_LEVEL)
111         level = 0;
112     else if (level < 0)
113         return AVERROR(ENOTSUP);
114
115     status = vdctx->get_proc_address(vdctx->device,
116                                      VDP_FUNC_ID_VIDEO_SURFACE_QUERY_CAPABILITIES,
117                                      &func);
118     if (status != VDP_STATUS_OK)
119         return vdpau_error(status);
120     else
121         surface_query_caps = func;
122
123     status = surface_query_caps(vdctx->device, VDP_CHROMA_TYPE_420, &supported,
124                                 &max_width, &max_height);
125     if (status != VDP_STATUS_OK)
126         return vdpau_error(status);
127     if (supported != VDP_TRUE ||
128         max_width < width || max_height < height)
129         return AVERROR(ENOTSUP);
130
131     status = vdctx->get_proc_address(vdctx->device,
132                                      VDP_FUNC_ID_DECODER_QUERY_CAPABILITIES,
133                                      &func);
134     if (status != VDP_STATUS_OK)
135         return vdpau_error(status);
136     else
137         decoder_query_caps = func;
138
139     status = decoder_query_caps(vdctx->device, profile, &supported, &max_level,
140                                 &max_mb, &max_width, &max_height);
141 #ifdef VDP_DECODER_PROFILE_H264_CONSTRAINED_BASELINE
142     if (status != VDP_STATUS_OK && profile == VDP_DECODER_PROFILE_H264_CONSTRAINED_BASELINE) {
143         /* Run-time backward compatibility for libvdpau 0.8 and earlier */
144         profile = VDP_DECODER_PROFILE_H264_MAIN;
145         status = decoder_query_caps(vdctx->device, profile, &supported,
146                                     &max_level, &max_mb,
147                                     &max_width, &max_height);
148     }
149 #endif
150     if (status != VDP_STATUS_OK)
151         return vdpau_error(status);
152
153     if (supported != VDP_TRUE || max_level < level ||
154         max_width < width || max_height < height)
155         return AVERROR(ENOTSUP);
156
157     status = vdctx->get_proc_address(vdctx->device, VDP_FUNC_ID_DECODER_CREATE,
158                                      &func);
159     if (status != VDP_STATUS_OK)
160         return vdpau_error(status);
161     else
162         create = func;
163
164     status = vdctx->get_proc_address(vdctx->device, VDP_FUNC_ID_DECODER_RENDER,
165                                      &func);
166     if (status != VDP_STATUS_OK)
167         return vdpau_error(status);
168     else
169         vdctx->render = func;
170
171     status = create(vdctx->device, profile, width, height, avctx->refs,
172                     &vdctx->decoder);
173     if (status == VDP_STATUS_OK) {
174         vdctx->width  = avctx->coded_width;
175         vdctx->height = avctx->coded_height;
176     }
177
178     return vdpau_error(status);
179 }
180
181 int ff_vdpau_common_uninit(AVCodecContext *avctx)
182 {
183     VDPAUContext *vdctx = avctx->internal->hwaccel_priv_data;
184     VdpDecoderDestroy *destroy;
185     void *func;
186     VdpStatus status;
187
188     if (vdctx->device == VDP_INVALID_HANDLE)
189         return 0; /* Decoder created and destroyed by user */
190     if (vdctx->width == UINT32_MAX && vdctx->height == UINT32_MAX)
191         return 0;
192
193     status = vdctx->get_proc_address(vdctx->device,
194                                      VDP_FUNC_ID_DECODER_DESTROY, &func);
195     if (status != VDP_STATUS_OK)
196         return vdpau_error(status);
197     else
198         destroy = func;
199
200     status = destroy(vdctx->decoder);
201     return vdpau_error(status);
202 }
203
204 static int ff_vdpau_common_reinit(AVCodecContext *avctx)
205 {
206     VDPAUHWContext *hwctx = avctx->hwaccel_context;
207     VDPAUContext *vdctx = avctx->internal->hwaccel_priv_data;
208
209     if (vdctx->device == VDP_INVALID_HANDLE)
210         return 0; /* Decoder created by user */
211     if (avctx->coded_width == vdctx->width &&
212         avctx->coded_height == vdctx->height && !hwctx->reset)
213         return 0;
214
215     avctx->hwaccel->uninit(avctx);
216     return avctx->hwaccel->init(avctx);
217 }
218
219 int ff_vdpau_common_start_frame(struct vdpau_picture_context *pic_ctx,
220                                 av_unused const uint8_t *buffer,
221                                 av_unused uint32_t size)
222 {
223     pic_ctx->bitstream_buffers_allocated = 0;
224     pic_ctx->bitstream_buffers_used      = 0;
225     pic_ctx->bitstream_buffers           = NULL;
226     return 0;
227 }
228
229 int ff_vdpau_common_end_frame(AVCodecContext *avctx, AVFrame *frame,
230                               struct vdpau_picture_context *pic_ctx)
231 {
232     VDPAUContext *vdctx = avctx->internal->hwaccel_priv_data;
233     AVVDPAUContext *hwctx = avctx->hwaccel_context;
234     VdpVideoSurface surf = ff_vdpau_get_surface_id(frame);
235     VdpStatus status;
236     int val;
237
238     val = ff_vdpau_common_reinit(avctx);
239     if (val < 0)
240         return val;
241
242 #if FF_API_BUFS_VDPAU
243 FF_DISABLE_DEPRECATION_WARNINGS
244     hwctx->info = pic_ctx->info;
245     hwctx->bitstream_buffers = pic_ctx->bitstream_buffers;
246     hwctx->bitstream_buffers_used = pic_ctx->bitstream_buffers_used;
247     hwctx->bitstream_buffers_allocated = pic_ctx->bitstream_buffers_allocated;
248 FF_ENABLE_DEPRECATION_WARNINGS
249 #endif
250
251     if (!hwctx->render && hwctx->render2) {
252         status = hwctx->render2(avctx, frame, (void *)&pic_ctx->info,
253                                 pic_ctx->bitstream_buffers_used, pic_ctx->bitstream_buffers);
254     } else
255     status = vdctx->render(vdctx->decoder, surf, (void *)&pic_ctx->info,
256                            pic_ctx->bitstream_buffers_used,
257                            pic_ctx->bitstream_buffers);
258
259     av_freep(&pic_ctx->bitstream_buffers);
260
261 #if FF_API_BUFS_VDPAU
262 FF_DISABLE_DEPRECATION_WARNINGS
263     hwctx->bitstream_buffers = NULL;
264     hwctx->bitstream_buffers_used = 0;
265     hwctx->bitstream_buffers_allocated = 0;
266 FF_ENABLE_DEPRECATION_WARNINGS
267 #endif
268
269     return vdpau_error(status);
270 }
271
272 #if CONFIG_H263_VDPAU_HWACCEL  || CONFIG_MPEG1_VDPAU_HWACCEL || \
273     CONFIG_MPEG2_VDPAU_HWACCEL || CONFIG_MPEG4_VDPAU_HWACCEL || \
274     CONFIG_VC1_VDPAU_HWACCEL   || CONFIG_WMV3_VDPAU_HWACCEL
275 int ff_vdpau_mpeg_end_frame(AVCodecContext *avctx)
276 {
277     MpegEncContext *s = avctx->priv_data;
278     Picture *pic = s->current_picture_ptr;
279     struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
280     int val;
281
282     val = ff_vdpau_common_end_frame(avctx, pic->f, pic_ctx);
283     if (val < 0)
284         return val;
285
286     ff_mpeg_draw_horiz_band(s, 0, s->avctx->height);
287     return 0;
288 }
289 #endif
290
291 int ff_vdpau_add_buffer(struct vdpau_picture_context *pic_ctx,
292                         const uint8_t *buf, uint32_t size)
293 {
294     VdpBitstreamBuffer *buffers = pic_ctx->bitstream_buffers;
295
296     buffers = av_fast_realloc(buffers, &pic_ctx->bitstream_buffers_allocated,
297                               (pic_ctx->bitstream_buffers_used + 1) * sizeof(*buffers));
298     if (!buffers)
299         return AVERROR(ENOMEM);
300
301     pic_ctx->bitstream_buffers = buffers;
302     buffers += pic_ctx->bitstream_buffers_used++;
303
304     buffers->struct_version  = VDP_BITSTREAM_BUFFER_VERSION;
305     buffers->bitstream       = buf;
306     buffers->bitstream_bytes = size;
307     return 0;
308 }
309
310 /* Obsolete non-hwaccel VDPAU support below... */
311
312 void ff_vdpau_h264_set_reference_frames(H264Context *h)
313 {
314     struct vdpau_render_state *render, *render_ref;
315     VdpReferenceFrameH264 *rf, *rf2;
316     H264Picture *pic;
317     int i, list, pic_frame_idx;
318
319     render = (struct vdpau_render_state *)h->cur_pic_ptr->f.data[0];
320     assert(render);
321
322     rf = &render->info.h264.referenceFrames[0];
323 #define H264_RF_COUNT FF_ARRAY_ELEMS(render->info.h264.referenceFrames)
324
325     for (list = 0; list < 2; ++list) {
326         H264Picture **lp = list ? h->long_ref : h->short_ref;
327         int ls = list ? 16 : h->short_ref_count;
328
329         for (i = 0; i < ls; ++i) {
330             pic = lp[i];
331             if (!pic || !pic->reference)
332                 continue;
333             pic_frame_idx = pic->long_ref ? pic->pic_id : pic->frame_num;
334
335             render_ref = (struct vdpau_render_state *)pic->f.data[0];
336             assert(render_ref);
337
338             rf2 = &render->info.h264.referenceFrames[0];
339             while (rf2 != rf) {
340                 if (
341                     (rf2->surface == render_ref->surface)
342                     && (rf2->is_long_term == pic->long_ref)
343                     && (rf2->frame_idx == pic_frame_idx)
344                 )
345                     break;
346                 ++rf2;
347             }
348             if (rf2 != rf) {
349                 rf2->top_is_reference    |= (pic->reference & PICT_TOP_FIELD)    ? VDP_TRUE : VDP_FALSE;
350                 rf2->bottom_is_reference |= (pic->reference & PICT_BOTTOM_FIELD) ? VDP_TRUE : VDP_FALSE;
351                 continue;
352             }
353
354             if (rf >= &render->info.h264.referenceFrames[H264_RF_COUNT])
355                 continue;
356
357             rf->surface             = render_ref->surface;
358             rf->is_long_term        = pic->long_ref;
359             rf->top_is_reference    = (pic->reference & PICT_TOP_FIELD)    ? VDP_TRUE : VDP_FALSE;
360             rf->bottom_is_reference = (pic->reference & PICT_BOTTOM_FIELD) ? VDP_TRUE : VDP_FALSE;
361             rf->field_order_cnt[0]  = pic->field_poc[0];
362             rf->field_order_cnt[1]  = pic->field_poc[1];
363             rf->frame_idx           = pic_frame_idx;
364
365             ++rf;
366         }
367     }
368
369     for (; rf < &render->info.h264.referenceFrames[H264_RF_COUNT]; ++rf) {
370         rf->surface             = VDP_INVALID_HANDLE;
371         rf->is_long_term        = 0;
372         rf->top_is_reference    = 0;
373         rf->bottom_is_reference = 0;
374         rf->field_order_cnt[0]  = 0;
375         rf->field_order_cnt[1]  = 0;
376         rf->frame_idx           = 0;
377     }
378 }
379
380 void ff_vdpau_add_data_chunk(uint8_t *data, const uint8_t *buf, int buf_size)
381 {
382     struct vdpau_render_state *render = (struct vdpau_render_state*)data;
383     assert(render);
384
385     render->bitstream_buffers= av_fast_realloc(
386         render->bitstream_buffers,
387         &render->bitstream_buffers_allocated,
388         sizeof(*render->bitstream_buffers)*(render->bitstream_buffers_used + 1)
389     );
390
391     render->bitstream_buffers[render->bitstream_buffers_used].struct_version  = VDP_BITSTREAM_BUFFER_VERSION;
392     render->bitstream_buffers[render->bitstream_buffers_used].bitstream       = buf;
393     render->bitstream_buffers[render->bitstream_buffers_used].bitstream_bytes = buf_size;
394     render->bitstream_buffers_used++;
395 }
396
397 #if CONFIG_H264_VDPAU_DECODER
398 void ff_vdpau_h264_picture_start(H264Context *h)
399 {
400     struct vdpau_render_state *render;
401     int i;
402
403     render = (struct vdpau_render_state *)h->cur_pic_ptr->f.data[0];
404     assert(render);
405
406     for (i = 0; i < 2; ++i) {
407         int foc = h->cur_pic_ptr->field_poc[i];
408         if (foc == INT_MAX)
409             foc = 0;
410         render->info.h264.field_order_cnt[i] = foc;
411     }
412
413     render->info.h264.frame_num = h->frame_num;
414 }
415
416 void ff_vdpau_h264_picture_complete(H264Context *h)
417 {
418     struct vdpau_render_state *render;
419
420     render = (struct vdpau_render_state *)h->cur_pic_ptr->f.data[0];
421     assert(render);
422
423     render->info.h264.slice_count = h->slice_num;
424     if (render->info.h264.slice_count < 1)
425         return;
426
427     render->info.h264.is_reference                           = (h->cur_pic_ptr->reference & 3) ? VDP_TRUE : VDP_FALSE;
428     render->info.h264.field_pic_flag                         = h->picture_structure != PICT_FRAME;
429     render->info.h264.bottom_field_flag                      = h->picture_structure == PICT_BOTTOM_FIELD;
430     render->info.h264.num_ref_frames                         = h->sps.ref_frame_count;
431     render->info.h264.mb_adaptive_frame_field_flag           = h->sps.mb_aff && !render->info.h264.field_pic_flag;
432     render->info.h264.constrained_intra_pred_flag            = h->pps.constrained_intra_pred;
433     render->info.h264.weighted_pred_flag                     = h->pps.weighted_pred;
434     render->info.h264.weighted_bipred_idc                    = h->pps.weighted_bipred_idc;
435     render->info.h264.frame_mbs_only_flag                    = h->sps.frame_mbs_only_flag;
436     render->info.h264.transform_8x8_mode_flag                = h->pps.transform_8x8_mode;
437     render->info.h264.chroma_qp_index_offset                 = h->pps.chroma_qp_index_offset[0];
438     render->info.h264.second_chroma_qp_index_offset          = h->pps.chroma_qp_index_offset[1];
439     render->info.h264.pic_init_qp_minus26                    = h->pps.init_qp - 26;
440     render->info.h264.num_ref_idx_l0_active_minus1           = h->pps.ref_count[0] - 1;
441     render->info.h264.num_ref_idx_l1_active_minus1           = h->pps.ref_count[1] - 1;
442     render->info.h264.log2_max_frame_num_minus4              = h->sps.log2_max_frame_num - 4;
443     render->info.h264.pic_order_cnt_type                     = h->sps.poc_type;
444     render->info.h264.log2_max_pic_order_cnt_lsb_minus4      = h->sps.poc_type ? 0 : h->sps.log2_max_poc_lsb - 4;
445     render->info.h264.delta_pic_order_always_zero_flag       = h->sps.delta_pic_order_always_zero_flag;
446     render->info.h264.direct_8x8_inference_flag              = h->sps.direct_8x8_inference_flag;
447     render->info.h264.entropy_coding_mode_flag               = h->pps.cabac;
448     render->info.h264.pic_order_present_flag                 = h->pps.pic_order_present;
449     render->info.h264.deblocking_filter_control_present_flag = h->pps.deblocking_filter_parameters_present;
450     render->info.h264.redundant_pic_cnt_present_flag         = h->pps.redundant_pic_cnt_present;
451     memcpy(render->info.h264.scaling_lists_4x4, h->pps.scaling_matrix4, sizeof(render->info.h264.scaling_lists_4x4));
452     memcpy(render->info.h264.scaling_lists_8x8[0], h->pps.scaling_matrix8[0], sizeof(render->info.h264.scaling_lists_8x8[0]));
453     memcpy(render->info.h264.scaling_lists_8x8[1], h->pps.scaling_matrix8[3], sizeof(render->info.h264.scaling_lists_8x8[0]));
454
455     ff_h264_draw_horiz_band(h, 0, h->avctx->height);
456     render->bitstream_buffers_used = 0;
457 }
458 #endif /* CONFIG_H264_VDPAU_DECODER */
459
460 #if CONFIG_MPEG_VDPAU_DECODER || CONFIG_MPEG1_VDPAU_DECODER
461 void ff_vdpau_mpeg_picture_complete(MpegEncContext *s, const uint8_t *buf,
462                                     int buf_size, int slice_count)
463 {
464     struct vdpau_render_state *render, *last, *next;
465     int i;
466
467     if (!s->current_picture_ptr) return;
468
469     render = (struct vdpau_render_state *)s->current_picture_ptr->f->data[0];
470     assert(render);
471
472     /* fill VdpPictureInfoMPEG1Or2 struct */
473     render->info.mpeg.picture_structure          = s->picture_structure;
474     render->info.mpeg.picture_coding_type        = s->pict_type;
475     render->info.mpeg.intra_dc_precision         = s->intra_dc_precision;
476     render->info.mpeg.frame_pred_frame_dct       = s->frame_pred_frame_dct;
477     render->info.mpeg.concealment_motion_vectors = s->concealment_motion_vectors;
478     render->info.mpeg.intra_vlc_format           = s->intra_vlc_format;
479     render->info.mpeg.alternate_scan             = s->alternate_scan;
480     render->info.mpeg.q_scale_type               = s->q_scale_type;
481     render->info.mpeg.top_field_first            = s->top_field_first;
482     render->info.mpeg.full_pel_forward_vector    = s->full_pel[0]; // MPEG-1 only.  Set 0 for MPEG-2
483     render->info.mpeg.full_pel_backward_vector   = s->full_pel[1]; // MPEG-1 only.  Set 0 for MPEG-2
484     render->info.mpeg.f_code[0][0]               = s->mpeg_f_code[0][0]; // For MPEG-1 fill both horiz. & vert.
485     render->info.mpeg.f_code[0][1]               = s->mpeg_f_code[0][1];
486     render->info.mpeg.f_code[1][0]               = s->mpeg_f_code[1][0];
487     render->info.mpeg.f_code[1][1]               = s->mpeg_f_code[1][1];
488     for (i = 0; i < 64; ++i) {
489         render->info.mpeg.intra_quantizer_matrix[i]     = s->intra_matrix[i];
490         render->info.mpeg.non_intra_quantizer_matrix[i] = s->inter_matrix[i];
491     }
492
493     render->info.mpeg.forward_reference          = VDP_INVALID_HANDLE;
494     render->info.mpeg.backward_reference         = VDP_INVALID_HANDLE;
495
496     switch(s->pict_type){
497     case  AV_PICTURE_TYPE_B:
498         next = (struct vdpau_render_state *)s->next_picture.f->data[0];
499         assert(next);
500         render->info.mpeg.backward_reference     = next->surface;
501         // no return here, going to set forward prediction
502     case  AV_PICTURE_TYPE_P:
503         last = (struct vdpau_render_state *)s->last_picture.f->data[0];
504         if (!last) // FIXME: Does this test make sense?
505             last = render; // predict second field from the first
506         render->info.mpeg.forward_reference      = last->surface;
507     }
508
509     ff_vdpau_add_data_chunk(s->current_picture_ptr->f->data[0], buf, buf_size);
510
511     render->info.mpeg.slice_count                = slice_count;
512
513     if (slice_count)
514         ff_mpeg_draw_horiz_band(s, 0, s->avctx->height);
515     render->bitstream_buffers_used               = 0;
516 }
517 #endif /* CONFIG_MPEG_VDPAU_DECODER || CONFIG_MPEG1_VDPAU_DECODER */
518
519 #if CONFIG_VC1_VDPAU_DECODER
520 void ff_vdpau_vc1_decode_picture(MpegEncContext *s, const uint8_t *buf,
521                                  int buf_size)
522 {
523     VC1Context *v = s->avctx->priv_data;
524     struct vdpau_render_state *render, *last, *next;
525
526     render = (struct vdpau_render_state *)s->current_picture.f->data[0];
527     assert(render);
528
529     /*  fill LvPictureInfoVC1 struct */
530     render->info.vc1.frame_coding_mode  = v->fcm ? v->fcm + 1 : 0;
531     render->info.vc1.postprocflag       = v->postprocflag;
532     render->info.vc1.pulldown           = v->broadcast;
533     render->info.vc1.interlace          = v->interlace;
534     render->info.vc1.tfcntrflag         = v->tfcntrflag;
535     render->info.vc1.finterpflag        = v->finterpflag;
536     render->info.vc1.psf                = v->psf;
537     render->info.vc1.dquant             = v->dquant;
538     render->info.vc1.panscan_flag       = v->panscanflag;
539     render->info.vc1.refdist_flag       = v->refdist_flag;
540     render->info.vc1.quantizer          = v->quantizer_mode;
541     render->info.vc1.extended_mv        = v->extended_mv;
542     render->info.vc1.extended_dmv       = v->extended_dmv;
543     render->info.vc1.overlap            = v->overlap;
544     render->info.vc1.vstransform        = v->vstransform;
545     render->info.vc1.loopfilter         = v->s.loop_filter;
546     render->info.vc1.fastuvmc           = v->fastuvmc;
547     render->info.vc1.range_mapy_flag    = v->range_mapy_flag;
548     render->info.vc1.range_mapy         = v->range_mapy;
549     render->info.vc1.range_mapuv_flag   = v->range_mapuv_flag;
550     render->info.vc1.range_mapuv        = v->range_mapuv;
551     /* Specific to simple/main profile only */
552     render->info.vc1.multires           = v->multires;
553     render->info.vc1.syncmarker         = v->resync_marker;
554     render->info.vc1.rangered           = v->rangered | (v->rangeredfrm << 1);
555     render->info.vc1.maxbframes         = v->s.max_b_frames;
556
557     render->info.vc1.deblockEnable      = v->postprocflag & 1;
558     render->info.vc1.pquant             = v->pq;
559
560     render->info.vc1.forward_reference  = VDP_INVALID_HANDLE;
561     render->info.vc1.backward_reference = VDP_INVALID_HANDLE;
562
563     if (v->bi_type)
564         render->info.vc1.picture_type = 4;
565     else
566         render->info.vc1.picture_type = s->pict_type - 1 + s->pict_type / 3;
567
568     switch(s->pict_type){
569     case  AV_PICTURE_TYPE_B:
570         next = (struct vdpau_render_state *)s->next_picture.f->data[0];
571         assert(next);
572         render->info.vc1.backward_reference = next->surface;
573         // no break here, going to set forward prediction
574     case  AV_PICTURE_TYPE_P:
575         last = (struct vdpau_render_state *)s->last_picture.f->data[0];
576         if (!last) // FIXME: Does this test make sense?
577             last = render; // predict second field from the first
578         render->info.vc1.forward_reference = last->surface;
579     }
580
581     ff_vdpau_add_data_chunk(s->current_picture_ptr->f->data[0], buf, buf_size);
582
583     render->info.vc1.slice_count          = 1;
584
585     ff_mpeg_draw_horiz_band(s, 0, s->avctx->height);
586     render->bitstream_buffers_used        = 0;
587 }
588 #endif /* (CONFIG_VC1_VDPAU_DECODER */
589
590 #if CONFIG_MPEG4_VDPAU_DECODER
591 void ff_vdpau_mpeg4_decode_picture(Mpeg4DecContext *ctx, const uint8_t *buf,
592                                    int buf_size)
593 {
594     MpegEncContext *s = &ctx->m;
595     struct vdpau_render_state *render, *last, *next;
596     int i;
597
598     if (!s->current_picture_ptr) return;
599
600     render = (struct vdpau_render_state *)s->current_picture_ptr->f->data[0];
601     assert(render);
602
603     /* fill VdpPictureInfoMPEG4Part2 struct */
604     render->info.mpeg4.trd[0]                            = s->pp_time;
605     render->info.mpeg4.trb[0]                            = s->pb_time;
606     render->info.mpeg4.trd[1]                            = s->pp_field_time >> 1;
607     render->info.mpeg4.trb[1]                            = s->pb_field_time >> 1;
608     render->info.mpeg4.vop_time_increment_resolution     = s->avctx->time_base.den;
609     render->info.mpeg4.vop_coding_type                   = 0;
610     render->info.mpeg4.vop_fcode_forward                 = s->f_code;
611     render->info.mpeg4.vop_fcode_backward                = s->b_code;
612     render->info.mpeg4.resync_marker_disable             = !ctx->resync_marker;
613     render->info.mpeg4.interlaced                        = !s->progressive_sequence;
614     render->info.mpeg4.quant_type                        = s->mpeg_quant;
615     render->info.mpeg4.quarter_sample                    = s->quarter_sample;
616     render->info.mpeg4.short_video_header                = s->avctx->codec->id == AV_CODEC_ID_H263;
617     render->info.mpeg4.rounding_control                  = s->no_rounding;
618     render->info.mpeg4.alternate_vertical_scan_flag      = s->alternate_scan;
619     render->info.mpeg4.top_field_first                   = s->top_field_first;
620     for (i = 0; i < 64; ++i) {
621         render->info.mpeg4.intra_quantizer_matrix[i]     = s->intra_matrix[i];
622         render->info.mpeg4.non_intra_quantizer_matrix[i] = s->inter_matrix[i];
623     }
624     render->info.mpeg4.forward_reference                 = VDP_INVALID_HANDLE;
625     render->info.mpeg4.backward_reference                = VDP_INVALID_HANDLE;
626
627     switch (s->pict_type) {
628     case AV_PICTURE_TYPE_B:
629         next = (struct vdpau_render_state *)s->next_picture.f->data[0];
630         assert(next);
631         render->info.mpeg4.backward_reference     = next->surface;
632         render->info.mpeg4.vop_coding_type        = 2;
633         // no break here, going to set forward prediction
634     case AV_PICTURE_TYPE_P:
635         last = (struct vdpau_render_state *)s->last_picture.f->data[0];
636         assert(last);
637         render->info.mpeg4.forward_reference      = last->surface;
638     }
639
640     ff_vdpau_add_data_chunk(s->current_picture_ptr->f->data[0], buf, buf_size);
641
642     ff_mpeg_draw_horiz_band(s, 0, s->avctx->height);
643     render->bitstream_buffers_used = 0;
644 }
645 #endif /* CONFIG_MPEG4_VDPAU_DECODER */
646
647 int av_vdpau_get_profile(AVCodecContext *avctx, VdpDecoderProfile *profile)
648 {
649 #define PROFILE(prof)                      \
650 do {                                       \
651     *profile = VDP_DECODER_PROFILE_##prof; \
652     return 0;                              \
653 } while (0)
654
655     switch (avctx->codec_id) {
656     case AV_CODEC_ID_MPEG1VIDEO:               PROFILE(MPEG1);
657     case AV_CODEC_ID_MPEG2VIDEO:
658         switch (avctx->profile) {
659         case FF_PROFILE_MPEG2_MAIN:            PROFILE(MPEG2_MAIN);
660         case FF_PROFILE_MPEG2_SIMPLE:          PROFILE(MPEG2_SIMPLE);
661         default:                               return AVERROR(EINVAL);
662         }
663     case AV_CODEC_ID_H263:                     PROFILE(MPEG4_PART2_ASP);
664     case AV_CODEC_ID_MPEG4:
665         switch (avctx->profile) {
666         case FF_PROFILE_MPEG4_SIMPLE:          PROFILE(MPEG4_PART2_SP);
667         case FF_PROFILE_MPEG4_ADVANCED_SIMPLE: PROFILE(MPEG4_PART2_ASP);
668         default:                               return AVERROR(EINVAL);
669         }
670     case AV_CODEC_ID_H264:
671         switch (avctx->profile & ~FF_PROFILE_H264_INTRA) {
672         case FF_PROFILE_H264_BASELINE:         PROFILE(H264_BASELINE);
673         case FF_PROFILE_H264_CONSTRAINED_BASELINE:
674         case FF_PROFILE_H264_MAIN:             PROFILE(H264_MAIN);
675         case FF_PROFILE_H264_HIGH:             PROFILE(H264_HIGH);
676 #ifdef VDP_DECODER_PROFILE_H264_EXTENDED
677         case FF_PROFILE_H264_EXTENDED:         PROFILE(H264_EXTENDED);
678 #endif
679         default:                               return AVERROR(EINVAL);
680         }
681     case AV_CODEC_ID_WMV3:
682     case AV_CODEC_ID_VC1:
683         switch (avctx->profile) {
684         case FF_PROFILE_VC1_SIMPLE:            PROFILE(VC1_SIMPLE);
685         case FF_PROFILE_VC1_MAIN:              PROFILE(VC1_MAIN);
686         case FF_PROFILE_VC1_ADVANCED:          PROFILE(VC1_ADVANCED);
687         default:                               return AVERROR(EINVAL);
688         }
689     }
690     return AVERROR(EINVAL);
691 #undef PROFILE
692 }
693
694 AVVDPAUContext *av_vdpau_alloc_context(void)
695 {
696     return av_mallocz(sizeof(AVVDPAUContext));
697 }
698
699 int av_vdpau_bind_context(AVCodecContext *avctx, VdpDevice device,
700                           VdpGetProcAddress *get_proc, unsigned flags)
701 {
702     VDPAUHWContext *hwctx;
703
704     if (flags & ~AV_HWACCEL_FLAG_IGNORE_LEVEL)
705         return AVERROR(EINVAL);
706
707     if (av_reallocp(&avctx->hwaccel_context, sizeof(*hwctx)))
708         return AVERROR(ENOMEM);
709
710     hwctx = avctx->hwaccel_context;
711
712     memset(hwctx, 0, sizeof(*hwctx));
713     hwctx->context.decoder  = VDP_INVALID_HANDLE;
714     hwctx->device           = device;
715     hwctx->get_proc_address = get_proc;
716     hwctx->flags            = flags;
717     hwctx->reset            = 1;
718     return 0;
719 }
720
721 /* @}*/