1 @chapter Filtering Introduction
2 @c man begin FILTERING INTRODUCTION
4 Filtering in FFmpeg is enabled through the libavfilter library.
6 Libavfilter is the filtering API of FFmpeg. It is the substitute of
7 the now deprecated 'vhooks' and started as a Google Summer of Code
10 Audio filtering integration into the main FFmpeg repository is a work in
11 progress, so audio API and ABI should not be considered stable yet.
13 In libavfilter, it is possible for filters to have multiple inputs and
15 To illustrate the sorts of things that are possible, we can
16 use a complex filter graph. For example, the following one:
19 input --> split --> fifo -----------------------> overlay --> output
22 +------> fifo --> crop --> vflip --------+
25 splits the stream in two streams, sends one stream through the crop filter
26 and the vflip filter before merging it back with the other stream by
27 overlaying it on top. You can use the following command to achieve this:
30 ffmpeg -i input -vf "[in] split [T1], fifo, [T2] overlay=0:H/2 [out]; [T1] fifo, crop=iw:ih/2:0:ih/2, vflip [T2]" output
33 The result will be that in output the top half of the video is mirrored
36 Filters are loaded using the @var{-vf} or @var{-af} option passed to
37 @command{ffmpeg} or to @command{ffplay}. Filters in the same linear
38 chain are separated by commas. In our example, @var{split, fifo,
39 overlay} are in one linear chain, and @var{fifo, crop, vflip} are in
40 another. The points where the linear chains join are labeled by names
41 enclosed in square brackets. In our example, that is @var{[T1]} and
42 @var{[T2]}. The special labels @var{[in]} and @var{[out]} are the points
43 where video is input and output.
45 Some filters take in input a list of parameters: they are specified
46 after the filter name and an equal sign, and are separated from each other
49 There exist so-called @var{source filters} that do not have an
50 audio/video input, and @var{sink filters} that will not have audio/video
53 @c man end FILTERING INTRODUCTION
56 @c man begin GRAPH2DOT
58 The @file{graph2dot} program included in the FFmpeg @file{tools}
59 directory can be used to parse a filter graph description and issue a
60 corresponding textual representation in the dot language.
67 to see how to use @file{graph2dot}.
69 You can then pass the dot description to the @file{dot} program (from
70 the graphviz suite of programs) and obtain a graphical representation
73 For example the sequence of commands:
75 echo @var{GRAPH_DESCRIPTION} | \
76 tools/graph2dot -o graph.tmp && \
77 dot -Tpng graph.tmp -o graph.png && \
81 can be used to create and display an image representing the graph
82 described by the @var{GRAPH_DESCRIPTION} string.
86 @chapter Filtergraph description
87 @c man begin FILTERGRAPH DESCRIPTION
89 A filtergraph is a directed graph of connected filters. It can contain
90 cycles, and there can be multiple links between a pair of
91 filters. Each link has one input pad on one side connecting it to one
92 filter from which it takes its input, and one output pad on the other
93 side connecting it to the one filter accepting its output.
95 Each filter in a filtergraph is an instance of a filter class
96 registered in the application, which defines the features and the
97 number of input and output pads of the filter.
99 A filter with no input pads is called a "source", a filter with no
100 output pads is called a "sink".
102 @anchor{Filtergraph syntax}
103 @section Filtergraph syntax
105 A filtergraph can be represented using a textual representation, which is
106 recognized by the @option{-filter}/@option{-vf} and @option{-filter_complex}
107 options in @command{ffmpeg} and @option{-vf} in @command{ffplay}, and by the
108 @code{avfilter_graph_parse()}/@code{avfilter_graph_parse2()} function defined in
109 @file{libavfilter/avfiltergraph.h}.
111 A filterchain consists of a sequence of connected filters, each one
112 connected to the previous one in the sequence. A filterchain is
113 represented by a list of ","-separated filter descriptions.
115 A filtergraph consists of a sequence of filterchains. A sequence of
116 filterchains is represented by a list of ";"-separated filterchain
119 A filter is represented by a string of the form:
120 [@var{in_link_1}]...[@var{in_link_N}]@var{filter_name}=@var{arguments}[@var{out_link_1}]...[@var{out_link_M}]
122 @var{filter_name} is the name of the filter class of which the
123 described filter is an instance of, and has to be the name of one of
124 the filter classes registered in the program.
125 The name of the filter class is optionally followed by a string
128 @var{arguments} is a string which contains the parameters used to
129 initialize the filter instance, and are described in the filter
132 The list of arguments can be quoted using the character "'" as initial
133 and ending mark, and the character '\' for escaping the characters
134 within the quoted text; otherwise the argument string is considered
135 terminated when the next special character (belonging to the set
136 "[]=;,") is encountered.
138 The name and arguments of the filter are optionally preceded and
139 followed by a list of link labels.
140 A link label allows to name a link and associate it to a filter output
141 or input pad. The preceding labels @var{in_link_1}
142 ... @var{in_link_N}, are associated to the filter input pads,
143 the following labels @var{out_link_1} ... @var{out_link_M}, are
144 associated to the output pads.
146 When two link labels with the same name are found in the
147 filtergraph, a link between the corresponding input and output pad is
150 If an output pad is not labelled, it is linked by default to the first
151 unlabelled input pad of the next filter in the filterchain.
152 For example in the filterchain:
154 nullsrc, split[L1], [L2]overlay, nullsink
156 the split filter instance has two output pads, and the overlay filter
157 instance two input pads. The first output pad of split is labelled
158 "L1", the first input pad of overlay is labelled "L2", and the second
159 output pad of split is linked to the second input pad of overlay,
160 which are both unlabelled.
162 In a complete filterchain all the unlabelled filter input and output
163 pads must be connected. A filtergraph is considered valid if all the
164 filter input and output pads of all the filterchains are connected.
166 Libavfilter will automatically insert scale filters where format
167 conversion is required. It is possible to specify swscale flags
168 for those automatically inserted scalers by prepending
169 @code{sws_flags=@var{flags};}
170 to the filtergraph description.
172 Follows a BNF description for the filtergraph syntax:
174 @var{NAME} ::= sequence of alphanumeric characters and '_'
175 @var{LINKLABEL} ::= "[" @var{NAME} "]"
176 @var{LINKLABELS} ::= @var{LINKLABEL} [@var{LINKLABELS}]
177 @var{FILTER_ARGUMENTS} ::= sequence of chars (eventually quoted)
178 @var{FILTER} ::= [@var{LINKNAMES}] @var{NAME} ["=" @var{ARGUMENTS}] [@var{LINKNAMES}]
179 @var{FILTERCHAIN} ::= @var{FILTER} [,@var{FILTERCHAIN}]
180 @var{FILTERGRAPH} ::= [sws_flags=@var{flags};] @var{FILTERCHAIN} [;@var{FILTERGRAPH}]
183 @c man end FILTERGRAPH DESCRIPTION
185 @chapter Audio Filters
186 @c man begin AUDIO FILTERS
188 When you configure your FFmpeg build, you can disable any of the
189 existing filters using @code{--disable-filters}.
190 The configure output will show the audio filters included in your
193 Below is a description of the currently available audio filters.
197 Convert the input audio format to the specified formats.
199 The filter accepts a string of the form:
200 "@var{sample_format}:@var{channel_layout}".
202 @var{sample_format} specifies the sample format, and can be a string or the
203 corresponding numeric value defined in @file{libavutil/samplefmt.h}. Use 'p'
204 suffix for a planar sample format.
206 @var{channel_layout} specifies the channel layout, and can be a string
207 or the corresponding number value defined in @file{libavutil/audioconvert.h}.
209 The special parameter "auto", signifies that the filter will
210 automatically select the output format depending on the output filter.
212 Some examples follow.
216 Convert input to float, planar, stereo:
222 Convert input to unsigned 8-bit, automatically select out channel layout:
230 Convert the input audio to one of the specified formats. The framework will
231 negotiate the most appropriate format to minimize conversions.
233 The filter accepts the following named parameters:
237 A comma-separated list of requested sample formats.
240 A comma-separated list of requested sample rates.
242 @item channel_layouts
243 A comma-separated list of requested channel layouts.
247 If a parameter is omitted, all values are allowed.
249 For example to force the output to either unsigned 8-bit or signed 16-bit stereo:
251 aformat=sample_fmts\=u8\,s16:channel_layouts\=stereo
256 Merge two or more audio streams into a single multi-channel stream.
258 The filter accepts the following named options:
263 Set the number of inputs. Default is 2.
267 If the channel layouts of the inputs are disjoint, and therefore compatible,
268 the channel layout of the output will be set accordingly and the channels
269 will be reordered as necessary. If the channel layouts of the inputs are not
270 disjoint, the output will have all the channels of the first input then all
271 the channels of the second input, in that order, and the channel layout of
272 the output will be the default value corresponding to the total number of
275 For example, if the first input is in 2.1 (FL+FR+LF) and the second input
276 is FC+BL+BR, then the output will be in 5.1, with the channels in the
277 following order: a1, a2, b1, a3, b2, b3 (a1 is the first channel of the
278 first input, b1 is the first channel of the second input).
280 On the other hand, if both input are in stereo, the output channels will be
281 in the default order: a1, a2, b1, b2, and the channel layout will be
282 arbitrarily set to 4.0, which may or may not be the expected value.
284 All inputs must have the same sample rate, and format.
286 If inputs do not have the same duration, the output will stop with the
289 Example: merge two mono files into a stereo stream:
291 amovie=left.wav [l] ; amovie=right.mp3 [r] ; [l] [r] amerge
294 Example: multiple merges:
297 amovie=input.mkv:si=0 [a0];
298 amovie=input.mkv:si=1 [a1];
299 amovie=input.mkv:si=2 [a2];
300 amovie=input.mkv:si=3 [a3];
301 amovie=input.mkv:si=4 [a4];
302 amovie=input.mkv:si=5 [a5];
303 [a0][a1][a2][a3][a4][a5] amerge=inputs=6" -c:a pcm_s16le output.mkv
308 Mixes multiple audio inputs into a single output.
312 ffmpeg -i INPUT1 -i INPUT2 -i INPUT3 -filter_complex amix=inputs=3:duration=first:dropout_transition=3 OUTPUT
314 will mix 3 input audio streams to a single output with the same duration as the
315 first input and a dropout transition time of 3 seconds.
317 The filter accepts the following named parameters:
321 Number of inputs. If unspecified, it defaults to 2.
324 How to determine the end-of-stream.
328 Duration of longest input. (default)
331 Duration of shortest input.
334 Duration of first input.
338 @item dropout_transition
339 Transition time, in seconds, for volume renormalization when an input
340 stream ends. The default value is 2 seconds.
346 Pass the audio source unchanged to the output.
350 Resample the input audio to the specified sample rate.
352 The filter accepts exactly one parameter, the output sample rate. If not
353 specified then the filter will automatically convert between its input
354 and output sample rates.
356 For example, to resample the input audio to 44100Hz:
361 @section asetnsamples
363 Set the number of samples per each output audio frame.
365 The last output packet may contain a different number of samples, as
366 the filter will flush all the remaining samples when the input audio
369 The filter accepts parameters as a list of @var{key}=@var{value} pairs,
374 @item nb_out_samples, n
375 Set the number of frames per each output audio frame. The number is
376 intended as the number of samples @emph{per each channel}.
377 Default value is 1024.
380 If set to 1, the filter will pad the last audio frame with zeroes, so
381 that the last frame will contain the same number of samples as the
382 previous ones. Default value is 1.
385 For example, to set the number of per-frame samples to 1234 and
386 disable padding for the last frame, use:
388 asetnsamples=n=1234:p=0
393 Show a line containing various information for each input audio frame.
394 The input audio is not modified.
396 The shown line contains a sequence of key/value pairs of the form
397 @var{key}:@var{value}.
399 A description of each shown parameter follows:
403 sequential number of the input frame, starting from 0
406 presentation TimeStamp of the input frame, expressed as a number of
407 time base units. The time base unit depends on the filter input pad, and
408 is usually 1/@var{sample_rate}.
411 presentation TimeStamp of the input frame, expressed as a number of
415 position of the frame in the input stream, -1 if this information in
416 unavailable and/or meaningless (for example in case of synthetic audio)
422 channel layout description
425 number of samples (per each channel) contained in the filtered frame
428 sample rate for the audio frame
431 Adler-32 checksum (printed in hexadecimal) of all the planes of the input frame
434 Adler-32 checksum (printed in hexadecimal) for each input frame plane,
435 expressed in the form "[@var{c0} @var{c1} @var{c2} @var{c3} @var{c4} @var{c5}
441 Split input audio into several identical outputs.
443 The filter accepts a single parameter which specifies the number of outputs. If
444 unspecified, it defaults to 2.
448 [in] asplit [out0][out1]
451 will create two separate outputs from the same input.
453 To create 3 or more outputs, you need to specify the number of
456 [in] asplit=3 [out0][out1][out2]
460 ffmpeg -i INPUT -filter_complex asplit=5 OUTPUT
462 will create 5 copies of the input audio.
467 Forward two audio streams and control the order the buffers are forwarded.
469 The argument to the filter is an expression deciding which stream should be
470 forwarded next: if the result is negative, the first stream is forwarded; if
471 the result is positive or zero, the second stream is forwarded. It can use
472 the following variables:
476 number of buffers forwarded so far on each stream
478 number of samples forwarded so far on each stream
480 current timestamp of each stream
483 The default value is @code{t1-t2}, which means to always forward the stream
484 that has a smaller timestamp.
486 Example: stress-test @code{amerge} by randomly sending buffers on the wrong
487 input, while avoiding too much of a desynchronization:
489 amovie=file.ogg [a] ; amovie=file.mp3 [b] ;
490 [a] [b] astreamsync=(2*random(1))-1+tanh(5*(t1-t2)) [a2] [b2] ;
498 The filter accepts exactly one parameter, the audio tempo. If not
499 specified then the filter will assume nominal 1.0 tempo. Tempo must
500 be in the [0.5, 2.0] range.
502 For example, to slow down audio to 80% tempo:
507 For example, to speed up audio to 125% tempo:
514 Make audio easier to listen to on headphones.
516 This filter adds `cues' to 44.1kHz stereo (i.e. audio CD format) audio
517 so that when listened to on headphones the stereo image is moved from
518 inside your head (standard for headphones) to outside and in front of
519 the listener (standard for speakers).
525 Mix channels with specific gain levels. The filter accepts the output
526 channel layout followed by a set of channels definitions.
528 This filter is also designed to remap efficiently the channels of an audio
531 The filter accepts parameters of the form:
532 "@var{l}:@var{outdef}:@var{outdef}:..."
536 output channel layout or number of channels
539 output channel specification, of the form:
540 "@var{out_name}=[@var{gain}*]@var{in_name}[+[@var{gain}*]@var{in_name}...]"
543 output channel to define, either a channel name (FL, FR, etc.) or a channel
544 number (c0, c1, etc.)
547 multiplicative coefficient for the channel, 1 leaving the volume unchanged
550 input channel to use, see out_name for details; it is not possible to mix
551 named and numbered input channels
554 If the `=' in a channel specification is replaced by `<', then the gains for
555 that specification will be renormalized so that the total is 1, thus
556 avoiding clipping noise.
558 @subsection Mixing examples
560 For example, if you want to down-mix from stereo to mono, but with a bigger
561 factor for the left channel:
563 pan=1:c0=0.9*c0+0.1*c1
566 A customized down-mix to stereo that works automatically for 3-, 4-, 5- and
569 pan=stereo: FL < FL + 0.5*FC + 0.6*BL + 0.6*SL : FR < FR + 0.5*FC + 0.6*BR + 0.6*SR
572 Note that @command{ffmpeg} integrates a default down-mix (and up-mix) system
573 that should be preferred (see "-ac" option) unless you have very specific
576 @subsection Remapping examples
578 The channel remapping will be effective if, and only if:
581 @item gain coefficients are zeroes or ones,
582 @item only one input per channel output,
585 If all these conditions are satisfied, the filter will notify the user ("Pure
586 channel mapping detected"), and use an optimized and lossless method to do the
589 For example, if you have a 5.1 source and want a stereo audio stream by
590 dropping the extra channels:
592 pan="stereo: c0=FL : c1=FR"
595 Given the same source, you can also switch front left and front right channels
596 and keep the input channel layout:
598 pan="5.1: c0=c1 : c1=c0 : c2=c2 : c3=c3 : c4=c4 : c5=c5"
601 If the input is a stereo audio stream, you can mute the front left channel (and
602 still keep the stereo channel layout) with:
607 Still with a stereo audio stream input, you can copy the right channel in both
608 front left and right:
610 pan="stereo: c0=FR : c1=FR"
613 @section silencedetect
615 Detect silence in an audio stream.
617 This filter logs a message when it detects that the input audio volume is less
618 or equal to a noise tolerance value for a duration greater or equal to the
619 minimum detected noise duration.
621 The printed times and duration are expressed in seconds.
625 Set silence duration until notification (default is 2 seconds).
628 Set noise tolerance. Can be specified in dB (in case "dB" is appended to the
629 specified value) or amplitude ratio. Default is -60dB, or 0.001.
632 Detect 5 seconds of silence with -50dB noise tolerance:
634 silencedetect=n=-50dB:d=5
637 Complete example with @command{ffmpeg} to detect silence with 0.0001 noise
638 tolerance in @file{silence.mp3}:
640 ffmpeg -f lavfi -i amovie=silence.mp3,silencedetect=noise=0.0001 -f null -
645 Adjust the input audio volume.
647 The filter accepts exactly one parameter @var{vol}, which expresses
648 how the audio volume will be increased or decreased.
650 Output values are clipped to the maximum value.
652 If @var{vol} is expressed as a decimal number, the output audio
653 volume is given by the relation:
655 @var{output_volume} = @var{vol} * @var{input_volume}
658 If @var{vol} is expressed as a decimal number followed by the string
659 "dB", the value represents the requested change in decibels of the
660 input audio power, and the output audio volume is given by the
663 @var{output_volume} = 10^(@var{vol}/20) * @var{input_volume}
666 Otherwise @var{vol} is considered an expression and its evaluated
667 value is used for computing the output audio volume according to the
670 Default value for @var{vol} is 1.0.
676 Half the input audio volume:
681 The above example is equivalent to:
687 Decrease input audio power by 12 decibels:
694 Synchronize audio data with timestamps by squeezing/stretching it and/or
695 dropping samples/adding silence when needed.
697 The filter accepts the following named parameters:
701 Enable stretching/squeezing the data to make it match the timestamps.
704 Minimum difference between timestamps and audio data (in seconds) to trigger
705 adding/dropping samples.
708 Maximum compensation in samples per second.
711 Assume the first pts should be this value.
712 This allows for padding/trimming at the start of stream. By default, no
713 assumption is made about the first frame's expected pts, so no padding or
714 trimming is done. For example, this could be set to 0 to pad the beginning with
715 silence if an audio stream starts after the video stream.
719 @section channelsplit
720 Split each channel in input audio stream into a separate output stream.
722 This filter accepts the following named parameters:
725 Channel layout of the input stream. Default is "stereo".
728 For example, assuming a stereo input MP3 file
730 ffmpeg -i in.mp3 -filter_complex channelsplit out.mkv
732 will create an output Matroska file with two audio streams, one containing only
733 the left channel and the other the right channel.
735 To split a 5.1 WAV file into per-channel files
737 ffmpeg -i in.wav -filter_complex
738 'channelsplit=channel_layout=5.1[FL][FR][FC][LFE][SL][SR]'
739 -map '[FL]' front_left.wav -map '[FR]' front_right.wav -map '[FC]'
740 front_center.wav -map '[LFE]' lfe.wav -map '[SL]' side_left.wav -map '[SR]'
745 Remap input channels to new locations.
747 This filter accepts the following named parameters:
750 Channel layout of the output stream.
753 Map channels from input to output. The argument is a comma-separated list of
754 mappings, each in the @code{@var{in_channel}-@var{out_channel}} or
755 @var{in_channel} form. @var{in_channel} can be either the name of the input
756 channel (e.g. FL for front left) or its index in the input channel layout.
757 @var{out_channel} is the name of the output channel or its index in the output
758 channel layout. If @var{out_channel} is not given then it is implicitly an
759 index, starting with zero and increasing by one for each mapping.
762 If no mapping is present, the filter will implicitly map input channels to
763 output channels preserving index.
765 For example, assuming a 5.1+downmix input MOV file
767 ffmpeg -i in.mov -filter 'channelmap=map=DL-FL\,DR-FR' out.wav
769 will create an output WAV file tagged as stereo from the downmix channels of
772 To fix a 5.1 WAV improperly encoded in AAC's native channel order
774 ffmpeg -i in.wav -filter 'channelmap=1\,2\,0\,5\,3\,4:channel_layout=5.1' out.wav
778 Join multiple input streams into one multi-channel stream.
780 The filter accepts the following named parameters:
784 Number of input streams. Defaults to 2.
787 Desired output channel layout. Defaults to stereo.
790 Map channels from inputs to output. The argument is a comma-separated list of
791 mappings, each in the @code{@var{input_idx}.@var{in_channel}-@var{out_channel}}
792 form. @var{input_idx} is the 0-based index of the input stream. @var{in_channel}
793 can be either the name of the input channel (e.g. FL for front left) or its
794 index in the specified input stream. @var{out_channel} is the name of the output
798 The filter will attempt to guess the mappings when those are not specified
799 explicitly. It does so by first trying to find an unused matching input channel
800 and if that fails it picks the first unused input channel.
802 E.g. to join 3 inputs (with properly set channel layouts)
804 ffmpeg -i INPUT1 -i INPUT2 -i INPUT3 -filter_complex join=inputs=3 OUTPUT
807 To build a 5.1 output from 6 single-channel streams:
809 ffmpeg -i fl -i fr -i fc -i sl -i sr -i lfe -filter_complex
810 'join=inputs=6:channel_layout=5.1:map=0.0-FL\,1.0-FR\,2.0-FC\,3.0-SL\,4.0-SR\,5.0-LFE'
815 Convert the audio sample format, sample rate and channel layout. This filter is
816 not meant to be used directly.
818 @c man end AUDIO FILTERS
820 @chapter Audio Sources
821 @c man begin AUDIO SOURCES
823 Below is a description of the currently available audio sources.
827 Buffer audio frames, and make them available to the filter chain.
829 This source is mainly intended for a programmatic use, in particular
830 through the interface defined in @file{libavfilter/asrc_abuffer.h}.
832 It accepts the following mandatory parameters:
833 @var{sample_rate}:@var{sample_fmt}:@var{channel_layout}
838 The sample rate of the incoming audio buffers.
841 The sample format of the incoming audio buffers.
842 Either a sample format name or its corresponging integer representation from
843 the enum AVSampleFormat in @file{libavutil/samplefmt.h}
846 The channel layout of the incoming audio buffers.
847 Either a channel layout name from channel_layout_map in
848 @file{libavutil/audioconvert.c} or its corresponding integer representation
849 from the AV_CH_LAYOUT_* macros in @file{libavutil/audioconvert.h}
855 abuffer=44100:s16p:stereo
858 will instruct the source to accept planar 16bit signed stereo at 44100Hz.
859 Since the sample format with name "s16p" corresponds to the number
860 6 and the "stereo" channel layout corresponds to the value 0x3, this is
868 Generate an audio signal specified by an expression.
870 This source accepts in input one or more expressions (one for each
871 channel), which are evaluated and used to generate a corresponding
874 It accepts the syntax: @var{exprs}[::@var{options}].
875 @var{exprs} is a list of expressions separated by ":", one for each
876 separate channel. In case the @var{channel_layout} is not
877 specified, the selected channel layout depends on the number of
878 provided expressions.
880 @var{options} is an optional sequence of @var{key}=@var{value} pairs,
883 The description of the accepted options follows.
887 @item channel_layout, c
888 Set the channel layout. The number of channels in the specified layout
889 must be equal to the number of specified expressions.
892 Set the minimum duration of the sourced audio. See the function
893 @code{av_parse_time()} for the accepted format.
894 Note that the resulting duration may be greater than the specified
895 duration, as the generated audio is always cut at the end of a
898 If not specified, or the expressed duration is negative, the audio is
899 supposed to be generated forever.
902 Set the number of samples per channel per each output frame,
906 Specify the sample rate, default to 44100.
909 Each expression in @var{exprs} can contain the following constants:
913 number of the evaluated sample, starting from 0
916 time of the evaluated sample expressed in seconds, starting from 0
935 Generate a sin signal with frequency of 440 Hz, set sample rate to
938 aevalsrc="sin(440*2*PI*t)::s=8000"
942 Generate a two channels signal, specify the channel layout (Front
943 Center + Back Center) explicitly:
945 aevalsrc="sin(420*2*PI*t):cos(430*2*PI*t)::c=FC|BC"
949 Generate white noise:
951 aevalsrc="-2+random(0)"
955 Generate an amplitude modulated signal:
957 aevalsrc="sin(10*2*PI*t)*sin(880*2*PI*t)"
961 Generate 2.5 Hz binaural beats on a 360 Hz carrier:
963 aevalsrc="0.1*sin(2*PI*(360-2.5/2)*t) : 0.1*sin(2*PI*(360+2.5/2)*t)"
970 Null audio source, return unprocessed audio frames. It is mainly useful
971 as a template and to be employed in analysis / debugging tools, or as
972 the source for filters which ignore the input data (for example the sox
975 It accepts an optional sequence of @var{key}=@var{value} pairs,
978 The description of the accepted options follows.
983 Specify the sample rate, and defaults to 44100.
985 @item channel_layout, cl
987 Specify the channel layout, and can be either an integer or a string
988 representing a channel layout. The default value of @var{channel_layout}
991 Check the channel_layout_map definition in
992 @file{libavcodec/audioconvert.c} for the mapping between strings and
993 channel layout values.
996 Set the number of samples per requested frames.
1000 Follow some examples:
1002 # set the sample rate to 48000 Hz and the channel layout to AV_CH_LAYOUT_MONO.
1003 anullsrc=r=48000:cl=4
1006 anullsrc=r=48000:cl=mono
1010 Buffer audio frames, and make them available to the filter chain.
1012 This source is not intended to be part of user-supplied graph descriptions but
1013 for insertion by calling programs through the interface defined in
1014 @file{libavfilter/buffersrc.h}.
1016 It accepts the following named parameters:
1020 Timebase which will be used for timestamps of submitted frames. It must be
1021 either a floating-point number or in @var{numerator}/@var{denominator} form.
1027 Name of the sample format, as returned by @code{av_get_sample_fmt_name()}.
1029 @item channel_layout
1030 Channel layout of the audio data, in the form that can be accepted by
1031 @code{av_get_channel_layout()}.
1034 All the parameters need to be explicitly defined.
1038 Synthesize a voice utterance using the libflite library.
1040 To enable compilation of this filter you need to configure FFmpeg with
1041 @code{--enable-libflite}.
1043 Note that the flite library is not thread-safe.
1045 The source accepts parameters as a list of @var{key}=@var{value} pairs,
1048 The description of the accepted parameters follows.
1053 If set to 1, list the names of the available voices and exit
1054 immediately. Default value is 0.
1057 Set the maximum number of samples per frame. Default value is 512.
1060 Set the filename containing the text to speak.
1063 Set the text to speak.
1066 Set the voice to use for the speech synthesis. Default value is
1067 @code{kal}. See also the @var{list_voices} option.
1074 Read from file @file{speech.txt}, and synthetize the text using the
1075 standard flite voice:
1077 flite=textfile=speech.txt
1081 Read the specified text selecting the @code{slt} voice:
1083 flite=text='So fare thee well, poor devil of a Sub-Sub, whose commentator I am':voice=slt
1087 Make @file{ffplay} speech the specified text, using @code{flite} and
1088 the @code{lavfi} device:
1090 ffplay -f lavfi flite='No more be grieved for which that thou hast done.'
1094 For more information about libflite, check:
1095 @url{http://www.speech.cs.cmu.edu/flite/}
1097 @c man end AUDIO SOURCES
1099 @chapter Audio Sinks
1100 @c man begin AUDIO SINKS
1102 Below is a description of the currently available audio sinks.
1104 @section abuffersink
1106 Buffer audio frames, and make them available to the end of filter chain.
1108 This sink is mainly intended for programmatic use, in particular
1109 through the interface defined in @file{libavfilter/buffersink.h}.
1111 It requires a pointer to an AVABufferSinkContext structure, which
1112 defines the incoming buffers' formats, to be passed as the opaque
1113 parameter to @code{avfilter_init_filter} for initialization.
1117 Null audio sink, do absolutely nothing with the input audio. It is
1118 mainly useful as a template and to be employed in analysis / debugging
1121 @section abuffersink
1122 This sink is intended for programmatic use. Frames that arrive on this sink can
1123 be retrieved by the calling program using the interface defined in
1124 @file{libavfilter/buffersink.h}.
1126 This filter accepts no parameters.
1128 @c man end AUDIO SINKS
1130 @chapter Video Filters
1131 @c man begin VIDEO FILTERS
1133 When you configure your FFmpeg build, you can disable any of the
1134 existing filters using @code{--disable-filters}.
1135 The configure output will show the video filters included in your
1138 Below is a description of the currently available video filters.
1140 @section alphaextract
1142 Extract the alpha component from the input as a grayscale video. This
1143 is especially useful with the @var{alphamerge} filter.
1147 Add or replace the alpha component of the primary input with the
1148 grayscale value of a second input. This is intended for use with
1149 @var{alphaextract} to allow the transmission or storage of frame
1150 sequences that have alpha in a format that doesn't support an alpha
1153 For example, to reconstruct full frames from a normal YUV-encoded video
1154 and a separate video created with @var{alphaextract}, you might use:
1156 movie=in_alpha.mkv [alpha]; [in][alpha] alphamerge [out]
1159 Since this filter is designed for reconstruction, it operates on frame
1160 sequences without considering timestamps, and terminates when either
1161 input reaches end of stream. This will cause problems if your encoding
1162 pipeline drops frames. If you're trying to apply an image as an
1163 overlay to a video stream, consider the @var{overlay} filter instead.
1167 Draw ASS (Advanced Substation Alpha) subtitles on top of input video
1168 using the libass library.
1170 To enable compilation of this filter you need to configure FFmpeg with
1171 @code{--enable-libass}.
1173 This filter accepts the syntax: @var{ass_filename}[:@var{options}],
1174 where @var{ass_filename} is the filename of the ASS file to read, and
1175 @var{options} is an optional sequence of @var{key}=@var{value} pairs,
1178 A description of the accepted options follows.
1182 Specifies the size of the original video, the video for which the ASS file
1183 was composed. Due to a misdesign in ASS aspect ratio arithmetic, this is
1184 necessary to correctly scale the fonts if the aspect ratio has been changed.
1187 For example, to render the file @file{sub.ass} on top of the input
1188 video, use the command:
1195 Compute the bounding box for the non-black pixels in the input frame
1198 This filter computes the bounding box containing all the pixels with a
1199 luminance value greater than the minimum allowed value.
1200 The parameters describing the bounding box are printed on the filter
1203 @section blackdetect
1205 Detect video intervals that are (almost) completely black. Can be
1206 useful to detect chapter transitions, commercials, or invalid
1207 recordings. Output lines contains the time for the start, end and
1208 duration of the detected black interval expressed in seconds.
1210 In order to display the output lines, you need to set the loglevel at
1211 least to the AV_LOG_INFO value.
1213 This filter accepts a list of options in the form of
1214 @var{key}=@var{value} pairs separated by ":". A description of the
1215 accepted options follows.
1218 @item black_min_duration, d
1219 Set the minimum detected black duration expressed in seconds. It must
1220 be a non-negative floating point number.
1222 Default value is 2.0.
1224 @item picture_black_ratio_th, pic_th
1225 Set the threshold for considering a picture "black".
1226 Express the minimum value for the ratio:
1228 @var{nb_black_pixels} / @var{nb_pixels}
1231 for which a picture is considered black.
1232 Default value is 0.98.
1234 @item pixel_black_th, pix_th
1235 Set the threshold for considering a pixel "black".
1237 The threshold expresses the maximum pixel luminance value for which a
1238 pixel is considered "black". The provided value is scaled according to
1239 the following equation:
1241 @var{absolute_threshold} = @var{luminance_minimum_value} + @var{pixel_black_th} * @var{luminance_range_size}
1244 @var{luminance_range_size} and @var{luminance_minimum_value} depend on
1245 the input video format, the range is [0-255] for YUV full-range
1246 formats and [16-235] for YUV non full-range formats.
1248 Default value is 0.10.
1251 The following example sets the maximum pixel threshold to the minimum
1252 value, and detects only black intervals of 2 or more seconds:
1254 blackdetect=d=2:pix_th=0.00
1259 Detect frames that are (almost) completely black. Can be useful to
1260 detect chapter transitions or commercials. Output lines consist of
1261 the frame number of the detected frame, the percentage of blackness,
1262 the position in the file if known or -1 and the timestamp in seconds.
1264 In order to display the output lines, you need to set the loglevel at
1265 least to the AV_LOG_INFO value.
1267 The filter accepts the syntax:
1269 blackframe[=@var{amount}:[@var{threshold}]]
1272 @var{amount} is the percentage of the pixels that have to be below the
1273 threshold, and defaults to 98.
1275 @var{threshold} is the threshold below which a pixel value is
1276 considered black, and defaults to 32.
1280 Apply boxblur algorithm to the input video.
1282 This filter accepts the parameters:
1283 @var{luma_radius}:@var{luma_power}:@var{chroma_radius}:@var{chroma_power}:@var{alpha_radius}:@var{alpha_power}
1285 Chroma and alpha parameters are optional, if not specified they default
1286 to the corresponding values set for @var{luma_radius} and
1289 @var{luma_radius}, @var{chroma_radius}, and @var{alpha_radius} represent
1290 the radius in pixels of the box used for blurring the corresponding
1291 input plane. They are expressions, and can contain the following
1295 the input width and height in pixels
1298 the input chroma image width and height in pixels
1301 horizontal and vertical chroma subsample values. For example for the
1302 pixel format "yuv422p" @var{hsub} is 2 and @var{vsub} is 1.
1305 The radius must be a non-negative number, and must not be greater than
1306 the value of the expression @code{min(w,h)/2} for the luma and alpha planes,
1307 and of @code{min(cw,ch)/2} for the chroma planes.
1309 @var{luma_power}, @var{chroma_power}, and @var{alpha_power} represent
1310 how many times the boxblur filter is applied to the corresponding
1313 Some examples follow:
1318 Apply a boxblur filter with luma, chroma, and alpha radius
1325 Set luma radius to 2, alpha and chroma radius to 0
1331 Set luma and chroma radius to a fraction of the video dimension
1333 boxblur=min(h\,w)/10:1:min(cw\,ch)/10:1
1338 @section colormatrix
1340 The colormatrix filter allows conversion between any of the following color
1341 space: BT.709 (@var{bt709}), BT.601 (@var{bt601}), SMPTE-240M (@var{smpte240m})
1342 and FCC (@var{fcc}).
1344 The syntax of the parameters is @var{source}:@var{destination}:
1347 colormatrix=bt601:smpte240m
1352 Copy the input source unchanged to the output. Mainly useful for
1357 Crop the input video to @var{out_w}:@var{out_h}:@var{x}:@var{y}:@var{keep_aspect}
1359 The @var{keep_aspect} parameter is optional, if specified and set to a
1360 non-zero value will force the output display aspect ratio to be the
1361 same of the input, by changing the output sample aspect ratio.
1363 The @var{out_w}, @var{out_h}, @var{x}, @var{y} parameters are
1364 expressions containing the following constants:
1368 the computed values for @var{x} and @var{y}. They are evaluated for
1372 the input width and height
1375 same as @var{in_w} and @var{in_h}
1378 the output (cropped) width and height
1381 same as @var{out_w} and @var{out_h}
1384 same as @var{iw} / @var{ih}
1387 input sample aspect ratio
1390 input display aspect ratio, it is the same as (@var{iw} / @var{ih}) * @var{sar}
1393 horizontal and vertical chroma subsample values. For example for the
1394 pixel format "yuv422p" @var{hsub} is 2 and @var{vsub} is 1.
1397 the number of input frame, starting from 0
1400 the position in the file of the input frame, NAN if unknown
1403 timestamp expressed in seconds, NAN if the input timestamp is unknown
1407 The @var{out_w} and @var{out_h} parameters specify the expressions for
1408 the width and height of the output (cropped) video. They are
1409 evaluated just at the configuration of the filter.
1411 The default value of @var{out_w} is "in_w", and the default value of
1412 @var{out_h} is "in_h".
1414 The expression for @var{out_w} may depend on the value of @var{out_h},
1415 and the expression for @var{out_h} may depend on @var{out_w}, but they
1416 cannot depend on @var{x} and @var{y}, as @var{x} and @var{y} are
1417 evaluated after @var{out_w} and @var{out_h}.
1419 The @var{x} and @var{y} parameters specify the expressions for the
1420 position of the top-left corner of the output (non-cropped) area. They
1421 are evaluated for each frame. If the evaluated value is not valid, it
1422 is approximated to the nearest valid value.
1424 The default value of @var{x} is "(in_w-out_w)/2", and the default
1425 value for @var{y} is "(in_h-out_h)/2", which set the cropped area at
1426 the center of the input image.
1428 The expression for @var{x} may depend on @var{y}, and the expression
1429 for @var{y} may depend on @var{x}.
1431 Follow some examples:
1433 # crop the central input area with size 100x100
1436 # crop the central input area with size 2/3 of the input video
1437 "crop=2/3*in_w:2/3*in_h"
1439 # crop the input video central square
1442 # delimit the rectangle with the top-left corner placed at position
1443 # 100:100 and the right-bottom corner corresponding to the right-bottom
1444 # corner of the input image.
1445 crop=in_w-100:in_h-100:100:100
1447 # crop 10 pixels from the left and right borders, and 20 pixels from
1448 # the top and bottom borders
1449 "crop=in_w-2*10:in_h-2*20"
1451 # keep only the bottom right quarter of the input image
1452 "crop=in_w/2:in_h/2:in_w/2:in_h/2"
1454 # crop height for getting Greek harmony
1455 "crop=in_w:1/PHI*in_w"
1458 "crop=in_w/2:in_h/2:(in_w-out_w)/2+((in_w-out_w)/2)*sin(n/10):(in_h-out_h)/2 +((in_h-out_h)/2)*sin(n/7)"
1460 # erratic camera effect depending on timestamp
1461 "crop=in_w/2:in_h/2:(in_w-out_w)/2+((in_w-out_w)/2)*sin(t*10):(in_h-out_h)/2 +((in_h-out_h)/2)*sin(t*13)"
1463 # set x depending on the value of y
1464 "crop=in_w/2:in_h/2:y:10+10*sin(n/10)"
1469 Auto-detect crop size.
1471 Calculate necessary cropping parameters and prints the recommended
1472 parameters through the logging system. The detected dimensions
1473 correspond to the non-black area of the input video.
1475 It accepts the syntax:
1477 cropdetect[=@var{limit}[:@var{round}[:@var{reset}]]]
1483 Threshold, which can be optionally specified from nothing (0) to
1484 everything (255), defaults to 24.
1487 Value which the width/height should be divisible by, defaults to
1488 16. The offset is automatically adjusted to center the video. Use 2 to
1489 get only even dimensions (needed for 4:2:2 video). 16 is best when
1490 encoding to most video codecs.
1493 Counter that determines after how many frames cropdetect will reset
1494 the previously detected largest video area and start over to detect
1495 the current optimal crop area. Defaults to 0.
1497 This can be useful when channel logos distort the video area. 0
1498 indicates never reset and return the largest area encountered during
1504 Suppress a TV station logo by a simple interpolation of the surrounding
1505 pixels. Just set a rectangle covering the logo and watch it disappear
1506 (and sometimes something even uglier appear - your mileage may vary).
1508 The filter accepts parameters as a string of the form
1509 "@var{x}:@var{y}:@var{w}:@var{h}:@var{band}", or as a list of
1510 @var{key}=@var{value} pairs, separated by ":".
1512 The description of the accepted parameters follows.
1517 Specify the top left corner coordinates of the logo. They must be
1521 Specify the width and height of the logo to clear. They must be
1525 Specify the thickness of the fuzzy edge of the rectangle (added to
1526 @var{w} and @var{h}). The default value is 4.
1529 When set to 1, a green rectangle is drawn on the screen to simplify
1530 finding the right @var{x}, @var{y}, @var{w}, @var{h} parameters, and
1531 @var{band} is set to 4. The default value is 0.
1535 Some examples follow.
1540 Set a rectangle covering the area with top left corner coordinates 0,0
1541 and size 100x77, setting a band of size 10:
1543 delogo=0:0:100:77:10
1547 As the previous example, but use named options:
1549 delogo=x=0:y=0:w=100:h=77:band=10
1556 Attempt to fix small changes in horizontal and/or vertical shift. This
1557 filter helps remove camera shake from hand-holding a camera, bumping a
1558 tripod, moving on a vehicle, etc.
1560 The filter accepts parameters as a string of the form
1561 "@var{x}:@var{y}:@var{w}:@var{h}:@var{rx}:@var{ry}:@var{edge}:@var{blocksize}:@var{contrast}:@var{search}:@var{filename}"
1563 A description of the accepted parameters follows.
1568 Specify a rectangular area where to limit the search for motion
1570 If desired the search for motion vectors can be limited to a
1571 rectangular area of the frame defined by its top left corner, width
1572 and height. These parameters have the same meaning as the drawbox
1573 filter which can be used to visualise the position of the bounding
1576 This is useful when simultaneous movement of subjects within the frame
1577 might be confused for camera motion by the motion vector search.
1579 If any or all of @var{x}, @var{y}, @var{w} and @var{h} are set to -1
1580 then the full frame is used. This allows later options to be set
1581 without specifying the bounding box for the motion vector search.
1583 Default - search the whole frame.
1586 Specify the maximum extent of movement in x and y directions in the
1587 range 0-64 pixels. Default 16.
1590 Specify how to generate pixels to fill blanks at the edge of the
1591 frame. An integer from 0 to 3 as follows:
1594 Fill zeroes at blank locations
1596 Original image at blank locations
1598 Extruded edge value at blank locations
1600 Mirrored edge at blank locations
1603 The default setting is mirror edge at blank locations.
1606 Specify the blocksize to use for motion search. Range 4-128 pixels,
1610 Specify the contrast threshold for blocks. Only blocks with more than
1611 the specified contrast (difference between darkest and lightest
1612 pixels) will be considered. Range 1-255, default 125.
1615 Specify the search strategy 0 = exhaustive search, 1 = less exhaustive
1616 search. Default - exhaustive search.
1619 If set then a detailed log of the motion search is written to the
1626 Draw a colored box on the input image.
1628 It accepts the syntax:
1630 drawbox=@var{x}:@var{y}:@var{width}:@var{height}:@var{color}
1636 Specify the top left corner coordinates of the box. Default to 0.
1639 Specify the width and height of the box, if 0 they are interpreted as
1640 the input width and height. Default to 0.
1643 Specify the color of the box to write, it can be the name of a color
1644 (case insensitive match) or a 0xRRGGBB[AA] sequence.
1647 Follow some examples:
1649 # draw a black box around the edge of the input image
1652 # draw a box with color red and an opacity of 50%
1653 drawbox=10:20:200:60:red@@0.5"
1658 Draw text string or text from specified file on top of video using the
1659 libfreetype library.
1661 To enable compilation of this filter you need to configure FFmpeg with
1662 @code{--enable-libfreetype}.
1664 The filter also recognizes strftime() sequences in the provided text
1665 and expands them accordingly. Check the documentation of strftime().
1667 The filter accepts parameters as a list of @var{key}=@var{value} pairs,
1670 The description of the accepted parameters follows.
1675 Used to draw a box around text using background color.
1676 Value should be either 1 (enable) or 0 (disable).
1677 The default value of @var{box} is 0.
1680 The color to be used for drawing box around text.
1681 Either a string (e.g. "yellow") or in 0xRRGGBB[AA] format
1682 (e.g. "0xff00ff"), possibly followed by an alpha specifier.
1683 The default value of @var{boxcolor} is "white".
1686 Set an expression which specifies if the text should be drawn. If the
1687 expression evaluates to 0, the text is not drawn. This is useful for
1688 specifying that the text should be drawn only when specific conditions
1691 Default value is "1".
1693 See below for the list of accepted constants and functions.
1696 If true, check and fix text coords to avoid clipping.
1699 The color to be used for drawing fonts.
1700 Either a string (e.g. "red") or in 0xRRGGBB[AA] format
1701 (e.g. "0xff000033"), possibly followed by an alpha specifier.
1702 The default value of @var{fontcolor} is "black".
1705 The font file to be used for drawing text. Path must be included.
1706 This parameter is mandatory.
1709 The font size to be used for drawing text.
1710 The default value of @var{fontsize} is 16.
1713 Flags to be used for loading the fonts.
1715 The flags map the corresponding flags supported by libfreetype, and are
1716 a combination of the following values:
1723 @item vertical_layout
1724 @item force_autohint
1727 @item ignore_global_advance_width
1729 @item ignore_transform
1736 Default value is "render".
1738 For more information consult the documentation for the FT_LOAD_*
1742 The color to be used for drawing a shadow behind the drawn text. It
1743 can be a color name (e.g. "yellow") or a string in the 0xRRGGBB[AA]
1744 form (e.g. "0xff00ff"), possibly followed by an alpha specifier.
1745 The default value of @var{shadowcolor} is "black".
1747 @item shadowx, shadowy
1748 The x and y offsets for the text shadow position with respect to the
1749 position of the text. They can be either positive or negative
1750 values. Default value for both is "0".
1753 The size in number of spaces to use for rendering the tab.
1757 Set the initial timecode representation in "hh:mm:ss[:;.]ff"
1758 format. It can be used with or without text parameter. @var{timecode_rate}
1759 option must be specified.
1761 @item timecode_rate, rate, r
1762 Set the timecode frame rate (timecode only).
1765 The text string to be drawn. The text must be a sequence of UTF-8
1767 This parameter is mandatory if no file is specified with the parameter
1771 A text file containing text to be drawn. The text must be a sequence
1772 of UTF-8 encoded characters.
1774 This parameter is mandatory if no text string is specified with the
1775 parameter @var{text}.
1777 If both @var{text} and @var{textfile} are specified, an error is thrown.
1780 The expressions which specify the offsets where text will be drawn
1781 within the video frame. They are relative to the top/left border of the
1784 The default value of @var{x} and @var{y} is "0".
1786 See below for the list of accepted constants and functions.
1789 The parameters for @var{x} and @var{y} are expressions containing the
1790 following constants and functions:
1794 input display aspect ratio, it is the same as (@var{w} / @var{h}) * @var{sar}
1797 horizontal and vertical chroma subsample values. For example for the
1798 pixel format "yuv422p" @var{hsub} is 2 and @var{vsub} is 1.
1801 the height of each text line
1809 @item max_glyph_a, ascent
1810 the maximum distance from the baseline to the highest/upper grid
1811 coordinate used to place a glyph outline point, for all the rendered
1813 It is a positive value, due to the grid's orientation with the Y axis
1816 @item max_glyph_d, descent
1817 the maximum distance from the baseline to the lowest grid coordinate
1818 used to place a glyph outline point, for all the rendered glyphs.
1819 This is a negative value, due to the grid's orientation, with the Y axis
1823 maximum glyph height, that is the maximum height for all the glyphs
1824 contained in the rendered text, it is equivalent to @var{ascent} -
1828 maximum glyph width, that is the maximum width for all the glyphs
1829 contained in the rendered text
1832 the number of input frame, starting from 0
1834 @item rand(min, max)
1835 return a random number included between @var{min} and @var{max}
1838 input sample aspect ratio
1841 timestamp expressed in seconds, NAN if the input timestamp is unknown
1844 the height of the rendered text
1847 the width of the rendered text
1850 the x and y offset coordinates where the text is drawn.
1852 These parameters allow the @var{x} and @var{y} expressions to refer
1853 each other, so you can for example specify @code{y=x/dar}.
1856 If libavfilter was built with @code{--enable-fontconfig}, then
1857 @option{fontfile} can be a fontconfig pattern or omitted.
1859 Some examples follow.
1864 Draw "Test Text" with font FreeSerif, using the default values for the
1865 optional parameters.
1868 drawtext="fontfile=/usr/share/fonts/truetype/freefont/FreeSerif.ttf: text='Test Text'"
1872 Draw 'Test Text' with font FreeSerif of size 24 at position x=100
1873 and y=50 (counting from the top-left corner of the screen), text is
1874 yellow with a red box around it. Both the text and the box have an
1878 drawtext="fontfile=/usr/share/fonts/truetype/freefont/FreeSerif.ttf: text='Test Text':\
1879 x=100: y=50: fontsize=24: fontcolor=yellow@@0.2: box=1: boxcolor=red@@0.2"
1882 Note that the double quotes are not necessary if spaces are not used
1883 within the parameter list.
1886 Show the text at the center of the video frame:
1888 drawtext="fontsize=30:fontfile=FreeSerif.ttf:text='hello world':x=(w-text_w)/2:y=(h-text_h-line_h)/2"
1892 Show a text line sliding from right to left in the last row of the video
1893 frame. The file @file{LONG_LINE} is assumed to contain a single line
1896 drawtext="fontsize=15:fontfile=FreeSerif.ttf:text=LONG_LINE:y=h-line_h:x=-50*t"
1900 Show the content of file @file{CREDITS} off the bottom of the frame and scroll up.
1902 drawtext="fontsize=20:fontfile=FreeSerif.ttf:textfile=CREDITS:y=h-20*t"
1906 Draw a single green letter "g", at the center of the input video.
1907 The glyph baseline is placed at half screen height.
1909 drawtext="fontsize=60:fontfile=FreeSerif.ttf:fontcolor=green:text=g:x=(w-max_glyph_w)/2:y=h/2-ascent"
1913 Show text for 1 second every 3 seconds:
1915 drawtext="fontfile=FreeSerif.ttf:fontcolor=white:x=100:y=x/dar:draw=lt(mod(t\\,3)\\,1):text='blink'"
1919 Use fontconfig to set the font. Note that the colons need to be escaped.
1921 drawtext='fontfile=Linux Libertine O-40\\:style=Semibold:text=FFmpeg'
1926 For more information about libfreetype, check:
1927 @url{http://www.freetype.org/}.
1929 For more information about fontconfig, check:
1930 @url{http://freedesktop.org/software/fontconfig/fontconfig-user.html}.
1934 Apply fade-in/out effect to input video.
1936 It accepts the parameters:
1937 @var{type}:@var{start_frame}:@var{nb_frames}[:@var{options}]
1939 @var{type} specifies if the effect type, can be either "in" for
1940 fade-in, or "out" for a fade-out effect.
1942 @var{start_frame} specifies the number of the start frame for starting
1943 to apply the fade effect.
1945 @var{nb_frames} specifies the number of frames for which the fade
1946 effect has to last. At the end of the fade-in effect the output video
1947 will have the same intensity as the input video, at the end of the
1948 fade-out transition the output video will be completely black.
1950 @var{options} is an optional sequence of @var{key}=@var{value} pairs,
1951 separated by ":". The description of the accepted options follows.
1958 @item start_frame, s
1959 See @var{start_frame}.
1962 See @var{nb_frames}.
1965 If set to 1, fade only alpha channel, if one exists on the input.
1969 A few usage examples follow, usable too as test scenarios.
1971 # fade in first 30 frames of video
1974 # fade out last 45 frames of a 200-frame video
1977 # fade in first 25 frames and fade out last 25 frames of a 1000-frame video
1978 fade=in:0:25, fade=out:975:25
1980 # make first 5 frames black, then fade in from frame 5-24
1983 # fade in alpha over first 25 frames of video
1984 fade=in:0:25:alpha=1
1989 Transform the field order of the input video.
1991 It accepts one parameter which specifies the required field order that
1992 the input interlaced video will be transformed to. The parameter can
1993 assume one of the following values:
1997 output bottom field first
1999 output top field first
2002 Default value is "tff".
2004 Transformation is achieved by shifting the picture content up or down
2005 by one line, and filling the remaining line with appropriate picture content.
2006 This method is consistent with most broadcast field order converters.
2008 If the input video is not flagged as being interlaced, or it is already
2009 flagged as being of the required output field order then this filter does
2010 not alter the incoming video.
2012 This filter is very useful when converting to or from PAL DV material,
2013 which is bottom field first.
2017 ffmpeg -i in.vob -vf "fieldorder=bff" out.dv
2022 Buffer input images and send them when they are requested.
2024 This filter is mainly useful when auto-inserted by the libavfilter
2027 The filter does not take parameters.
2031 Convert the input video to one of the specified pixel formats.
2032 Libavfilter will try to pick one that is supported for the input to
2035 The filter accepts a list of pixel format names, separated by ":",
2036 for example "yuv420p:monow:rgb24".
2038 Some examples follow:
2040 # convert the input video to the format "yuv420p"
2043 # convert the input video to any of the formats in the list
2044 format=yuv420p:yuv444p:yuv410p
2049 Convert the video to specified constant framerate by duplicating or dropping
2050 frames as necessary.
2052 This filter accepts the following named parameters:
2056 Desired output framerate.
2063 Apply a frei0r effect to the input video.
2065 To enable compilation of this filter you need to install the frei0r
2066 header and configure FFmpeg with @code{--enable-frei0r}.
2068 The filter supports the syntax:
2070 @var{filter_name}[@{:|=@}@var{param1}:@var{param2}:...:@var{paramN}]
2073 @var{filter_name} is the name to the frei0r effect to load. If the
2074 environment variable @env{FREI0R_PATH} is defined, the frei0r effect
2075 is searched in each one of the directories specified by the colon
2076 separated list in @env{FREIOR_PATH}, otherwise in the standard frei0r
2077 paths, which are in this order: @file{HOME/.frei0r-1/lib/},
2078 @file{/usr/local/lib/frei0r-1/}, @file{/usr/lib/frei0r-1/}.
2080 @var{param1}, @var{param2}, ... , @var{paramN} specify the parameters
2081 for the frei0r effect.
2083 A frei0r effect parameter can be a boolean (whose values are specified
2084 with "y" and "n"), a double, a color (specified by the syntax
2085 @var{R}/@var{G}/@var{B}, @var{R}, @var{G}, and @var{B} being float
2086 numbers from 0.0 to 1.0) or by an @code{av_parse_color()} color
2087 description), a position (specified by the syntax @var{X}/@var{Y},
2088 @var{X} and @var{Y} being float numbers) and a string.
2090 The number and kind of parameters depend on the loaded effect. If an
2091 effect parameter is not specified the default value is set.
2093 Some examples follow:
2097 Apply the distort0r effect, set the first two double parameters:
2099 frei0r=distort0r:0.5:0.01
2103 Apply the colordistance effect, takes a color as first parameter:
2105 frei0r=colordistance:0.2/0.3/0.4
2106 frei0r=colordistance:violet
2107 frei0r=colordistance:0x112233
2111 Apply the perspective effect, specify the top left and top right image
2114 frei0r=perspective:0.2/0.2:0.8/0.2
2118 For more information see:
2119 @url{http://frei0r.dyne.org}
2123 Fix the banding artifacts that are sometimes introduced into nearly flat
2124 regions by truncation to 8bit color depth.
2125 Interpolate the gradients that should go where the bands are, and
2128 This filter is designed for playback only. Do not use it prior to
2129 lossy compression, because compression tends to lose the dither and
2130 bring back the bands.
2132 The filter takes two optional parameters, separated by ':':
2133 @var{strength}:@var{radius}
2135 @var{strength} is the maximum amount by which the filter will change
2136 any one pixel. Also the threshold for detecting nearly flat
2137 regions. Acceptable values range from .51 to 255, default value is
2138 1.2, out-of-range values will be clipped to the valid range.
2140 @var{radius} is the neighborhood to fit the gradient to. A larger
2141 radius makes for smoother gradients, but also prevents the filter from
2142 modifying the pixels near detailed regions. Acceptable values are
2143 8-32, default value is 16, out-of-range values will be clipped to the
2147 # default parameters
2156 Flip the input video horizontally.
2158 For example to horizontally flip the input video with @command{ffmpeg}:
2160 ffmpeg -i in.avi -vf "hflip" out.avi
2165 High precision/quality 3d denoise filter. This filter aims to reduce
2166 image noise producing smooth images and making still images really
2167 still. It should enhance compressibility.
2169 It accepts the following optional parameters:
2170 @var{luma_spatial}:@var{chroma_spatial}:@var{luma_tmp}:@var{chroma_tmp}
2174 a non-negative float number which specifies spatial luma strength,
2177 @item chroma_spatial
2178 a non-negative float number which specifies spatial chroma strength,
2179 defaults to 3.0*@var{luma_spatial}/4.0
2182 a float number which specifies luma temporal strength, defaults to
2183 6.0*@var{luma_spatial}/4.0
2186 a float number which specifies chroma temporal strength, defaults to
2187 @var{luma_tmp}*@var{chroma_spatial}/@var{luma_spatial}
2192 Modify the hue and/or the saturation of the input.
2194 This filter accepts the following optional named options:
2198 Specify the hue angle as a number of degrees. It accepts a float
2199 number or an expression, and defaults to 0.0.
2202 Specify the hue angle as a number of degrees. It accepts a float
2203 number or an expression, and defaults to 0.0.
2206 Specify the saturation in the [-10,10] range. It accepts a float number and
2210 The options can also be set using the syntax: @var{hue}:@var{saturation}
2212 In this case @var{hue} is expressed in degrees.
2214 Some examples follow:
2217 Set the hue to 90 degrees and the saturation to 1.0:
2223 Same command but expressing the hue in radians:
2229 Same command without named options, hue must be expressed in degrees:
2235 Note that "h:s" syntax does not support expressions for the values of
2236 h and s, so the following example will issue an error:
2244 Interlaceing detect filter. This filter tries to detect if the input is
2245 interlaced or progressive. Top or bottom field first.
2247 @section lut, lutrgb, lutyuv
2249 Compute a look-up table for binding each pixel component input value
2250 to an output value, and apply it to input video.
2252 @var{lutyuv} applies a lookup table to a YUV input video, @var{lutrgb}
2253 to an RGB input video.
2255 These filters accept in input a ":"-separated list of options, which
2256 specify the expressions used for computing the lookup table for the
2257 corresponding pixel component values.
2259 The @var{lut} filter requires either YUV or RGB pixel formats in
2260 input, and accepts the options:
2263 first pixel component
2265 second pixel component
2267 third pixel component
2269 fourth pixel component, corresponds to the alpha component
2272 The exact component associated to each option depends on the format in
2275 The @var{lutrgb} filter requires RGB pixel formats in input, and
2276 accepts the options:
2288 The @var{lutyuv} filter requires YUV pixel formats in input, and
2289 accepts the options:
2292 Y/luminance component
2301 The expressions can contain the following constants and functions:
2305 the input width and height
2308 input value for the pixel component
2311 the input value clipped in the @var{minval}-@var{maxval} range
2314 maximum value for the pixel component
2317 minimum value for the pixel component
2320 the negated value for the pixel component value clipped in the
2321 @var{minval}-@var{maxval} range , it corresponds to the expression
2322 "maxval-clipval+minval"
2325 the computed value in @var{val} clipped in the
2326 @var{minval}-@var{maxval} range
2328 @item gammaval(gamma)
2329 the computed gamma correction value of the pixel component value
2330 clipped in the @var{minval}-@var{maxval} range, corresponds to the
2332 "pow((clipval-minval)/(maxval-minval)\,@var{gamma})*(maxval-minval)+minval"
2336 All expressions default to "val".
2338 Some examples follow:
2340 # negate input video
2341 lutrgb="r=maxval+minval-val:g=maxval+minval-val:b=maxval+minval-val"
2342 lutyuv="y=maxval+minval-val:u=maxval+minval-val:v=maxval+minval-val"
2344 # the above is the same as
2345 lutrgb="r=negval:g=negval:b=negval"
2346 lutyuv="y=negval:u=negval:v=negval"
2351 # remove chroma components, turns the video into a graytone image
2352 lutyuv="u=128:v=128"
2354 # apply a luma burning effect
2357 # remove green and blue components
2360 # set a constant alpha channel value on input
2361 format=rgba,lutrgb=a="maxval-minval/2"
2363 # correct luminance gamma by a 0.5 factor
2364 lutyuv=y=gammaval(0.5)
2369 Apply an MPlayer filter to the input video.
2371 This filter provides a wrapper around most of the filters of
2374 This wrapper is considered experimental. Some of the wrapped filters
2375 may not work properly and we may drop support for them, as they will
2376 be implemented natively into FFmpeg. Thus you should avoid
2377 depending on them when writing portable scripts.
2379 The filters accepts the parameters:
2380 @var{filter_name}[:=]@var{filter_params}
2382 @var{filter_name} is the name of a supported MPlayer filter,
2383 @var{filter_params} is a string containing the parameters accepted by
2386 The list of the currently supported filters follows:
2434 The parameter syntax and behavior for the listed filters are the same
2435 of the corresponding MPlayer filters. For detailed instructions check
2436 the "VIDEO FILTERS" section in the MPlayer manual.
2438 Some examples follow:
2440 # adjust gamma, brightness, contrast
2443 # tweak hue and saturation
2447 See also mplayer(1), @url{http://www.mplayerhq.hu/}.
2453 This filter accepts an integer in input, if non-zero it negates the
2454 alpha component (if available). The default value in input is 0.
2458 Force libavfilter not to use any of the specified pixel formats for the
2459 input to the next filter.
2461 The filter accepts a list of pixel format names, separated by ":",
2462 for example "yuv420p:monow:rgb24".
2464 Some examples follow:
2466 # force libavfilter to use a format different from "yuv420p" for the
2467 # input to the vflip filter
2468 noformat=yuv420p,vflip
2470 # convert the input video to any of the formats not contained in the list
2471 noformat=yuv420p:yuv444p:yuv410p
2476 Pass the video source unchanged to the output.
2480 Apply video transform using libopencv.
2482 To enable this filter install libopencv library and headers and
2483 configure FFmpeg with @code{--enable-libopencv}.
2485 The filter takes the parameters: @var{filter_name}@{:=@}@var{filter_params}.
2487 @var{filter_name} is the name of the libopencv filter to apply.
2489 @var{filter_params} specifies the parameters to pass to the libopencv
2490 filter. If not specified the default values are assumed.
2492 Refer to the official libopencv documentation for more precise
2494 @url{http://opencv.willowgarage.com/documentation/c/image_filtering.html}
2496 Follows the list of supported libopencv filters.
2501 Dilate an image by using a specific structuring element.
2502 This filter corresponds to the libopencv function @code{cvDilate}.
2504 It accepts the parameters: @var{struct_el}:@var{nb_iterations}.
2506 @var{struct_el} represents a structuring element, and has the syntax:
2507 @var{cols}x@var{rows}+@var{anchor_x}x@var{anchor_y}/@var{shape}
2509 @var{cols} and @var{rows} represent the number of columns and rows of
2510 the structuring element, @var{anchor_x} and @var{anchor_y} the anchor
2511 point, and @var{shape} the shape for the structuring element, and
2512 can be one of the values "rect", "cross", "ellipse", "custom".
2514 If the value for @var{shape} is "custom", it must be followed by a
2515 string of the form "=@var{filename}". The file with name
2516 @var{filename} is assumed to represent a binary image, with each
2517 printable character corresponding to a bright pixel. When a custom
2518 @var{shape} is used, @var{cols} and @var{rows} are ignored, the number
2519 or columns and rows of the read file are assumed instead.
2521 The default value for @var{struct_el} is "3x3+0x0/rect".
2523 @var{nb_iterations} specifies the number of times the transform is
2524 applied to the image, and defaults to 1.
2526 Follow some example:
2528 # use the default values
2531 # dilate using a structuring element with a 5x5 cross, iterate two times
2532 ocv=dilate=5x5+2x2/cross:2
2534 # read the shape from the file diamond.shape, iterate two times
2535 # the file diamond.shape may contain a pattern of characters like this:
2541 # the specified cols and rows are ignored (but not the anchor point coordinates)
2542 ocv=0x0+2x2/custom=diamond.shape:2
2547 Erode an image by using a specific structuring element.
2548 This filter corresponds to the libopencv function @code{cvErode}.
2550 The filter accepts the parameters: @var{struct_el}:@var{nb_iterations},
2551 with the same syntax and semantics as the @ref{dilate} filter.
2555 Smooth the input video.
2557 The filter takes the following parameters:
2558 @var{type}:@var{param1}:@var{param2}:@var{param3}:@var{param4}.
2560 @var{type} is the type of smooth filter to apply, and can be one of
2561 the following values: "blur", "blur_no_scale", "median", "gaussian",
2562 "bilateral". The default value is "gaussian".
2564 @var{param1}, @var{param2}, @var{param3}, and @var{param4} are
2565 parameters whose meanings depend on smooth type. @var{param1} and
2566 @var{param2} accept integer positive values or 0, @var{param3} and
2567 @var{param4} accept float values.
2569 The default value for @var{param1} is 3, the default value for the
2570 other parameters is 0.
2572 These parameters correspond to the parameters assigned to the
2573 libopencv function @code{cvSmooth}.
2578 Overlay one video on top of another.
2580 It takes two inputs and one output, the first input is the "main"
2581 video on which the second input is overlayed.
2583 It accepts the parameters: @var{x}:@var{y}[:@var{options}].
2585 @var{x} is the x coordinate of the overlayed video on the main video,
2586 @var{y} is the y coordinate. @var{x} and @var{y} are expressions containing
2587 the following parameters:
2590 @item main_w, main_h
2591 main input width and height
2594 same as @var{main_w} and @var{main_h}
2596 @item overlay_w, overlay_h
2597 overlay input width and height
2600 same as @var{overlay_w} and @var{overlay_h}
2603 @var{options} is an optional list of @var{key}=@var{value} pairs,
2606 The description of the accepted options follows.
2610 If set to 1, force the filter to accept inputs in the RGB
2611 color space. Default value is 0.
2614 Be aware that frames are taken from each input video in timestamp
2615 order, hence, if their initial timestamps differ, it is a a good idea
2616 to pass the two inputs through a @var{setpts=PTS-STARTPTS} filter to
2617 have them begin in the same zero timestamp, as it does the example for
2618 the @var{movie} filter.
2620 Follow some examples:
2622 # draw the overlay at 10 pixels from the bottom right
2623 # corner of the main video.
2624 overlay=main_w-overlay_w-10:main_h-overlay_h-10
2626 # insert a transparent PNG logo in the bottom left corner of the input
2627 ffmpeg -i input -i logo -filter_complex 'overlay=10:main_h-overlay_h-10' output
2629 # insert 2 different transparent PNG logos (second logo on bottom
2631 ffmpeg -i input -i logo1 -i logo2 -filter_complex
2632 'overlay=10:H-h-10,overlay=W-w-10:H-h-10' output
2634 # add a transparent color layer on top of the main video,
2635 # WxH specifies the size of the main input to the overlay filter
2636 color=red@.3:WxH [over]; [in][over] overlay [out]
2638 # play an original video and a filtered version (here with the deshake filter)
2640 ffplay input.avi -vf 'split[a][b]; [a]pad=iw*2:ih[src]; [b]deshake[filt]; [src][filt]overlay=w'
2642 # the previous example is the same as:
2643 ffplay input.avi -vf 'split[b], pad=iw*2[src], [b]deshake, [src]overlay=w'
2646 You can chain together more overlays but the efficiency of such
2647 approach is yet to be tested.
2651 Add paddings to the input image, and places the original input at the
2652 given coordinates @var{x}, @var{y}.
2654 It accepts the following parameters:
2655 @var{width}:@var{height}:@var{x}:@var{y}:@var{color}.
2657 The parameters @var{width}, @var{height}, @var{x}, and @var{y} are
2658 expressions containing the following constants:
2662 the input video width and height
2665 same as @var{in_w} and @var{in_h}
2668 the output width and height, that is the size of the padded area as
2669 specified by the @var{width} and @var{height} expressions
2672 same as @var{out_w} and @var{out_h}
2675 x and y offsets as specified by the @var{x} and @var{y}
2676 expressions, or NAN if not yet specified
2679 same as @var{iw} / @var{ih}
2682 input sample aspect ratio
2685 input display aspect ratio, it is the same as (@var{iw} / @var{ih}) * @var{sar}
2688 horizontal and vertical chroma subsample values. For example for the
2689 pixel format "yuv422p" @var{hsub} is 2 and @var{vsub} is 1.
2692 Follows the description of the accepted parameters.
2697 Specify the size of the output image with the paddings added. If the
2698 value for @var{width} or @var{height} is 0, the corresponding input size
2699 is used for the output.
2701 The @var{width} expression can reference the value set by the
2702 @var{height} expression, and vice versa.
2704 The default value of @var{width} and @var{height} is 0.
2708 Specify the offsets where to place the input image in the padded area
2709 with respect to the top/left border of the output image.
2711 The @var{x} expression can reference the value set by the @var{y}
2712 expression, and vice versa.
2714 The default value of @var{x} and @var{y} is 0.
2718 Specify the color of the padded area, it can be the name of a color
2719 (case insensitive match) or a 0xRRGGBB[AA] sequence.
2721 The default value of @var{color} is "black".
2725 Some examples follow:
2728 # Add paddings with color "violet" to the input video. Output video
2729 # size is 640x480, the top-left corner of the input video is placed at
2731 pad=640:480:0:40:violet
2733 # pad the input to get an output with dimensions increased bt 3/2,
2734 # and put the input video at the center of the padded area
2735 pad="3/2*iw:3/2*ih:(ow-iw)/2:(oh-ih)/2"
2737 # pad the input to get a squared output with size equal to the maximum
2738 # value between the input width and height, and put the input video at
2739 # the center of the padded area
2740 pad="max(iw\,ih):ow:(ow-iw)/2:(oh-ih)/2"
2742 # pad the input to get a final w/h ratio of 16:9
2743 pad="ih*16/9:ih:(ow-iw)/2:(oh-ih)/2"
2745 # for anamorphic video, in order to set the output display aspect ratio,
2746 # it is necessary to use sar in the expression, according to the relation:
2747 # (ih * X / ih) * sar = output_dar
2748 # X = output_dar / sar
2749 pad="ih*16/9/sar:ih:(ow-iw)/2:(oh-ih)/2"
2751 # double output size and put the input video in the bottom-right
2752 # corner of the output padded area
2753 pad="2*iw:2*ih:ow-iw:oh-ih"
2756 @section pixdesctest
2758 Pixel format descriptor test filter, mainly useful for internal
2759 testing. The output video should be equal to the input video.
2763 format=monow, pixdesctest
2766 can be used to test the monowhite pixel format descriptor definition.
2770 Suppress a TV station logo, using an image file to determine which
2771 pixels comprise the logo. It works by filling in the pixels that
2772 comprise the logo with neighboring pixels.
2774 This filter requires one argument which specifies the filter bitmap
2775 file, which can be any image format supported by libavformat. The
2776 width and height of the image file must match those of the video
2777 stream being processed.
2779 Pixels in the provided bitmap image with a value of zero are not
2780 considered part of the logo, non-zero pixels are considered part of
2781 the logo. If you use white (255) for the logo and black (0) for the
2782 rest, you will be safe. For making the filter bitmap, it is
2783 recommended to take a screen capture of a black frame with the logo
2784 visible, and then using a threshold filter followed by the erode
2785 filter once or twice.
2787 If needed, little splotches can be fixed manually. Remember that if
2788 logo pixels are not covered, the filter quality will be much
2789 reduced. Marking too many pixels as part of the logo does not hurt as
2790 much, but it will increase the amount of blurring needed to cover over
2791 the image and will destroy more information than necessary, and extra
2792 pixels will slow things down on a large logo.
2796 Scale the input video to @var{width}:@var{height}[:@var{interl}=@{1|-1@}] and/or convert the image format.
2798 The scale filter forces the output display aspect ratio to be the same
2799 of the input, by changing the output sample aspect ratio.
2801 The parameters @var{width} and @var{height} are expressions containing
2802 the following constants:
2806 the input width and height
2809 same as @var{in_w} and @var{in_h}
2812 the output (cropped) width and height
2815 same as @var{out_w} and @var{out_h}
2818 same as @var{iw} / @var{ih}
2821 input sample aspect ratio
2824 input display aspect ratio, it is the same as (@var{iw} / @var{ih}) * @var{sar}
2827 horizontal and vertical chroma subsample values. For example for the
2828 pixel format "yuv422p" @var{hsub} is 2 and @var{vsub} is 1.
2831 If the input image format is different from the format requested by
2832 the next filter, the scale filter will convert the input to the
2835 If the value for @var{width} or @var{height} is 0, the respective input
2836 size is used for the output.
2838 If the value for @var{width} or @var{height} is -1, the scale filter will
2839 use, for the respective output size, a value that maintains the aspect
2840 ratio of the input image.
2842 The default value of @var{width} and @var{height} is 0.
2844 Valid values for the optional parameter @var{interl} are:
2848 force interlaced aware scaling
2851 select interlaced aware scaling depending on whether the source frames
2852 are flagged as interlaced or not
2855 Unless @var{interl} is set to one of the above options, interlaced scaling will not be used.
2857 Some examples follow:
2859 # scale the input video to a size of 200x100.
2862 # scale the input to 2x
2864 # the above is the same as
2867 # scale the input to 2x with forced interlaced scaling
2868 scale=2*iw:2*ih:interl=1
2870 # scale the input to half size
2873 # increase the width, and set the height to the same size
2876 # seek for Greek harmony
2880 # increase the height, and set the width to 3/2 of the height
2883 # increase the size, but make the size a multiple of the chroma
2884 scale="trunc(3/2*iw/hsub)*hsub:trunc(3/2*ih/vsub)*vsub"
2886 # increase the width to a maximum of 500 pixels, keep the same input aspect ratio
2887 scale='min(500\, iw*3/2):-1'
2891 Select frames to pass in output.
2893 It accepts in input an expression, which is evaluated for each input
2894 frame. If the expression is evaluated to a non-zero value, the frame
2895 is selected and passed to the output, otherwise it is discarded.
2897 The expression can contain the following constants:
2901 the sequential number of the filtered frame, starting from 0
2904 the sequential number of the selected frame, starting from 0
2906 @item prev_selected_n
2907 the sequential number of the last selected frame, NAN if undefined
2910 timebase of the input timestamps
2913 the PTS (Presentation TimeStamp) of the filtered video frame,
2914 expressed in @var{TB} units, NAN if undefined
2917 the PTS (Presentation TimeStamp) of the filtered video frame,
2918 expressed in seconds, NAN if undefined
2921 the PTS of the previously filtered video frame, NAN if undefined
2923 @item prev_selected_pts
2924 the PTS of the last previously filtered video frame, NAN if undefined
2926 @item prev_selected_t
2927 the PTS of the last previously selected video frame, NAN if undefined
2930 the PTS of the first video frame in the video, NAN if undefined
2933 the time of the first video frame in the video, NAN if undefined
2936 the type of the filtered frame, can assume one of the following
2948 @item interlace_type
2949 the frame interlace type, can assume one of the following values:
2952 the frame is progressive (not interlaced)
2954 the frame is top-field-first
2956 the frame is bottom-field-first
2960 1 if the filtered frame is a key-frame, 0 otherwise
2963 the position in the file of the filtered frame, -1 if the information
2964 is not available (e.g. for synthetic video)
2967 value between 0 and 1 to indicate a new scene; a low value reflects a low
2968 probability for the current frame to introduce a new scene, while a higher
2969 value means the current frame is more likely to be one (see the example below)
2973 The default value of the select expression is "1".
2975 Some examples follow:
2978 # select all frames in input
2981 # the above is the same as:
2987 # select only I-frames
2988 select='eq(pict_type\,I)'
2990 # select one frame every 100
2991 select='not(mod(n\,100))'
2993 # select only frames contained in the 10-20 time interval
2994 select='gte(t\,10)*lte(t\,20)'
2996 # select only I frames contained in the 10-20 time interval
2997 select='gte(t\,10)*lte(t\,20)*eq(pict_type\,I)'
2999 # select frames with a minimum distance of 10 seconds
3000 select='isnan(prev_selected_t)+gte(t-prev_selected_t\,10)'
3003 Complete example to create a mosaic of the first scenes:
3006 ffmpeg -i video.avi -vf select='gt(scene\,0.4)',scale=160:120,tile -frames:v 1 preview.png
3009 Comparing @var{scene} against a value between 0.3 and 0.5 is generally a sane
3012 @section setdar, setsar
3014 The @code{setdar} filter sets the Display Aspect Ratio for the filter
3017 This is done by changing the specified Sample (aka Pixel) Aspect
3018 Ratio, according to the following equation:
3020 @var{DAR} = @var{HORIZONTAL_RESOLUTION} / @var{VERTICAL_RESOLUTION} * @var{SAR}
3023 Keep in mind that the @code{setdar} filter does not modify the pixel
3024 dimensions of the video frame. Also the display aspect ratio set by
3025 this filter may be changed by later filters in the filterchain,
3026 e.g. in case of scaling or if another "setdar" or a "setsar" filter is
3029 The @code{setsar} filter sets the Sample (aka Pixel) Aspect Ratio for
3030 the filter output video.
3032 Note that as a consequence of the application of this filter, the
3033 output display aspect ratio will change according to the equation
3036 Keep in mind that the sample aspect ratio set by the @code{setsar}
3037 filter may be changed by later filters in the filterchain, e.g. if
3038 another "setsar" or a "setdar" filter is applied.
3040 The @code{setdar} and @code{setsar} filters accept a parameter string
3041 which represents the wanted aspect ratio. The parameter can
3042 be a floating point number string, an expression, or a string of the form
3043 @var{num}:@var{den}, where @var{num} and @var{den} are the numerator
3044 and denominator of the aspect ratio. If the parameter is not
3045 specified, it is assumed the value "0:1".
3047 For example to change the display aspect ratio to 16:9, specify:
3052 The example above is equivalent to:
3057 To change the sample aspect ratio to 10:11, specify:
3064 Force field for the output video frame.
3066 The @code{setfield} filter marks the interlace type field for the
3067 output frames. It does not change the input frame, but only sets the
3068 corresponding property, which affects how the frame is treated by
3069 following filters (e.g. @code{fieldorder} or @code{yadif}).
3071 It accepts a string parameter, which can assume the following values:
3074 Keep the same field property.
3077 Mark the frame as bottom-field-first.
3080 Mark the frame as top-field-first.
3083 Mark the frame as progressive.
3086 @section asetpts, setpts
3088 Change the PTS (presentation timestamp) of the input frames.
3090 @code{asetpts} works on audio frames, @code{setpts} on video frames.
3092 Accept in input an expression evaluated through the eval API, which
3093 can contain the following constants:
3097 the presentation timestamp in input
3100 the count of the input frame, starting from 0.
3102 @item NB_CONSUMED_SAMPLES
3103 the number of consumed samples, not including the current frame (only
3107 the number of samples in the current frame (only audio)
3113 the PTS of the first video frame
3116 tell if the current frame is interlaced
3122 original position in the file of the frame, or undefined if undefined
3123 for the current frame
3133 Some examples follow:
3136 # start counting PTS from zero
3148 # fixed rate 25 fps with some jitter
3149 setpts='1/(25*TB) * (N + 0.05 * sin(N*2*PI/25))'
3151 # apply an offset of 10 seconds to the input PTS
3155 @section settb, asettb
3157 Set the timebase to use for the output frames timestamps.
3158 It is mainly useful for testing timebase configuration.
3160 It accepts in input an arithmetic expression representing a rational.
3161 The expression can contain the constants "AVTB" (the
3162 default timebase), "intb" (the input timebase) and "sr" (the sample rate,
3165 The default value for the input is "intb".
3167 Follow some examples.
3170 # set the timebase to 1/25
3173 # set the timebase to 1/10
3176 #set the timebase to 1001/1000
3179 #set the timebase to 2*intb
3182 #set the default timebase value
3188 Show a line containing various information for each input video frame.
3189 The input video is not modified.
3191 The shown line contains a sequence of key/value pairs of the form
3192 @var{key}:@var{value}.
3194 A description of each shown parameter follows:
3198 sequential number of the input frame, starting from 0
3201 Presentation TimeStamp of the input frame, expressed as a number of
3202 time base units. The time base unit depends on the filter input pad.
3205 Presentation TimeStamp of the input frame, expressed as a number of
3209 position of the frame in the input stream, -1 if this information in
3210 unavailable and/or meaningless (for example in case of synthetic video)
3216 sample aspect ratio of the input frame, expressed in the form
3220 size of the input frame, expressed in the form
3221 @var{width}x@var{height}
3224 interlaced mode ("P" for "progressive", "T" for top field first, "B"
3225 for bottom field first)
3228 1 if the frame is a key frame, 0 otherwise
3231 picture type of the input frame ("I" for an I-frame, "P" for a
3232 P-frame, "B" for a B-frame, "?" for unknown type).
3233 Check also the documentation of the @code{AVPictureType} enum and of
3234 the @code{av_get_picture_type_char} function defined in
3235 @file{libavutil/avutil.h}.
3238 Adler-32 checksum (printed in hexadecimal) of all the planes of the input frame
3240 @item plane_checksum
3241 Adler-32 checksum (printed in hexadecimal) of each plane of the input frame,
3242 expressed in the form "[@var{c0} @var{c1} @var{c2} @var{c3}]"
3247 Pass the images of input video on to next video filter as multiple
3251 ffmpeg -i in.avi -vf "slicify=32" out.avi
3254 The filter accepts the slice height as parameter. If the parameter is
3255 not specified it will use the default value of 16.
3257 Adding this in the beginning of filter chains should make filtering
3258 faster due to better use of the memory cache.
3262 Split input video into several identical outputs.
3264 The filter accepts a single parameter which specifies the number of outputs. If
3265 unspecified, it defaults to 2.
3269 ffmpeg -i INPUT -filter_complex split=5 OUTPUT
3271 will create 5 copies of the input video.
3275 [in] split [splitout1][splitout2];
3276 [splitout1] crop=100:100:0:0 [cropout];
3277 [splitout2] pad=200:200:100:100 [padout];
3280 will create two separate outputs from the same input, one cropped and
3285 Scale the input by 2x and smooth using the Super2xSaI (Scale and
3286 Interpolate) pixel art scaling algorithm.
3288 Useful for enlarging pixel art images without reducing sharpness.
3294 Select the most representative frame in a given sequence of consecutive frames.
3296 It accepts as argument the frames batch size to analyze (default @var{N}=100);
3297 in a set of @var{N} frames, the filter will pick one of them, and then handle
3298 the next batch of @var{N} frames until the end.
3300 Since the filter keeps track of the whole frames sequence, a bigger @var{N}
3301 value will result in a higher memory usage, so a high value is not recommended.
3303 The following example extract one picture each 50 frames:
3308 Complete example of a thumbnail creation with @command{ffmpeg}:
3310 ffmpeg -i in.avi -vf thumbnail,scale=300:200 -frames:v 1 out.png
3315 Tile several successive frames together.
3317 It accepts as argument the tile size (i.e. the number of lines and columns)
3318 in the form "@var{w}x@var{h}".
3320 For example, produce 8×8 PNG tiles of all keyframes (@option{-skip_frame
3323 ffmpeg -skip_frame nokey -i file.avi -vf 'scale=128:72,tile=8x8' -an -vsync 0 keyframes%03d.png
3325 The @option{-vsync 0} is necessary to prevent @command{ffmpeg} from
3326 duplicating each output frame to accomodate the originally detected frame
3331 Perform various types of temporal field interlacing.
3333 Frames are counted starting from 1, so the first input frame is
3336 This filter accepts a single parameter specifying the mode. Available
3341 Move odd frames into the upper field, even into the lower field,
3342 generating a double height frame at half framerate.
3345 Only output even frames, odd frames are dropped, generating a frame with
3346 unchanged height at half framerate.
3349 Only output odd frames, even frames are dropped, generating a frame with
3350 unchanged height at half framerate.
3353 Expand each frame to full height, but pad alternate lines with black,
3354 generating a frame with double height at the same input framerate.
3356 @item interleave_top, 4
3357 Interleave the upper field from odd frames with the lower field from
3358 even frames, generating a frame with unchanged height at half framerate.
3360 @item interleave_bottom, 5
3361 Interleave the lower field from odd frames with the upper field from
3362 even frames, generating a frame with unchanged height at half framerate.
3364 @item interlacex2, 6
3365 Double frame rate with unchanged height. Frames are inserted each
3366 containing the second temporal field from the previous input frame and
3367 the first temporal field from the next input frame. This mode relies on
3368 the top_field_first flag. Useful for interlaced video displays with no
3369 field synchronisation.
3372 Numeric values are deprecated but are accepted for backward
3373 compatibility reasons.
3375 Default mode is @code{merge}.
3379 Transpose rows with columns in the input video and optionally flip it.
3381 It accepts a parameter representing an integer, which can assume the
3386 Rotate by 90 degrees counterclockwise and vertically flip (default), that is:
3394 Rotate by 90 degrees clockwise, that is:
3402 Rotate by 90 degrees counterclockwise, that is:
3410 Rotate by 90 degrees clockwise and vertically flip, that is:
3420 Sharpen or blur the input video.
3422 It accepts the following parameters:
3423 @var{luma_msize_x}:@var{luma_msize_y}:@var{luma_amount}:@var{chroma_msize_x}:@var{chroma_msize_y}:@var{chroma_amount}
3425 Negative values for the amount will blur the input video, while positive
3426 values will sharpen. All parameters are optional and default to the
3427 equivalent of the string '5:5:1.0:5:5:0.0'.
3432 Set the luma matrix horizontal size. It can be an integer between 3
3433 and 13, default value is 5.
3436 Set the luma matrix vertical size. It can be an integer between 3
3437 and 13, default value is 5.
3440 Set the luma effect strength. It can be a float number between -2.0
3441 and 5.0, default value is 1.0.
3443 @item chroma_msize_x
3444 Set the chroma matrix horizontal size. It can be an integer between 3
3445 and 13, default value is 5.
3447 @item chroma_msize_y
3448 Set the chroma matrix vertical size. It can be an integer between 3
3449 and 13, default value is 5.
3452 Set the chroma effect strength. It can be a float number between -2.0
3453 and 5.0, default value is 0.0.
3458 # Strong luma sharpen effect parameters
3461 # Strong blur of both luma and chroma parameters
3462 unsharp=7:7:-2:7:7:-2
3464 # Use the default values with @command{ffmpeg}
3465 ffmpeg -i in.avi -vf "unsharp" out.mp4
3470 Flip the input video vertically.
3473 ffmpeg -i in.avi -vf "vflip" out.avi
3478 Deinterlace the input video ("yadif" means "yet another deinterlacing
3481 It accepts the optional parameters: @var{mode}:@var{parity}:@var{auto}.
3483 @var{mode} specifies the interlacing mode to adopt, accepts one of the
3488 output 1 frame for each frame
3490 output 1 frame for each field
3492 like 0 but skips spatial interlacing check
3494 like 1 but skips spatial interlacing check
3499 @var{parity} specifies the picture field parity assumed for the input
3500 interlaced video, accepts one of the following values:
3504 assume top field first
3506 assume bottom field first
3508 enable automatic detection
3511 Default value is -1.
3512 If interlacing is unknown or decoder does not export this information,
3513 top field first will be assumed.
3515 @var{auto} specifies if deinterlacer should trust the interlaced flag
3516 and only deinterlace frames marked as interlaced
3520 deinterlace all frames
3522 only deinterlace frames marked as interlaced
3527 @c man end VIDEO FILTERS
3529 @chapter Video Sources
3530 @c man begin VIDEO SOURCES
3532 Below is a description of the currently available video sources.
3536 Buffer video frames, and make them available to the filter chain.
3538 This source is mainly intended for a programmatic use, in particular
3539 through the interface defined in @file{libavfilter/vsrc_buffer.h}.
3541 It accepts a list of options in the form of @var{key}=@var{value} pairs
3542 separated by ":". A descroption of the accepted options follows.
3547 Specify the size (width and height) of the buffered video frames.
3550 A string representing the pixel format of the buffered video frames.
3551 It may be a number corresponding to a pixel format, or a pixel format
3555 Specify the timebase assumed by the timestamps of the buffered frames.
3558 Specify the frame rate expected for the video stream.
3561 Specify the sample aspect ratio assumed by the video frames.
3564 Specify the optional parameters to be used for the scale filter which
3565 is automatically inserted when an input change is detected in the
3566 input size or format.
3571 buffer=size=320x240:pix_fmt=yuv410p:time_base=1/24:pixel_aspect=1/1
3574 will instruct the source to accept video frames with size 320x240 and
3575 with format "yuv410p", assuming 1/24 as the timestamps timebase and
3576 square pixels (1:1 sample aspect ratio).
3577 Since the pixel format with name "yuv410p" corresponds to the number 6
3578 (check the enum PixelFormat definition in @file{libavutil/pixfmt.h}),
3579 this example corresponds to:
3581 buffer=size=320x240:pixfmt=6:time_base=1/24:pixel_aspect=1/1
3584 Alternatively, the options can be specified as a flat string, but this
3585 syntax is deprecated:
3587 @var{width}:@var{height}:@var{pix_fmt}:@var{time_base.num}:@var{time_base.den}:@var{pixel_aspect.num}:@var{pixel_aspect.den}[:@var{sws_param}]
3591 Create a pattern generated by an elementary cellular automaton.
3593 The initial state of the cellular automaton can be defined through the
3594 @option{filename}, and @option{pattern} options. If such options are
3595 not specified an initial state is created randomly.
3597 At each new frame a new row in the video is filled with the result of
3598 the cellular automaton next generation. The behavior when the whole
3599 frame is filled is defined by the @option{scroll} option.
3601 This source accepts a list of options in the form of
3602 @var{key}=@var{value} pairs separated by ":". A description of the
3603 accepted options follows.
3607 Read the initial cellular automaton state, i.e. the starting row, from
3609 In the file, each non-whitespace character is considered an alive
3610 cell, a newline will terminate the row, and further characters in the
3611 file will be ignored.
3614 Read the initial cellular automaton state, i.e. the starting row, from
3615 the specified string.
3617 Each non-whitespace character in the string is considered an alive
3618 cell, a newline will terminate the row, and further characters in the
3619 string will be ignored.
3622 Set the video rate, that is the number of frames generated per second.
3625 @item random_fill_ratio, ratio
3626 Set the random fill ratio for the initial cellular automaton row. It
3627 is a floating point number value ranging from 0 to 1, defaults to
3630 This option is ignored when a file or a pattern is specified.
3632 @item random_seed, seed
3633 Set the seed for filling randomly the initial row, must be an integer
3634 included between 0 and UINT32_MAX. If not specified, or if explicitly
3635 set to -1, the filter will try to use a good random seed on a best
3639 Set the cellular automaton rule, it is a number ranging from 0 to 255.
3640 Default value is 110.
3643 Set the size of the output video.
3645 If @option{filename} or @option{pattern} is specified, the size is set
3646 by default to the width of the specified initial state row, and the
3647 height is set to @var{width} * PHI.
3649 If @option{size} is set, it must contain the width of the specified
3650 pattern string, and the specified pattern will be centered in the
3653 If a filename or a pattern string is not specified, the size value
3654 defaults to "320x518" (used for a randomly generated initial state).
3657 If set to 1, scroll the output upward when all the rows in the output
3658 have been already filled. If set to 0, the new generated row will be
3659 written over the top row just after the bottom row is filled.
3662 @item start_full, full
3663 If set to 1, completely fill the output with generated rows before
3664 outputting the first frame.
3665 This is the default behavior, for disabling set the value to 0.
3668 If set to 1, stitch the left and right row edges together.
3669 This is the default behavior, for disabling set the value to 0.
3672 @subsection Examples
3676 Read the initial state from @file{pattern}, and specify an output of
3679 cellauto=f=pattern:s=200x400
3683 Generate a random initial row with a width of 200 cells, with a fill
3686 cellauto=ratio=2/3:s=200x200
3690 Create a pattern generated by rule 18 starting by a single alive cell
3691 centered on an initial row with width 100:
3693 cellauto=p=@@:s=100x400:full=0:rule=18
3697 Specify a more elaborated initial pattern:
3699 cellauto=p='@@@@ @@ @@@@':s=100x400:full=0:rule=18
3706 Generate a Mandelbrot set fractal, and progressively zoom towards the
3707 point specified with @var{start_x} and @var{start_y}.
3709 This source accepts a list of options in the form of
3710 @var{key}=@var{value} pairs separated by ":". A description of the
3711 accepted options follows.
3716 Set the terminal pts value. Default value is 400.
3719 Set the terminal scale value.
3720 Must be a floating point value. Default value is 0.3.
3723 Set the inner coloring mode, that is the algorithm used to draw the
3724 Mandelbrot fractal internal region.
3726 It shall assume one of the following values:
3731 Show time until convergence.
3733 Set color based on point closest to the origin of the iterations.
3738 Default value is @var{mincol}.
3741 Set the bailout value. Default value is 10.0.
3744 Set the maximum of iterations performed by the rendering
3745 algorithm. Default value is 7189.
3748 Set outer coloring mode.
3749 It shall assume one of following values:
3751 @item iteration_count
3752 Set iteration cound mode.
3753 @item normalized_iteration_count
3754 set normalized iteration count mode.
3756 Default value is @var{normalized_iteration_count}.
3759 Set frame rate, expressed as number of frames per second. Default
3763 Set frame size. Default value is "640x480".
3766 Set the initial scale value. Default value is 3.0.
3769 Set the initial x position. Must be a floating point value between
3770 -100 and 100. Default value is -0.743643887037158704752191506114774.
3773 Set the initial y position. Must be a floating point value between
3774 -100 and 100. Default value is -0.131825904205311970493132056385139.
3779 Generate various test patterns, as generated by the MPlayer test filter.
3781 The size of the generated video is fixed, and is 256x256.
3782 This source is useful in particular for testing encoding features.
3784 This source accepts an optional sequence of @var{key}=@var{value} pairs,
3785 separated by ":". The description of the accepted options follows.
3790 Specify the frame rate of the sourced video, as the number of frames
3791 generated per second. It has to be a string in the format
3792 @var{frame_rate_num}/@var{frame_rate_den}, an integer number, a float
3793 number or a valid video frame rate abbreviation. The default value is
3797 Set the video duration of the sourced video. The accepted syntax is:
3802 See also the function @code{av_parse_time()}.
3804 If not specified, or the expressed duration is negative, the video is
3805 supposed to be generated forever.
3809 Set the number or the name of the test to perform. Supported tests are:
3824 Default value is "all", which will cycle through the list of all tests.
3827 For example the following:
3832 will generate a "dc_luma" test pattern.
3836 Provide a frei0r source.
3838 To enable compilation of this filter you need to install the frei0r
3839 header and configure FFmpeg with @code{--enable-frei0r}.
3841 The source supports the syntax:
3843 @var{size}:@var{rate}:@var{src_name}[@{=|:@}@var{param1}:@var{param2}:...:@var{paramN}]
3846 @var{size} is the size of the video to generate, may be a string of the
3847 form @var{width}x@var{height} or a frame size abbreviation.
3848 @var{rate} is the rate of the video to generate, may be a string of
3849 the form @var{num}/@var{den} or a frame rate abbreviation.
3850 @var{src_name} is the name to the frei0r source to load. For more
3851 information regarding frei0r and how to set the parameters read the
3852 section @ref{frei0r} in the description of the video filters.
3854 For example, to generate a frei0r partik0l source with size 200x200
3855 and frame rate 10 which is overlayed on the overlay filter main input:
3857 frei0r_src=200x200:10:partik0l=1234 [overlay]; [in][overlay] overlay
3862 Generate a life pattern.
3864 This source is based on a generalization of John Conway's life game.
3866 The sourced input represents a life grid, each pixel represents a cell
3867 which can be in one of two possible states, alive or dead. Every cell
3868 interacts with its eight neighbours, which are the cells that are
3869 horizontally, vertically, or diagonally adjacent.
3871 At each interaction the grid evolves according to the adopted rule,
3872 which specifies the number of neighbor alive cells which will make a
3873 cell stay alive or born. The @option{rule} option allows to specify
3876 This source accepts a list of options in the form of
3877 @var{key}=@var{value} pairs separated by ":". A description of the
3878 accepted options follows.
3882 Set the file from which to read the initial grid state. In the file,
3883 each non-whitespace character is considered an alive cell, and newline
3884 is used to delimit the end of each row.
3886 If this option is not specified, the initial grid is generated
3890 Set the video rate, that is the number of frames generated per second.
3893 @item random_fill_ratio, ratio
3894 Set the random fill ratio for the initial random grid. It is a
3895 floating point number value ranging from 0 to 1, defaults to 1/PHI.
3896 It is ignored when a file is specified.
3898 @item random_seed, seed
3899 Set the seed for filling the initial random grid, must be an integer
3900 included between 0 and UINT32_MAX. If not specified, or if explicitly
3901 set to -1, the filter will try to use a good random seed on a best
3907 A rule can be specified with a code of the kind "S@var{NS}/B@var{NB}",
3908 where @var{NS} and @var{NB} are sequences of numbers in the range 0-8,
3909 @var{NS} specifies the number of alive neighbor cells which make a
3910 live cell stay alive, and @var{NB} the number of alive neighbor cells
3911 which make a dead cell to become alive (i.e. to "born").
3912 "s" and "b" can be used in place of "S" and "B", respectively.
3914 Alternatively a rule can be specified by an 18-bits integer. The 9
3915 high order bits are used to encode the next cell state if it is alive
3916 for each number of neighbor alive cells, the low order bits specify
3917 the rule for "borning" new cells. Higher order bits encode for an
3918 higher number of neighbor cells.
3919 For example the number 6153 = @code{(12<<9)+9} specifies a stay alive
3920 rule of 12 and a born rule of 9, which corresponds to "S23/B03".
3922 Default value is "S23/B3", which is the original Conway's game of life
3923 rule, and will keep a cell alive if it has 2 or 3 neighbor alive
3924 cells, and will born a new cell if there are three alive cells around
3928 Set the size of the output video.
3930 If @option{filename} is specified, the size is set by default to the
3931 same size of the input file. If @option{size} is set, it must contain
3932 the size specified in the input file, and the initial grid defined in
3933 that file is centered in the larger resulting area.
3935 If a filename is not specified, the size value defaults to "320x240"
3936 (used for a randomly generated initial grid).
3939 If set to 1, stitch the left and right grid edges together, and the
3940 top and bottom edges also. Defaults to 1.
3943 Set cell mold speed. If set, a dead cell will go from @option{death_color} to
3944 @option{mold_color} with a step of @option{mold}. @option{mold} can have a
3945 value from 0 to 255.
3948 Set the color of living (or new born) cells.
3951 Set the color of dead cells. If @option{mold} is set, this is the first color
3952 used to represent a dead cell.
3955 Set mold color, for definitely dead and moldy cells.
3958 @subsection Examples
3962 Read a grid from @file{pattern}, and center it on a grid of size
3965 life=f=pattern:s=300x300
3969 Generate a random grid of size 200x200, with a fill ratio of 2/3:
3971 life=ratio=2/3:s=200x200
3975 Specify a custom rule for evolving a randomly generated grid:
3981 Full example with slow death effect (mold) using @command{ffplay}:
3983 ffplay -f lavfi life=s=300x200:mold=10:r=60:ratio=0.1:death_color=#C83232:life_color=#00ff00,scale=1200:800:flags=16
3987 @section color, nullsrc, rgbtestsrc, smptebars, testsrc
3989 The @code{color} source provides an uniformly colored input.
3991 The @code{nullsrc} source returns unprocessed video frames. It is
3992 mainly useful to be employed in analysis / debugging tools, or as the
3993 source for filters which ignore the input data.
3995 The @code{rgbtestsrc} source generates an RGB test pattern useful for
3996 detecting RGB vs BGR issues. You should see a red, green and blue
3997 stripe from top to bottom.
3999 The @code{smptebars} source generates a color bars pattern, based on
4000 the SMPTE Engineering Guideline EG 1-1990.
4002 The @code{testsrc} source generates a test video pattern, showing a
4003 color pattern, a scrolling gradient and a timestamp. This is mainly
4004 intended for testing purposes.
4006 These sources accept an optional sequence of @var{key}=@var{value} pairs,
4007 separated by ":". The description of the accepted options follows.
4012 Specify the color of the source, only used in the @code{color}
4013 source. It can be the name of a color (case insensitive match) or a
4014 0xRRGGBB[AA] sequence, possibly followed by an alpha specifier. The
4015 default value is "black".
4018 Specify the size of the sourced video, it may be a string of the form
4019 @var{width}x@var{height}, or the name of a size abbreviation. The
4020 default value is "320x240".
4023 Specify the frame rate of the sourced video, as the number of frames
4024 generated per second. It has to be a string in the format
4025 @var{frame_rate_num}/@var{frame_rate_den}, an integer number, a float
4026 number or a valid video frame rate abbreviation. The default value is
4030 Set the sample aspect ratio of the sourced video.
4033 Set the video duration of the sourced video. The accepted syntax is:
4035 [-]HH[:MM[:SS[.m...]]]
4038 See also the function @code{av_parse_time()}.
4040 If not specified, or the expressed duration is negative, the video is
4041 supposed to be generated forever.
4044 Set the number of decimals to show in the timestamp, only used in the
4045 @code{testsrc} source.
4047 The displayed timestamp value will correspond to the original
4048 timestamp value multiplied by the power of 10 of the specified
4049 value. Default value is 0.
4052 For example the following:
4054 testsrc=duration=5.3:size=qcif:rate=10
4057 will generate a video with a duration of 5.3 seconds, with size
4058 176x144 and a frame rate of 10 frames per second.
4060 The following graph description will generate a red source
4061 with an opacity of 0.2, with size "qcif" and a frame rate of 10
4064 color=c=red@@0.2:s=qcif:r=10
4067 If the input content is to be ignored, @code{nullsrc} can be used. The
4068 following command generates noise in the luminance plane by employing
4069 the @code{mp=geq} filter:
4071 nullsrc=s=256x256, mp=geq=random(1)*255:128:128
4074 @c man end VIDEO SOURCES
4076 @chapter Video Sinks
4077 @c man begin VIDEO SINKS
4079 Below is a description of the currently available video sinks.
4083 Buffer video frames, and make them available to the end of the filter
4086 This sink is mainly intended for a programmatic use, in particular
4087 through the interface defined in @file{libavfilter/buffersink.h}.
4089 It does not require a string parameter in input, but you need to
4090 specify a pointer to a list of supported pixel formats terminated by
4091 -1 in the opaque parameter provided to @code{avfilter_init_filter}
4092 when initializing this sink.
4096 Null video sink, do absolutely nothing with the input video. It is
4097 mainly useful as a template and to be employed in analysis / debugging
4100 @c man end VIDEO SINKS
4102 @chapter Multimedia Filters
4103 @c man begin MULTIMEDIA FILTERS
4105 Below is a description of the currently available multimedia filters.
4109 Concatenate audio and video streams, joining them together one after the
4112 The filter works on segments of synchronized video and audio streams. All
4113 segments must have the same number of streams of each type, and that will
4114 also be the number of streams at output.
4116 The filter accepts the following named parameters:
4120 Set the number of segments. Default is 2.
4123 Set the number of output video streams, that is also the number of video
4124 streams in each segment. Default is 1.
4127 Set the number of output audio streams, that is also the number of video
4128 streams in each segment. Default is 0.
4132 The filter has @var{v}+@var{a} outputs: first @var{v} video outputs, then
4133 @var{a} audio outputs.
4135 There are @var{n}×(@var{v}+@var{a}) inputs: first the inputs for the first
4136 segment, in the same order as the outputs, then the inputs for the second
4139 Related streams do not always have exactly the same duration, for various
4140 reasons including codec frame size or sloppy authoring. For that reason,
4141 related synchronized streams (e.g. a video and its audio track) should be
4142 concatenated at once. The concat filter will use the duration of the longest
4143 stream in each segment (except the last one), and if necessary pad shorter
4144 audio streams with silence.
4146 For this filter to work correctly, all segments must start at timestamp 0.
4148 All corresponding streams must have the same parameters in all segments; the
4149 filtering system will automatically select a common pixel format for video
4150 streams, and a common sample format, sample rate and channel layout for
4151 audio streams, but other settings, such as resolution, must be converted
4152 explicitly by the user.
4154 Different frame rates are acceptable but will result in variable frame rate
4155 at output; be sure to configure the output file to handle it.
4160 Concatenate an opening, an episode and an ending, all in bilingual version
4161 (video in stream 0, audio in streams 1 and 2):
4163 ffmpeg -i opening.mkv -i episode.mkv -i ending.mkv -filter_complex \
4164 '[0:0] [0:1] [0:2] [1:0] [1:1] [1:2] [2:0] [2:1] [2:2]
4165 concat=n=3:v=1:a=2 [v] [a1] [a2]' \
4166 -map '[v]' -map '[a1]' -map '[a2]' output.mkv
4170 Concatenate two parts, handling audio and video separately, using the
4171 (a)movie sources, and adjusting the resolution:
4173 movie=part1.mp4, scale=512:288 [v1] ; amovie=part1.mp4 [a1] ;
4174 movie=part2.mp4, scale=512:288 [v2] ; amovie=part2.mp4 [a2] ;
4175 [v1] [v2] concat [outv] ; [a1] [a2] concat=v=0:a=1 [outa]
4177 Note that a desync will happen at the stitch if the audio and video streams
4178 do not have exactly the same duration in the first file.
4184 Convert input audio to a video output, representing the samples waves.
4186 The filter accepts the following named parameters:
4190 Set the number of samples which are printed on the same column. A
4191 larger value will decrease the frame rate. Must be a positive
4192 integer. This option can be set only if the value for @var{rate}
4193 is not explicitly specified.
4196 Set the (approximate) output frame rate. This is done by setting the
4197 option @var{n}. Default value is "25".
4200 Specify the video size for the output. Default value is "600x240".
4203 Some examples follow.
4206 Output the input file audio and the corresponding video representation
4209 amovie=a.mp3,asplit[out0],showwaves[out1]
4213 Create a synthetic signal and show it with showwaves, forcing a
4214 framerate of 30 frames per second:
4216 aevalsrc=sin(1*2*PI*t)*sin(880*2*PI*t):cos(2*PI*200*t),asplit[out0],showwaves=r=30[out1]
4220 @c man end MULTIMEDIA FILTERS
4222 @chapter Multimedia Sources
4223 @c man begin MULTIMEDIA SOURCES
4225 Below is a description of the currently available multimedia sources.
4229 This is the same as @ref{src_movie} source, except it selects an audio
4235 Read audio and/or video stream(s) from a movie container.
4237 It accepts the syntax: @var{movie_name}[:@var{options}] where
4238 @var{movie_name} is the name of the resource to read (not necessarily
4239 a file but also a device or a stream accessed through some protocol),
4240 and @var{options} is an optional sequence of @var{key}=@var{value}
4241 pairs, separated by ":".
4243 The description of the accepted options follows.
4247 @item format_name, f
4248 Specifies the format assumed for the movie to read, and can be either
4249 the name of a container or an input device. If not specified the
4250 format is guessed from @var{movie_name} or by probing.
4252 @item seek_point, sp
4253 Specifies the seek point in seconds, the frames will be output
4254 starting from this seek point, the parameter is evaluated with
4255 @code{av_strtod} so the numerical value may be suffixed by an IS
4256 postfix. Default value is "0".
4259 Specifies the streams to read. Several streams can be specified, separated
4260 by "+". The source will then have as many outputs, in the same order. The
4261 syntax is explained in the @ref{Stream specifiers} chapter. Two special
4262 names, "dv" and "da" specify respectively the default (best suited) video
4263 and audio stream. Default is "dv", or "da" if the filter is called as
4266 @item stream_index, si
4267 Specifies the index of the video stream to read. If the value is -1,
4268 the best suited video stream will be automatically selected. Default
4269 value is "-1". Deprecated. If the filter is called "amovie", it will select
4270 audio instead of video.
4273 Specifies how many times to read the stream in sequence.
4274 If the value is less than 1, the stream will be read again and again.
4275 Default value is "1".
4277 Note that when the movie is looped the source timestamps are not
4278 changed, so it will generate non monotonically increasing timestamps.
4281 This filter allows to overlay a second video on top of main input of
4282 a filtergraph as shown in this graph:
4284 input -----------> deltapts0 --> overlay --> output
4287 movie --> scale--> deltapts1 -------+
4290 Some examples follow.
4294 Skip 3.2 seconds from the start of the avi file in.avi, and overlay it
4295 on top of the input labelled as "in":
4297 movie=in.avi:seek_point=3.2, scale=180:-1, setpts=PTS-STARTPTS [movie];
4298 [in] setpts=PTS-STARTPTS, [movie] overlay=16:16 [out]
4302 Read from a video4linux2 device, and overlay it on top of the input
4305 movie=/dev/video0:f=video4linux2, scale=180:-1, setpts=PTS-STARTPTS [movie];
4306 [in] setpts=PTS-STARTPTS, [movie] overlay=16:16 [out]
4310 Read the first video stream and the audio stream with id 0x81 from
4311 dvd.vob; the video is connected to the pad named "video" and the audio is
4312 connected to the pad named "audio":
4314 movie=dvd.vob:s=v:0+#0x81 [video] [audio]
4318 @c man end MULTIMEDIA SOURCES