| @@ -33,19 +33,19 @@ debug mode or a NoDaemon option is specified in the configuration | |||||
| file. | file. | ||||
| This documentation covers only the streaming aspects of avserver / | This documentation covers only the streaming aspects of avserver / | ||||
| ffmpeg. All questions about parameters for ffmpeg, codec questions, | |||||
| etc. are not covered here. Read @file{ffmpeg-doc.html} for more | |||||
| avconv. All questions about parameters for avconv, codec questions, | |||||
| etc. are not covered here. Read @file{avconv.html} for more | |||||
| information. | information. | ||||
| @section How does it work? | @section How does it work? | ||||
| avserver receives prerecorded files or FFM streams from some ffmpeg | |||||
| avserver receives prerecorded files or FFM streams from some avconv | |||||
| instance as input, then streams them over RTP/RTSP/HTTP. | instance as input, then streams them over RTP/RTSP/HTTP. | ||||
| An avserver instance will listen on some port as specified in the | An avserver instance will listen on some port as specified in the | ||||
| configuration file. You can launch one or more instances of ffmpeg and | |||||
| configuration file. You can launch one or more instances of avconv and | |||||
| send one or more FFM streams to the port where avserver is expecting | send one or more FFM streams to the port where avserver is expecting | ||||
| to receive them. Alternately, you can make avserver launch such ffmpeg | |||||
| to receive them. Alternately, you can make avserver launch such avconv | |||||
| instances at startup. | instances at startup. | ||||
| Input streams are called feeds, and each one is specified by a <Feed> | Input streams are called feeds, and each one is specified by a <Feed> | ||||
| @@ -107,11 +107,11 @@ LAME is important as it allows for streaming audio to Windows Media Player. | |||||
| Don't ask why the other audio types do not work. | Don't ask why the other audio types do not work. | ||||
| As a simple test, just run the following two command lines where INPUTFILE | As a simple test, just run the following two command lines where INPUTFILE | ||||
| is some file which you can decode with ffmpeg: | |||||
| is some file which you can decode with avconv: | |||||
| @example | @example | ||||
| ./avserver -f doc/avserver.conf & | ./avserver -f doc/avserver.conf & | ||||
| ./ffmpeg -i INPUTFILE http://localhost:8090/feed1.ffm | |||||
| ./avconv -i INPUTFILE http://localhost:8090/feed1.ffm | |||||
| @end example | @end example | ||||
| At this point you should be able to go to your Windows machine and fire up | At this point you should be able to go to your Windows machine and fire up | ||||
| @@ -130,7 +130,7 @@ The same is true of AVI files. | |||||
| @section What happens next? | @section What happens next? | ||||
| You should edit the avserver.conf file to suit your needs (in terms of | You should edit the avserver.conf file to suit your needs (in terms of | ||||
| frame rates etc). Then install avserver and ffmpeg, write a script to start | |||||
| frame rates etc). Then install avserver and avconv, write a script to start | |||||
| them up, and off you go. | them up, and off you go. | ||||
| @section Troubleshooting | @section Troubleshooting | ||||
| @@ -138,13 +138,13 @@ them up, and off you go. | |||||
| @subsection I don't hear any audio, but video is fine. | @subsection I don't hear any audio, but video is fine. | ||||
| Maybe you didn't install LAME, or got your ./configure statement wrong. Check | Maybe you didn't install LAME, or got your ./configure statement wrong. Check | ||||
| the ffmpeg output to see if a line referring to MP3 is present. If not, then | |||||
| the avconv output to see if a line referring to MP3 is present. If not, then | |||||
| your configuration was incorrect. If it is, then maybe your wiring is not | your configuration was incorrect. If it is, then maybe your wiring is not | ||||
| set up correctly. Maybe the sound card is not getting data from the right | set up correctly. Maybe the sound card is not getting data from the right | ||||
| input source. Maybe you have a really awful audio interface (like I do) | input source. Maybe you have a really awful audio interface (like I do) | ||||
| that only captures in stereo and also requires that one channel be flipped. | that only captures in stereo and also requires that one channel be flipped. | ||||
| If you are one of these people, then export 'AUDIO_FLIP_LEFT=1' before | If you are one of these people, then export 'AUDIO_FLIP_LEFT=1' before | ||||
| starting ffmpeg. | |||||
| starting avconv. | |||||
| @subsection The audio and video loose sync after a while. | @subsection The audio and video loose sync after a while. | ||||
| @@ -250,7 +250,7 @@ Use @file{configfile} instead of @file{/etc/avserver.conf}. | |||||
| @item -n | @item -n | ||||
| Enable no-launch mode. This option disables all the Launch directives | Enable no-launch mode. This option disables all the Launch directives | ||||
| within the various <Stream> sections. Since avserver will not launch | within the various <Stream> sections. Since avserver will not launch | ||||
| any ffmpeg instances, you will have to launch them manually. | |||||
| any avconv instances, you will have to launch them manually. | |||||
| @item -d | @item -d | ||||
| Enable debug mode. This option increases log verbosity, directs log | Enable debug mode. This option increases log verbosity, directs log | ||||
| messages to stdout and causes avserver to run in the foreground | messages to stdout and causes avserver to run in the foreground | ||||
| @@ -265,7 +265,7 @@ rather than as a daemon. | |||||
| @c man begin SEEALSO | @c man begin SEEALSO | ||||
| avconv(1), avplay(1), avprobe(1), the @file{ffmpeg/doc/avserver.conf} | |||||
| avconv(1), avplay(1), avprobe(1), the @file{avserver.conf} | |||||
| example and the Libav HTML documentation | example and the Libav HTML documentation | ||||
| @c man end | @c man end | ||||
| @@ -132,8 +132,8 @@ libavcodec libraries. To see the list of available AVOptions, use the | |||||
| @option{-help} option. They are separated into two categories: | @option{-help} option. They are separated into two categories: | ||||
| @table @option | @table @option | ||||
| @item generic | @item generic | ||||
| These options can be set for any container, codec or device. Generic options are | |||||
| listed under AVFormatContext options for containers/devices and under | |||||
| These options can be set for any container, codec or device. Generic options | |||||
| are listed under AVFormatContext options for containers/devices and under | |||||
| AVCodecContext options for codecs. | AVCodecContext options for codecs. | ||||
| @item private | @item private | ||||
| These options are specific to the given container, device or codec. Private | These options are specific to the given container, device or codec. Private | ||||
| @@ -144,14 +144,14 @@ For example to write an ID3v2.3 header instead of a default ID3v2.4 to | |||||
| an MP3 file, use the @option{id3v2_version} private option of the MP3 | an MP3 file, use the @option{id3v2_version} private option of the MP3 | ||||
| muxer: | muxer: | ||||
| @example | @example | ||||
| ffmpeg -i input.flac -id3v2_version 3 out.mp3 | |||||
| avconv -i input.flac -id3v2_version 3 out.mp3 | |||||
| @end example | @end example | ||||
| All codec AVOptions are obviously per-stream, so the chapter on stream | All codec AVOptions are obviously per-stream, so the chapter on stream | ||||
| specifiers applies to them | specifiers applies to them | ||||
| Note -nooption syntax cannot be used for boolean AVOptions, use -option | |||||
| 0/-option 1. | |||||
| Note @option{-nooption} syntax cannot be used for boolean AVOptions, | |||||
| use @option{-option 0}/@option{-option 1}. | |||||
| Note2 old undocumented way of specifying per-stream AVOptions by prepending | Note2 old undocumented way of specifying per-stream AVOptions by prepending | ||||
| v/a/s to the options name is now obsolete and will be removed soon. | v/a/s to the options name is now obsolete and will be removed soon. | ||||
| @@ -34,7 +34,7 @@ JPEG image. The individual frames can be extracted without loss, | |||||
| e.g. by | e.g. by | ||||
| @example | @example | ||||
| ffmpeg -i ../some_mjpeg.avi -vcodec copy frames_%d.jpg | |||||
| avconv -i ../some_mjpeg.avi -c:v copy frames_%d.jpg | |||||
| @end example | @end example | ||||
| Unfortunately, these chunks are incomplete JPEG images, because | Unfortunately, these chunks are incomplete JPEG images, because | ||||
| @@ -57,9 +57,9 @@ stream (carrying the AVI1 header ID and lacking a DHT segment) to | |||||
| produce fully qualified JPEG images. | produce fully qualified JPEG images. | ||||
| @example | @example | ||||
| ffmpeg -i mjpeg-movie.avi -vcodec copy -vbsf mjpeg2jpeg frame_%d.jpg | |||||
| avconv -i mjpeg-movie.avi -c:v copy -vbsf mjpeg2jpeg frame_%d.jpg | |||||
| exiftran -i -9 frame*.jpg | exiftran -i -9 frame*.jpg | ||||
| ffmpeg -i frame_%d.jpg -vcodec copy rotated.avi | |||||
| avconv -i frame_%d.jpg -c:v copy rotated.avi | |||||
| @end example | @end example | ||||
| @section mjpega_dump_header | @section mjpega_dump_header | ||||
| @@ -42,10 +42,10 @@ specify card number or identifier, device number and subdevice number | |||||
| To see the list of cards currently recognized by your system check the | To see the list of cards currently recognized by your system check the | ||||
| files @file{/proc/asound/cards} and @file{/proc/asound/devices}. | files @file{/proc/asound/cards} and @file{/proc/asound/devices}. | ||||
| For example to capture with @file{ffmpeg} from an ALSA device with | |||||
| For example to capture with @command{avconv} from an ALSA device with | |||||
| card id 0, you may run the command: | card id 0, you may run the command: | ||||
| @example | @example | ||||
| ffmpeg -f alsa -i hw:0 alsaout.wav | |||||
| avconv -f alsa -i hw:0 alsaout.wav | |||||
| @end example | @end example | ||||
| For more information see: | For more information see: | ||||
| @@ -72,14 +72,14 @@ For more detailed information read the file | |||||
| Documentation/fb/framebuffer.txt included in the Linux source tree. | Documentation/fb/framebuffer.txt included in the Linux source tree. | ||||
| To record from the framebuffer device @file{/dev/fb0} with | To record from the framebuffer device @file{/dev/fb0} with | ||||
| @file{ffmpeg}: | |||||
| @command{avconv}: | |||||
| @example | @example | ||||
| ffmpeg -f fbdev -r 10 -i /dev/fb0 out.avi | |||||
| avconv -f fbdev -r 10 -i /dev/fb0 out.avi | |||||
| @end example | @end example | ||||
| You can take a single screenshot image with the command: | You can take a single screenshot image with the command: | ||||
| @example | @example | ||||
| ffmpeg -f fbdev -vframes 1 -r 1 -i /dev/fb0 screenshot.jpeg | |||||
| avconv -f fbdev -frames:v 1 -r 1 -i /dev/fb0 screenshot.jpeg | |||||
| @end example | @end example | ||||
| See also @url{http://linux-fbdev.sourceforge.net/}, and fbset(1). | See also @url{http://linux-fbdev.sourceforge.net/}, and fbset(1). | ||||
| @@ -109,10 +109,10 @@ To list the JACK clients and their properties you can invoke the command | |||||
| @file{jack_lsp}. | @file{jack_lsp}. | ||||
| Follows an example which shows how to capture a JACK readable client | Follows an example which shows how to capture a JACK readable client | ||||
| with @file{ffmpeg}. | |||||
| with @command{avconv}. | |||||
| @example | @example | ||||
| # Create a JACK writable client with name "ffmpeg". | |||||
| $ ffmpeg -f jack -i ffmpeg -y out.wav | |||||
| # Create a JACK writable client with name "libav". | |||||
| $ avconv -f jack -i libav -y out.wav | |||||
| # Start the sample jack_metro readable client. | # Start the sample jack_metro readable client. | ||||
| $ jack_metro -b 120 -d 0.2 -f 4000 | $ jack_metro -b 120 -d 0.2 -f 4000 | ||||
| @@ -123,11 +123,11 @@ system:capture_1 | |||||
| system:capture_2 | system:capture_2 | ||||
| system:playback_1 | system:playback_1 | ||||
| system:playback_2 | system:playback_2 | ||||
| ffmpeg:input_1 | |||||
| libav:input_1 | |||||
| metro:120_bpm | metro:120_bpm | ||||
| # Connect metro to the ffmpeg writable client. | |||||
| $ jack_connect metro:120_bpm ffmpeg:input_1 | |||||
| # Connect metro to the avconv writable client. | |||||
| $ jack_connect metro:120_bpm libav:input_1 | |||||
| @end example | @end example | ||||
| For more information read: | For more information read: | ||||
| @@ -145,10 +145,10 @@ The filename to provide to the input device is the device node | |||||
| representing the OSS input device, and is usually set to | representing the OSS input device, and is usually set to | ||||
| @file{/dev/dsp}. | @file{/dev/dsp}. | ||||
| For example to grab from @file{/dev/dsp} using @file{ffmpeg} use the | |||||
| For example to grab from @file{/dev/dsp} using @command{avconv} use the | |||||
| command: | command: | ||||
| @example | @example | ||||
| ffmpeg -f oss -i /dev/dsp /tmp/oss.wav | |||||
| avconv -f oss -i /dev/dsp /tmp/oss.wav | |||||
| @end example | @end example | ||||
| For more information about OSS see: | For more information about OSS see: | ||||
| @@ -248,10 +248,10 @@ The filename to provide to the input device is the device node | |||||
| representing the sndio input device, and is usually set to | representing the sndio input device, and is usually set to | ||||
| @file{/dev/audio0}. | @file{/dev/audio0}. | ||||
| For example to grab from @file{/dev/audio0} using @file{ffmpeg} use the | |||||
| For example to grab from @file{/dev/audio0} using @command{avconv} use the | |||||
| command: | command: | ||||
| @example | @example | ||||
| ffmpeg -f sndio -i /dev/audio0 /tmp/oss.wav | |||||
| avconv -f sndio -i /dev/audio0 /tmp/oss.wav | |||||
| @end example | @end example | ||||
| @section video4linux and video4linux2 | @section video4linux and video4linux2 | ||||
| @@ -290,7 +290,7 @@ avplay -f video4linux2 /dev/video0 | |||||
| # Grab and record the input of a video4linux2 device, autoadjust size, | # Grab and record the input of a video4linux2 device, autoadjust size, | ||||
| # frame rate value defaults to 0/0 so it is read from the video4linux2 | # frame rate value defaults to 0/0 so it is read from the video4linux2 | ||||
| # driver. | # driver. | ||||
| ffmpeg -f video4linux2 -i /dev/video0 out.mpeg | |||||
| avconv -f video4linux2 -i /dev/video0 out.mpeg | |||||
| @end example | @end example | ||||
| @section vfwcap | @section vfwcap | ||||
| @@ -326,12 +326,12 @@ Check the X11 documentation (e.g. man X) for more detailed information. | |||||
| Use the @file{dpyinfo} program for getting basic information about the | Use the @file{dpyinfo} program for getting basic information about the | ||||
| properties of your X11 display (e.g. grep for "name" or "dimensions"). | properties of your X11 display (e.g. grep for "name" or "dimensions"). | ||||
| For example to grab from @file{:0.0} using @file{ffmpeg}: | |||||
| For example to grab from @file{:0.0} using @command{avconv}: | |||||
| @example | @example | ||||
| ffmpeg -f x11grab -r 25 -s cif -i :0.0 out.mpg | |||||
| avconv -f x11grab -r 25 -s cif -i :0.0 out.mpg | |||||
| # Grab at position 10,20. | # Grab at position 10,20. | ||||
| ffmpeg -f x11grab -r 25 -s cif -i :0.0+10,20 out.mpg | |||||
| avconv -f x11grab -r 25 -s cif -i :0.0+10,20 out.mpg | |||||
| @end example | @end example | ||||
| @subsection @var{follow_mouse} AVOption | @subsection @var{follow_mouse} AVOption | ||||
| @@ -348,10 +348,10 @@ zero) to the edge of region. | |||||
| For example: | For example: | ||||
| @example | @example | ||||
| ffmpeg -f x11grab -follow_mouse centered -r 25 -s cif -i :0.0 out.mpg | |||||
| avconv -f x11grab -follow_mouse centered -r 25 -s cif -i :0.0 out.mpg | |||||
| # Follows only when the mouse pointer reaches within 100 pixels to edge | # Follows only when the mouse pointer reaches within 100 pixels to edge | ||||
| ffmpeg -f x11grab -follow_mouse 100 -r 25 -s cif -i :0.0 out.mpg | |||||
| avconv -f x11grab -follow_mouse 100 -r 25 -s cif -i :0.0 out.mpg | |||||
| @end example | @end example | ||||
| @subsection @var{show_region} AVOption | @subsection @var{show_region} AVOption | ||||
| @@ -367,10 +367,10 @@ being grabbed if only a portion of the screen is grabbed. | |||||
| For example: | For example: | ||||
| @example | @example | ||||
| ffmpeg -f x11grab -show_region 1 -r 25 -s cif -i :0.0+10,20 out.mpg | |||||
| avconv -f x11grab -show_region 1 -r 25 -s cif -i :0.0+10,20 out.mpg | |||||
| # With follow_mouse | # With follow_mouse | ||||
| ffmpeg -f x11grab -follow_mouse centered -show_region 1 -r 25 -s cif -i :0.0 out.mpg | |||||
| avconv -f x11grab -follow_mouse centered -show_region 1 -r 25 -s cif -i :0.0 out.mpg | |||||
| @end example | @end example | ||||
| @c man end INPUT DEVICES | @c man end INPUT DEVICES | ||||
| @@ -35,20 +35,20 @@ CRC=0x@var{CRC}, where @var{CRC} is a hexadecimal number 0-padded to | |||||
| For example to compute the CRC of the input, and store it in the file | For example to compute the CRC of the input, and store it in the file | ||||
| @file{out.crc}: | @file{out.crc}: | ||||
| @example | @example | ||||
| ffmpeg -i INPUT -f crc out.crc | |||||
| avconv -i INPUT -f crc out.crc | |||||
| @end example | @end example | ||||
| You can print the CRC to stdout with the command: | You can print the CRC to stdout with the command: | ||||
| @example | @example | ||||
| ffmpeg -i INPUT -f crc - | |||||
| avconv -i INPUT -f crc - | |||||
| @end example | @end example | ||||
| You can select the output format of each frame with @file{ffmpeg} by | |||||
| You can select the output format of each frame with @command{avconv} by | |||||
| specifying the audio and video codec and format. For example to | specifying the audio and video codec and format. For example to | ||||
| compute the CRC of the input audio converted to PCM unsigned 8-bit | compute the CRC of the input audio converted to PCM unsigned 8-bit | ||||
| and the input video converted to MPEG-2 video, use the command: | and the input video converted to MPEG-2 video, use the command: | ||||
| @example | @example | ||||
| ffmpeg -i INPUT -acodec pcm_u8 -vcodec mpeg2video -f crc - | |||||
| avconv -i INPUT -c:a pcm_u8 -c:v mpeg2video -f crc - | |||||
| @end example | @end example | ||||
| See also the @ref{framecrc} muxer. | See also the @ref{framecrc} muxer. | ||||
| @@ -71,21 +71,21 @@ number 0-padded to 8 digits containing the CRC of the decoded frame. | |||||
| For example to compute the CRC of each decoded frame in the input, and | For example to compute the CRC of each decoded frame in the input, and | ||||
| store it in the file @file{out.crc}: | store it in the file @file{out.crc}: | ||||
| @example | @example | ||||
| ffmpeg -i INPUT -f framecrc out.crc | |||||
| avconv -i INPUT -f framecrc out.crc | |||||
| @end example | @end example | ||||
| You can print the CRC of each decoded frame to stdout with the command: | You can print the CRC of each decoded frame to stdout with the command: | ||||
| @example | @example | ||||
| ffmpeg -i INPUT -f framecrc - | |||||
| avconv -i INPUT -f framecrc - | |||||
| @end example | @end example | ||||
| You can select the output format of each frame with @file{ffmpeg} by | |||||
| You can select the output format of each frame with @command{avconv} by | |||||
| specifying the audio and video codec and format. For example, to | specifying the audio and video codec and format. For example, to | ||||
| compute the CRC of each decoded input audio frame converted to PCM | compute the CRC of each decoded input audio frame converted to PCM | ||||
| unsigned 8-bit and of each decoded input video frame converted to | unsigned 8-bit and of each decoded input video frame converted to | ||||
| MPEG-2 video, use the command: | MPEG-2 video, use the command: | ||||
| @example | @example | ||||
| ffmpeg -i INPUT -acodec pcm_u8 -vcodec mpeg2video -f framecrc - | |||||
| avconv -i INPUT -c:a pcm_u8 -c:v mpeg2video -f framecrc - | |||||
| @end example | @end example | ||||
| See also the @ref{crc} muxer. | See also the @ref{crc} muxer. | ||||
| @@ -119,26 +119,26 @@ The pattern "img%%-%d.jpg" will specify a sequence of filenames of the | |||||
| form @file{img%-1.jpg}, @file{img%-2.jpg}, ..., @file{img%-10.jpg}, | form @file{img%-1.jpg}, @file{img%-2.jpg}, ..., @file{img%-10.jpg}, | ||||
| etc. | etc. | ||||
| The following example shows how to use @file{ffmpeg} for creating a | |||||
| The following example shows how to use @command{avconv} for creating a | |||||
| sequence of files @file{img-001.jpeg}, @file{img-002.jpeg}, ..., | sequence of files @file{img-001.jpeg}, @file{img-002.jpeg}, ..., | ||||
| taking one image every second from the input video: | taking one image every second from the input video: | ||||
| @example | @example | ||||
| ffmpeg -i in.avi -r 1 -f image2 'img-%03d.jpeg' | |||||
| avconv -i in.avi -vsync 1 -r 1 -f image2 'img-%03d.jpeg' | |||||
| @end example | @end example | ||||
| Note that with @file{ffmpeg}, if the format is not specified with the | |||||
| Note that with @command{avconv}, if the format is not specified with the | |||||
| @code{-f} option and the output filename specifies an image file | @code{-f} option and the output filename specifies an image file | ||||
| format, the image2 muxer is automatically selected, so the previous | format, the image2 muxer is automatically selected, so the previous | ||||
| command can be written as: | command can be written as: | ||||
| @example | @example | ||||
| ffmpeg -i in.avi -r 1 'img-%03d.jpeg' | |||||
| avconv -i in.avi -vsync 1 -r 1 'img-%03d.jpeg' | |||||
| @end example | @end example | ||||
| Note also that the pattern must not necessarily contain "%d" or | Note also that the pattern must not necessarily contain "%d" or | ||||
| "%0@var{N}d", for example to create a single image file | "%0@var{N}d", for example to create a single image file | ||||
| @file{img.jpeg} from the input video you can employ the command: | @file{img.jpeg} from the input video you can employ the command: | ||||
| @example | @example | ||||
| ffmpeg -i in.avi -f image2 -vframes 1 img.jpeg | |||||
| avconv -i in.avi -f image2 -frames:v 1 img.jpeg | |||||
| @end example | @end example | ||||
| @section mpegts | @section mpegts | ||||
| @@ -171,7 +171,7 @@ and @code{service_name}. If they are not set the default for | |||||
| @code{service_name} is "Service01". | @code{service_name} is "Service01". | ||||
| @example | @example | ||||
| ffmpeg -i file.mpg -acodec copy -vcodec copy \ | |||||
| avconv -i file.mpg -c copy \ | |||||
| -mpegts_original_network_id 0x1122 \ | -mpegts_original_network_id 0x1122 \ | ||||
| -mpegts_transport_stream_id 0x3344 \ | -mpegts_transport_stream_id 0x3344 \ | ||||
| -mpegts_service_id 0x5566 \ | -mpegts_service_id 0x5566 \ | ||||
| @@ -189,19 +189,19 @@ Null muxer. | |||||
| This muxer does not generate any output file, it is mainly useful for | This muxer does not generate any output file, it is mainly useful for | ||||
| testing or benchmarking purposes. | testing or benchmarking purposes. | ||||
| For example to benchmark decoding with @file{ffmpeg} you can use the | |||||
| For example to benchmark decoding with @command{avconv} you can use the | |||||
| command: | command: | ||||
| @example | @example | ||||
| ffmpeg -benchmark -i INPUT -f null out.null | |||||
| avconv -benchmark -i INPUT -f null out.null | |||||
| @end example | @end example | ||||
| Note that the above command does not read or write the @file{out.null} | Note that the above command does not read or write the @file{out.null} | ||||
| file, but specifying the output file is required by the @file{ffmpeg} | |||||
| file, but specifying the output file is required by the @command{avconv} | |||||
| syntax. | syntax. | ||||
| Alternatively you can write the command as: | Alternatively you can write the command as: | ||||
| @example | @example | ||||
| ffmpeg -benchmark -i INPUT -f null - | |||||
| avconv -benchmark -i INPUT -f null - | |||||
| @end example | @end example | ||||
| @section matroska | @section matroska | ||||
| @@ -264,7 +264,7 @@ Both eyes laced in one Block, Right-eye view is first | |||||
| For example a 3D WebM clip can be created using the following command line: | For example a 3D WebM clip can be created using the following command line: | ||||
| @example | @example | ||||
| ffmpeg -i sample_left_right_clip.mpg -an -vcodec libvpx -metadata STEREO_MODE=left_right -y stereo_clip.webm | |||||
| avconv -i sample_left_right_clip.mpg -an -c:v libvpx -metadata STEREO_MODE=left_right -y stereo_clip.webm | |||||
| @end example | @end example | ||||
| @c man end MUXERS | @c man end MUXERS | ||||
| @@ -67,10 +67,10 @@ File access protocol. | |||||
| Allow to read from or read to a file. | Allow to read from or read to a file. | ||||
| For example to read from a file @file{input.mpeg} with @file{ffmpeg} | |||||
| For example to read from a file @file{input.mpeg} with @command{avconv} | |||||
| use the command: | use the command: | ||||
| @example | @example | ||||
| ffmpeg -i file:input.mpeg output.mpeg | |||||
| avconv -i file:input.mpeg output.mpeg | |||||
| @end example | @end example | ||||
| The ff* tools default to the file protocol, that is a resource | The ff* tools default to the file protocol, that is a resource | ||||
| @@ -109,10 +109,10 @@ be used to test muxers without writing an actual file. | |||||
| Some examples follow. | Some examples follow. | ||||
| @example | @example | ||||
| # Write the MD5 hash of the encoded AVI file to the file output.avi.md5. | # Write the MD5 hash of the encoded AVI file to the file output.avi.md5. | ||||
| ffmpeg -i input.flv -f avi -y md5:output.avi.md5 | |||||
| avconv -i input.flv -f avi -y md5:output.avi.md5 | |||||
| # Write the MD5 hash of the encoded AVI file to stdout. | # Write the MD5 hash of the encoded AVI file to stdout. | ||||
| ffmpeg -i input.flv -f avi -y md5: | |||||
| avconv -i input.flv -f avi -y md5: | |||||
| @end example | @end example | ||||
| Note that some formats (typically MOV) require the output protocol to | Note that some formats (typically MOV) require the output protocol to | ||||
| @@ -134,18 +134,18 @@ pipe (e.g. 0 for stdin, 1 for stdout, 2 for stderr). If @var{number} | |||||
| is not specified, by default the stdout file descriptor will be used | is not specified, by default the stdout file descriptor will be used | ||||
| for writing, stdin for reading. | for writing, stdin for reading. | ||||
| For example to read from stdin with @file{ffmpeg}: | |||||
| For example to read from stdin with @command{avconv}: | |||||
| @example | @example | ||||
| cat test.wav | ffmpeg -i pipe:0 | |||||
| cat test.wav | avconv -i pipe:0 | |||||
| # ...this is the same as... | # ...this is the same as... | ||||
| cat test.wav | ffmpeg -i pipe: | |||||
| cat test.wav | avconv -i pipe: | |||||
| @end example | @end example | ||||
| For writing to stdout with @file{ffmpeg}: | |||||
| For writing to stdout with @command{avconv}: | |||||
| @example | @example | ||||
| ffmpeg -i test.wav -f avi pipe:1 | cat > test.avi | |||||
| avconv -i test.wav -f avi pipe:1 | cat > test.avi | |||||
| # ...this is the same as... | # ...this is the same as... | ||||
| ffmpeg -i test.wav -f avi pipe: | cat > test.avi | |||||
| avconv -i test.wav -f avi pipe: | cat > test.avi | |||||
| @end example | @end example | ||||
| Note that some formats (typically MOV), require the output protocol to | Note that some formats (typically MOV), require the output protocol to | ||||
| @@ -219,9 +219,9 @@ meaning as specified for the RTMP native protocol. | |||||
| See the librtmp manual page (man 3 librtmp) for more information. | See the librtmp manual page (man 3 librtmp) for more information. | ||||
| For example, to stream a file in real-time to an RTMP server using | For example, to stream a file in real-time to an RTMP server using | ||||
| @file{ffmpeg}: | |||||
| @command{avconv}: | |||||
| @example | @example | ||||
| ffmpeg -re -i myfile -f flv rtmp://myserver/live/mystream | |||||
| avconv -re -i myfile -f flv rtmp://myserver/live/mystream | |||||
| @end example | @end example | ||||
| To play the same stream using @file{avplay}: | To play the same stream using @file{avplay}: | ||||
| @@ -249,7 +249,7 @@ The required syntax for a RTSP url is: | |||||
| rtsp://@var{hostname}[:@var{port}]/@var{path} | rtsp://@var{hostname}[:@var{port}]/@var{path} | ||||
| @end example | @end example | ||||
| The following options (set on the @file{avconv}/@file{avplay} command | |||||
| The following options (set on the @command{avconv}/@file{avplay} command | |||||
| line, or set in code via @code{AVOption}s or in @code{avformat_open_input}), | line, or set in code via @code{AVOption}s or in @code{avformat_open_input}), | ||||
| are supported: | are supported: | ||||
| @@ -310,7 +310,7 @@ avplay -rtsp_transport http rtsp://server/video.mp4 | |||||
| To send a stream in realtime to a RTSP server, for others to watch: | To send a stream in realtime to a RTSP server, for others to watch: | ||||
| @example | @example | ||||
| ffmpeg -re -i @var{input} -f rtsp -muxdelay 0.1 rtsp://server/live.sdp | |||||
| avconv -re -i @var{input} -f rtsp -muxdelay 0.1 rtsp://server/live.sdp | |||||
| @end example | @end example | ||||
| @section sap | @section sap | ||||
| @@ -362,19 +362,19 @@ Example command lines follow. | |||||
| To broadcast a stream on the local subnet, for watching in VLC: | To broadcast a stream on the local subnet, for watching in VLC: | ||||
| @example | @example | ||||
| ffmpeg -re -i @var{input} -f sap sap://224.0.0.255?same_port=1 | |||||
| avconv -re -i @var{input} -f sap sap://224.0.0.255?same_port=1 | |||||
| @end example | @end example | ||||
| Similarly, for watching in avplay: | Similarly, for watching in avplay: | ||||
| @example | @example | ||||
| ffmpeg -re -i @var{input} -f sap sap://224.0.0.255 | |||||
| avconv -re -i @var{input} -f sap sap://224.0.0.255 | |||||
| @end example | @end example | ||||
| And for watching in avplay, over IPv6: | And for watching in avplay, over IPv6: | ||||
| @example | @example | ||||
| ffmpeg -re -i @var{input} -f sap sap://[ff0e::1:2:3:4] | |||||
| avconv -re -i @var{input} -f sap sap://[ff0e::1:2:3:4] | |||||
| @end example | @end example | ||||
| @subsection Demuxer | @subsection Demuxer | ||||
| @@ -420,7 +420,7 @@ tcp://@var{hostname}:@var{port}[?@var{options}] | |||||
| Listen for an incoming connection | Listen for an incoming connection | ||||
| @example | @example | ||||
| ffmpeg -i @var{input} -f @var{format} tcp://@var{hostname}:@var{port}?listen | |||||
| avconv -i @var{input} -f @var{format} tcp://@var{hostname}:@var{port}?listen | |||||
| avplay tcp://@var{hostname}:@var{port} | avplay tcp://@var{hostname}:@var{port} | ||||
| @end example | @end example | ||||
| @@ -472,21 +472,21 @@ For receiving, this gives the benefit of only receiving packets from | |||||
| the specified peer address/port. | the specified peer address/port. | ||||
| @end table | @end table | ||||
| Some usage examples of the udp protocol with @file{ffmpeg} follow. | |||||
| Some usage examples of the udp protocol with @command{avconv} follow. | |||||
| To stream over UDP to a remote endpoint: | To stream over UDP to a remote endpoint: | ||||
| @example | @example | ||||
| ffmpeg -i @var{input} -f @var{format} udp://@var{hostname}:@var{port} | |||||
| avconv -i @var{input} -f @var{format} udp://@var{hostname}:@var{port} | |||||
| @end example | @end example | ||||
| To stream in mpegts format over UDP using 188 sized UDP packets, using a large input buffer: | To stream in mpegts format over UDP using 188 sized UDP packets, using a large input buffer: | ||||
| @example | @example | ||||
| ffmpeg -i @var{input} -f mpegts udp://@var{hostname}:@var{port}?pkt_size=188&buffer_size=65535 | |||||
| avconv -i @var{input} -f mpegts udp://@var{hostname}:@var{port}?pkt_size=188&buffer_size=65535 | |||||
| @end example | @end example | ||||
| To receive over UDP from a remote endpoint: | To receive over UDP from a remote endpoint: | ||||
| @example | @example | ||||
| ffmpeg -i udp://[@var{multicast-address}]:@var{port} | |||||
| avconv -i udp://[@var{multicast-address}]:@var{port} | |||||
| @end example | @end example | ||||
| @c man end PROTOCOLS | @c man end PROTOCOLS | ||||