Update the last basic tutorials

This commit is contained in:
Olivier Crête 2016-06-15 20:00:24 -04:00
parent ab17d47d29
commit 0a973622a9
10 changed files with 458 additions and 619 deletions

View file

@ -1,5 +1,10 @@
# Basic tutorial 15: Clutter integration # Basic tutorial 15: Clutter integration
> ![Warning](images/icons/emoticons/warning.png)
>
> **THIS TUTORIAL HAS NOT BEEN UPDATED, CLUTTER IS DEPRECATED, DON'T USE IT**
# Goal # Goal
“[Clutter](https://clutter-project.org/) is an open source software “[Clutter](https://clutter-project.org/) is an open source software

View file

@ -1,6 +1,6 @@
# Basic tutorial 11: Debugging tools # Basic tutorial 11: Debugging tools
# Goal ## Goal
Sometimes things wont go as expected and the error messages retrieved Sometimes things wont go as expected and the error messages retrieved
from the bus (if any) just dont provide enough information. Luckily, from the bus (if any) just dont provide enough information. Luckily,
@ -13,7 +13,7 @@ hint what the problem might be. This tutorial shows:
- How to get pipeline graphs - How to get pipeline graphs
# Printing debug information ## Printing debug information
### The debug log ### The debug log
@ -22,7 +22,7 @@ the code where a particularly interesting piece of information is
printed to the console, along with time stamping, process, category, printed to the console, along with time stamping, process, category,
source code file, function and element information. source code file, function and element information.
The debug output is controlled with the `GST_DEBUG` environment The debug output is controlled with the `GST_DEBUG` environment
variable. Heres an example with variable. Heres an example with
`GST_DEBUG=2`: `GST_DEBUG=2`:
@ -40,54 +40,43 @@ at once.
The first category is the Debug Level, which is a number specifying the The first category is the Debug Level, which is a number specifying the
amount of desired output: amount of desired output:
<table> | # | Name | Description |
<tbody> |---|---------|---|
<tr class="odd"> | 0 | none | No debug information is output. |
<td><span style="color: rgb(192,192,192);">0</span></td> | 1 | ERROR | Logs all fatal errors. These are errors that do not allow the core or elements to perform the requested action. The application can still recover if programmed to handle the conditions that triggered the error. |
</tr> | 2 | WARNING | Logs all warnings. Typically these are non-fatal, but user-visible problems are expected to happen. |
<tr class="even"> | 3 | FIXME | Logs all "fixme" messages. Those typically that a codepath that is known to be incomplete has been triggered. It may work in most cases, but mauy cause problems in specific instances. |
<td><span style="color: rgb(192,192,192);">1</span></td> | 4 | INFO | Logs all informational messages. These are typically used for events in the system that only happen once, or are important and rare enough to be logged at this level. |
</tr> | 5 | DEBUG | Logs all debug messages. These are general debug messages for events that happen only a limited number of times during an object's lifetime; these include setup, teardown, change of parameters, ... |
<tr class="odd"> | 6 | LOG | Logs all log messages. These are messages for events that happen repeatedly during an object's lifetime; these include streaming and steady-state conditions. This is used for log messages that happen on every buffer in an element for example. |
<td><span style="color: rgb(192,192,192);">2</span></td> | 7 | TRACE | Logs all trace messages. Those are message that happen very very often. This is for example is each each time the reference count of a GstMiniObject, such as a GstBuffer or GstEvent, is modified. |
</tr> | 8 | MEMDUMP | Logs all memory dump messages. This is the heaviest logging and may include dumping the content of blocks of memory. |
<tr class="even">
<td><span style="color: rgb(192,192,192);">3</span></td>
</tr>
<tr class="odd">
<td><span style="color: rgb(192,192,192);">4</span></td>
</tr>
<tr class="even">
<td><span style="color: rgb(192,192,192);">5</span></td>
</tr>
</tbody>
</table>
To enable debug output, set the `GST_DEBUG` environment variable to the To enable debug output, set the `GST_DEBUG` environment variable to the
desired debug level. All levels below that will also be shown (i.e., if desired debug level. All levels below that will also be shown (i.e., if
you set `GST_DEBUG=2`, you will get both `ERROR` and you set `GST_DEBUG=2`, you will get both `ERROR` and
`WARNING` messages). `WARNING` messages).
Furthermore, each plugin or part of the GStreamer defines its own Furthermore, each plugin or part of the GStreamer defines its own
category, so you can specify a debug level for each individual category. category, so you can specify a debug level for each individual category.
For example, `GST_DEBUG=2,audiotestsrc:5`, will use Debug Level 5 for For example, `GST_DEBUG=2,audiotestsrc:6`, will use Debug Level 6 for
the `audiotestsrc` element, and 2 for all the others. the `audiotestsrc` element, and 2 for all the others.
The `GST_DEBUG` environment variable, then, is a comma-separated list of The `GST_DEBUG` environment variable, then, is a comma-separated list of
*category*:*level* pairs, with an optional *level* at the beginning, *category*:*level* pairs, with an optional *level* at the beginning,
representing the default debug level for all categories. representing the default debug level for all categories.
The `'*'` wildcard is also available. For example The `'*'` wildcard is also available. For example
`GST_DEBUG=2,audio*:5` will use Debug Level 5 for all categories `GST_DEBUG=2,audio*:6` will use Debug Level 5 for all categories
starting with the word `audio`. `GST_DEBUG=*:2` is equivalent to starting with the word `audio`. `GST_DEBUG=*:2` is equivalent to
`GST_DEBUG=2`. `GST_DEBUG=2`.
Use `gst-launch-1.0 --gst-debug-help` to obtain the list of all Use `gst-launch-1.0 --gst-debug-help` to obtain the list of all
registered categories. Bear in mind that each plugin registers its own registered categories. Bear in mind that each plugin registers its own
categories, so, when installing or removing plugins, this list can categories, so, when installing or removing plugins, this list can
change. change.
Use `GST_DEBUG` when the error information posted on the GStreamer bus Use `GST_DEBUG` when the error information posted on the GStreamer bus
does not help you nail down a problem. It is common practice to redirect does not help you nail down a problem. It is common practice to redirect
the output log to a file, and then examine it later, searching for the output log to a file, and then examine it later, searching for
specific messages. specific messages.
@ -99,48 +88,18 @@ is:
0:00:00.868050000 1592 09F62420 WARN filesrc gstfilesrc.c:1044:gst_file_src_start:<filesrc0> error: No such file "non-existing-file.webm" 0:00:00.868050000 1592 09F62420 WARN filesrc gstfilesrc.c:1044:gst_file_src_start:<filesrc0> error: No such file "non-existing-file.webm"
``` ```
<table>
<thead> | Example | Explained |
<tr class="header"> |---------------------|-----------|
<th><code>0:00:00.868050000</code></th> | `0:00:00.868050000` | Time stamp in HH:MM:SS.sssssssss format since the start of the program |
<th>Time stamp in HH:MM:SS.sssssssss format since the start of the program</th> | `1592` | Process ID from which the message was issued. Useful when your problem involves multiple processes |
</tr> | `09F62420` | Thread ID from which the message was issued. Useful when your problem involves multiple threads |
</thead> | `WARN` | Debug level of the message |
<tbody> | `filesrc` | Debug Category of the message |
<tr class="odd"> | `gstfilesrc.c:1044` | Source file and line in the GStreamer source code where this message is printed |
<td><code>1592</code></td> | `gst_file_src_start`| Function from which the message was issued |
<td>Process ID from which the message was issued. Useful when your problem involves multiple processes</td> | `&lt;filesrc0&gt;` | Name of the object that issued the message. It can be an element, a Pad, or something else. Useful when you have multiple elements of the same kind and need to distinguish among them. Naming your elements with the name property will make this debug output more readable (otherwise, GStreamer assigns each new element a unique name). |
</tr> | `error: No such file &quot;non-existing-file.webm&quot;` | The actual message.|
<tr class="even">
<td><code>09F62420</code></td>
<td><span>Thread ID from which the message was issued. Useful when your problem involves multiple threads</span></td>
</tr>
<tr class="odd">
<td><code>WARN</code></td>
<td>Debug level of the message</td>
</tr>
<tr class="even">
<td><code>filesrc</code></td>
<td>Debug Category of the message</td>
</tr>
<tr class="odd">
<td><code>gstfilesrc.c:1044</code></td>
<td>Source file and line in the GStreamer source code where this message is printed</td>
</tr>
<tr class="even">
<td><code>gst_file_src_start</code></td>
<td>Function from which the message was issued</td>
</tr>
<tr class="odd">
<td><code>&lt;filesrc0&gt;</code></td>
<td>Name of the object that issued the message. It can be an element, a Pad, or something else. Useful when you have multiple elements of the same kind and need to distinguish among them. Naming your elements with the name property will make this debug output more readable (otherwise, GStreamer assigns each new element a unique name).</td>
</tr>
<tr class="even">
<td><code>error: No such file &quot;non-existing-file.webm&quot;</code></td>
<td>The actual message.</td>
</tr>
</tbody>
</table>
### Adding your own debug information ### Adding your own debug information
@ -150,8 +109,8 @@ have all debug output in the same file and the temporal relationship
between different messages is preserved. between different messages is preserved.
To do so, use the `GST_ERROR()`, `GST_WARNING()`, `GST_INFO()`, To do so, use the `GST_ERROR()`, `GST_WARNING()`, `GST_INFO()`,
`GST_LOG()` and `GST_DEBUG()` macros. They accept the same parameters as `GST_LOG()` and `GST_DEBUG()` macros. They accept the same parameters as
`printf`, and they use the `default` category (`default` will be shown `printf`, and they use the `default` category (`default` will be shown
as the Debug category in the output log). as the Debug category in the output log).
To change the category to something more meaningful, add these two lines To change the category to something more meaningful, add these two lines
@ -171,34 +130,34 @@ GST_DEBUG_CATEGORY_INIT (my_category, "my category", 0, "This is my very own");
This registers a new category (this is, for the duration of your This registers a new category (this is, for the duration of your
application: it is not stored in any file), and sets it as the default application: it is not stored in any file), and sets it as the default
category for your code. See the documentation category for your code. See the documentation
for `GST_DEBUG_CATEGORY_INIT()`. for `GST_DEBUG_CATEGORY_INIT()`.
### Getting pipeline graphs ### Getting pipeline graphs
For those cases where your pipeline starts to grow too large and you For those cases where your pipeline starts to grow too large and you
lose track of what is connected with what, GStreamer has the capability lose track of what is connected with what, GStreamer has the capability
to output graph files. These are `.dot` files, readable with free to output graph files. These are `.dot` files, readable with free
programs like [GraphViz](http://www.graphviz.org), that describe the programs like [GraphViz](http://www.graphviz.org), that describe the
topology of your pipeline, along with the caps negotiated in each link. topology of your pipeline, along with the caps negotiated in each link.
This is also very handy when using all-in-one elements like `playbin` This is also very handy when using all-in-one elements like `playbin`
 or `uridecodebin`, which instantiate several elements inside them. Use or `uridecodebin`, which instantiate several elements inside them. Use
the `.dot` files to learn what pipeline they have created inside (and the `.dot` files to learn what pipeline they have created inside (and
learn a bit of GStreamer along the way). learn a bit of GStreamer along the way).
To obtain `.dot` files, simply set To obtain `.dot` files, simply set
the `GST_DEBUG_DUMP_DOT_DIR` environment variable to point to the the `GST_DEBUG_DUMP_DOT_DIR` environment variable to point to the
folder where you want the files to be placed. `gst-launch-1.0` will create folder where you want the files to be placed. `gst-launch-1.0` will create
a `.dot` file at each state change, so you can see the evolution of the a `.dot` file at each state change, so you can see the evolution of the
caps negotiation. Unset the variable to disable this facility. From caps negotiation. Unset the variable to disable this facility. From
within your application, you can use the within your application, you can use the
`GST_DEBUG_BIN_TO_DOT_FILE()` and `GST_DEBUG_BIN_TO_DOT_FILE()` and
`GST_DEBUG_BIN_TO_DOT_FILE_WITH_TS()` macros to generate `.dot` files `GST_DEBUG_BIN_TO_DOT_FILE_WITH_TS()` macros to generate `.dot` files
at your convenience. at your convenience.
Here you have an example of the kind of pipelines that playbin Here you have an example of the kind of pipelines that playbin
generates. It is very complex because `playbin` can handle many generates. It is very complex because `playbin` can handle many
different cases: Your manual pipelines normally do not need to be this different cases: Your manual pipelines normally do not need to be this
long. If your manual pipeline is starting to get very big, consider long. If your manual pipeline is starting to get very big, consider
using `playbin`. using `playbin`.
@ -208,20 +167,15 @@ using `playbin`.
To download the full-size picture, use the attachments link at the top To download the full-size picture, use the attachments link at the top
of this page (It's the paperclip icon). of this page (It's the paperclip icon).
# Conclusion ## Conclusion
This tutorial has shown: This tutorial has shown:
- How to get more debug information from GStreamer using the - How to get more debug information from GStreamer using the
`GST_DEBUG` environment variable. `GST_DEBUG` environment variable.
- How to print your own debug information into the GStreamer log with - How to print your own debug information into the GStreamer log with
the `GST_ERROR()` macro and relatives. the `GST_ERROR()` macro and relatives.
- How to get pipeline graphs with the - How to get pipeline graphs with the
`GST_DEBUG_DUMP_DOT_DIR` environment variable. `GST_DEBUG_DUMP_DOT_DIR` environment variable.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!
## Attachments:
![](images/icons/bullet_blue.gif)
[playbin.png](attachments/327830/2424840.png) (image/png)

View file

@ -1,44 +1,46 @@
# Basic tutorial 10: GStreamer tools # Basic tutorial 10: GStreamer tools
# Goal ## Goal
GStreamer (and the GStreamer SDK) come with a set of tools which range GStreamer comes with a set of tools which range from handy to
from handy to absolutely essential. There is no code in this tutorial, absolutely essential. There is no code in this tutorial, just sit back
just sit back and relax, and we will teach you: and relax, and we will teach you:
- How to build and run GStreamer pipelines from the command line, - How to build and run GStreamer pipelines from the command line,
without using C at all\! without using C at all!
- How to find out what GStreamer elements you have available and their - How to find out what GStreamer elements you have available and their
capabilities. capabilities.
- How to discover the internal structure of media files. - How to discover the internal structure of media files.
# Introduction ## Introduction
These tools are available in the bin directory of the SDK. You need to These tools are available in the bin directory of the SDK. You need to
move to this directory to execute them, because it is not added to the move to this directory to execute them, because it is not added to the
systems `PATH` environment variable (to avoid polluting it too much). systems `PATH` environment variable (to avoid polluting it too much).
Just open a terminal (or console window) and go to the `bin` directory Just open a terminal (or console window) and go to the `bin` directory
of your GStreamer SDK installation (Read again the [Installing the of your GStreamer SDK installation (Read again the [Installing the
SDK](Installing%2Bthe%2BSDK.html) section to find our where this is), SDK](sdk-installing.html) section to find our where this is),
and you are ready to start typing the commands given in this tutorial. and you are ready to start typing the commands given in this tutorial.
<table>
<tbody> > ![Information](images/icons/emoticons/information.png)
<tr class="odd"> >
<td><img src="images/icons/emoticons/information.png" width="16" height="16" /></td> >On Linux, though, you can use the provided
<td><p>On Linux, though, you can use the provided <code>/opt/gstreamer-sdk/bin/gst-sdk-shell</code> script to enter the GStreamer SDK shell environment, in which the <code>bin</code> directory is in the path. In this environment, you can use the GStreamer tools from any folder.</p></td> >`/opt/gstreamer-sdk/bin/gst-sdk-shell` script to enter the
</tr> >GStreamer SDK shell environment, in which the <code>bin</code>
</tbody> >directory is in the path. In this environment, you can use the
</table> >GStreamer tools from any folder.
**FIXME: What is this now? Just refer to /usr/bin of the distro??**
In order to allow for multiple versions of GStreamer to coexists in the In order to allow for multiple versions of GStreamer to coexists in the
same system, these tools are versioned, this is, a GStreamer version same system, these tools are versioned, this is, a GStreamer version
number is appended to their name. This version of the SDK is based on number is appended to their name. This version of the SDK is based on
GStreamer 1.0, so the tools are called `gst-launch-1.0`, GStreamer 1.0, so the tools are called `gst-launch-1.0`,
`gst-inspect-1.0` and `gst-discoverer-1.0` `gst-inspect-1.0` and `gst-discoverer-1.0`
# `gst-launch-1.0` ## `gst-launch-1.0`
This tool accepts a textual description of a pipeline, instantiates it, This tool accepts a textual description of a pipeline, instantiates it,
and sets it to the PLAYING state. It allows you to quickly check if a and sets it to the PLAYING state. It allows you to quickly check if a
@ -51,60 +53,60 @@ up to a certain level. In any case, it is extremely handy to test
pipelines quickly, and is used by GStreamer developers around the world pipelines quickly, and is used by GStreamer developers around the world
on a daily basis. on a daily basis.
Please note that `gst-launch-1.0` is primarily a debugging tool for Please note that `gst-launch-1.0` is primarily a debugging tool for
developers. You should not build applications on top of it. Instead, use developers. You should not build applications on top of it. Instead, use
the `gst_parse_launch()` function of the GStreamer API as an easy way to the `gst_parse_launch()` function of the GStreamer API as an easy way to
construct pipelines from pipeline descriptions. construct pipelines from pipeline descriptions.
Although the rules to construct pipeline descriptions are very simple, Although the rules to construct pipeline descriptions are very simple,
the concatenation of multiple elements can quickly make such the concatenation of multiple elements can quickly make such
descriptions resemble black magic. Fear not, for everyone learns the descriptions resemble black magic. Fear not, for everyone learns the
`gst-launch-1.0` syntax, eventually. `gst-launch-1.0` syntax, eventually.
The command line for gst-launch-1.0 consists of a list of options followed The command line for gst-launch-1.0 consists of a list of options followed
by a PIPELINE-DESCRIPTION. Some simplified instructions are given next, by a PIPELINE-DESCRIPTION. Some simplified instructions are given next,
se the complete documentation at [the reference page](gst-launch-1.0.html) se the complete documentation at [the reference page](gst-launch.md)
for `gst-launch-1.0`. for `gst-launch-1.0`.
#### Elements ### Elements
In simple form, a PIPELINE-DESCRIPTION is a list of element types In simple form, a PIPELINE-DESCRIPTION is a list of element types
separated by exclamation marks (\!). Go ahead and type in the following separated by exclamation marks (!). Go ahead and type in the following
command: command:
``` ```
gst-launch-1.0 videotestsrc ! ffmpegcolorspace ! autovideosink gst-launch-1.0 videotestsrc ! videoconvert ! autovideosink
``` ```
You should see a windows with an animated video pattern. Use CTRL+C on You should see a windows with an animated video pattern. Use CTRL+C on
the terminal to stop the program. the terminal to stop the program.
This instantiates a new element of type `videotestsrc` (an element which This instantiates a new element of type `videotestsrc` (an element which
generates a sample video pattern), an `ffmpegcolorspace` (an element generates a sample video pattern), an `videoconvert` (an element
which does color space conversion, making sure other elements can which does raw video format conversion, making sure other elements can
understand each other), and an `autovideosink` (a window to which video understand each other), and an `autovideosink` (a window to which video
is rendered). Then, GStreamer tries to link the output of each element is rendered). Then, GStreamer tries to link the output of each element
to the input of the element appearing on its right in the description. to the input of the element appearing on its right in the description.
If more than one input or output Pad is available, the Pad Caps are used If more than one input or output Pad is available, the Pad Caps are used
to find two compatible Pads. to find two compatible Pads.
#### Properties ### Properties
Properties may be appended to elements, in the form Properties may be appended to elements, in the form
*property=value *(multiple properties can be specified, separated by *property=value *(multiple properties can be specified, separated by
spaces). Use the `gst-inspect-1.0` tool (explained next) to find out the spaces). Use the `gst-inspect-1.0` tool (explained next) to find out the
available properties for an available properties for an
element. element.
``` ```
gst-launch-1.0 videotestsrc pattern=11 ! ffmpegcolorspace ! autovideosink gst-launch-1.0 videotestsrc pattern=11 ! videoconvert ! autovideosink
``` ```
You should see a static video pattern, made of circles. You should see a static video pattern, made of circles.
#### Named elements ### Named elements
Elements can be named using the `name` property, in this way complex Elements can be named using the `name` property, in this way complex
pipelines involving branches can be created. Names allow linking to pipelines involving branches can be created. Names allow linking to
elements created previously in the description, and are indispensable to elements created previously in the description, and are indispensable to
use elements with multiple output pads, like demuxers or tees, for use elements with multiple output pads, like demuxers or tees, for
@ -114,7 +116,7 @@ Named elements are referred to using their name followed by a
dot. dot.
``` ```
gst-launch-1.0 videotestsrc ! ffmpegcolorspace ! tee name=t ! queue ! autovideosink t. ! queue ! autovideosink gst-launch-1.0 videotestsrc ! videoconvert ! tee name=t ! queue ! autovideosink t. ! queue ! autovideosink
``` ```
You should see two video windows, showing the same sample video pattern. You should see two video windows, showing the same sample video pattern.
@ -122,42 +124,36 @@ If you see only one, try to move it, since it is probably on top of the
second window. second window.
This example instantiates a `videotestsrc`, linked to a This example instantiates a `videotestsrc`, linked to a
`ffmpegcolorspace`, linked to a `tee` (Remember from [Basic tutorial 7: `videoconvert`, linked to a `tee` (Remember from [](sdk-basic-tutorial-multithreading-and-pad-availability.md) that
Multithreading and Pad a `tee` copies to each of its output pads everything coming through its
Availability](Basic%2Btutorial%2B7%253A%2BMultithreading%2Band%2BPad%2BAvailability.html) that input pad). The `tee` is named simply t (using the `name` property)
a `tee` copies to each of its output pads everything coming through its and then linked to a `queue` and an `autovideosink`. The same `tee` is
input pad). The `tee` is named simply t (using the `name` property)
and then linked to a `queue` and an `autovideosink`. The same `tee` is
referred to using t. (mind the dot) and then linked to a second referred to using t. (mind the dot) and then linked to a second
`queue` and a second `autovideosink`. `queue` and a second `autovideosink`.
To learn why the queues are necessary read [Basic tutorial 7: To learn why the queues are necessary read [](sdk-basic-tutorial-multithreading-and-pad-availability.md).
Multithreading and Pad
Availability](Basic%2Btutorial%2B7%253A%2BMultithreading%2Band%2BPad%2BAvailability.html).
#### Pads ### Pads
Instead of letting GStreamer choose which Pad to use when linking two Instead of letting GStreamer choose which Pad to use when linking two
elements, you may want to specify the Pads directly. You can do this by elements, you may want to specify the Pads directly. You can do this by
adding a dot plus the Pad name after the name of the element (it must be adding a dot plus the Pad name after the name of the element (it must be
a named element). Learn the names of the Pads of an element by using a named element). Learn the names of the Pads of an element by using
the `gst-inspect-1.0` tool. the `gst-inspect-1.0` tool.
This is useful, for example, when you want to retrieve one particular This is useful, for example, when you want to retrieve one particular
stream out of a stream out of a
demuxer: demuxer:
``` ```
gst-launch-1.0.exe souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux name=d d.video_00 ! matroskamux ! filesink location=sintel_video.mkv gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux name=d d.video_00 ! matroskamux ! filesink location=sintel_video.mkv
``` ```
This fetches a media file from the internet using `souphttpsrc`, which This fetches a media file from the internet using `souphttpsrc`, which
is in webm format (a special kind of Matroska container, see [Basic is in webm format (a special kind of Matroska container, see [](sdk-basic-tutorial-concepts.md)). We
tutorial 2: GStreamer
concepts](Basic%2Btutorial%2B2%253A%2BGStreamer%2Bconcepts.html)). We
then open the container using `matroskademux`. This media contains both then open the container using `matroskademux`. This media contains both
audio and video, so `matroskademux` will create two output Pads, named audio and video, so `matroskademux` will create two output Pads, named
`video_00` and `audio_00`. We link `video_00` to a `matroskamux` element `video_00` and `audio_00`. We link `video_00` to a `matroskamux` element
to re-pack the video stream into a new container, and finally link it to to re-pack the video stream into a new container, and finally link it to
a `filesink`, which will write the stream into a file named a `filesink`, which will write the stream into a file named
"sintel\_video.mkv" (the `location` property specifies the name of the "sintel\_video.mkv" (the `location` property specifies the name of the
@ -168,20 +164,20 @@ new matroska file with the video. If we wanted to keep only the
audio: audio:
``` ```
gst-launch-1.0.exe souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux name=d d.audio_00 ! vorbisparse ! matroskamux ! filesink location=sintel_audio.mka gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux name=d d.audio_00 ! vorbisparse ! matroskamux ! filesink location=sintel_audio.mka
``` ```
The `vorbisparse` element is required to extract some information from The `vorbisparse` element is required to extract some information from
the stream and put it in the Pad Caps, so the next element, the stream and put it in the Pad Caps, so the next element,
`matroskamux`, knows how to deal with the stream. In the case of video `matroskamux`, knows how to deal with the stream. In the case of video
this was not necessary, because `matroskademux` already extracted this this was not necessary, because `matroskademux` already extracted this
information and added it to the Caps. information and added it to the Caps.
Note that in the above two examples no media has been decoded or played. Note that in the above two examples no media has been decoded or played.
We have just moved from one container to another (demultiplexing and We have just moved from one container to another (demultiplexing and
re-multiplexing again). re-multiplexing again).
#### Caps filters ### Caps filters
When an element has more than one output pad, it might happen that the When an element has more than one output pad, it might happen that the
link to the next element is ambiguous: the next element may have more link to the next element is ambiguous: the next element may have more
@ -197,10 +193,10 @@ pipeline:
gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux ! filesink location=test gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux ! filesink location=test
``` ```
This is the same media file and demuxer as in the previous example. The This is the same media file and demuxer as in the previous example. The
input Pad Caps of `filesink` are `ANY`, meaning that it can accept any input Pad Caps of `filesink` are `ANY`, meaning that it can accept any
kind of media. Which one of the two output pads of `matroskademux` will kind of media. Which one of the two output pads of `matroskademux` will
be linked against the filesink? `video_00` or `audio_00`? You cannot be linked against the filesink? `video_00` or `audio_00`? You cannot
know. know.
You can remove this ambiguity, though, by using named pads, as in the You can remove this ambiguity, though, by using named pads, as in the
@ -213,34 +209,33 @@ gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trail
A Caps Filter behaves like a pass-through element which does nothing and A Caps Filter behaves like a pass-through element which does nothing and
only accepts media with the given Caps, effectively resolving the only accepts media with the given Caps, effectively resolving the
ambiguity. In this example, between `matroskademux` and `matroskamux` we ambiguity. In this example, between `matroskademux` and `matroskamux` we
added a `video/x-vp8` Caps Filter to specify that we are interested in added a `video/x-vp8` Caps Filter to specify that we are interested in
the output pad of `matroskademux` which can produce this kind of video. the output pad of `matroskademux` which can produce this kind of video.
To find out the Caps an element accepts and produces, use the To find out the Caps an element accepts and produces, use the
`gst-inspect-1.0` tool. To find out the Caps contained in a particular file, `gst-inspect-1.0` tool. To find out the Caps contained in a particular file,
use the `gst-discoverer-1.0` tool. To find out the Caps an element is use the `gst-discoverer-1.0` tool. To find out the Caps an element is
producing for a particular pipeline, run `gst-launch-1.0` as usual, with the producing for a particular pipeline, run `gst-launch-1.0` as usual, with the
`v` option to print Caps information. `v` option to print Caps information.
#### Examples ### Examples
Play a media file using `playbin` (as in [Basic tutorial 1: Hello Play a media file using `playbin` (as in [](sdk-basic-tutorial-hello-world.md)):
world\!](Basic%2Btutorial%2B1%253A%2BHello%2Bworld%2521.html)):
``` ```
gst-launch-1.0 playbin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm gst-launch-1.0 playbin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm
``` ```
A fully operation playback pipeline, with audio and video (more or less A fully operation playback pipeline, with audio and video (more or less
the same pipeline that `playbin` will create the same pipeline that `playbin` will create
internally): internally):
``` ```
gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux name=d ! queue ! vp8dec ! ffmpegcolorspace ! autovideosink d. ! queue ! vorbisdec ! audioconvert ! audioresample ! autoaudiosink gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! matroskademux name=d ! queue ! vp8dec ! videoconvert ! autovideosink d. ! queue ! vorbisdec ! audioconvert ! audioresample ! autoaudiosink
``` ```
A transcoding pipeline, which opens the webm container and decodes both A transcoding pipeline, which opens the webm container and decodes both
streams (via uridecodebin), then re-encodes the audio and video branches streams (via uridecodebin), then re-encodes the audio and video branches
with a different codec, and puts them back together in an Ogg container with a different codec, and puts them back together in an Ogg container
(just for the sake of (just for the sake of
@ -250,20 +245,20 @@ it).
gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm name=d ! queue ! theoraenc ! oggmux name=m ! filesink location=sintel.ogg d. ! queue ! audioconvert ! audioresample ! flacenc ! m. gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm name=d ! queue ! theoraenc ! oggmux name=m ! filesink location=sintel.ogg d. ! queue ! audioconvert ! audioresample ! flacenc ! m.
``` ```
A rescaling pipeline. The `videoscale` element performs a rescaling A rescaling pipeline. The `videoscale` element performs a rescaling
operation whenever the frame size is different in the input and the operation whenever the frame size is different in the input and the
output caps. The output caps are set by the Caps Filter to output caps. The output caps are set by the Caps Filter to
320x200. 320x200.
``` ```
gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! queue ! videoscale ! video/x-raw-yuv,width=320,height=200 ! ffmpegcolorspace ! autovideosink gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! queue ! videoscale ! video/x-raw-yuv,width=320,height=200 ! videoconvert ! autovideosink
``` ```
This short description of `gst-launch-1.0` should be enough to get you This short description of `gst-launch-1.0` should be enough to get you
started. Remember that you have the [complete documentation available started. Remember that you have the [complete documentation available
here](gst-launch-1.0.html). here](gst-launch.md).
# `gst-inspect-1.0` ## `gst-inspect-1.0`
This tool has three modes of operation: This tool has three modes of operation:
@ -279,120 +274,114 @@ Let's see an example of the third mode:
``` ```
gst-inspect-1.0 vp8dec gst-inspect-1.0 vp8dec
 
Factory Details: Factory Details:
Long name: On2 VP8 Decoder Rank primary (256)
Class: Codec/Decoder/Video Long-name On2 VP8 Decoder
Description: Decode VP8 video streams Klass Codec/Decoder/Video
Author(s): David Schleef <ds@entropywave.com> Description Decode VP8 video streams
Rank: primary (256) Author David Schleef <ds@entropywave.com>, Sebastian Dröge <sebastian.droege@collabora.co.uk>
Plugin Details: Plugin Details:
Name: vp8 Name vpx
Description: VP8 plugin Description VP8 plugin
Filename: I:\gstreamer-sdk\2012.5\x86\lib\gstreamer-1.0\libgstvp8.dll Filename /usr/lib64/gstreamer-1.0/libgstvpx.so
Version: 0.10.23 Version 1.6.4
License: LGPL License LGPL
Source module: gst-plugins-bad Source module gst-plugins-good
Source release date: 2012-02-20 Source release date 2016-04-14
Binary package: GStreamer Bad Plug-ins (GStreamer SDK) Binary package Fedora GStreamer-plugins-good package
Origin URL: http://www.gstreamer.com Origin URL http://download.fedoraproject.org
GObject GObject
+----GstObject +----GInitiallyUnowned
+----GstElement +----GstObject
+----GstBaseVideoCodec +----GstElement
+----GstBaseVideoDecoder +----GstVideoDecoder
+----GstVP8Dec +----GstVP8Dec
Pad Templates: Pad Templates:
SRC template: 'src'
Availability: Always
Capabilities:
video/x-raw-yuv
format: I420
width: [ 1, 2147483647 ]
height: [ 1, 2147483647 ]
framerate: [ 0/1, 2147483647/1 ]
SINK template: 'sink' SINK template: 'sink'
Availability: Always Availability: Always
Capabilities: Capabilities:
video/x-vp8 video/x-vp8
SRC template: 'src'
Availability: Always
Capabilities:
video/x-raw
format: I420
width: [ 1, 2147483647 ]
height: [ 1, 2147483647 ]
framerate: [ 0/1, 2147483647/1 ]
Element Flags: Element Flags:
no flags set no flags set
Element Implementation: Element Implementation:
Has change_state() function: gst_base_video_decoder_change_state Has change_state() function: gst_video_decoder_change_state
Has custom save_thyself() function: gst_element_save_thyself
Has custom restore_thyself() function: gst_element_restore_thyself
Element has no clocking capabilities. Element has no clocking capabilities.
Element has no indexing capabilities.
Element has no URI handling capabilities. Element has no URI handling capabilities.
Pads: Pads:
SRC: 'src'
Implementation:
Has custom eventfunc(): gst_base_video_decoder_src_event
Has custom queryfunc(): gst_base_video_decoder_src_query
Provides query types:
(1): position (Current position)
(2): duration (Total duration)
(8): convert (Converting between formats)
Has custom iterintlinkfunc(): gst_pad_iterate_internal_links_default
Has getcapsfunc(): gst_pad_get_fixed_caps_func
Has acceptcapsfunc(): gst_pad_acceptcaps_default
Pad Template: 'src'
SINK: 'sink' SINK: 'sink'
Implementation:
Has chainfunc(): gst_base_video_decoder_chain
Has custom eventfunc(): gst_base_video_decoder_sink_event
Has custom queryfunc(): gst_base_video_decoder_sink_query
Has custom iterintlinkfunc(): gst_pad_iterate_internal_links_default
Has setcapsfunc(): gst_base_video_decoder_sink_setcaps
Has acceptcapsfunc(): gst_pad_acceptcaps_default
Pad Template: 'sink' Pad Template: 'sink'
SRC: 'src'
Pad Template: 'src'
Element Properties: Element Properties:
name : The name of the object name : The name of the object
flags: readable, writable flags: readable, writable
String. Default: "vp8dec0" String. Default: "vp8dec0"
parent : The parent of the object
flags: readable, writable
Object of type "GstObject"
post-processing : Enable post processing post-processing : Enable post processing
flags: readable, writable flags: readable, writable
Boolean. Default: false Boolean. Default: false
post-processing-flags: Flags to control post processing post-processing-flags: Flags to control post processing
flags: readable, writable flags: readable, writable
Flags "GstVP8DecPostProcessingFlags" Default: 0x00000003, "demacroblock+deblock" Flags "GstVP8DecPostProcessingFlags" Default: 0x00000403, "mfqe+demacroblock+deblock"
(0x00000001): deblock - Deblock (0x00000001): deblock - Deblock
(0x00000002): demacroblock - Demacroblock (0x00000002): demacroblock - Demacroblock
(0x00000004): addnoise - Add noise (0x00000004): addnoise - Add noise
(0x00000400): mfqe - Multi-frame quality enhancement
deblocking-level : Deblocking level deblocking-level : Deblocking level
flags: readable, writable flags: readable, writable
Unsigned Integer. Range: 0 - 16 Default: 4 Unsigned Integer. Range: 0 - 16 Default: 4
noise-level : Noise level noise-level : Noise level
flags: readable, writable flags: readable, writable
Unsigned Integer. Range: 0 - 16 Default: 0   Unsigned Integer. Range: 0 - 16 Default: 0
threads : Maximum number of decoding threads
flags: readable, writable
Unsigned Integer. Range: 1 - 16 Default: 0
``` ```
The most relevant sections are: The most relevant sections are:
- Pad Templates (line 25): This lists all the kinds of Pads this - Pad Templates: This lists all the kinds of Pads this
element can have, along with their capabilities. This is where you element can have, along with their capabilities. This is where you
look to find out if an element can link with another one. In this look to find out if an element can link with another one. In this
case, it has only one sink pad template, accepting only case, it has only one sink pad template, accepting only
`video/x-vp8` (encoded video data in VP8 format) and only one source `video/x-vp8` (encoded video data in VP8 format) and only one source
pad template, producing `video/x-raw-yuv` (decoded video data). pad template, producing `video/x-raw` (decoded video data).
- Element Properties (line 70): This lists the properties of the - Element Properties: This lists the properties of the
element, along with their type and accepted values. element, along with their type and accepted values.
For more information, you can check the [documentation For more information, you can check the [documentation
page](http://gst-inspect-1.0) of `gst-inspect-1.0`. page](gst-inspect.md) of `gst-inspect-1.0`.
# `gst-discoverer-1.0` ## `gst-discoverer-1.0`
This tool is a wrapper around the `GstDiscoverer` object shown in [Basic This tool is a wrapper around the `GstDiscoverer` object shown in [](sdk-basic-tutorial-media-information-gathering.md).
tutorial 9: Media information
gathering](Basic%2Btutorial%2B9%253A%2BMedia%2Binformation%2Bgathering.html).
It accepts a URI from the command line and prints all information It accepts a URI from the command line and prints all information
regarding the media that GStreamer can extract. It is useful to find out regarding the media that GStreamer can extract. It is useful to find out
what container and codecs have been used to produce the media, and what container and codecs have been used to produce the media, and
therefore what elements you need to put in a pipeline to play it. therefore what elements you need to put in a pipeline to play it.
Use `gst-discoverer-1.0 --help` to obtain the list of available options, Use `gst-discoverer-1.0 --help` to obtain the list of available options,
which basically control the amount of verbosity of the output. which basically control the amount of verbosity of the output.
Let's see an Let's see an
@ -438,7 +427,7 @@ Properties:
Duration: 0:00:52.250000000 Duration: 0:00:52.250000000
Seekable: yes Seekable: yes
Tags: Tags:
video codec: On2 VP8 video codec: VP8 video
language code: en language code: en
container format: Matroska container format: Matroska
application name: ffmpeg2theora-0.24 application name: ffmpeg2theora-0.24
@ -449,15 +438,15 @@ Properties:
bitrate: 80000 bitrate: 80000
``` ```
# Conclusion ## Conclusion
This tutorial has shown: This tutorial has shown:
- How to build and run GStreamer pipelines from the command line using - How to build and run GStreamer pipelines from the command line using
the `gst-launch-1.0` tool. the `gst-launch-1.0` tool.
- How to find out what GStreamer elements you have available and their - How to find out what GStreamer elements you have available and their
capabilities, using the `gst-inspect-1.0` tool. capabilities, using the `gst-inspect-1.0` tool.
- How to discover the internal structure of media files, using - How to discover the internal structure of media files, using
`gst-discoverer-1.0`. `gst-discoverer-1.0`.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!

View file

@ -1,19 +1,18 @@
# Basic tutorial 14: Handy elements # Basic tutorial 14: Handy elements
# Goal ## Goal
This tutorial gives a list of handy GStreamer elements that are worth This tutorial gives a list of handy GStreamer elements that are worth
knowing. They range from powerful all-in-one elements that allow you to knowing. They range from powerful all-in-one elements that allow you to
build complex pipelines easily (like `playbin`), to little helper build complex pipelines easily (like `playbin`), to little helper
elements which are extremely useful when debugging. elements which are extremely useful when debugging.
For simplicity, the following examples are given using the For simplicity, the following examples are given using the
`gst-launch-1.0` tool (Learn about it in [Basic tutorial 10: GStreamer `gst-launch-1.0` tool (Learn about it in
tools](Basic%2Btutorial%2B10%253A%2BGStreamer%2Btools.html)). Use the [](sdk-basic-tutorial-gstreamer-tools.md)). Use the `-v` command line
`-v` command line parameter if you want to see the Pad Caps that are parameter if you want to see the Pad Caps that are being negotiated.
being negotiated.
# Bins ## Bins
These are Bin elements which you treat as a single element and they take These are Bin elements which you treat as a single element and they take
care of instantiating all the necessary internal pipeline to accomplish care of instantiating all the necessary internal pipeline to accomplish
@ -24,76 +23,75 @@ their task.
This element has been extensively used throughout the tutorials. It This element has been extensively used throughout the tutorials. It
manages all aspects of media playback, from source to display, passing manages all aspects of media playback, from source to display, passing
through demuxing and decoding. It is so flexible and has so many options through demuxing and decoding. It is so flexible and has so many options
that a whole set of tutorials are devoted to it. See the [Playback that a whole set of tutorials are devoted to it. See the [](sdk-playback-tutorials.md) for more details.
tutorials](Playback%2Btutorials.html) for more details.
### `uridecodebin` ### `uridecodebin`
This element decodes data from a URI into raw media. It selects a source This element decodes data from a URI into raw media. It selects a source
element that can handle the given URI scheme and connects it to element that can handle the given URI scheme and connects it to
a `decodebin2` element. It acts like a demuxer, so it offers as many a `decodebin` element. It acts like a demuxer, so it offers as many
source pads as streams are found in the source pads as streams are found in the
media. media.
``` bash ``` bash
gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! ffmpegcolorspace ! autovideosink gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! videoconvert ! autovideosink
``` ```
``` bash ``` bash
gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! audioconvert ! autoaudiosink gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! audioconvert ! autoaudiosink
``` ```
### `decodebin2` ### `decodebin`
This element automatically constructs a decoding pipeline using This element automatically constructs a decoding pipeline using
available decoders and demuxers via auto-plugging until raw media is available decoders and demuxers via auto-plugging until raw media is
obtained.  It is used internally by `uridecodebin` which is often more obtained. It is used internally by `uridecodebin` which is often more
convenient to use, as it creates a suitable source element as well. It convenient to use, as it creates a suitable source element as well. It
replaces the old `decodebin` element. It acts like a demuxer, so it replaces the old `decodebin` element. It acts like a demuxer, so it
offers as many source pads as streams are found in the offers as many source pads as streams are found in the
media. media.
``` bash ``` bash
gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! decodebin2 ! autovideosink gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! decodebin ! autovideosink
``` ```
# File input/output ## File input/output
### `filesrc` ### `filesrc`
This element reads a local file and produces media with `ANY` Caps. If This element reads a local file and produces media with `ANY` Caps. If
you want to obtain the correct Caps for the media, explore the stream by you want to obtain the correct Caps for the media, explore the stream by
using a `typefind` element or by setting the `typefind` property using a `typefind` element or by setting the `typefind` property
of `filesrc` to of `filesrc` to
`TRUE`. `TRUE`.
``` c ``` c
gst-launch-1.0 filesrc location=f:\\media\\sintel\\sintel_trailer-480p.webm ! decodebin2 ! autovideosink gst-launch-1.0 filesrc location=f:\\media\\sintel\\sintel_trailer-480p.webm ! decodebin ! autovideosink
``` ```
### `filesink` ### `filesink`
This element writes to a file all the media it receives. Use the This element writes to a file all the media it receives. Use the
`location` property to specify the file `location` property to specify the file
name. name.
``` ```
gst-launch-1.0 audiotestsrc ! vorbisenc ! oggmux ! filesink location=test.ogg gst-launch-1.0 audiotestsrc ! vorbisenc ! oggmux ! filesink location=test.ogg
``` ```
# Network ## Network
### `souphttpsrc` ### `souphttpsrc`
This element receives data as a client over the network via HTTP using This element receives data as a client over the network via HTTP using
the SOUP library. Set the URL to retrieve through the `location` the [libsoup](https://wiki.gnome.org/Projects/libsoup) library. Set the URL to retrieve through the `location`
property. property.
``` bash ``` bash
gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! decodebin2 ! autovideosink gst-launch-1.0 souphttpsrc location=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! decodebin ! autovideosink
``` ```
# Test media generation ## Test media generation
These elements are very useful to check if other parts of the pipeline These elements are very useful to check if other parts of the pipeline
are working, by replacing the source by one of these test sources which are working, by replacing the source by one of these test sources which
@ -102,24 +100,24 @@ are “guaranteed” to work.
### `videotestsrc` ### `videotestsrc`
This element produces a video pattern (selectable among many different This element produces a video pattern (selectable among many different
options with the `pattern` property). Use it to test video pipelines. options with the `pattern` property). Use it to test video pipelines.
``` bash ``` bash
gst-launch-1.0 videotestsrc ! ffmpegcolorspace ! autovideosink gst-launch-1.0 videotestsrc ! videoconvert ! autovideosink
``` ```
### `audiotestsrc` ### `audiotestsrc`
This element produces an audio wave (selectable among many different This element produces an audio wave (selectable among many different
options with the `wave` property). Use it to test video pipelines. options with the `wave` property). Use it to test video pipelines.
``` bash ``` bash
gst-launch-1.0 audiotestsrc ! audioconvert ! autoaudiosink gst-launch-1.0 audiotestsrc ! audioconvert ! autoaudiosink
``` ```
# Video adapters ## Video adapters
### `ffmpegcolorspace` ### `videoconvert`
This element converts from one color space (e.g. RGB) to another one This element converts from one color space (e.g. RGB) to another one
(e.g. YUV). It can also convert between different YUV formats (e.g. (e.g. YUV). It can also convert between different YUV formats (e.g.
@ -130,13 +128,13 @@ When not needed, because its upstream and downstream elements can
already understand each other, it acts in pass-through mode having already understand each other, it acts in pass-through mode having
minimal impact on the performance. minimal impact on the performance.
As a rule of thumb, always use `ffmpegcolorspace` whenever you use As a rule of thumb, always use `videoconvert` whenever you use
elements whose Caps are unknown at design time, like `autovideosink`, or elements whose Caps are unknown at design time, like `autovideosink`, or
that can vary depending on external factors, like decoding a that can vary depending on external factors, like decoding a
user-provided file. user-provided file.
``` bash ``` bash
gst-launch-1.0 videotestsrc ! ffmpegcolorspace ! autovideosink gst-launch-1.0 videotestsrc ! videoconvert ! autovideosink
``` ```
### `videorate` ### `videorate`
@ -156,7 +154,7 @@ rate is unknown at design time, just in
case. case.
``` c ``` c
gst-launch-1.0 videotestsrc ! video/x-raw-rgb,framerate=30/1 ! videorate ! video/x-raw-rgb,framerate=1/1 ! ffmpegcolorspace ! autovideosink gst-launch-1.0 videotestsrc ! video/x-raw,framerate=30/1 ! videorate ! video/x-raw,framerate=1/1 ! videoconvert ! autovideosink
``` ```
### `videoscale` ### `videoscale`
@ -172,15 +170,15 @@ and RGB formats and is therefore generally able to operate anywhere in a
pipeline. pipeline.
If the video is to be output to a window whose size is controlled by the If the video is to be output to a window whose size is controlled by the
user, it is a good idea to use a `videoscale` element, since not all user, it is a good idea to use a `videoscale` element, since not all
video sinks are capable of performing scaling video sinks are capable of performing scaling
operations. operations.
``` bash ``` bash
gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! videoscale ! video/x-raw-yuv,width=178,height=100 ! ffmpegcolorspace ! autovideosink gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm ! videoscale ! video/x-raw,width=178,height=100 ! videoconvert ! autovideosink
``` ```
# Audio adapters ## Audio adapters
### `audioconvert` ### `audioconvert`
@ -189,7 +187,7 @@ formats. It supports integer to float conversion, width/depth
conversion, signedness and endianness conversion and channel conversion, signedness and endianness conversion and channel
transformations. transformations.
Like `ffmpegcolorspace` does for video, you use this to solve Like `videoconvert` does for video, you use this to solve
negotiation problems with audio, and it is generally safe to use it negotiation problems with audio, and it is generally safe to use it
liberally, since this element does nothing if it is not needed. liberally, since this element does nothing if it is not needed.
@ -215,7 +213,7 @@ gst-launch-1.0 uridecodebin uri=http://docs.gstreamer.com/media/sintel_trailer-4
This element takes an incoming stream of time-stamped raw audio frames This element takes an incoming stream of time-stamped raw audio frames
and produces a perfect stream by inserting or dropping samples as and produces a perfect stream by inserting or dropping samples as
needed. It does not allow the sample rate to be changed needed. It does not allow the sample rate to be changed
as `videorate` does, it just fills gaps and removes overlapped samples as `videorate` does, it just fills gaps and removes overlapped samples
so the output stream is continuous and “clean”. so the output stream is continuous and “clean”.
It is useful in situations where the timestamps are going to be lost It is useful in situations where the timestamps are going to be lost
@ -223,13 +221,14 @@ It is useful in situations where the timestamps are going to be lost
will require all samples to be present. It is cumbersome to exemplify will require all samples to be present. It is cumbersome to exemplify
this, so no example is given. this, so no example is given.
# Multithreading ![Warning](images/icons/emoticons/warning.png)
Most of the time, `audiorate` is not what you want.
## Multithreading
### `queue` ### `queue`
Queues have been explained in [Basic tutorial 7: Multithreading and Pad Queues have been explained in [](sdk-basic-tutorial-multithreading-and-pad-availability.md). Basically, a queue performs two tasks:
Availability](Basic%2Btutorial%2B7%253A%2BMultithreading%2Band%2BPad%2BAvailability.html).
Basically, a queue performs two tasks:
- Data is queued until a selected limit is reached. Any attempt to - Data is queued until a selected limit is reached. Any attempt to
push more buffers into the queue blocks the pushing thread until push more buffers into the queue blocks the pushing thread until
@ -237,14 +236,13 @@ Basically, a queue performs two tasks:
- The queue creates a new thread on the source Pad to decouple the - The queue creates a new thread on the source Pad to decouple the
processing on sink and source Pads. processing on sink and source Pads.
Additionally, `queue` triggers signals when it is about to become empty Additionally, `queue` triggers signals when it is about to become empty
or full (according to some configurable thresholds), and can be or full (according to some configurable thresholds), and can be
instructed to drop buffers instead of blocking when it is full. instructed to drop buffers instead of blocking when it is full.
As a rule of thumb, prefer the simpler `queue` element As a rule of thumb, prefer the simpler `queue` element
over `queue2` whenever network buffering is not a concern to you. over `queue2` whenever network buffering is not a concern to you.
See [Basic tutorial 7: Multithreading and Pad See [](sdk-basic-tutorial-multithreading-and-pad-availability.md)
Availability](Basic%2Btutorial%2B7%253A%2BMultithreading%2Band%2BPad%2BAvailability.html)
for an example. for an example.
### `queue2` ### `queue2`
@ -254,16 +252,15 @@ goals but follows a different implementation approach, which results in
different features. Unfortunately, it is often not easy to tell which different features. Unfortunately, it is often not easy to tell which
queue is the best choice. queue is the best choice.
`queue2` performs the two tasks listed above for `queue`, and, `queue2` performs the two tasks listed above for `queue`, and,
additionally, is able to store the received data (or part of it) on a additionally, is able to store the received data (or part of it) on a
disk file, for later retrieval. It also replaces the signals with the disk file, for later retrieval. It also replaces the signals with the
more general and convenient buffering messages described in [Basic more general and convenient buffering messages described in
tutorial 12: Streaming](Basic%2Btutorial%2B12%253A%2BStreaming.html). [](sdk-basic-tutorial-streaming.md).
As a rule of thumb, prefer `queue2` over `queue` whenever network As a rule of thumb, prefer `queue2` over `queue` whenever network
buffering is a concern to you. See [Basic tutorial 12: buffering is a concern to you. See [](sdk-basic-tutorial-streaming.md)
Streaming](Basic%2Btutorial%2B12%253A%2BStreaming.html) for an example for an example (`queue2` is hidden inside `playbin`).
(`queue2` is hidden inside `playbin`).
### `multiqueue` ### `multiqueue`
@ -271,46 +268,42 @@ This element provides queues for multiple streams simultaneously, and
eases their management, by allowing some queues to grow if no data is eases their management, by allowing some queues to grow if no data is
being received on other streams, or by allowing some queues to drop data being received on other streams, or by allowing some queues to drop data
if they are not connected to anything (instead of returning an error, as if they are not connected to anything (instead of returning an error, as
a simpler queue would do). Additionally, it synchronizes the different a simpler queue would do). Additionally, it synchronizes the different
streams, ensuring that none of them goes too far ahead of the others. streams, ensuring that none of them goes too far ahead of the others.
This is an advanced element. It is found inside `decodebin2`, but you This is an advanced element. It is found inside `decodebin`, but you
will rarely need to instantiate it yourself in a normal playback will rarely need to instantiate it yourself in a normal playback
application. application.
### `tee` ### `tee`
[Basic tutorial 7: Multithreading and Pad [](sdk-basic-tutorial-multithreading-and-pad-availability.md) already
Availability](Basic%2Btutorial%2B7%253A%2BMultithreading%2Band%2BPad%2BAvailability.html) already showed how to use a `tee` element, which splits data to multiple pads.
showed how to use a `tee` element, which splits data to multiple pads.
Splitting the data flow is useful, for example, when capturing a video Splitting the data flow is useful, for example, when capturing a video
where the video is shown on the screen and also encoded and written to a where the video is shown on the screen and also encoded and written to a
file. Another example is playing music and hooking up a visualization file. Another example is playing music and hooking up a visualization
module. module.
One needs to use separate `queue` elements in each branch to provide One needs to use separate `queue` elements in each branch to provide
separate threads for each branch. Otherwise a blocked dataflow in one separate threads for each branch. Otherwise a blocked dataflow in one
branch would stall the other branch would stall the other
branches. branches.
``` ```
gst-launch-1.0 audiotestsrc ! tee name=t ! queue ! audioconvert ! autoaudiosink t. ! queue ! wavescope ! ffmpegcolorspace ! autovideosink gst-launch-1.0 audiotestsrc ! tee name=t ! queue ! audioconvert ! autoaudiosink t. ! queue ! wavescope ! videoconvert ! autovideosink
``` ```
# Capabilities ## Capabilities
### `capsfilter` ### `capsfilter`
[](sdk-basic-tutorial-gstreamer-tools.md) already
[Basic tutorial 10: GStreamer
tools](Basic%2Btutorial%2B10%253A%2BGStreamer%2Btools.html) already
explained how to use Caps filters with `gst-launch-1.0`. When building a explained how to use Caps filters with `gst-launch-1.0`. When building a
pipeline programmatically, Caps filters are implemented with pipeline programmatically, Caps filters are implemented with
the `capsfilter` element. This element does not modify data as such, the `capsfilter` element. This element does not modify data as such,
but enforces limitations on the data but enforces limitations on the data format.
format.
``` bash ``` bash
gst-launch-1.0 videotestsrc ! video/x-raw-gray ! ffmpegcolorspace ! autovideosink gst-launch-1.0 videotestsrc ! video/x-raw, format=GRAY8 ! videoconvert ! autovideosink
``` ```
### `typefind` ### `typefind`
@ -320,20 +313,19 @@ typefind functions in the order of their rank. Once the type has been
detected it sets its source Pad Caps to the found media type and emits detected it sets its source Pad Caps to the found media type and emits
the `have-type` signal. the `have-type` signal.
It is instantiated internally by `decodebin2`, and you can use it too to It is instantiated internally by `decodebin`, and you can use it too to
find the media type, although you can normally use the find the media type, although you can normally use the
`GstDiscoverer` which provides more information (as seen in [Basic `GstDiscoverer` which provides more information (as seen in
tutorial 9: Media information [](sdk-basic-tutorial-media-information-gathering.md)).
gathering](Basic%2Btutorial%2B9%253A%2BMedia%2Binformation%2Bgathering.html)).
# Debugging ## Debugging
### `fakesink` ### `fakesink`
This sink element simply swallows any data fed to it. It is useful when This sink element simply swallows any data fed to it. It is useful when
debugging, to replace your normal sinks and rule them out of the debugging, to replace your normal sinks and rule them out of the
equation. It can be very verbose when combined with the `-v` switch equation. It can be very verbose when combined with the `-v` switch
of `gst-launch-1.0`, so use the `silent` property to remove any unwanted of `gst-launch-1.0`, so use the `silent` property to remove any unwanted
noise. noise.
``` ```
@ -344,7 +336,7 @@ gst-launch-1.0 audiotestsrc num-buffers=1000 ! fakesink sync=false
This is a dummy element that passes incoming data through unmodified. It This is a dummy element that passes incoming data through unmodified. It
has several useful diagnostic functions, such as offset and timestamp has several useful diagnostic functions, such as offset and timestamp
checking, or buffer dropping. Read its documentation to learn all the checking, or buffer dropping. Read its documentation to learn all the
things this seemingly harmless element can things this seemingly harmless element can
do. do.
@ -352,11 +344,11 @@ do.
gst-launch-1.0 audiotestsrc ! identity drop-probability=0.1 ! audioconvert ! autoaudiosink gst-launch-1.0 audiotestsrc ! identity drop-probability=0.1 ! audioconvert ! autoaudiosink
``` ```
# Conclusion ## Conclusion
This tutorial has listed a few elements which are worth knowing, due to This tutorial has listed a few elements which are worth knowing, due to
their usefulness in the day-to-day work with GStreamer. Some are their usefulness in the day-to-day work with GStreamer. Some are
valuable for production pipelines, whereas others are only needed for valuable for production pipelines, whereas others are only needed for
debugging purposes. debugging purposes.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!

View file

@ -1,6 +1,6 @@
# Basic tutorial 9: Media information gathering # Basic tutorial 9: Media information gathering
# Goal ## Goal
Sometimes you might want to quickly find out what kind of media a file Sometimes you might want to quickly find out what kind of media a file
(or URI) contains, or if you will be able to play the media at all. You (or URI) contains, or if you will be able to play the media at all. You
@ -12,9 +12,9 @@ shows:
- How to find out if a URI is playable - How to find out if a URI is playable
# Introduction ## Introduction
`GstDiscoverer` is a utility object found in the `pbutils` library `GstDiscoverer` is a utility object found in the `pbutils` library
(Plug-in Base utilities) that accepts a URI or list of URIs, and returns (Plug-in Base utilities) that accepts a URI or list of URIs, and returns
information about them. It can work in synchronous or asynchronous information about them. It can work in synchronous or asynchronous
modes. modes.
@ -29,9 +29,8 @@ The recovered information includes codec descriptions, stream topology
(number of streams and sub-streams) and available metadata (like the (number of streams and sub-streams) and available metadata (like the
audio language). audio language).
![](images/icons/grey_arrow_down.gif)As an example, this is the result As an example, this is the result
of discovering http://docs.gstreamer.com/media/sintel\_trailer-480p.webm of discovering http://docs.gstreamer.com/media/sintel\_trailer-480p.webm
(Click to expand)
Duration: 0:00:52.250000000 Duration: 0:00:52.250000000
Tags: Tags:
@ -66,15 +65,14 @@ The following code tries to discover the URI provided through the
command line, and outputs the retrieved information (If no URI is command line, and outputs the retrieved information (If no URI is
provided it uses a default one). provided it uses a default one).
This is a simplified version of what the `gst-discoverer-1.0` tool does This is a simplified version of what the `gst-discoverer-1.0` tool does
([Basic tutorial 10: GStreamer ([](sdk-basic-tutorial-gstreamer-tools.md)), which is
tools](Basic%2Btutorial%2B10%253A%2BGStreamer%2Btools.html)), which is
an application that only displays data, but does not perform any an application that only displays data, but does not perform any
playback. playback.
# The GStreamer Discoverer ## The GStreamer Discoverer
Copy this code into a text file named `basic-tutorial-9.c` (or find it Copy this code into a text file named `basic-tutorial-9.c` (or find it
in the SDK installation). in the SDK installation).
**basic-tutorial-9.c** **basic-tutorial-9.c**
@ -298,31 +296,22 @@ int main (int argc, char **argv) {
} }
``` ```
<table>
<tbody>
<tr class="odd">
<td><img src="images/icons/emoticons/information.png" width="16" height="16" /></td>
<td><div id="expander-2049220294" class="expand-container">
<div id="expander-control-2049220294" class="expand-control">
<span class="expand-control-icon"><img src="images/icons/grey_arrow_down.gif" class="expand-control-image" /></span><span class="expand-control-text">Need help? (Click to expand)</span>
</div>
<div id="expander-content-2049220294" class="expand-content">
<p>If you need help to compile this code, refer to the <strong>Building the tutorials</strong> section for your platform: <a href="Installing%2Bon%2BLinux.html#InstallingonLinux-Build">Linux</a>, <a href="Installing%2Bon%2BMac%2BOS%2BX.html#InstallingonMacOSX-Build">Mac OS X</a> or <a href="Installing%2Bon%2BWindows.html#InstallingonWindows-Build">Windows</a>, or use this specific command on Linux:</p>
<div class="panel" style="border-width: 1px;">
<div class="panelContent">
<p><code>gcc basic-tutorial-9.c -o basic-tutorial-9 `pkg-config --cflags --libs gstreamer-pbutils-1.0 gstreamer-1.0`</code></p>
</div>
</div>
<p>If you need help to run this code, refer to the <strong>Running the tutorials</strong> section for your platform: <a href="Installing%2Bon%2BLinux.html#InstallingonLinux-Run">Linux</a>, <a href="Installing%2Bon%2BMac%2BOS%2BX.html#InstallingonMacOSX-Run">Mac OS X</a> or <a href="Installing%2Bon%2BWindows.html#InstallingonWindows-Run">Windows</a></p>
<p><span>This tutorial opens the URI passed as the first parameter in the command line (or a default URI if none is provided) and outputs information about it on the screen. If the media is located on the Internet, the application might take a bit to react depending on your connection speed.</span></p>
<p>Required libraries: <code>gstreamer-pbutils-1.0 gstreamer-1.0</code></p>
</div>
</div></td>
</tr>
</tbody>
</table>
# Walkthrough > ![Information](images/icons/emoticons/information.png)
> Need help?
>
> If you need help to compile this code, refer to the **Building the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Build), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Build) or [Windows](sdk-installing-on-windows.md#InstallingonWindows-Build), or use this specific command on Linux:
>
> ``gcc basic-tutorial-9.c -o basic-tutorial-9 `pkg-config --cflags --libs gstreamer-1.0 gstreamer-pbutils-1.0` ``
>
>If you need help to run this code, refer to the **Running the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Run), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Run) or [Windows](sdk-installing-on-windows.md#InstallingonWindows-Run).
>
> This tutorial opens the URI passed as the first parameter in the command line (or a default URI if none is provided) and outputs information about it on the screen. If the media is located on the Internet, the application might take a bit to react depending on your connection speed.
>
> Required libraries: `gstreamer-pbutils-1.0` `gstreamer-1.0`
## Walkthrough
These are the main steps to use the `GstDiscoverer`: These are the main steps to use the `GstDiscoverer`:
@ -336,9 +325,9 @@ if (!data.discoverer) {
} }
``` ```
`gst_discoverer_new()` creates a new Discoverer object. The first `gst_discoverer_new()` creates a new Discoverer object. The first
parameter is the timeout per file, in nanoseconds (use the parameter is the timeout per file, in nanoseconds (use the
`GST_SECOND` macro for simplicity). `GST_SECOND` macro for simplicity).
``` c ``` c
/* Connect to the interesting signals */ /* Connect to the interesting signals */
@ -354,7 +343,7 @@ snippet for their callbacks.
gst_discoverer_start (data.discoverer); gst_discoverer_start (data.discoverer);
``` ```
`gst_discoverer_start()` launches the discovering process, but we have `gst_discoverer_start()` launches the discovering process, but we have
not provided any URI to discover yet. This is done not provided any URI to discover yet. This is done
next: next:
@ -367,7 +356,7 @@ if (!gst_discoverer_discover_uri_async (data.discoverer, uri)) {
} }
``` ```
`gst_discoverer_discover_uri_async()` enqueues the provided URI for `gst_discoverer_discover_uri_async()` enqueues the provided URI for
discovery. Multiple URIs can be enqueued with this function. As the discovery. Multiple URIs can be enqueued with this function. As the
discovery process for each of them finishes, the registered callback discovery process for each of them finishes, the registered callback
functions will be fired functions will be fired
@ -380,8 +369,8 @@ g_main_loop_run (data.loop);
``` ```
The usual GLib main loop is instantiated and executed. We will get out The usual GLib main loop is instantiated and executed. We will get out
of it when `g_main_loop_quit()` is called from the of it when `g_main_loop_quit()` is called from the
`on_finished_cb` callback. `on_finished_cb` callback.
``` c ``` c
/* Stop the discoverer process */ /* Stop the discoverer process */
@ -389,7 +378,7 @@ gst_discoverer_stop (data.discoverer);
``` ```
Once we are done with the discoverer, we stop it with Once we are done with the discoverer, we stop it with
`gst_discoverer_stop()` and unref it with `g_object_unref()`. `gst_discoverer_stop()` and unref it with `g_object_unref()`.
Let's review now the callbacks we have Let's review now the callbacks we have
registered: registered:
@ -408,11 +397,11 @@ static void on_discovered_cb (GstDiscoverer *discoverer, GstDiscovererInfo *info
``` ```
We got here because the Discoverer has finished working on one URI, and We got here because the Discoverer has finished working on one URI, and
provides us a `GstDiscovererInfo` structure with all the information. provides us a `GstDiscovererInfo` structure with all the information.
The first step is to retrieve the particular URI this call refers to (in The first step is to retrieve the particular URI this call refers to (in
case we had multiple discover process running, which is not the case in case we had multiple discover process running, which is not the case in
this example) with `gst_discoverer_info_get_uri()` and the discovery this example) with `gst_discoverer_info_get_uri()` and the discovery
result with `gst_discoverer_info_get_result()`. result with `gst_discoverer_info_get_result()`.
``` c ``` c
@ -451,15 +440,15 @@ if (result != GST_DISCOVERER_OK) {
} }
``` ```
As the code shows, any result other than `GST_DISCOVERER_OK` means that As the code shows, any result other than `GST_DISCOVERER_OK` means that
there has been some kind of problem, and this URI cannot be played. The there has been some kind of problem, and this URI cannot be played. The
reasons can vary, but the enum values are quite explicit reasons can vary, but the enum values are quite explicit
(`GST_DISCOVERER_BUSY` can only happen when in synchronous mode, which (`GST_DISCOVERER_BUSY` can only happen when in synchronous mode, which
is not used in this example). is not used in this example).
If no error happened, information can be retrieved from the If no error happened, information can be retrieved from the
`GstDiscovererInfo` structure with the different `GstDiscovererInfo` structure with the different
`gst_discoverer_info_get_*` methods (like, `gst_discoverer_info_get_*` methods (like,
`gst_discoverer_info_get_duration()`, for example). `gst_discoverer_info_get_duration()`, for example).
Bits of information which are made of lists, like tags and stream info, Bits of information which are made of lists, like tags and stream info,
@ -474,10 +463,10 @@ if (tags) {
``` ```
Tags are metadata (labels) attached to the media. They can be examined Tags are metadata (labels) attached to the media. They can be examined
with `gst_tag_list_foreach()`, which will call `print_tag_foreach` for with `gst_tag_list_foreach()`, which will call `print_tag_foreach` for
each tag found (the list could also be traversed manually, for example, each tag found (the list could also be traversed manually, for example,
or a specific tag could be searched for with or a specific tag could be searched for with
`gst_tag_list_get_string()`). The code for `print_tag_foreach` is pretty `gst_tag_list_get_string()`). The code for `print_tag_foreach` is pretty
much self-explicative. much self-explicative.
``` c ``` c
@ -492,10 +481,10 @@ print_topology (sinfo, 1);
gst_discoverer_stream_info_unref (sinfo); gst_discoverer_stream_info_unref (sinfo);
``` ```
`gst_discoverer_info_get_stream_info()` returns `gst_discoverer_info_get_stream_info()` returns
a `GstDiscovererStreamInfo` structure that is parsed in a `GstDiscovererStreamInfo` structure that is parsed in
the `print_topology` function, and then discarded the `print_topology` function, and then discarded
with `gst_discoverer_stream_info_unref()`. with `gst_discoverer_stream_info_unref()`.
``` c ``` c
/* Print information regarding a stream and its substreams, if any */ /* Print information regarding a stream and its substreams, if any */
@ -524,25 +513,25 @@ static void print_topology (GstDiscovererStreamInfo *info, gint depth) {
} }
``` ```
The `print_stream_info` function's code is also pretty much The `print_stream_info` function's code is also pretty much
self-explicative: it prints the stream's capabilities and then the self-explicative: it prints the stream's capabilities and then the
associated caps, using `print_tag_foreach` too. associated caps, using `print_tag_foreach` too.
Then, `print_topology` looks for the next element to display. If Then, `print_topology` looks for the next element to display. If
`gst_discoverer_stream_info_get_next()` returns a non-NULL stream info, `gst_discoverer_stream_info_get_next()` returns a non-NULL stream info,
it refers to our descendant and that should be displayed. Otherwise, if it refers to our descendant and that should be displayed. Otherwise, if
we are a container, recursively call `print_topology` on each of our we are a container, recursively call `print_topology` on each of our
children obatined with `gst_discoverer_container_info_get_streams()`. children obatined with `gst_discoverer_container_info_get_streams()`.
Otherwise, we are a final stream, and do not need to recurse (This part Otherwise, we are a final stream, and do not need to recurse (This part
of the Discoverer API is admittedly a bit obscure). of the Discoverer API is admittedly a bit obscure).
# Conclusion ## Conclusion
This tutorial has shown: This tutorial has shown:
- How to recover information regarding a URI using the `GstDiscoverer` - How to recover information regarding a URI using the `GstDiscoverer`
- How to find out if a URI is playable by looking at the return code - How to find out if a URI is playable by looking at the return code
obtained with `gst_discoverer_info_get_result()`. obtained with `gst_discoverer_info_get_result()`.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!

View file

@ -1,4 +1,4 @@
# Basic tutorial 7: Multithreading and Pad Availability # Basic tutorial 7: Multithreading and Pad Availability
## Goal ## Goal
@ -184,7 +184,7 @@ int main(int argc, char *argv[]) {
> ``gcc basic-tutorial-7.c -o basic-tutorial-7 `pkg-config --cflags --libs gstreamer-1.0` `` > ``gcc basic-tutorial-7.c -o basic-tutorial-7 `pkg-config --cflags --libs gstreamer-1.0` ``
> >
>If you need help to run this code, refer to the **Running the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Run), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Run) or [Windows](sdk-installing-on-windows.md#InstallingonWindows-Run). >If you need help to run this code, refer to the **Running the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Run), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Run) or [Windows](sdk-installing-on-windows.md#InstallingonWindows-Run).
> >
> This tutorial plays an audible tone through the audio card and opens a window with a waveform representation of the tone. The waveform should be a sinusoid, but due to the refreshing of the window might not appear so. > This tutorial plays an audible tone through the audio card and opens a window with a waveform representation of the tone. The waveform should be a sinusoid, but due to the refreshing of the window might not appear so.
> >
> Required libraries: `gstreamer-1.0` > Required libraries: `gstreamer-1.0`

View file

@ -1,55 +1,57 @@
# Basic tutorial 16: Platform-specific elements # Basic tutorial 16: Platform-specific elements
# Goal ## Goal
Even though GStreamer is a multiplatform framework, not all the elements Even though GStreamer is a multiplatform framework, not all the elements
are available on all platforms. For example, the audio and video sinks are available on all platforms. For example, the video sinks
depend heavily on the underlying windowing system, and a different one depend heavily on the underlying windowing system, and a different one
needs to be selected depending on the platform. You normally do not need needs to be selected depending on the platform. You normally do not need
to worry about this when using elements like `playbin` or to worry about this when using elements like `playbin` or
`autovideosink`, but, for those cases when you need to use one of the `autovideosink`, but, for those cases when you need to use one of the
sinks that are only available on specific platforms, this tutorial hints sinks that are only available on specific platforms, this tutorial hints
you some of their peculiarities. you some of their peculiarities.
# Linux ## Cross Platform
### `glimagesink`
This video sink is based on
[OpenGL](http://en.wikipedia.org/wiki/OpenGL) or [OpenGL ES](http://en.wikipedia.org/wiki/OpenGL ES). It supports rescaling
and filtering of the scaled image to alleviate aliasing. It implements
the VideoOverlay interface, so the video window can be re-parented
(embedded inside other windows). This is the video sink recommended on
most platforms. In particular, on Android and iOS, it is the only
available video sink. It can be decomposed into
`glupload ! glcolorconvert ! glimagesinkelement` to insert further OpenGL
hardware accelerated processing into the pipeline.
## Linux
### `ximagesink` ### `ximagesink`
A standard X-based video sink. It implements the XOverlay interface, so A standard RGB only X-based video sink. It implements the VideoOverlay
the video window can be re-parented (embedded inside other windows). It interface, so the video window can be re-parented (embedded inside
does not support scaling; it has to be performed by different means other windows). It does not support scaling or color formats other
(using the `videoscale` element, for example). than RGB; it has to be performed by different means (using the
`videoscale` element, for example).
### `xvimagesink` ### `xvimagesink`
An X-based video sink, using the [X Video An X-based video sink, using the [X Video
Extension](http://en.wikipedia.org/wiki/X_video_extension) (Xv). It Extension](http://en.wikipedia.org/wiki/X_video_extension) (Xv). It
implements the XOverlay interface, so the video window can be implements the VideoOverlay interface, so the video window can be
re-parented (embedded inside other windows). It can perform scaling re-parented (embedded inside other windows). It can perform scaling
efficiently, on the GPU. It is only available if the hardware and efficiently, on the GPU. It is only available if the hardware and
corresponding drivers support the Xv extension. corresponding drivers support the Xv extension.
### `cluttersink`
This is a GStreamer video sink element that sends data to a
[ClutterTexture](http://developer.gnome.org/clutter-gst/stable/ClutterGstVideoTexture.html)
which can then be used in any [Clutter](https://clutter-project.org/)
scene (See [Basic tutorial 15: Clutter
integration](Basic%2Btutorial%2B15%253A%2BClutter%2Bintegration.html)). Clutter
is a multiplatform library, so this sink is available on every platform
(as long as Clutter is installed). Clutter achieves
platform-independence by using [OpenGL](http://www.opengl.org) as the
rendering backend, so OpenGL must also be available in the system.
### `alsasink` ### `alsasink`
This audio sink outputs to the sound card via This audio sink outputs to the sound card via
[ALSA](http://www.alsa-project.org/) (Advanced Linux Sound [ALSA](http://www.alsa-project.org/) (Advanced Linux Sound
Architecture). This sink is available on almost every Linux platform. It Architecture). This sink is available on almost every Linux platform. It
is often seen as a “low level” interface to the sound card, and can be is often seen as a “low level” interface to the sound card, and can be
complicated to configure (See the comment on [Playback tutorial 9: complicated to configure (See the comment on
Digital audio [](sdk-playback-tutorial-digital-audio-pass-through.md)).
pass-through](Playback%2Btutorial%2B9%253A%2BDigital%2Baudio%2Bpass-through.html)).
### `pulsesink` ### `pulsesink`
@ -58,29 +60,18 @@ server. It is a higher level abstraction of the sound card than ALSA,
and is therefore easier to use and offers more advanced features. It has and is therefore easier to use and offers more advanced features. It has
been known to be unstable on some older Linux distributions, though. been known to be unstable on some older Linux distributions, though.
# Mac OS X ## Mac OS X
### `osxvideosink` ### `osxvideosink`
This is the only video sink available to GStreamer on Mac OS X. This is the video sink available to GStreamer on Mac OS X. It is also
possible to draw using `glimagesink` using OpenGL.
### `cluttersink`
This is a GStreamer video sink element that sends data to
a [ClutterTexture](http://developer.gnome.org/clutter-gst/stable/ClutterGstVideoTexture.html) which
can then be used in any [Clutter](https://clutter-project.org/) scene
(See [Basic tutorial 15: Clutter
integration](Basic%2Btutorial%2B15%253A%2BClutter%2Bintegration.html)). Clutter
is a multiplatform library, so this sink is available on every platform
(as long as Clutter is installed). Clutter achieves
platform-independence by using [OpenGL](http://www.opengl.org/) as the
rendering backend, so OpenGL must also be available in the system.
### `osxaudiosink` ### `osxaudiosink`
This is the only audio sink available to GStreamer on Mac OS X. This is the only audio sink available to GStreamer on Mac OS X.
# Windows ## Windows
### `directdrawsink` ### `directdrawsink`
@ -92,15 +83,15 @@ rescaling and filtering of the scaled image to alleviate aliasing.
### `dshowvideosink` ### `dshowvideosink`
This video sink is based on [Direct This video sink is based on [Direct
Show](http://en.wikipedia.org/wiki/Direct_Show).  It can use different Show](http://en.wikipedia.org/wiki/Direct_Show). It can use different
rendering back-ends, like rendering back-ends, like
[EVR](http://en.wikipedia.org/wiki/Enhanced_Video_Renderer), [EVR](http://en.wikipedia.org/wiki/Enhanced_Video_Renderer),
[VMR9](http://en.wikipedia.org/wiki/Direct_Show#Video_rendering_filters) [VMR9](http://en.wikipedia.org/wiki/Direct_Show#Video_rendering_filters)
or or
[VMR7](http://en.wikipedia.org/wiki/Direct_Show#Video_rendering_filters), [VMR7](http://en.wikipedia.org/wiki/Direct_Show#Video_rendering_filters),
EVR only being available on Windows Vista or more recent. It supports EVR only being available on Windows Vista or more recent. It supports
rescaling and filtering of the scaled image to alleviate aliasing. It rescaling and filtering of the scaled image to alleviate aliasing. It
implements the XOverlay interface, so the video window can be implements the VideoOverlay interface, so the video window can be
re-parented (embedded inside other windows). re-parented (embedded inside other windows).
### `d3dvideosink` ### `d3dvideosink`
@ -108,22 +99,10 @@ re-parented (embedded inside other windows).
This video sink is based on This video sink is based on
[Direct3D](http://en.wikipedia.org/wiki/Direct3D) and its the most [Direct3D](http://en.wikipedia.org/wiki/Direct3D) and its the most
recent Windows video sink. It supports rescaling and filtering of the recent Windows video sink. It supports rescaling and filtering of the
scaled image to alleviate aliasing. It implements the XOverlay scaled image to alleviate aliasing. It implements the VideoOverlay
interface, so the video window can be re-parented (embedded inside other interface, so the video window can be re-parented (embedded inside other
windows). windows).
### `cluttersink`
This is a GStreamer video sink element that sends data to
a [ClutterTexture](http://developer.gnome.org/clutter-gst/stable/ClutterGstVideoTexture.html) which
can then be used in any [Clutter](https://clutter-project.org/) scene
(See [Basic tutorial 15: Clutter
integration](Basic%2Btutorial%2B15%253A%2BClutter%2Bintegration.html)). Clutter
is a multiplatform library, so this sink is available on every platform
(as long as Clutter is installed). Clutter achieves
platform-independence by using [OpenGL](http://www.opengl.org/) as the
rendering backend, so OpenGL must also be available in the system.
### `directsoundsink` ### `directsoundsink`
This is the default audio sink for Windows, based on [Direct This is the default audio sink for Windows, based on [Direct
@ -132,50 +111,44 @@ all Windows versions.
### `dshowdecwrapper` ### `dshowdecwrapper`
[Direct Show](http://en.wikipedia.org/wiki/Direct_Show) is a multimedia [Direct Show](http://en.wikipedia.org/wiki/Direct_Show) is a multimedia
framework similar to GStreamer. They are different enough, though, so framework similar to GStreamer. They are different enough, though, so
that their pipelines cannot be interconnected. However, through this that their pipelines cannot be interconnected. However, through this
element, GStreamer can benefit from the decoding elements present in element, GStreamer can benefit from the decoding elements present in
Direct Show. `dshowdecwrapper` wraps multiple Direct Show decoders so Direct Show. `dshowdecwrapper` wraps multiple Direct Show decoders so
they can be embedded in a GStreamer pipeline. Use the `gst-inspect-1.0` tool they can be embedded in a GStreamer pipeline. Use the `gst-inspect-1.0` tool
(see [Basic tutorial 10: GStreamer (see [Basic tutorial 10: GStreamer
tools](Basic%2Btutorial%2B10%253A%2BGStreamer%2Btools.html)) to see the tools](Basic%2Btutorial%2B10%253A%2BGStreamer%2Btools.html)) to see the
available decoders. available decoders.
# Android ## Android
### `eglglessink`
This video sink is based on [OpenGL
ES](http://en.wikipedia.org/wiki/OpenGL_ES) and
[EGL](http://en.wikipedia.org/wiki/EGL_%28OpenGL%29). It supports
rescaling and filtering of the scaled image to alleviate aliasing. It
implements the XOverlay interface, so the video window can be
re-parented (embedded inside other windows).
### `openslessink` ### `openslessink`
This is the only audio sink available to GStreamer on Android. It is This is the only audio sink available to GStreamer on Android. It is
based on [OpenSL based on [OpenSL ES](http://en.wikipedia.org/wiki/OpenSL_ES).
ES](http://en.wikipedia.org/wiki/OpenSL_ES).
### `openslessrc`
This is the only audio source available to GStreamer on Android. It is
based on [OpenSL ES](http://en.wikipedia.org/wiki/OpenSL_ES).
### `androidmedia` ### `androidmedia`
[android.media.MediaCodec](http://developer.android.com/reference/android/media/MediaCodec.html) [android.media.MediaCodec](http://developer.android.com/reference/android/media/MediaCodec.html)
is an Android specific API to access the codecs that are available on is an Android specific API to access the codecs that are available on
the device, including hardware codecs. It is available since API level the device, including hardware codecs. It is available since API level
16 (JellyBean) and GStreamer can use it via the androidmedia plugin for 16 (JellyBean) and GStreamer can use it via the androidmedia plugin
audio and video decoding. for audio and video decoding. On Android, attaching the hardware
decoder to the `glimagesink` element can produce a high performance
zero-copy decodebin pipeline.
# iOS ### `ahcsrc`
### `eglglessink` This video source can capture from the cameras on Android devices, it is part
of the androidmedia plugin and uses the [android.hardware.Camera API](https://developer.android.com/reference/android/hardware/Camera.html).
This video sink is based on [OpenGL ## iOS
ES](http://en.wikipedia.org/wiki/OpenGL_ES) and [EGL](http://en.wikipedia.org/wiki/EGL_%28OpenGL%29).
It supports rescaling and filtering of the scaled image to alleviate
aliasing. It implements the XOverlay interface, so the video window can
be re-parented (embedded inside other windows).
### `osxaudiosink` ### `osxaudiosink`
@ -185,23 +158,23 @@ This is the only audio sink available to GStreamer on iOS.
Source element to read iOS assets, this is, documents stored in the Source element to read iOS assets, this is, documents stored in the
Library (like photos, music and videos). It can be instantiated Library (like photos, music and videos). It can be instantiated
automatically by `playbin` when URIs use the automatically by `playbin` when URIs use the
`assets-library://` scheme. `assets-library://` scheme.
### `iosavassetsrc` ### `iosavassetsrc`
Source element to read and decode iOS audiovisual assets, this is, Source element to read and decode iOS audiovisual assets, this is,
documents stored in the Library (like photos, music and videos). It can documents stored in the Library (like photos, music and videos). It can
be instantiated automatically by `playbin` when URIs use the be instantiated automatically by `playbin` when URIs use the
`ipod-library://` scheme. Decoding is performed by the system, so `ipod-library://` scheme. Decoding is performed by the system, so
dedicated hardware will be used if available. dedicated hardware will be used if available.
# Conclusion ## Conclusion
This tutorial has shown a few specific details about some GStreamer This tutorial has shown a few specific details about some GStreamer
elements which are not available on all platforms. You do not have to elements which are not available on all platforms. You do not have to
worry about them when using multiplatform elements like `playbin` or worry about them when using multiplatform elements like `playbin` or
`autovideosink`, but it is good to know their personal quirks if `autovideosink`, but it is good to know their personal quirks if
instancing them manually. instancing them manually.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!

View file

@ -1,6 +1,6 @@
# Basic tutorial 13: Playback speed # Basic tutorial 13: Playback speed
# Goal ## Goal
Fast-forward, reverse-playback and slow-motion are all techniques Fast-forward, reverse-playback and slow-motion are all techniques
collectively known as *trick modes* and they all have in common that collectively known as *trick modes* and they all have in common that
@ -12,7 +12,7 @@ shows:
forward and backwards. forward and backwards.
- How to advance a video frame-by-frame - How to advance a video frame-by-frame
# Introduction ## Introduction
Fast-forward is the technique that plays a media at a speed higher than Fast-forward is the technique that plays a media at a speed higher than
its normal (intended) speed; whereas slow-motion uses a speed lower than its normal (intended) speed; whereas slow-motion uses a speed lower than
@ -30,45 +30,37 @@ media besides changing the subsequent playback rate (only to positive
values). Seek Events, additionally, allow jumping to any position in the values). Seek Events, additionally, allow jumping to any position in the
stream and set positive and negative playback rates. stream and set positive and negative playback rates.
In [Basic tutorial 4: Time In [](sdk-basic-tutorial-time-management.md) seek
management](Basic%2Btutorial%2B4%253A%2BTime%2Bmanagement.html) seek
events have already been shown, using a helper function to hide their events have already been shown, using a helper function to hide their
complexity. This tutorial explains a bit more how to use these events. complexity. This tutorial explains a bit more how to use these events.
Step Events are a more convenient way of changing the playback rate, due Step Events are a more convenient way of changing the playback rate, due
to the reduced number of parameters needed to create them; however, to the reduced number of parameters needed to create them; however,
their implementation in GStreamer still needs a bit more polishing their implementation in GStreamer still needs a bit more polishing
so Seek Events are used in this tutorial instead. so Seek Events are used in this tutorial instead.
**FIXME: Is that even true ???**
To use these events, they are created and then passed onto the pipeline, To use these events, they are created and then passed onto the pipeline,
where they propagate upstream until they reach an element that can where they propagate upstream until they reach an element that can
handle them. If an event is passed onto a bin element like `playbin`, handle them. If an event is passed onto a bin element like `playbin`,
it will simply feed the event to all its sinks, which will result in it will simply feed the event to all its sinks, which will result in
multiple seeks being performed. The common approach is to retrieve one multiple seeks being performed. The common approach is to retrieve one
of `playbin`s sinks through the `video-sink` or of `playbin`s sinks through the `video-sink` or
`audio-sink` properties and feed the event directly into the sink. `audio-sink` properties and feed the event directly into the sink.
Frame stepping is a technique that allows playing a video frame by Frame stepping is a technique that allows playing a video frame by
frame. It is implemented by pausing the pipeline, and then sending Step frame. It is implemented by pausing the pipeline, and then sending Step
Events to skip one frame each time. Events to skip one frame each time.
# A trick mode player ## A trick mode player
Copy this code into a text file named `basic-tutorial-13.c`. Copy this code into a text file named `basic-tutorial-13.c`.
<table>
<tbody>
<tr class="odd">
<td><img src="images/icons/emoticons/information.png" width="16" height="16" /></td>
<td><p>This tutorial is included in the SDK since release 2012.7. If you cannot find it in the downloaded code, please install the latest release of the GStreamer SDK.</p></td>
</tr>
</tbody>
</table>
**basic-tutorial-13.c** **basic-tutorial-13.c**
``` c ``` c
#include <string.h> #include <string.h>
#include <stdio.h>
#include <gst/gst.h> #include <gst/gst.h>
typedef struct _CustomData { typedef struct _CustomData {
@ -184,7 +176,7 @@ int main(int argc, char *argv[]) {
data.pipeline = gst_parse_launch ("playbin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm", NULL); data.pipeline = gst_parse_launch ("playbin uri=http://docs.gstreamer.com/media/sintel_trailer-480p.webm", NULL);
/* Add a keyboard watch so we get notified of keystrokes */ /* Add a keyboard watch so we get notified of keystrokes */
#ifdef _WIN32 #ifdef G_OS_WIN32
io_stdin = g_io_channel_win32_new_fd (fileno (stdin)); io_stdin = g_io_channel_win32_new_fd (fileno (stdin));
#else #else
io_stdin = g_io_channel_unix_new (fileno (stdin)); io_stdin = g_io_channel_unix_new (fileno (stdin));
@ -216,34 +208,24 @@ int main(int argc, char *argv[]) {
} }
``` ```
<table>
<tbody>
<tr class="odd">
<td><img src="images/icons/emoticons/information.png" width="16" height="16" /></td>
<td><div id="expander-1662010270" class="expand-container">
<div id="expander-control-1662010270" class="expand-control">
<span class="expand-control-icon"><img src="images/icons/grey_arrow_down.gif" class="expand-control-image" /></span><span class="expand-control-text">Need help? (Click to expand)</span>
</div>
<div id="expander-content-1662010270" class="expand-content">
<p>If you need help to compile this code, refer to the <strong>Building the tutorials</strong> section for your platform: <a href="Installing%2Bon%2BLinux.html#InstallingonLinux-Build">Linux</a>, <a href="Installing%2Bon%2BMac%2BOS%2BX.html#InstallingonMacOSX-Build">Mac OS X</a> or <a href="Installing%2Bon%2BWindows.html#InstallingonWindows-Build">Windows</a>, or use this specific command on Linux:</p>
<div class="panel" style="border-width: 1px;">
<div class="panelContent">
<p><code>gcc basic-tutorial-13.c -o basic-tutorial-13 `pkg-config --cflags --libs gstreamer-1.0`</code></p>
</div>
</div>
<p>If you need help to run this code, refer to the <strong>Running the tutorials</strong> section for your platform: <a href="Installing%2Bon%2BLinux.html#InstallingonLinux-Run">Linux</a>, <a href="Installing%2Bon%2BMac%2BOS%2BX.html#InstallingonMacOSX-Run">Mac OS X</a> or <a href="Installing%2Bon%2BWindows.html#InstallingonWindows-Run">Windows</a></p>
<p><span>This tutorial opens a window and displays a movie, with accompanying audio. The media is fetched from the Internet, so the window might take a few seconds to appear, depending on your connection speed. The console shows the available commands, composed of a single upper-case or lower-case letter, which you should input followed by the Enter key.</span></p>
<p>Required libraries: <code>gstreamer-1.0</code></p>
</div>
</div></td>
</tr>
</tbody>
</table>
# Walkthrough > ![Information](images/icons/emoticons/information.png)
> Need help?
>
> If you need help to compile this code, refer to the **Building the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Build), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Build) or [Windows](sdk-installing-on-windows.mdb#InstallingonWindows-Build), or use this specific command on Linux:
>
> `` gcc basic-tutorial-13.c -o basic-tutorial-13 `pkg-config --cflags --libs gstreamer-1.0` ``
>
>If you need help to run this code, refer to the **Running the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Run), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Run) or [Windows](sdk-installing-on-windows.md#InstallingonWindows-Run).
>
> This tutorial opens a window and displays a movie, with accompanying audio. The media is fetched from the Internet, so the window might take a few seconds to appear, depending on your connection speed. The console shows the available commands, composed of a single upper-case or lower-case letter, which you should input followed by the Enter key.
>
> Required libraries: `gstreamer-1.0`
There is nothing new in the initialization code in the main function:  a ## Walkthrough
`playbin` pipeline is instantiated, an I/O watch is installed to track
There is nothing new in the initialization code in the main function: a
`playbin` pipeline is instantiated, an I/O watch is installed to track
keystrokes and a GLib main loop is executed. keystrokes and a GLib main loop is executed.
Then, in the keyboard handler function: Then, in the keyboard handler function:
@ -265,7 +247,7 @@ static gboolean handle_keyboard (GIOChannel *source, GIOCondition cond, CustomDa
break; break;
``` ```
Pause / Playing toggle is handled with `gst_element_set_state()` as in Pause / Playing toggle is handled with `gst_element_set_state()` as in
previous tutorials. previous tutorials.
``` c ``` c
@ -285,18 +267,17 @@ case 'd':
Use S and s to double or halve the current playback rate, and d to Use S and s to double or halve the current playback rate, and d to
reverse the current playback direction. In both cases, the reverse the current playback direction. In both cases, the
`rate` variable is updated and `send_seek_event` is called. Lets `rate` variable is updated and `send_seek_event` is called. Lets
review this function. review this function.
``` c ``` c
/* Send seek event to change rate */ /* Send seek event to change rate */
static void send_seek_event (CustomData *data) { static void send_seek_event (CustomData *data) {
gint64 position; gint64 position;
GstFormat format = GST_FORMAT_TIME;
GstEvent *seek_event; GstEvent *seek_event;
/* Obtain the current position, needed for the seek event */ /* Obtain the current position, needed for the seek event */
if (!gst_element_query_position (data->pipeline, &format, &position)) { if (!gst_element_query_position (data->pipeline, GST_FORMAT_TIME, &position)) {
g_printerr ("Unable to retrieve current position.\n"); g_printerr ("Unable to retrieve current position.\n");
return; return;
} }
@ -336,7 +317,7 @@ if (data->video_sink == NULL) {
As explained in the Introduction, to avoid performing multiple Seeks, As explained in the Introduction, to avoid performing multiple Seeks,
the Event is sent to only one sink, in this case, the video sink. It is the Event is sent to only one sink, in this case, the video sink. It is
obtained from `playbin` through the `video-sink` property. It is read obtained from `playbin` through the `video-sink` property. It is read
at this time instead at initialization time because the actual sink may at this time instead at initialization time because the actual sink may
change depending on the media contents, and this wont be known until change depending on the media contents, and this wont be known until
the pipeline is PLAYING and some media has been read. the pipeline is PLAYING and some media has been read.
@ -369,36 +350,25 @@ A new Step Event is created with `gst_event_new_step()`, whose
parameters basically specify the amount to skip (1 frame in the example) parameters basically specify the amount to skip (1 frame in the example)
and the new rate (which we do not change). and the new rate (which we do not change).
The video sink is grabbed from `playbin` in case we didnt have it yet, The video sink is grabbed from `playbin` in case we didnt have it yet,
just like before. just like before.
And with this we are done. When testing this tutorial, keep in mind that And with this we are done. When testing this tutorial, keep in mind that
backward playback is not optimal in many elements. backward playback is not optimal in many elements.
<table> > ![Warning](images/icons/emoticons/warning.png)
<tbody> >
<tr class="odd"> >Changing the playback rate might only work with local files. If you cannot modify it, try changing the URI passed to `playbin` in line 114 to a local URI, starting with `file:///`
<td><img src="images/icons/emoticons/warning.png" width="16" height="16" /></td>
<td><p>Changing the playback rate might only work with local files. If you cannot modify it, try changing the URI passed to <code>playbin</code> in line 114 to a local URI, starting with <code>file:///</code></p></td>
</tr>
</tbody>
</table> </table>
# Conclusion ## Conclusion
This tutorial has shown: This tutorial has shown:
- How to change the playback rate using a Seek Event, created with - How to change the playback rate using a Seek Event, created with
`gst_event_new_seek()` and fed to the pipeline `gst_event_new_seek()` and fed to the pipeline
with `gst_element_send_event()`. with `gst_element_send_event()`.
- How to advance a video frame-by-frame by using Step Events, created - How to advance a video frame-by-frame by using Step Events, created
with `gst_event_new_step()`. with `gst_event_new_step()`.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!
## Attachments:
![](images/icons/bullet_blue.gif)
[basic-tutorial-13.c](attachments/327800/2424883.c) (text/plain)
![](images/icons/bullet_blue.gif)
[vs2010.zip](attachments/327800/2424884.zip) (application/zip)

View file

@ -12,8 +12,7 @@ any time, in a variety of ways. This tutorial shows:
- How to access and manipulate this data. - How to access and manipulate this data.
[Playback tutorial 3: Short-cutting the [](sdk-playback-tutorial-short-cutting-the-pipeline.md) explains
pipeline](sdk-playback-tutorial-short-cutting-the-pipeline.md) explains
how to achieve the same goals in a playbin-based pipeline. how to achieve the same goals in a playbin-based pipeline.
## Introduction ## Introduction
@ -69,8 +68,7 @@ this simplified vision should suffice for now.
As an example, a `filesrc` (a GStreamer element that reads files) As an example, a `filesrc` (a GStreamer element that reads files)
produces buffers with the “ANY” caps and no time-stamping information. produces buffers with the “ANY” caps and no time-stamping information.
After demuxing (see [Basic tutorial 3: Dynamic After demuxing (see [](sdk-basic-tutorial-dynamic-pipelines.md))
pipelines](sdk-basic-tutorial-dynamic-pipelines.md))
buffers can have some specific caps, for example “video/x-h264”. After buffers can have some specific caps, for example “video/x-h264”. After
decoding, each buffer will contain a single video frame with raw caps decoding, each buffer will contain a single video frame with raw caps
(for example, “video/x-raw-yuv”) and very precise time stamps indicating (for example, “video/x-raw-yuv”) and very precise time stamps indicating
@ -78,8 +76,7 @@ when should that frame be displayed.
### This tutorial ### This tutorial
This tutorial expands [Basic tutorial 7: Multithreading and Pad This tutorial expands [](sdk-basic-tutorial-multithreading-and-pad-availability.md) in
Availability](sdk-basic-tutorial-multithreading-and-pad-availability.md) in
two ways: firstly, the `audiotestsrc` is replaced by an `appsrc` that two ways: firstly, the `audiotestsrc` is replaced by an `appsrc` that
will generate the audio data. Secondly, a new branch is added to the will generate the audio data. Secondly, a new branch is added to the
`tee` so data going into the audio sink and the wave display is also `tee` so data going into the audio sink and the wave display is also
@ -498,9 +495,7 @@ gst_buffer_unref (buffer);
``` ```
Once we have the buffer ready, we pass it to `appsrc` with the Once we have the buffer ready, we pass it to `appsrc` with the
`push-buffer` action signal (see information box at the end of [Playback `push-buffer` action signal (see information box at the end of [](sdk-playback-tutorial-playbin-usage.md)), and then
tutorial 1: Playbin
usage](sdk-playback-tutorial-playbin-usage.md)), and then
`gst_buffer_unref()` it since we no longer need it. `gst_buffer_unref()` it since we no longer need it.
``` c ``` c
@ -539,8 +534,7 @@ This tutorial has shown how applications can:
- Manipulate this data by accessing the `GstBuffer`. - Manipulate this data by accessing the `GstBuffer`.
In a playbin-based pipeline, the same goals are achieved in a slightly In a playbin-based pipeline, the same goals are achieved in a slightly
different way. [Playback tutorial 3: Short-cutting the different way. [](sdk-playback-tutorial-short-cutting-the-pipeline.md) shows
pipeline](sdk-playback-tutorial-short-cutting-the-pipeline.md) shows
how to do it. how to do it.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon\!

View file

@ -1,6 +1,6 @@
# Basic tutorial 12: Streaming # Basic tutorial 12: Streaming
# Goal ## Goal
Playing media straight from the Internet without storing it locally is Playing media straight from the Internet without storing it locally is
known as Streaming. We have been doing it throughout the tutorials known as Streaming. We have been doing it throughout the tutorials
@ -11,7 +11,7 @@ particular:
- How to enable buffering (to alleviate network problems) - How to enable buffering (to alleviate network problems)
- How to recover from interruptions (lost clock) - How to recover from interruptions (lost clock)
# Introduction ## Introduction
When streaming, media chunks are decoded and queued for presentation as When streaming, media chunks are decoded and queued for presentation as
soon as they arrive form the network. This means that if a chunk is soon as they arrive form the network. This means that if a chunk is
@ -26,7 +26,7 @@ waiting.
As it turns out, this solution is already implemented in GStreamer, but As it turns out, this solution is already implemented in GStreamer, but
the previous tutorials have not been benefiting from it. Some elements, the previous tutorials have not been benefiting from it. Some elements,
like the `queue2` and `multiqueue` found inside `playbin`, are capable like the `queue2` and `multiqueue` found inside `playbin`, are capable
of building this buffer and post bus messages regarding the buffer level of building this buffer and post bus messages regarding the buffer level
(the state of the queue). An application wanting to have more network (the state of the queue). An application wanting to have more network
resilience, then, should listen to these messages and pause playback if resilience, then, should listen to these messages and pause playback if
@ -45,18 +45,9 @@ When the clock is lost, the application receives a message on the bus;
to select a new one, the application just needs to set the pipeline to to select a new one, the application just needs to set the pipeline to
PAUSED and then to PLAYING again. PAUSED and then to PLAYING again.
# A network-resilient example ## A network-resilient example
Copy this code into a text file named `basic-tutorial-12.c`. Copy this code into a text file named `basic-tutorial-12.c`.
<table>
<tbody>
<tr class="odd">
<td><img src="images/icons/emoticons/information.png" width="16" height="16" /></td>
<td><p>This tutorial is included in the SDK since release 2012.7. If you cannot find it in the downloaded code, please install the latest release of the GStreamer SDK.</p></td>
</tr>
</tbody>
</table>
**basic-tutorial-12.c** **basic-tutorial-12.c**
@ -162,35 +153,24 @@ int main(int argc, char *argv[]) {
} }
``` ```
<table> > ![Information](images/icons/emoticons/information.png)
<tbody> > Need help?
<tr class="odd"> >
<td><img src="images/icons/emoticons/information.png" width="16" height="16" /></td> > If you need help to compile this code, refer to the **Building the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Build), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Build) or [Windows](sdk-installing-on-windows.mdb#InstallingonWindows-Build), or use this specific command on Linux:
<td><div id="expander-8150053" class="expand-container"> >
<div id="expander-control-8150053" class="expand-control"> > `` gcc basic-tutorial-12.c -o basic-tutorial-12 `pkg-config --cflags --libs gstreamer-1.0` ``
<span class="expand-control-icon"><img src="images/icons/grey_arrow_down.gif" class="expand-control-image" /></span><span class="expand-control-text">Need help? (Click to expand)</span> >
</div> >If you need help to run this code, refer to the **Running the tutorials** section for your platform: [Linux](sdk-installing-on-linux.md#InstallingonLinux-Run), [Mac OS X](sdk-installing-on-mac-osx.md#InstallingonMacOSX-Run) or [Windows](sdk-installing-on-windows.md#InstallingonWindows-Run).
<div id="expander-content-8150053" class="expand-content"> >
<p>If you need help to compile this code, refer to the <strong>Building the tutorials</strong> section for your platform: <a href="Installing%2Bon%2BLinux.html#InstallingonLinux-Build">Linux</a>, <a href="Installing%2Bon%2BMac%2BOS%2BX.html#InstallingonMacOSX-Build">Mac OS X</a> or <a href="Installing%2Bon%2BWindows.html#InstallingonWindows-Build">Windows</a>, or use this specific command on Linux:</p> > This tutorial opens a window and displays a movie, with accompanying audio. The media is fetched from the Internet, so the window might take a few seconds to appear, depending on your connection speed. In the console window, you should see a buffering message, and playback should only start when the buffering reaches 100%. This percentage might not change at all if your connection is fast enough and buffering is not required.
<div class="panel" style="border-width: 1px;"> >
<div class="panelContent"> > Required libraries: `gstreamer-1.0`
<p><code>gcc basic-tutorial-10.c -o basic-tutorial-10 `pkg-config --cflags --libs gstreamer-1.0`</code></p>
</div>
</div>
<p>If you need help to run this code, refer to the <strong>Running the tutorials</strong> section for your platform: <a href="Installing%2Bon%2BLinux.html#InstallingonLinux-Run">Linux</a>, <a href="Installing%2Bon%2BMac%2BOS%2BX.html#InstallingonMacOSX-Run">Mac OS X</a> or <a href="Installing%2Bon%2BWindows.html#InstallingonWindows-Run">Windows</a></p>
<p>This tutorial opens a window and displays a movie, with accompanying audio. The media is fetched from the Internet, so the window might take a few seconds to appear, depending on your connection speed. In the console window, you should see a buffering message, and playback should only start when the buffering reaches 100%. This percentage might not change at all if your connection is fast enough and buffering is not required.</p>
<p>Required libraries: <code>gstreamer-1.0</code></p>
</div>
</div></td>
</tr>
</tbody>
</table>
# Walkthrough ## Walkthrough
The only special thing this tutorial does is react to certain messages; The only special thing this tutorial does is react to certain messages;
therefore, the initialization code is very simple and should be therefore, the initialization code is very simple and should be
self-explanative by now. The only new bit is the detection of live self-explanative by now. The only new bit is the detection of live
streams: streams:
``` c ``` c
@ -208,14 +188,14 @@ if (ret == GST_STATE_CHANGE_FAILURE) {
Live streams cannot be paused, so they behave in PAUSED state as if they Live streams cannot be paused, so they behave in PAUSED state as if they
were in the PLAYING state. Setting live streams to PAUSED succeeds, but were in the PLAYING state. Setting live streams to PAUSED succeeds, but
returns `GST_STATE_CHANGE_NO_PREROLL`, instead of returns `GST_STATE_CHANGE_NO_PREROLL`, instead of
`GST_STATE_CHANGE_SUCCESS` to indicate that this is a live stream. We `GST_STATE_CHANGE_SUCCESS` to indicate that this is a live stream. We
are receiving the NO\_PROROLL return code even though we are trying to are receiving the NO\_PROROLL return code even though we are trying to
set the pipeline to PLAYING, because state changes happen progressively set the pipeline to PLAYING, because state changes happen progressively
(from NULL to READY, to PAUSED and then to PLAYING). (from NULL to READY, to PAUSED and then to PLAYING).
We care about live streams because we want to disable buffering for We care about live streams because we want to disable buffering for
them, so we take note of the result of `gst_element_set_state()` in the them, so we take note of the result of `gst_element_set_state()` in the
`is_live` variable. `is_live` variable.
Lets now review the interesting parts of the message parsing callback: Lets now review the interesting parts of the message parsing callback:
@ -239,7 +219,7 @@ case GST_MESSAGE_BUFFERING: {
First, if this is a live source, ignore buffering messages. First, if this is a live source, ignore buffering messages.
We parse the buffering message with `gst_message_parse_buffering()` to We parse the buffering message with `gst_message_parse_buffering()` to
retrieve the buffering level. retrieve the buffering level.
Then, we print the buffering level on the console and set the pipeline Then, we print the buffering level on the console and set the pipeline
@ -264,7 +244,7 @@ For the second network issue, the loss of clock, we simply set the
pipeline to PAUSED and back to PLAYING, so a new clock is selected, pipeline to PAUSED and back to PLAYING, so a new clock is selected,
waiting for new media chunks to be received if necessary. waiting for new media chunks to be received if necessary.
# Conclusion ## Conclusion
This tutorial has described how to add network resilience to your This tutorial has described how to add network resilience to your
application with two very simple precautions: application with two very simple precautions:
@ -275,11 +255,4 @@ application with two very simple precautions:
Handling these messages improves the applications response to network Handling these messages improves the applications response to network
problems, increasing the overall playback smoothness. problems, increasing the overall playback smoothness.
It has been a pleasure having you here, and see you soon\! It has been a pleasure having you here, and see you soon!
## Attachments:
![](images/icons/bullet_blue.gif)
[basic-tutorial-12.c](attachments/327806/2424843.c) (text/plain)
![](images/icons/bullet_blue.gif)
[vs2010.zip](attachments/327806/2424844.zip) (application/zip)