Live streaming ( play && publish ) over rtmp/rtmfp with codecs ( h264 && PCMU ) on Android/iOS

I have a question: On what operating systems(android,ios) can be played and published stream with codecs H264 and PCMU with rtmp and rtmfp protocols(live streaming)?    
     As far as I found out with the codecs can be played on Android(rtmp protocol). On IOS video is not displayed, as I understood AIR environment cuting it.
Another question regarding video texture. Will be included in future releases support for live playing h264+PCMU on IOS?

On iOS, you'll need to be playing a HLS stream for h264 to decode when streaming from a remote server.

Similar Messages

  • Use HTTP Dynamic Streaming (HDS) and HTTP Live Streaming (HLS) to serve live streams to clients over HTTP

    I have created a live stream of a video and it gets stored in live folder.
    Now i need to use HTTP Dynamic Streaming (HDS) and HTTP Live Streaming (HLS) to serve live streams to clients over HTTP, publish the streams to the HTTP Live Packager service on Flash Media Server.
    So what necessary steps do I  need to follow to do that ??

    You need to generate a manifest file using Configurator tool and placed it under the webroot directory.
    C:\Program Files\Adobe\Flash Media Server 4.5\tools\f4mconfig\configurator

  • Archiving live stream at FMS and injecting metadata: VP6 good h264 not

    When I record a live stream at FMS, one in which I've injected  metadata in my main.asc file, the archived file plays back fine.  The  metadata plays back too.  I'm able to retreive it just fine - if I  encode VP6.
    If I encode h.264 the file plays back but  the metadata does not.  The fact that the archived file is created and  plays back tells me things are wired correctly.  The only thing I  changed is the format.
    According to FMS docs (http://help.adobe.com/en_US/FlashMediaServer/3.5_SS_ASD/WS5b3ccc516d4fbf351e63e3d11a11afc9 5e-7e42.html#WS5b3ccc516d4fbf351e63e3d11a11afc95e-7f35)
    ..."The recording format is determined by the filename you pass to the Stream.get()
    method."
    So my record code looks like the following:
    application.onPublish = function(client, stream) {
         trace("onPublish");
         s = Stream.get("mp4:streamname.f4v");
         if(s){
             s.record();
         this.doRepublish(this.nc, stream);
    My code that injects the data in to the stream looks like this:
    Client.prototype.sendDataEvent = function(data) {
         trace("Call to sendDataEvent...");
         this.newStream = Stream.get("mp4:streamname.f4v");
         this.newStream.send("onTextData",data);
    All must be wired  correctly because the metadata comes through during the live stream.  On  play back of the archive though, the metadata doesn't appear to be  there.
    Any thoughts?
    Thanks

    My apologies on the s.play() confusion.  I had been trying different versions of the code and posted the one without it.
    Whether I include s.play() or not the file gets created.  Here are the various versions of the onPublish() function I've tried (differences in red):
    1.
    application.onPublish = function(client, stream) {
        trace("onPublish");   
        s = Stream.get("mp4:streamname.f4v");
        if(s){
            s.record();
            s.play("mp4:streamname.f4v");
        this.doRepublish(this.nc, stream);
    2.
    application.onPublish = function(client, stream) {
        trace("onPublish");   
        s = Stream.get("mp4:streamname.f4v");
        if(s){
            s.record();
            s.play(stream);
        this.doRepublish(this.nc, stream);
    3.
    application.onPublish = function(client, stream) {
         trace("onPublish");   
         s = Stream.get("mp4:streamname.f4v");
         if(s){
             s.record();
         this.doRepublish(this.nc, stream);
    All produce the same result - an archived file called mp4:streamname.f4v in my streams folder.  This file plays back fine but does not play back the commands.
    On your other question, about things working fine for VP6, it works fine for FLV.  A file called streamname.flv is produced.  This file plays back fine and does indeed play back commands baked into the file as well.  This is what makes me believe the code is not the problem.  If it works perfectly for one format, there would seem to be very little I could do in my code to break things for the other.
    Can you try this using the record() code snippets in the live docs Stream.record() section?
    http://help.adobe.com/en_US/FlashMediaServer/3.5_SS_ASD/WS5b3ccc516d4fbf351e63e3d11a11afc9 5e-7e42.html#WS5b3ccc516d4fbf351e63e3d11a11afc95e-7f35
    All you'd need is the code snippets there to record your live stream and another server side function to inject commands into that live stream. Here is that function:
    Client.prototype.sendDataEvent = function(data) {
        trace("Call to sendDataEvent...");
        this.newStream = Stream.get("mp4:streamname.f4v");
        this.newStream.send("onTextData",data);
    Do something simple like call onTextData and pass some text in the data parameter.  Then on the client side viewer, handle the onTextData method.  It will receive the text.  Display it in a text area or something.
    If you record while injecting this text into your stream, the text should display on playback of the archived file.  It will if you encode VP6/FLV, but not if you encode H.264/F4V.
    Let me know what you discover.

  • Playing mp3 over RTMP (Flash Media Server)

    I've got some mp3 files that are served from a Flash Media Server.  I want to write a Flex application to play them.  Do I need to use NetConnection/NetStream or can I do it with just the Sound class?  I've got a sample app that can connect to the stream with NetConnection/NetStream but I don't know how to connect the netstream data to something that will actually play the sound.
    Will I be able to read the id3 data of the files with this?

    I actually got it working , let me know if that helps
    Create NetStream and attach it to Video Control, then play();
    @return void
      private function connectStream():void {
       // Assign a new Netstream Connection
    if(_nsStream == null){
      nsStream = new NetStream(ncConnection);
      nsStream.addEventListener(NetStatusEvent.NETSTATUS, netStatusHandler);
      nsStream.addEventListener( AsyncErrorEvent.ASYNCERROR,
    handleCatchAllErrors );
      nsStream.addEventListener( IOErrorEvent.IOERROR, handleCatchAllErrors );
      _nsStream.client = {onMetaData:handleMetaData, close:handleCloseStream};
      _nsStream.bufferTime = _videoData.buffer;
       // Assign a new ID3 Netstream Connection
    if( id3ns == null && _videoData.mediaType=="mp3") {
      id3ns = new NetStream(_ncConnection);
      id3ns.addEventListener( AsyncErrorEvent.ASYNC_ERROR,
    handleCatchAllErrors );
      id3ns.addEventListener( IOErrorEvent.IO_ERROR, handleCatchAllErrors );
      id3ns.addEventListener( NetStatusEvent.NET_STATUS, netStatusHandler );
        //HACK: we are required to pass in static duration via XML till we get
    server-side actionscript working
      metadata.duration=videoData.duration;
        id3ns.play(_id3_path);
      id3ns.client = {onId3:handleOnId3};
        _id3Field = new TextField();
      _id3Field.width = _video.width;
      _id3Field.wordWrap = true;
      _id3Field.multiline = true;
      _id3Field.antiAliasType = flash.text.AntiAliasType.ADVANCED;
      //_id3Field.defaultTextFormat = _ccTextFormat;
      _id3Field.selectable = false;
      this.addChild(_id3Field);
       _video.smoothing = _playerData.smoothing;
    video.attachNetStream(nsStream);
        nsStream.play(videoData.file);
      _isPlaying = true;
      createSoundObject();
       this.addEventListener(Event.ENTER_FRAME, onEnterFrame);
    trace("[MEDIA FILE] "+_videoData.file);
    Get MP3 data from ID3 tag
    @return void
    private function handleOnId3( obj:Object ):void{
    if(obj["artist"] != undefined){
      _id3Field.text = "by: "obj["artist"].toString() + "\n";
    if(obj["songtitle"] != undefined){
      _id3Field.text = "title: "obj["songtitle"].toString();
    //if we wan to inject media with other data, add it here
    for( var b:String in obj ) {
      //_id3Field.text += b + ": " + obj[ b ];
      //trace(b + ": " + obj[ b ]);

  • Listening to live streaming talk radio over your iphone.

    Well I live in South Florida and there is a talk radio station that actually has their live broadcast streaming on their website. Well since the iphone is of course capable of using the internet, I was wondering if there is anyway possible to get like windows media player or some kind of plug-in or something for safari that will allow the iphone to do this.

    Tbo6652 wrote:
    Adobe, the makers of Flash, i've heard, is trying to make an application for the iPhone...since this is only something i heard(and not from Adobe themselves) i cant say if its true or not...
    if it is true, then it should be available this june or soon after, thought the AppStore
    cross your fingers
    I do recall Adobe saying that they have the SDK. That doesn't meant that they can or will develop flash for the iPhone.

  • Problem of republishing remote rtmp live stream with AMF

    Hi, Guys
    I am trying recieving one live stream from a remote rtmp server and republishing this stream in AMF.
    a remote rtmp (rtmp://192.168.1.100/live/remotestream)
    AMF instance(codes in application/app/main.asc) recieve and republish to another AMF instance.
    I can watch with url(rtmp://192.168.1.100/live/remotestream ) in a player (ex. OSMF player).
    But,no data from remote rtmp streaming can be recieved in AMF.
    Here is the log:
    NetConnection.Connect.Success
    Sending error message: Method not found (onBWDone).
    mystream.onStatus: NetStream.Publish.Start
    mystream.onStatus: NetStream.Play.Reset
    My codes in application/app/main.asc
    application.onAppStart = function()
              mystream = Stream.get("myvideo");
              myRemoteConn = new NetConnection();
              myRemoteConn.connect("rtmp://192.168.1.100/live");
              myRemoteConn.onStatus =function(info){
               trace(info.code);
               if(info.code == "NetConnection.Connect.Success"){
                    mystream.play("remotestream", -1, -1, true, myRemoteConn);
                    //mystream.play("sample", 0, -1, true);<-----if play a local flv video file ,it works.republish fine.
              nc = new NetConnection();
              nc.connect("rtmp://localhost/demo");
              ns = new NetStream(nc); 
              ns.setBufferTime(2);
        mystream.onStatus = function(sinfo)
            trace("mystream.onStatus: "+sinfo.code);
            if(sinfo.code == "NetStream.Publish.Start")
                                  attach_retVal = ns.attach(mystream);
                if(attach_retVal==true){
                        ns.publish(mystream.name,"live");

    On the Macintosh side there is the Camtwist app which allows you to do this and it works smoothly with FMLE. On the PC side I´m not sure but there just got to be more apps capable of this than on the mac side.

  • Problem playing remote live stream from FMS

    Hello all,
    I'm having problems playing remote live streams from FMS (server-side) that I can play with any other player (client-side)
    Example of server-side application which plays a remote live stream:
    nc = new NetConnection();
    nc.connect("rtmp://remoteserver/live");
    stream = Stream.get("localStream");
    stream.play("remoteStream.flv", -1, -1, true, nc);
    This code works only sometimes.
    Most of the times, FMS is able to connect to remote string and the localStream dispatches events:
       NetStream.Publish.Strart
       NetStream.Play.Reset
    In this case, FMS is publishing the local stream but it is not playing the remoteStrem on it.
    The rest of the times, localStream dispatches events:
       NetStream.Publish.Strart
       NetStream.Play.Reset
       NetStream.Play.Start
       NetStream.Data.Start
    In this case FMS plays the remoteStream correctly.
    Any hint to solve this issue?
    Regards.

    Thanks, I tried your code and it works when playing a remote live stream on another FMS.
    But the remote live stream is not on a FMS but on a Wowza Server which re-streams an RTSP stream over RTMP.
    With this code:
    var nc;
    var myStream;
    application.onAppStart = function(){
         nc = new NetConnection();
         myStream = Stream.get("localstream");
         myStream.onStatus= function(info){
               trace(info.code);
         nc.onStatus = function(info){
               trace (info.code);
               if(info.code == "NetConnection.Connect.Success" ){
                    myStream.play("remoteLive.sdp", -1, -1, true, nc);
         nc.connect("rtmp://remoteServer/live");
    Every live stream player I tried was able to play the remote stream "remoteLive.sdp", but FMS play it only sometimes. This is the output log of the FMS application:
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset        <--------------------------------- Stuck there, stream won't play
    Unloaded application instance wowza4/_definst_ <--- Reload app
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset        <--------------------------------- Stuck there, stream won't play
    Unloaded application instance wowza4/_definst_ <--- Reload app
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset
    NetStream.Play.Start         <--------------------------------- Stream is playing just fine
    NetStream.Data.Start
    Unloaded application instance wowza4/_definst_ <--- Reload app
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset       <--------------------------------- Stuck there, stream won't play
    Any idea? Why FMS won't play it everytime?
    Regards

  • Adobe Air 3.0 iOS streaming H264/Speex over RTMP -- No Video

    Problem:
    H264/Speex RTMP stream from Flash Media Server doesn't display video.  Audio plays fine.
    Conditions:
    Adobe Air 3.0
    iOS device (iPad 2, iOS4.3)
    App settings:
    <renderMode>direct</renderMode>
    Flash Builder 4.5.1 compiler settings:
    -swf-version=13
    -target-player=11.0.0
    I have tried using both stageVideo and the regular Video object to render an H264 signal streaming from Flash Media Server over rtmp, but with no luck.  I can hear the audio, but the video is never rendered.
    I CAN see H263 video when streamed over RTMP using this setup (with just the Video object). 
    I can also stream a locally stored mp4 (H264/AAC) file over rtmp from the iOS device and play it locally just fine (using a stageVideo object).
    I have attempted this with both stageVideo (which works fine when streaming an mp4 file from the iOS device) and with the regular Video object (the regular Video object handles H263 just fine streaming down from FMS over rtmp).  I've also played around with backgroundAlpha = "0" for this case, all with no luck.
    I've followed all the online instructions and tutorials to get this working for the past week, but have not had any luck.  I'm happy to post more code, but this approach is fairly straight forward and has been described multiple times around these forums and in the adobe blog posts.  It goes like this:
         - instantiate NetConnection (_nc) to FMS and wait for successful connection:
                   _ns = new NetConnection();  
                   _nc.connect(serverAddress);
         - instantiate NetStream (_ns) and connect using NetConnection: 
                   _ns = new NetStream(_nc);
         - setup listener for StageVideoAvailabilityEvent; if stageVideo becomes available, attach _ns: 
                  _stageVideo = stage.stageVideos[0];
                  _stageVideo = _stageVideo.attachNetStream(_ns);
         - if stageVideo ISN'T available, fall back to Video object:
                   _video = new Video();
                   _video.attachNetStream(_ns);
                   stage.addChild(_video);
         - play netStream app:
                   _ns.play(appName);
    My renderMode is set correctly ("direct"), and from what I can tell, this isn't a backgroundAlpha issue.
    I have seen this problem described elsewhere on this forums and on the net. 
    I'd like to know from Adobe: are there any compatibility issues with this approach?  There doesn't seem to be any type of compatibility chart listing which video codecs and transport protocols are available on the different mobile platforms.  Since mobile deployment is so heaviliy fragmented, a descriptive chart or table like this would be helpful for those of us developing -- at least for the main mobile platforms predominantly in use.
    Other forum posts similar to this problem:
    http://forums.adobe.com/message/3981541#3981541
    http://forums.adobe.com/message/3954578#3954578
    Thanks in advance!

    I do hope someone from Adobe is "hearing" this, guys. The lack of RTMP-based H.264 video on the Air for iOS is a major problem, indeed.
    As Fabio Sonnati mentioned in http://sonnati.wordpress.com/2011/04/26/air-2-6-for-ios-and-video-play back/, AIR for iOS does support HTTP streaming (via HLS) of h.264 videos. However, when streaming via RTMP, AIR for iOS only supports VP6 and Spark – a couple of old, retired codecs.
    While HTTP streaming (HLS) seems to be a good option for those who simply want to “play a video” in iOS, I do believe it has some severe limitations, especially for live-communications. I’d like to share some of these thoughts with you.
    1. HLS has ridiculously high latency for live videos (around 40 seconds), when compared to RTMP. Although this may not be a problem for on-demand videos, it sure is a great problem for anyone doing serious live-communications applications (such as webconferencing, live webcasting with audience interaction or Skype-like video chats), which require near-zero latency.
    2. Perhaps someone can correct me on this (hopefully!), but as far as I know, HTTP streaming will not allow cuepoints to be read from videos. This is particularly painful for anyone doing video-triggered actions, such as slide changes (for webinar apps), subtitling or live closed captioning, etc. I read somewhere that OSMF player allows cuepoints (or "temporal metadata". See http://blogs.adobe.com/osmf/2009/11/cue_point_support_in_osmf.html), but I haven't been able to test it myself.
    3. Although HLS it is quite compatible with firewalls (since it flows through port 80), RTMP with tunnelling also flows through port 80 or 443, which adds great compatibility, even on very restricted networks. Our experience with very large clients proves that, hands down.
    In other words, HTTP/HLS streaming is Ok. But it simply does *not* fit into every shoe that RTMP does. We do believe that RTMP remains as our best option for live streaming or serious streaming-oriented *apps* (in which things more complex than “mere video playing in a window” actually happen).
    That all said, I do believe we should let Adobe know about this need. The fact that RTMP streaming in AIR for iOS is limited to VP6 and Spark, which are two “dead” codecs, still puts us, Air developers, in a very fragile position in terms of what we can accomplish with video in iOS.
    I’m sure some of you cheered when you heard about Flash Player 11 having h.264 video encoding. This, (plus the echo cancellation feature that came in 10.3) opened great doors for great Unified-Communication applications to be developed for Flash/Air. Now, it’s undeniable that clients want those applications running on tablets, especially the iPad.
    Not being able to use h.264 via RTMP on iOS is certainly a huge step backwards. Anyone shares this same opinion? What do you guys believe to be the best option to let Adobe really know about this need? Is this limitation a simple lack-of-a-feature (which can be fixed by Adobe) or is this some imposed thing by Apple?
    Just one final note: Air for Android does *not* have the same limitation. It does allow RTMP streaming of h.264.
    Thanks for your attention,
    Helder Conde

  • Live Stream Issues with FP 11.2

    We're seeing lots of bugs within the NetStream when connecting live streams and publishing/playing the stream via direct peer connections.
    Specifically when attaching a new Camera the receiving peer can no longer see the video stream. Also when toggling publish(bull) publish(streamName) from the publisher side and similarly toggling play(null) and play(streamName) from the subscriber side the streams don't play (render video) anymore even though i can see through the netwrok that the bytes are coming through.
    Along with this we seem to be getting new kinds of NetStatus errors ever since we started testing - "NetConnection.Call.BadVersion" was one we've seen but can't make any sense out of given we're using DIRECT peer to peer connections for the netstream classes.
    Lastly on general we're seeing lots of instability issues with on the h264settings applied to the nestream is this still in BETA? Instibility = crashing browser mostly when showing multiple video such as in a group video chat.
    Are any of you seeing this as well?

    Hi Damorgan,
    As you suggested, I did fresh 11.2.0.2 installation and find the mentioned components are valid from v$registry.
    Only desupported component is UltraSearch that I was asked to remove during upgrade and I did so.
    COMP_NAME VERSION STATUS
    OLAP Catalog 11.2.0.2.0 VALID
    Spatia 11.2.0.2.0 VALID
    OLAP Analytic Workspace 11.2.0.2.0 VALID
    Also the metalink note: ID 270588.1 indicates that mentioned components exists for 11gR2.
    DBUA was asking to install these component prior upgrading. My first note has that DBUA instruction.
    Is there any way to select mentioned components with "software only" installation prior upgrading?
    Thanks.

  • How to secure Flash Media Streaming Server 3.5 (live stream connection)

    I have Flash Media Streaming Server 3.5 (not Interactive) running on RHEL5.5 x86_64 Linux.
    All is working well, however how do I prevent unauthorized access to connecting to the live stream and streaming content?
    How can I setup the server to require a user and password to stream live media to the server?
    Any help would be great, I am new to this product and I have been reading some documentation but I have not found a clear cut answer on how to force a username and password to connect to the server to stream live content only.
    I am using the Adobe FMS Apache install, what files need changing?
    I have the following files:
    /opt/adobe/fms/applications/live
    total 44
    -rwxr-xr-x 1 fms fms 1035 Jul 27 14:52 allowedHTMLdomains.txt
    -rwxr-xr-x 1 fms fms  992 Jul 27 14:52 allowedSWFdomains.txt
    -rwxr-xr-x 1 fms fms 1550 Jul 27 14:52 Application.xml
    -rwxr-xr-x 1 fms fms 4371 Jul 27 14:52 main.far
    -rwxr-xr-x 1 fms fms 2292 Jul 27 14:52 readme.txt
    I want to lock down a person from connecting to the server on the public internet and starting a live stream?
    Can this be done with a user name and password?

    If you are concerned about anyone starting publishing to your stream then we have security update on this. Please upgrade to 3.5.4 and now you can use FMS Authentication Add-in (used with FMLE as publisher). If you are facing problems with publisher which are not FMLE, please enable SWF Verification so that only valid SWF's connect to your FMS.
    You can find 3.5.4 update here : http://www.adobe.com/support/flashmediaserver/downloads_updaters.html
    You can find FMS Authentication Add-in on FMLE Download page: https://www.adobe.com/cfusion/entitlement/index.cfm?e=fmle3 (when you click on Agree - it will prompt for your Adobe ID - then when you reach Download page - you will get Optional Download of Authentication Add-in)

  • Custom metada and record live stream

    I tried to inject my own metadata into a live stream (webcam) recorded on FMS 4.5.1
    with no success. I was able to change only the standard metadata but not all my custom ones were recorded
    despite of all example on adobe livedocs, no luck.(as2 and as3)
    this example
    http://help.adobe.com/en_US/flashmediaserver/devguide/WS5b3ccc516d4fbf351e63e3d11a0773d56e -7ff6Dev.html
    doesn't work at all for custom metadat and EVEN recroded special metadata like createdby and creationdata
    example:
    metaData.createdby = String(streamTitle + " of "+fullName);
    metaData.creationdate = String(now.toString());
    metaData.tags = {createdby:metaData.createdby, creationdate:metaData.creationdate};
    metaData.width = Number(devicesQuality.width);
    metaData.height = Number(devicesQuality.height);
    nodeSendStream.send("@setDataFrame", "onMetaData", metaData);
    once the stream is recorded I can see only metData.width and height. createdby, creationdate and tags don't exist.
    notice: I encode and stream from FP11.1 in H2564 with speex or pcmu
    Thanks

    well,
    I exactly respected the example of the link you gave by modifying the Application.xml, but if it was a segment problem
    how can you explain that the NON custom metadas are successfully modified ?
    please check this link
    http://help.adobe.com/en_US/flashmediaserver/devguide/WS5b3ccc516d4fbf351e63e3d11a0773d56e -7ff6Dev.html
    I even tried the official property called createdby and creationdate without success. I was able to change only  width and height
    the code I use from the client is exactly the example of adobe manual, in AS2 and AS3
    Thanks

  • Flash Player Support for HTTP Live Streaming?

    Does Flash Player 11 / Air for TV 3.0 support HTTP Live Streaming ?

    If you need help with posting messages in the forum there is a test forum available at http://forums.adobe.com/community/general/forum_comments/testing

  • How to integrate flash media server 4.0 live streaming for iOS devices ?

    Hi All,
    I have website which has live streaming module its working fine, same module i want to integrate for iOS devices. For live video streaming we are using FMS 4.0. So please let me know how we can integrate this for iOS devices using flash media server 4.0.
    Thanks in Adavnce
    Mohammad Sharique

    You need to place the crossdomain.xml in the webroot folder. Create a text file in the webroot folder using notepad, and call it crossdomain.xml. The text below will give you a wide open access policy, which is fine for testing.
    <?xml version="1.0"?>
    <!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
    <cross-domain-policy>
              <allow-access-from domain="*" />
    </cross-domain-policy>
    For debugging HTTP streaming I recommend you get hold of something like Charles or Fiddler. These will greatly assist in pinpointing any issues.

  • Publish Live Stream with VP6 or H264

    Hi all,
            I am publishing live stream using web camera but the quality with respect to bandwidth is not good.
            I want to publish the live image using VP6 or H264 codec, how it can be done, please help.
           It is good with FMLE but trying to do in as3.
    Thanx in advance.

    FMLE can only publish via rtmp or rtmpt.

  • How to play http live streaming with VLC using FMS 4.0

    Hello,
    i'm using FMS 4 and i need to play my http live streaming event with VLC
    I have had the following result with firefox using the url : http://myip:8134/live/events/livepkgr.old/events/_definst_/liveevent.f4m
    <manifest>
    <id>
                        livepkgr.old/events/_definst_/liveevent
              </id><mimeType>
              </mimeType><streamType>
                        live
              </streamType><duration>
                        0
              </duration><media streamId="livestream1" bitrate="100" url="../../../../streams/livepkgr.old/streams/_definst_/livestream1/livestream1" bootstrapInfoId="bootstrap7580"><metadata>
                                  AgAKb25NZXRhRGF0YQgAAAAAAAhkdXJhdGlvbgAAAAAAAAAAAAAFd2lkdGgAQHQAAAAAAAAABmhlaWdodABAbgAAA AAAAAAMdmlkZW9jb2RlY2lkAgAESDI2NAAKYXZjcHJvZmlsZQBAU0AAAAAAAAAIYXZjbGV2ZWwAQD8AAAAAAAAADnZ pZGVvZnJhbWVyYXRlAP/4AAAAAAAAAAl0cmFja2luZm8KAAAAAgMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlA ECPQAAAAAAAAAhsYW5ndWFnZQIAA2VuZwAACQMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlAECPQAAAAAAAAAh sYW5ndWFnZQIAA2VuZwAACQANc21wdGVUaW1lSW5mbwMAC3NtcHRlT2Zmc2V0AAAAAAAAAAAAAAAJAAdjdXN0ZGVmC gAAAAAAAAk=
                        </metadata></media><media streamId="livestream2" bitrate="200"><metadata>
                        </metadata></media><media streamId="livestream3" bitrate="350"><metadata>
                        </metadata></media><bootstrapInfo profile="named" url="../../../../streams/livepkgr.old/streams/_definst_/livestream/livestream.bootstrap" id="bootstrap3136">
              </bootstrapInfo><media streamId="livestream" url="../../../../streams/livepkgr.old/streams/_definst_/livestream/livestream" bootstrapInfoId="bootstrap3136"><metadata>
                                  AgAKb25NZXRhRGF0YQgAAAAAAAhkdXJhdGlvbgAAAAAAAAAAAAAFd2lkdGgAQHQAAAAAAAAABmhlaWdodABAbgAAA AAAAAAMdmlkZW9jb2RlY2lkAgAESDI2NAAKYXZjcHJvZmlsZQBAU0AAAAAAAAAIYXZjbGV2ZWwAQD8AAAAAAAAADnZ pZGVvZnJhbWVyYXRlAP/4AAAAAAAAAAl0cmFja2luZm8KAAAAAgMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlA ECPQAAAAAAAAAhsYW5ndWFnZQIAA2VuZwAACQMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlAECPQAAAAAAAAAh sYW5ndWFnZQIAA2VuZwAACQANc21wdGVUaW1lSW5mbwMAC3NtcHRlT2Zmc2V0AAAAAAAAAAAAAAAJAAdjdXN0ZGVmC gAAAAAAAAk=
                        </metadata></media><bootstrapInfo profile="named" url="../../../../streams/livepkgr.old/streams/_definst_/livestream1/livestream1.bootstrap " id="bootstrap7580">
              </bootstrapInfo>
    </manifest>
    How can i play that event with VLC?
    Regards,

    Problem solved  - thanks to JayCharles. I forgot to set the live attribute. Here's the working code:
    <?xml version="1.0"?>
    <!-- controls\videodisplay\VideoDisplayFMS.mxml -->
    <mx:Application xmlns:mx="http://www.adobe.com/2006/mxml">
        <mx:HBox>
             <mx:VideoDisplay
                 live="true"
                 autoBandWidthDetection="false"
                 source="rtmp://localhost/live/livestream"/>
         </mx:HBox>
    </mx:Application>

Maybe you are looking for