How to play rtmp live streaming through VideoElement??

Hi , i'm trying to play live streaming by using VideoElement
like this :
<s:VideoElement>
<s:StreamingVideoSource serverURI="rtmp://192.168.0.136/tv" live="true">
<s:StreamItem streamName="stream" />
</s:StreamingVideoSource>
</s:VideoElement>
but nothing occurred ........did i use the wrong syntax @@a  or....?

Anyone?

Similar Messages

  • How to play http live streaming with VLC using FMS 4.0

    Hello,
    i'm using FMS 4 and i need to play my http live streaming event with VLC
    I have had the following result with firefox using the url : http://myip:8134/live/events/livepkgr.old/events/_definst_/liveevent.f4m
    <manifest>
    <id>
                        livepkgr.old/events/_definst_/liveevent
              </id><mimeType>
              </mimeType><streamType>
                        live
              </streamType><duration>
                        0
              </duration><media streamId="livestream1" bitrate="100" url="../../../../streams/livepkgr.old/streams/_definst_/livestream1/livestream1" bootstrapInfoId="bootstrap7580"><metadata>
                                  AgAKb25NZXRhRGF0YQgAAAAAAAhkdXJhdGlvbgAAAAAAAAAAAAAFd2lkdGgAQHQAAAAAAAAABmhlaWdodABAbgAAA AAAAAAMdmlkZW9jb2RlY2lkAgAESDI2NAAKYXZjcHJvZmlsZQBAU0AAAAAAAAAIYXZjbGV2ZWwAQD8AAAAAAAAADnZ pZGVvZnJhbWVyYXRlAP/4AAAAAAAAAAl0cmFja2luZm8KAAAAAgMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlA ECPQAAAAAAAAAhsYW5ndWFnZQIAA2VuZwAACQMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlAECPQAAAAAAAAAh sYW5ndWFnZQIAA2VuZwAACQANc21wdGVUaW1lSW5mbwMAC3NtcHRlT2Zmc2V0AAAAAAAAAAAAAAAJAAdjdXN0ZGVmC gAAAAAAAAk=
                        </metadata></media><media streamId="livestream2" bitrate="200"><metadata>
                        </metadata></media><media streamId="livestream3" bitrate="350"><metadata>
                        </metadata></media><bootstrapInfo profile="named" url="../../../../streams/livepkgr.old/streams/_definst_/livestream/livestream.bootstrap" id="bootstrap3136">
              </bootstrapInfo><media streamId="livestream" url="../../../../streams/livepkgr.old/streams/_definst_/livestream/livestream" bootstrapInfoId="bootstrap3136"><metadata>
                                  AgAKb25NZXRhRGF0YQgAAAAAAAhkdXJhdGlvbgAAAAAAAAAAAAAFd2lkdGgAQHQAAAAAAAAABmhlaWdodABAbgAAA AAAAAAMdmlkZW9jb2RlY2lkAgAESDI2NAAKYXZjcHJvZmlsZQBAU0AAAAAAAAAIYXZjbGV2ZWwAQD8AAAAAAAAADnZ pZGVvZnJhbWVyYXRlAP/4AAAAAAAAAAl0cmFja2luZm8KAAAAAgMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlA ECPQAAAAAAAAAhsYW5ndWFnZQIAA2VuZwAACQMABmxlbmd0aAAAAAAAAAAAAAAJdGltZXNjYWxlAECPQAAAAAAAAAh sYW5ndWFnZQIAA2VuZwAACQANc21wdGVUaW1lSW5mbwMAC3NtcHRlT2Zmc2V0AAAAAAAAAAAAAAAJAAdjdXN0ZGVmC gAAAAAAAAk=
                        </metadata></media><bootstrapInfo profile="named" url="../../../../streams/livepkgr.old/streams/_definst_/livestream1/livestream1.bootstrap " id="bootstrap7580">
              </bootstrapInfo>
    </manifest>
    How can i play that event with VLC?
    Regards,

    Problem solved  - thanks to JayCharles. I forgot to set the live attribute. Here's the working code:
    <?xml version="1.0"?>
    <!-- controls\videodisplay\VideoDisplayFMS.mxml -->
    <mx:Application xmlns:mx="http://www.adobe.com/2006/mxml">
        <mx:HBox>
             <mx:VideoDisplay
                 live="true"
                 autoBandWidthDetection="false"
                 source="rtmp://localhost/live/livestream"/>
         </mx:HBox>
    </mx:Application>

  • Could you play a live stream from safari to apple tv without going through itunes?

    could you play a live stream from safari to apple tv without going through itunes?

    There is no way to stream Safari content, with or without iTunes, to your AppleTV.
    The only way would be to use a mirroring feature given that you have an AppleTV 2nd gen or newer and a 2011 or newer MBP running mountain lion.
    If not, some third party software like AirParrot can also gives you way to mirror your MBP display on the TV.

  • Unable to play Level3 live stream with OSMF v1.0

    Hi,
    I'm trying to play a live stream hosted by level3. When adding the mp4: prefix to my rtmp stream url I am able to connect but get a mediaError:
    "Failed to play  (stream ID: 1)".
    parallelElement.addChild(mediaFactory.createMediaElement(new URLResource(config.rtmpUrl)));
    rtmpUrl: rtmp://link-to-level3-server.net/app/mp4:name/streamname/live
    I've tested the same stream in flowplayer and jwplayer; both work like a charm. I've noticed that Level3 is listed as a planned CDN  provider on the osmf page. Do I have to wait for the level3 plugin to be released or am I missing someting here?

    I just found this post: http://forums.adobe.com/message/2782150#2782150
    Adding a streamingURLResource like below doesn't give any mediaErrors.
    var resource:StreamingURLResource = new StreamingURLResource(config.rtmpUrl);
    resource.urlIncludesFMSApplicationInstance = true;
    parallelElement.addChild(new VideoElement(resource));
    I have to test it when I'm back at the office but it seems promising.
    Keep you posted!

  • Problem playing remote live stream from FMS

    Hello all,
    I'm having problems playing remote live streams from FMS (server-side) that I can play with any other player (client-side)
    Example of server-side application which plays a remote live stream:
    nc = new NetConnection();
    nc.connect("rtmp://remoteserver/live");
    stream = Stream.get("localStream");
    stream.play("remoteStream.flv", -1, -1, true, nc);
    This code works only sometimes.
    Most of the times, FMS is able to connect to remote string and the localStream dispatches events:
       NetStream.Publish.Strart
       NetStream.Play.Reset
    In this case, FMS is publishing the local stream but it is not playing the remoteStrem on it.
    The rest of the times, localStream dispatches events:
       NetStream.Publish.Strart
       NetStream.Play.Reset
       NetStream.Play.Start
       NetStream.Data.Start
    In this case FMS plays the remoteStream correctly.
    Any hint to solve this issue?
    Regards.

    Thanks, I tried your code and it works when playing a remote live stream on another FMS.
    But the remote live stream is not on a FMS but on a Wowza Server which re-streams an RTSP stream over RTMP.
    With this code:
    var nc;
    var myStream;
    application.onAppStart = function(){
         nc = new NetConnection();
         myStream = Stream.get("localstream");
         myStream.onStatus= function(info){
               trace(info.code);
         nc.onStatus = function(info){
               trace (info.code);
               if(info.code == "NetConnection.Connect.Success" ){
                    myStream.play("remoteLive.sdp", -1, -1, true, nc);
         nc.connect("rtmp://remoteServer/live");
    Every live stream player I tried was able to play the remote stream "remoteLive.sdp", but FMS play it only sometimes. This is the output log of the FMS application:
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset        <--------------------------------- Stuck there, stream won't play
    Unloaded application instance wowza4/_definst_ <--- Reload app
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset        <--------------------------------- Stuck there, stream won't play
    Unloaded application instance wowza4/_definst_ <--- Reload app
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset
    NetStream.Play.Start         <--------------------------------- Stream is playing just fine
    NetStream.Data.Start
    Unloaded application instance wowza4/_definst_ <--- Reload app
    NetConnection.Connect.Success
    NetStream.Publish.Start
    NetStream.Play.Reset       <--------------------------------- Stuck there, stream won't play
    Any idea? Why FMS won't play it everytime?
    Regards

  • Anyone know how to play Windows Live Messenger or MSN in iPhone4 3G? thanks..

    anyone knoes how to play Windows Live Messenger in iPhone4 3G

    what do you mean?
    how to install and use msn live messenger on iphone?
    or when you say play do you mean some game that use live msg?
    if the former then you just get the live messenger official client from app store and type in your login
    if the later then doubt you can

  • Disk Management for RTMP live streams

    Hello,
    Can anyone tell me if there is anything similar to DiskManagement functionality (available for HTTP streams) available to RTMP live streams in AMS?
    I'm having trouble with INDEX files that are being created for the live+DVR streams (we are serving 24/7 streaming), even though I delete RAW DVR files which are old, INDEX files constantly grow and FMS crashes CPU after few days.
    Any suggestions are appreciated.

    Hi,
    The methods that are mentioned as missing as expected by FMLE as it makes calls to them. They are defined in the sample live application that comes with the installation. The code for these can be picked up from the samples folder under the FMS installation from the application live. You can add these in all of your server side scripts to make them more relevant.
    Thank you !

  • How do I use multiple cameras to live stream through FME?

    I am looking to live stream pool tournaments from multiple angles but don't know what software or hardware I might need. Anybody have any good how to advice or links they might share? I stream through Ustream.tv if that makes a difference. Should I look for something different? Thanks

    I am working on getting just the counter working by using
    the program posted previously, and I am running into issues. Periodically I get
    the error:
    Error -200141
    occurred at DAQmx Read (Counter DBL 1Chan 1Samp).vi
    Possible reason(s):
    Data was overwritten
    before it could be read by the system.
    If Data Transfer
    Mechanism is Interrupts, try using DMA. Otherwise, divide the input signal
    before taking the measurement.
    It seems to work better if I use cascaded counters, but I need timer 0 for
    analog channels when I run this code along with the program for the other
    measurements.
    I have tried averaging, and selecting different values for the millisecond
    timer, and these did not seem to have an effect.
    I tried different DAQms configurations and "Counter DBL 1Samp" seemed
    to work the best.
    The program will work for a while and then it will give me the above error
    message.
    If I use counter 0 as a cascaded counter input, the program runs fine. If I run
    this with other analog channels, it errors out because the analog channels use
    counter 0.
    If I use counter 1 as a cascaded counter input, it seems to work better than a
    single channel, but it will still error out with the above error.
    If I use only counter 1, I get the error above even faster.
    Also, none of the
    configurations give measurements outside the While Loop.
    The only place I can add a speed dial for the front panel is within the While
    Loop.
    Is there someway to get the signal to continuously send out of the while loop?
    I thought if I could get the signal out of the while loop, I could condition it
    anyway I wanted without the program erroring out.
    Any suggestions would be much appreciated.
    Thank you.
    Attachments:
    Counter_error.jpg ‏45 KB

  • IOS RTMP live stream audio

    Hi guys hopefully some of you will be able to help me.
    I am trying to stream audio from a live RTMP feed using the NetConnection and NetStream classes. I've managed to get my app running no problem on Android, however I am having some major difficulties getting it to play the audio back on iPad. Interestingly it works in the device emulators when debugging, however I'm assuming this is not really an accurate representation. I've tried streaming the RTMP in both AAC and MP3, but with no luck from either. I can verify through debug that it has connected to the stream, however I just get no audio playing.
    Everything I've read about seems to suggest that this is possible on IOS as I'm only interested in audio and not video. Can anyone help?
    Code sample below (it's quick and dirty! ).
    THanks in advance!
    <?xml version="1.0" encoding="utf-8"?>
    <s:View xmlns:fx="http://ns.adobe.com/mxml/2009"
                        xmlns:s="library://ns.adobe.com/flex/spark" title="Audio" creationComplete="init()">
              <s:layout>
                        <s:VerticalLayout paddingLeft="10" paddingRight="10"
                                                                  paddingTop="10" paddingBottom="10"/>
              </s:layout>
              <fx:Script>
                        <![CDATA[
                                  import flash.media.Video;
                                  import flash.net.NetConnection;
                                  import flash.net.NetStream;
                                  import mx.core.UIComponent;
                                  private var vid:Video;
                                  private var videoHolder:UIComponent;
                                  private var nc:NetConnection;
                                  private var defaultURL:String="[STREAM]";
                                  private var streamName:String="[STREAMNAME]";
                                  private var ns:NetStream;
                                  private var msg:Boolean;
                                  private var intervalMonitorBufferLengthEverySecond:uint;
                                  private function init():void
                                            vid=new Video();
                                            vid.width=864;
                                            vid.height=576;
                                            vid.smoothing = true;                           
                                            //Attach the video to the stage             
                                            videoHolder = new UIComponent();
                                            videoHolder.addChild(vid);
                                            addEventListener(SecurityErrorEvent.SECURITY_ERROR, onSecurityError);
                                            grpVideo.addElement(videoHolder);
                                            connect();
                                  public function onSecurityError(e:SecurityError):void
                                            trace("Security error: ");
                                  public function connect():void
                                            nc = new NetConnection();
                                            nc.client = this;
                                            nc.addEventListener(NetStatusEvent.NET_STATUS, netStatusHandler);
                                            nc.objectEncoding = flash.net.ObjectEncoding.AMF0;
                                            nc.connect(defaultURL);      
                                  public function netStatusHandler(e:NetStatusEvent):void
                                            switch (e.info.code) {
                                                      case "NetConnection.Connect.Success":
                                                                trace("audio - Connected successfully");
                                                                createNS();                
                                                                break;
                                                      case "NetConnection.Connect.Closed":
                                                                trace("audio - Connection closed");                
                                                                //connect();
                                                                break; 
                                                      case "NetConnection.Connect.Failed":
                                                                trace("audio - Connection failed");                
                                                                break;
                                                      case "NetConnection.Connect.Rejected":
                                                                trace("audio - Connection rejected");                                  
                                                                break; 
                                                      case "NetConnection.Connect.AppShutdown":
                                                                trace("audio - App shutdown");                                 
                                                                break;         
                                                      case "NetConnection.Connect.InvalidApp":
                                                                trace("audio - Connection invalid app");                                   
                                                                break; 
                                                      default:
                                                                trace("audio - " + e.info.code + "-" + e.info.description);
                                                                break;                                                                                                    
                                  public function createNS():void
                                            trace("Creating NetStream");
                                            ns=new NetStream(nc);
                                            //nc.call("FCSubscribe", null, "live_production"); // Only use this if your CDN requires it
                                            ns.addEventListener(NetStatusEvent.NET_STATUS, netStreamStatusHandler);
                                            vid.attachNetStream(ns);
                                            //Handle onMetaData and onCuePoint event callbacks: solution at http://tinyurl.com/mkadas
                                            //See another solution at http://www.adobe.com/devnet/flash/quickstart/metadata_cue_points/
                                            var infoClient:Object = new Object();
                                            infoClient.onMetaData = function oMD():void {};
                                            infoClient.onCuePoint = function oCP():void {};        
                                            ns.client = infoClient;
                                            ns.bufferTime = 0;   
                                            ns.play(streamName);  
                                            ns.addEventListener(AsyncErrorEvent.ASYNC_ERROR, asyncErrorHandler);
                                            function asyncErrorHandler(event:AsyncErrorEvent):void {
                                                      trace(event.text);
                                            intervalMonitorBufferLengthEverySecond = setInterval(monPlayback, 1000);
                                  public function netStreamStatusHandler(e:NetStatusEvent):void
                                            switch (e.info.code) {
                                                      case "NetStream.Buffer.Empty":
                                                                trace("audio - Buffer empty: ");
                                                                break;
                                                      case "NetStream.Buffer.Full":
                                                                trace("audio - Buffer full:");
                                                                break;
                                                      case "NetStream.Play.Start":
                                                                trace("audio - Play start:");
                                                                break;
                                                      default:
                                                                trace("audio - " + e.info.code + "-" + e.info.description);
                                                                break;
                                  public function monPlayback():void {
                                            // Print current buffer length
                                            trace("audio - Buffer length: " + ns.bufferLength);
                                            trace("audio - FPS: " + ns.currentFPS);
                                            trace("audio - Live delay: " + ns.liveDelay);
                                  public function onBWDone():void {
                                            //Do nothing
                                  public function onFCSubscribe(info:Object):void {      
                                            // Do nothing. Prevents error if connecting to CDN.    
                                  public function onFCUnsubscribe(info:Object):void {    
                                            // Do nothing. Prevents error if connecting to CDN.    
                        ]]>
              </fx:Script>
              <s:Group id="grpVideo">
              </s:Group>
    </s:View>

    Just an update on this for anyone coming along after me.
    I've managed to get this working on MP3 but not AAC (I guess AAC just isn't supported?).
    My problem was the buffertime. The docs seemed to indicate it shoudl be set to 0 for live streaming, however switching it to "1" solved my problem on the particular stream I was pointing at.
    So essentially justchanged the above line:
                                            ns.bufferTime = 0;  
    to
                                            ns.bufferTime = 1;  
    Would be great to find out if anyone has gotten AAC working however...

  • Simple HTML example on how to display a live stream?

    Hi all
    Is there anyone who has a HTML example (or can give an URL with a page) that has the Adobe Flash player displaying a live stream (and not a file)?
    There is one in the installation under "Flash Media Server Start Screen" but it was too complex for me to understand. I just need a simple example on how to display a live flash stream in Adobes own web player.
    Thank you very much :-)

    Hi,
    Here is the html code.
    <html>
    <head>
    <title>Flash Media Player</title>
    </head>
    <object width="550" height="400">
    <param name="movie" value="player.swf">
    <embed src="somefilename.swf" width="550" height="400">
    </embed>
    </object>
    </body>
    </html>
    And here is the action script code of player.swf.
    import flash.net.NetConnection;
    import flash.events.NetStatusEvent;
    import flash.net.NetStream;
    import flash.media.Video;
    import flash.events.MouseEvent;
    import flash.media.Camera;
    startBtn.addEventListener(MouseEvent.CLICK,mouseHandler);
    var nc:NetConnection;
    var ns:NetStream;
    var vid:Video;
    var cam:Camera=Camera.getCamera();
    function mouseHandler(evt:MouseEvent){
        nc=new NetConnection();
        nc.addEventListener(NetStatusEvent.NET_STATUS,statusHandler);
        nc.connect("rtmp://localhost/sample");
    function statusHandler(evt:NetStatusEvent){
        trace(evt.info.code);
        if(evt.info.code=="NetConnection.Connect.Success"){
            ns=new NetStream(nc);
            ns.addEventListener(NetStatusEvent.NET_STATUS,streamHandler);
            ns.play("LiveStream");
            vid=new Video();
            vid.attachNetStream(ns);
            vid.height=400;
            vid.width=600;
            addChild(vid);
    function streamHandler(evt:NetStatusEvent){
        trace(evt.info.code);
    Please do the following before using the actionscript code:
    Add a button called start in fla and keep its instance name startBtn.
    create an application called sample in FMS and publish a live stream called "LiveStream".

  • Using netStream.time on rtmp live stream (broadcaster!=receiver)

    I am trying to synchronize a live stream (which is
    broadcasted from the flash player plugin) with some scripted
    actions. The problem is, that the displayed frames of the rtmp
    stream do not correlate to the netStream.time property on the
    receiver side. In fact, i do not understand at all, how the
    property is rendered on the receiver, since i can not recognize any
    dependencies to e.g. bufferLength.
    Does anybody now, how it is calculated? I tried with Red5 and
    Wowza and they seem to behave similar. I guess, FMS would not make
    any difference (if so: lease let me know!), since i assume that the
    property is rendered during the encoding process i.e. by the
    plugin.

    Hello Jay,
    thank you for your answer! NetStream.send() seems to be at
    least a possibility to solve my problem with a workaround.
    I just want to synchronize the time information of the up-
    and downstream: both should have the same time information when
    they show the same visual content (which i called "frame" in the
    former post).
    What i would suppose the netstream.time to be is that if i
    shake my camera on upstream.time=10.0 then downstream.time should
    equal 10.0 when this camera shaking is played back. This is the
    behaviour that i am used to from streaming prerecorded FLVs. But
    with live streams things work out diffferently.
    In fact my downstream.time is bigger than upstream.time. And
    i can not imagine how this can be. If i streamed up for let's say
    20 seconds and i start the downstream after about 10 seconds, i
    would expect my downstream.time to start with 10 seconds and then
    increase continually. But against this, my downstream.time starts
    with something above 20. How can this be? Or back to my initial
    question: how is the downstream.time rendered?
    This behaviour seems not to be dependent on the
    downstream.bufferTime.
    With netStream.send() i could send the upstream.time
    information via the upstream to render an offset for
    downstream.time on receiver side. This should work (have to check
    it), but is a workaround, no "clean" solution.

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder which does not have firewire out/input? it comes only with a component video output, USB, HDMI and composite RCA output?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder ..that doesn't have firewire out/input?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • Problem of republishing remote rtmp live stream with AMF

    Hi, Guys
    I am trying recieving one live stream from a remote rtmp server and republishing this stream in AMF.
    a remote rtmp (rtmp://192.168.1.100/live/remotestream)
    AMF instance(codes in application/app/main.asc) recieve and republish to another AMF instance.
    I can watch with url(rtmp://192.168.1.100/live/remotestream ) in a player (ex. OSMF player).
    But,no data from remote rtmp streaming can be recieved in AMF.
    Here is the log:
    NetConnection.Connect.Success
    Sending error message: Method not found (onBWDone).
    mystream.onStatus: NetStream.Publish.Start
    mystream.onStatus: NetStream.Play.Reset
    My codes in application/app/main.asc
    application.onAppStart = function()
              mystream = Stream.get("myvideo");
              myRemoteConn = new NetConnection();
              myRemoteConn.connect("rtmp://192.168.1.100/live");
              myRemoteConn.onStatus =function(info){
               trace(info.code);
               if(info.code == "NetConnection.Connect.Success"){
                    mystream.play("remotestream", -1, -1, true, myRemoteConn);
                    //mystream.play("sample", 0, -1, true);<-----if play a local flv video file ,it works.republish fine.
              nc = new NetConnection();
              nc.connect("rtmp://localhost/demo");
              ns = new NetStream(nc); 
              ns.setBufferTime(2);
        mystream.onStatus = function(sinfo)
            trace("mystream.onStatus: "+sinfo.code);
            if(sinfo.code == "NetStream.Publish.Start")
                                  attach_retVal = ns.attach(mystream);
                if(attach_retVal==true){
                        ns.publish(mystream.name,"live");

    On the Macintosh side there is the Camtwist app which allows you to do this and it works smoothly with FMLE. On the PC side I´m not sure but there just got to be more apps capable of this than on the mac side.

  • How to use Youtube Live Streaming API ??

    I want to make a WPF application that gets the video from my ip camera and sends it to my youtube channel live. I look around all of the websites but there's no example how can i live stream a video to Youtube with c#. There are examples in google's website
    but they were written with PHP, Java and Phyton but I don't know this programming languages so i couldn't use the API.
    I tried to write a little bit but it didn't work. Here's my code that i wrote looking through the Java example.
    string devkey = "AIzaSyCbxm6g9orAw9PF3MkzTb_0PGbpD3Xo1Qg";
    string username = "MyYoutubeChannelEmailAdress";
    string password = "MyPassword";
    YouTubeRequestSettings youtubereqsetting = new YouTubeRequestSettings("API Project", devkey, username, password);
    YouTubeRequest youtubereq = new YouTubeRequest(youtubereqsetting);
    LiveBroadcastSnippet broadcastSnippet = new LiveBroadcastSnippet();
    broadcastSnippet.Title = "Test Live Stream";
    broadcastSnippet.ScheduledStartTime = new DateTime(2015, 3, 12, 19, 00, 00);
    broadcastSnippet.ScheduledEndTime = new DateTime(2015, 3, 12, 20, 00, 00);
    LiveBroadcastStatus status = new LiveBroadcastStatus();
    status.PrivacyStatus = "Private";
    LiveBroadcast broadcast = new LiveBroadcast();
    broadcast.Kind = "youtube#liveBroadcast";
    broadcast.Snippet = broadcastSnippet;
    broadcast.Status = status;
    Google.Apis.YouTube.v3.LiveBroadcastsResource.InsertRequest liveBroadcastInsert = new Google.Apis.YouTube.v3.LiveBroadcastsResource.InsertRequest(service, broadcast, "");
    LiveBroadcast returnLiveBroadcast = liveBroadcastInsert.Execute();

    This article should get you going:
    https://developers.google.com/youtube/v3/code_samples/dotnet
    Also, questions on YouTube API issues should get the best service here:
    https://groups.google.com/forum/#!forum/google-api-dotnet-client

Maybe you are looking for

  • REQ: Add 'Fit-width' or 'Fit-visible' view mode, in which the zoom level is automatically set based on the width of the _current_ page.

    Currently, the default zoom level when viewing a page of a PDF file is automatically calculated based on the width of the "widest" page of the document. This means that, if the document contains one page that is wider than the others or is in landsca

  • On the contrary....

    Ive had the 1gb shuffle since a little before they hit the shelves and NEVER had a problem. Im perfectly happy with it!! Even thou the nano makes me wanna give up women completly I dont feel a need (urge) to upgrade.......yet. Thank you Apple and mr.

  • Iweb '08 and FTP client

    Bear with me I am new to some of this...I have iweb '08, can I use a FTP client to publish? I read in one place you can only do this with iweb '09, is that the case? Thanks

  • Multihreading seems to run slower than multiple JVM

    Hello, I have a 4-core machine and I use windows XP. I build a java application that can launch a thread doing a series of heaving calculation. The thread takes 8sec to finish and the task manager indicate that 30% of CPU is in use. If I start 2 thre

  • How to retrieve st03n data from swnc_collector_get_aggregates

    Hi experts,    I am trying to retrieve Dialogue response time from -ST03N  using 'swnc_collector_get_aggregates'. This FM returns too many tables. which is the correct table that has all the values of dialogue response time. Please help. Answers will