Live Streaming w/ RTMPE Only?

I'm hoping there's a simple answer to this. I'm attempting to stream a live event using FMLE and FMS. The first thing that throws me off is that I cannot push the stream to the server using RTMPE, only RTMP[T]. I assume this is because the the encryption is handled by the server and not the encoder. However, for the sake of security I would like to disable RTMP connections from the flash player to the server. Yet it would seem that if I incorporate the server side scripting described here http://kb2.adobe.com/cps/405/kb405456.html (namely the changes to my main.asc file), that my server would reject an RTMP connection from FMLE? Is that correct? If so, where do I go from here? Can I broadcast a live stream using RTMPE only? Do I need a different encoder or something, or am I just missing a very simple point?

You're correct that FMLE can't publish RTMPE.
What you can do is, in your logic for inspecting the client protocol for the connection, make an exception for FMLE user-agents. Based on the example code in the doc you linked to, you could do something like:
application.onConnect = function(clientObj) {
   if(clientObj.protocol == "rtmpe" || clientObj.protocol == "rtmpte" || client.agent.indexOf("FMLE") > -1 || client.agent.indexOf("FME") > -1){
     return true;
  return false

Similar Messages

  • How to package multibitrate live stream as audio-only for Apple compliance

    Hi
    From AMS documentation, I can package vod content as audio-only by appending "audio-only" to url for instance:
    http://<ams server ip>/hls-vod/audio-only-aac/vod.m3u8
    Similarly I can package live content (single bitrate) as audio-only for instance:
    http://<ams server ip>/hls-live/audio-only-aac/livepkgr/_definst_/liveevent/livestream1.m3u8
    In my case, I have 3 bitrates being published for a live audio-video event. The variant playlist for this event is created by hls_http module by refering to Manifest.xml.
    I could not figure out a way to insert an audio-only stream to this variant playlist created by hls_http module for live stream.
    Please let me know if I am missing something because this seems to be a pretty obvious requirement for live streaming to IOS devices.
    Thanks,
    Nitin

    who can answer my question?Is the question a difficult one?

  • Live streaming, no video only audio

    Hello
    I have 3 pc: in the first I'm encoding a camera with Adobe
    Flash Encoder 2.5 and send the signal to the second PC
    in the second PC have installed Flash Media Server
    in the third have a player.
    Every thing is connected and in the player i can hear only
    the audio, but not the video. Tryed with different players.
    Any help
    Sincerly
    Saimir

    Sounds to me like a h.264 issue. If you're using the h.264
    codec in FME, you need to make sure you're using the latest version
    of FMS3. H264 support was fixed in 3.0.2. You also need Flashplayer
    9.0.115 or better.
    Try switching the video codec in FME to VP6. If that makes
    the problem go away, then you know it's an issue with either FMS or
    your flashplayer being outdated.

  • Publish Live stream to FMS on RTMPE protocol

    Hi,
    I am trying to publish Live Stream to FMS using FMLE or third party encoders.
    Is it possible to have RTMPE or RTMPTE protocol used between encoder and FMS?
    Very urgent.
    thanks,
    Pooja Jain

    FMLE can only publish via rtmp or rtmpt.

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder which does not have firewire out/input? it comes only with a component video output, USB, HDMI and composite RCA output?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • Live Stream only spools, never connects.

    Up until several weeks ago, I was able to watch Live Stream. Now, it only 'spools' and never connects. The chat message box says 'disconnected.' Firefox is my browser.

    I know this is a problem with JWPlayer which is very widely used. When videos get converted to Mpeg4 (h.264) flash, the time information is often at the end of the video. This causes the player to try to buffer the entire video before it starts playing.
    Interestingly, Quicktime Pro 7 and Final Cut Pro put the time information where it belongs, but most of the other tools to convert do not. I spent weeks trying to solve this problem for videos on my site www.dfdtv.com . The solution was right in front of me all the time, QT Pro.
    You can find out for sure by just starting the video in question and just waiting for a long time. If it is the time atom problem, it will eventually start to play.
    All of our videos' are now h.264 and if created with QT Pro they stream great using JWplayer/Firefox (and underneath it all Adobe Flash).

  • Creating an audio-only stream from a live stream

    I am using Flash Media Live Encoder (3.2) to send multiple bitrate streams to an instance of Adobe Media Server 5. In addition to the video streams, I am trying to generate an audio-only stream.
    I am using the default 'livepkgr' application. The input stream contains H.264 video + AAC audio.
    The stream name that I set in FMLE is set to this:
    livestream%i?adbe-live-event=liveevent&adbe-audio-stream-name=livestream1_audio_only&adbe- audio-stream-src=livestream1
    I created a variant playlist to include the audio-only stream. It looks like this:
    #EXTM3U
    #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=150000
    /hls-live/livepkgr/_definst_/liveevent/livestream1.m3u8
    #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=500000
    /hls-live/livepkgr/_definst_/liveevent/livestream2.m3u8
    #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=700000
    /hls-live/livepkgr/_definst_/liveevent/livestream3.m3u8
    #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=64000,CODECS="mp4a.40.2"
    /hls-live/livepkgr/_definst_/liveevent/livestream1_audio_only.m3u8
    However, when I retrieve the audio-only playlist it contains a list of .ts files. If it was working correctly I would expect there to be a list of .aac files. Here's an extract from the playlist (livestream1_audio_only.m3u8):
    #EXTM3U
    #EXT-X-MEDIA-SEQUENCE:1927
    #EXT-X-ALLOW-CACHE:NO
    #EXT-X-VERSION:2
    #EXT-X-TARGETDURATION:8
    #EXTINF:8,
    ../../../../hls-live/streams/livepkgr/events/_definst_/liveevent/livestream1_audio_onlyNum1927.ts
    #EXTINF:8,
    ../../../../hls-live/streams/livepkgr/events/_definst_/liveevent/livestream1_audio_onlyNum1928.ts
    #EXTINF:8,
    ../../../../hls-live/streams/livepkgr/events/_definst_/liveevent/livestream1_audio_onlyNum1929.ts
    #EXTINF:8,
    ../../../../hls-live/streams/livepkgr/events/_definst_/liveevent/livestream1_audio_onlyNum1930.ts
    Any ideas why it doesn't seem to be working?
    Thanks

    Hi
    No !
    You first have to back-engineer the DVD to something the Mac can use eg StreamingDV
    (I use Roxio Toast™ to do this) then convert this via QuickTime to .aiff 16-bit 44.1 kHz
    then burn this onto a CD via eg iTunes.
    An Analog copy from a DVD-player to the Mac Line-in and into an audiuo editor as Audacity (free on internet) then use this for the CD is easier and about as good as. (I can't differ)
    Yours Bengt W

  • HTTP Live Streaming only works in Safari

    I have used the facillity in the share menu to export for HTTP Live streaming. I have uploaded the reulting directory to my wordpress site and can get the video to work in Safari. When I check itwith other browsers though it does not appear. Is there a reason for this?

    Because your not live streaming.  But seriously, it's a specialized codec you shouldn't use unless a client requires it.  Use plain old H.264 for the web.

  • Live streaming event is not streaming, only buffering

    http://www.adobe.com/uk/events/create-now-tour/?promoid=KNNXM
    is the link to the live streaming event from Adobe World Tour. I am in North Yorkshire. The streaming is not working. It is buffering all the time.
    I missed the whole bit about lightroom, and I am appauled that Adobe does not stream with a working stream...
    Who can do something about this? Of course it is now 20.10 in the evening, the event started at 19 PM... I have been able to watch maybe 5 minutes...
    Nobody there at Adobe UK and the chatagent was useless as well...
    AAAGGGRRRRRR

    hi mate,
    WP does not support Adobe Flash and even Android has stopped supporting it so theres nothing you can do abnout this excpet perhaps stream it on a PC with flash player installed.

  • HDS live streaming to Flash not working

    Adobe Flash Media Server 4.5.5 r4013
    Windows 2008
    Sources:
    http://help.adobe.com/en_US/flashmediaserver/devguide/WSd391de4d9c7bd609-52e437a812a3725df a0-8000.html
    http://www.adobe.com/devnet/adobe-media-server/articles/live-multi-bitrate-video-http-flas h-ios.html
    Live streaming a single or multi-bitrate video over HTTP to Flash does not work. I have followed the instructions on the 2 sources listed above repeatedly, but I can’t get live streaming over HTTP to Flash to work. Live streaming to iOS over HTTP works with no problems (single and multi-bitrate streams).
    I have tried the troubleshooting steps from the following:
    http://help.adobe.com/en_US/flashmediaserver/devguide/WS0432746db30523c21e63e3d12efac195bd -8000.html
    Troubleshoot live streaming (HTTP)
    1.      Services window (Windows): Flash Media Server (FMS), Flash Media Administration Server, and FMSHttpd services are running. ✓
    2.      Verified that the request URL is correct. ✓
    3.      Configured ports:
    a.      Configure Apache to use port 80. Open rootinstall/Apache2.2/conf/httpd.conf in a text editor. Change the line Listen 8134 to Listen 80.
    b.     Configure Flash Media Server not to use port 80. Open rootinstall/conf/fms.ini in a text editor. Remove 80 from the ADAPTOR.HOSTPORT parameter so the parameter looks like the following: ADAPTOR.HOSTPORT = :1935 ✓
    4.      Placed a crossdomain.xml file to the rootinstall/webroot directory. ✓
    5.      In Flash Media Live Encoder, select the Encoding Options tab, choose Output from the Panel options menu, and verify the following:
    a) The value of FMS URL is rtmp://fms-dns-or-ip/livepkgr. If you’re testing on the same server as Flash Media Server, you can use the value localhost for fms-dns-or-ip. ✓
    b) For a single stream, the value of Stream is livestream?adbe-live-event=liveevent. ✓
    c) For adaptive bitrate streaming, the value of Stream is livestream%i?adbe-live-event=liveevent. ✓
    Flash Media Live Encoder uses this value to create unique stream names. To use another encoder, provide your own unique stream names, for example, livestream1?adbe-live-event=liveevent, livestream2?adbe-live-event=liveevent.
    The encoder is showing all 3 streams being published and streaming.
    6. Check Administration Console: the livepkgr application and the 3 streams are running. ✓
    7. Check the logs for errors. Flash Media Server logs are located in the rootinstall/logs folder. The master.xx.log file and the core.xx.log file show startup failures. Apache logs are located in the rootinstall/Apache2.2/logs folder. X
    a)   core00.log: these errors did not occur every time that I tried playing the live stream but these are the only relevant errors in the logs.
    1. 7968 (w)2611179     Warning from libf4f.dll: [Utils] [livestream2] Discarded all queued Media Messages received before first Video Keyframe Message
    2. 7968 (w)2611179     Warning from libf4f.dll: [Utils] [livestream3] Discarded all queued Media Messages received before first Video Keyframe Message
    b) edge00.log:
    13:33:57 4492          (w)2641213 Connection rejected by server. Reason : [ Server.Reject ] : (_defaultRoot_, _defaultVHost_) : Application (hds-live) is not defined.          -
    c) Apache-Error:
    1.     [warn]  Checking if stream is disabled but bootstrap path in event file is empty for event:livepkgr/events/_definst_/liveevent stream name:livestream
    2.     [warn] bootstrap path is in event file is empty for event:livepkgr/events/_definst_/liveevent stream name:livestream1
    As I mentioned, everything works on iOS and FMS seems to be creating all of the stream segments and meta files:
    a.     The 3 streams are being created in: HD:\Program Files\Adobe\Flash Media Server 4.5\applications\livepkgr\streams\_definst_
    b.    FMS is creating the following files in each stream folder (livestream1, livestream2, livestream 3):
    1. livestream1.bootstrap
    2. livestream1.control
    3. livestream1.meta
    4. .f4f segments
    5. .f4x segments
    The appropriate files are also being created in the HD:\Program Files\Adobe\Flash Media Server 4.5\applications\livepkgr\events\_definst_\liveevent folder, in which I have the following Manifest.xml and Event.xml files:
    <manifest xmlns="http://ns.adobe.com/f4m/1.0">
      <media streamId="livestream1" bitrate="200" />
      <media streamId="livestream2" bitrate="500" />
      <media streamId="livestream3" bitrate="1000" />
    </manifest>
    <Event>
      <EventID>liveevent</EventID>
      <Recording>
    <FragmentDuration>4000</FragmentDuration>
    <SegmentDuration>16000</SegmentDuration>
        <DiskManagementDuration>3</DiskManagementDuration>
      </Recording>
    </Event>
    I’ve tried clearing the contents of both streams\_definst_ and events\_definst_\liveevent (keeping the xml files) after restarting the encoder, and creating a different event definst for the streams (liveevent2 for example).
    We have an event in 2 weeks that we would like to stream to both Flash and iOS. Any help in solving this problem will be greatly appreciated.

    One step closer:
    Changed the crossdomain.xml file (more permissive settings).
    Changed the encoding on FMLE to vp6. Working somewhat (don't know what I did to make it start streaming through hds).
    But at least now I can get the individual streams in the set manifest file to work:
    http://localhost/hds-live/livepkgr/_definst_/livevent/livestream1.f4m
    http://localhost/hds-live/livepkgr/_definst_/livevent/livestream2.f4m
    http://localhost/hds-live/livepkgr/_definst_/livevent/livestream3.f4m
    BUT when I try to play the streams through the set manifest file from http://localhost/liveevent.f4m I'm getting the following error:
    "The F4m document contains errors URL missing from Media tag." I'll search the forums to see if anyone else has come across this problem.
    I used the f4m config tool to make the file. These are the file's contents:
    <manifest xmlns="http://ns.adobe.com/f4m/2.0">
      <baseURL>http://localhost/hds-live/livepkgr/_definst_/liveevent/</baseURL>
      <media href="livestream1.f4m " bitrate="200"/>
      <media href="livestream2.f4m " bitrate="500"/>
      <media href="livestream3.f4m " bitrate="1000"/>
    </manifest>
    Thanks

  • How do I use multiple cameras to live stream through FME?

    I am looking to live stream pool tournaments from multiple angles but don't know what software or hardware I might need. Anybody have any good how to advice or links they might share? I stream through Ustream.tv if that makes a difference. Should I look for something different? Thanks

    I am working on getting just the counter working by using
    the program posted previously, and I am running into issues. Periodically I get
    the error:
    Error -200141
    occurred at DAQmx Read (Counter DBL 1Chan 1Samp).vi
    Possible reason(s):
    Data was overwritten
    before it could be read by the system.
    If Data Transfer
    Mechanism is Interrupts, try using DMA. Otherwise, divide the input signal
    before taking the measurement.
    It seems to work better if I use cascaded counters, but I need timer 0 for
    analog channels when I run this code along with the program for the other
    measurements.
    I have tried averaging, and selecting different values for the millisecond
    timer, and these did not seem to have an effect.
    I tried different DAQms configurations and "Counter DBL 1Samp" seemed
    to work the best.
    The program will work for a while and then it will give me the above error
    message.
    If I use counter 0 as a cascaded counter input, the program runs fine. If I run
    this with other analog channels, it errors out because the analog channels use
    counter 0.
    If I use counter 1 as a cascaded counter input, it seems to work better than a
    single channel, but it will still error out with the above error.
    If I use only counter 1, I get the error above even faster.
    Also, none of the
    configurations give measurements outside the While Loop.
    The only place I can add a speed dial for the front panel is within the While
    Loop.
    Is there someway to get the signal to continuously send out of the while loop?
    I thought if I could get the signal out of the while loop, I could condition it
    anyway I wanted without the program erroring out.
    Any suggestions would be much appreciated.
    Thank you.
    Attachments:
    Counter_error.jpg ‏45 KB

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder ..that doesn't have firewire out/input?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • About Flash P2P live streaming from non-webcam sources

    Hello, I am a university student. Our lab is attempting to work on a p2p live streaming using flash p2p features. The media source is not from a webcam but a file from a certain server, which is not directly supported by any of the 4 methods flash player offers(posting, direct routing , object replication,and multicast.) As some of the forum threads mentioned, our method is to use NetStream.publish() a stream and use send("callbackname", data)  to all subscribers that have joined in a NetGroup. So we can use p2p transmission. Now here are our questions:
    1. We know from MAX 2009 that flash p2p camera video multicasting implements pull-push mechanism inside which makes full use of p2p features like buffer map exchange. If we use NetStream.send() API to send non-webcam source data, will it also make use of this pull-push mechanism to spread data among all peers(subscribers) with proper data exchange?
    or more detailedly,
    I saw the P2P Gaming Libs at flashrealtime.com. It uses DIRECT_CONNECTIONS when creating publish-used NetStream in order to send data with lowest latency. My question is if I do not use DIRECT_CONNECTIONS, than when the NetGroup grow large (1000), will those peers that are not direct neighbors to the publisher in the same group relay data to each other using pull-push via buffer map?
    2. I have written a sample app and use send() to deliver data only (NetStream.bufferTime = 0) without camera video and audio. When the publisher sends the data  in a "for()" stantence for 20 times, the subscribers can only receive about 8 of them. But when I set a timer and periodically send them(e.g., 500ms), the subscribers can receive them correctly. My questions are: Is the packet loss caused by UDP unreliability? Can this be solved by setting the NetStream.dataReliable = ture? Can this totally be solved by setting a timer? How to set the delaytime property in timer? Are all data sent are orderedly received?
    I'm very appreciated if you can answer my questions? thanks.

    1. NetStream.send() data is delivered in the order it was sent, whether in client-server, DIRECT_CONNECTIONS, or multicast mode.
    2. the lowest latency will be achieved with DIRECT_CONNECTIONS.  however, that isn't scalable to large numbers of recipients.  the P2P multicast system trades low latency for scalability.
    3. there are no plans to support a natural NetStream streaming source that's not a camera/microphone.  Flash Media Server can perform that function.
    4. you should never need to implement "bitmap exchange"; the object replication system handles all the communication needed to accomplish its function. typically you will divide a file into segments, giving each one an index number. object replication uses ranges internally, so the index numbers should be contiguous.  nodes that have indices use NetGroup.addHaveObjects().  nodes that want indices they don't have use NetGroup.addWantObjects().  nodes that have objects will send them to nodes that want the objects.  when a node receives an object, it's automatically removed from its "want" set; once the node decides the object is acceptable, it adds it to its "have" set, at which point it can answer requests from other nodes who still "want" it.  eventually every node will have all the pieces and not want any.  for an example, please see Tom Krcha's articles on using object replication:
       http://www.flashrealtime.com/file-share-object-replication-flash-p2p/
       http://www.flashrealtime.com/video-on-demand-over-p2p-in-flash-player-101-with-object-repl ication/
    5. please see Tom Krcha's articles.  for VOD you will probably want to use the "LOWEST_FIRST" replication strategy, as that may allow you to start playing earlier.  for general file replication you'll get better sharing if you use RAREST_FIRST.

  • Pls playlist of mp3 live stream doesn't work properly, what to do?

    Can others confirm the following issue with a live mp3 stream on version 7.0.2.16 on windows xp (sp2):
    Go to wnyc.org and click on the AM or FM live stream (32k) links at the top left of the page. They should link to either of the following urls:
    http://www.wnyc.org/stream/fm.pls
    http://www.wnyc.org/stream/am.pls
    Open these playlists with the current iTunes for windows. You should hear a 15 second sponsor pre-roll for this NPR station, and the live stream should start after that. I only hear the pre-roll, and then iTunes skips to the next item in the library.
    Do you only hear the sponsor pre-roll, and then nothing OR does it work properly for you? If not, what to do?

    Reboot your computer in Safe Mode and then try to reset all Skype settings.
    Go to Windows Start and in the Search/Run box type %appdata% and then press Enter or click the OK button. The Windows File Explorer will pop up. There locate a folder named “Skype”. Rename this folder to something different, e.g. Skype_old.
    Next go to Windows Start and in the Search/Run box type %temp%\skype and then press Enter or click the OK button. Delete the DbTemp folder.
    Reboot in normal mode and try to start Skype.
    N.B. If needed, you will still be able to re-establish your call and chat history. All data is still saved in the Skype_old folder.

  • Trying to add metadata to a live stream completely fails

    Hi.  I've been following this document http://help.adobe.com/en_US/flashmediaserver/devguide/WS5b3ccc516d4fbf351e63e3d11a0773d56e -7ff6Dev.html for adding metadata to an FMS live stream from a Flash widget.
    I've created a stripped-down version of the code (listed below); the same as the original code except the controls are removed.  It works fine only if I comment out the ns.send calls.  If I leave in the ns.send calls, no clients are ever able to view anything.  I'm using a stock FMS 4.5 install on Amazon EC2 -- the AMI is ami-904f08c2.  And I'm compiling the swf using Flex SDK 4.6 on Linux with the command "mxmlc -compiler.library-path+=./playerglobal11_0.swc -swf-version=13 -static-link-runtime-shared-libraries Broadcaster.as".
    package {
        import flash.display.MovieClip;
        import flash.net.NetConnection;
        import flash.events.NetStatusEvent; 
        import flash.events.MouseEvent;
        import flash.events.AsyncErrorEvent;
        import flash.net.NetStream;
        import flash.media.Video;
        import flash.media.Camera;
        import flash.media.Microphone;
        public class Broadcaster extends MovieClip {
            private var nc:NetConnection;
            private var ns:NetStream;
            private var nsPlayer:NetStream;
            private var vid:Video;
            private var vidPlayer:Video;
            private var cam:Camera;
            private var mic:Microphone;
            private var myMetadata:Object;
            public function Broadcaster(){
                setupUI();
                nc = new NetConnection();
                nc.addEventListener(NetStatusEvent.NET_STATUS, onNetStatus);
                nc.connect("rtmp://myserver/live");
             *  Clear the MetaData associated with the stream
            private function clearHandler(event:MouseEvent):void {
                if (ns){
                    trace("Clearing MetaData");
                    ns.send("@clearDataFrame", "onMetaData");
            private function startHandler(event:MouseEvent):void {
                displayPlaybackVideo();
            private function onNetStatus(event:NetStatusEvent):void {
                trace(event.target + ": " + event.info.code);
                switch (event.info.code)
                    case "NetConnection.Connect.Success":
                        publishCamera();
                        displayPublishingVideo();
                        break;
                    case "NetStream.Publish.Start":
                        sendMetadata();
                        break;
            private function asyncErrorHandler(event:AsyncErrorEvent):void {
                trace(event.text);
            private function sendMetadata():void {
                trace("sendMetaData() called")
                myMetadata = new Object();
                myMetadata.customProp = "Welcome to the Live feed of YOUR LIFE, already in progress.";
                ns.send("@setDataFrame", "onMetaData", myMetadata);
            private function publishCamera():void {
                cam = Camera.getCamera();
                mic = Microphone.getMicrophone();
                ns = new NetStream(nc);
                ns.client = this;
                ns.addEventListener(NetStatusEvent.NET_STATUS, onNetStatus);
                ns.addEventListener(AsyncErrorEvent.ASYNC_ERROR, asyncErrorHandler);
                ns.attachCamera(cam);
                ns.attachAudio(mic);
                ns.publish("livestream", "live");
            private function displayPublishingVideo():void {
                vid = new Video(cam.width, cam.height);
                vid.x = 10;
                vid.y = 10;
                vid.attachCamera(cam);
                addChild(vid); 
            private function displayPlaybackVideo():void {
                nsPlayer = new NetStream(nc);
                nsPlayer.client = this;
                nsPlayer.addEventListener(NetStatusEvent.NET_STATUS, onNetStatus);
                nsPlayer.addEventListener(AsyncErrorEvent.ASYNC_ERROR, asyncErrorHandler);
                nsPlayer.play("myCamera", 0);
                vidPlayer = new Video(cam.width, cam.height);
                vidPlayer.x = cam.width + 100;
                vidPlayer.y = 10;
                vidPlayer.attachNetStream(nsPlayer);
                addChild(vidPlayer);
            private function setupUI():void {
            public function onMetaData(info:Object):void {

    Also, emitting other events in the ns.send calls works; eg. if I do ns.send("blah", "onMetaData", myMetadata), nothing happens (because there's no "blah" function to do anything), but at least this doesn't cause the entire stream to fail.

Maybe you are looking for