Live stream to start with keyframe

Hi,
In a .3gp file, it always start streaming at the first frame (zero seconds in to the stream), so the player always received an I frame first and displays the video correctly from then on.
In a live stream on the Quicktime server it start streaming at current frame (eg. 13 seconds after the encoder has started), so the server sends 5 P frames to the client first then an I frame. The client would display a distorted image for 5 seconds until an I frame was received.
Is there any setting on the QuickTime server to always send I frames (key frames) first for a live stream?
QuickTime Streaming Server Software Version: 5.5.4   Mac OS X (10.0.x)  

Big thanks for your reply
Really helped me on my long understanding way to upgrade my live chat application to
work with IOS devices.
So lets continue... make this important post to be a good helper for all developers on their way to stream to IOS devices.
If you can I clarify this - Nitin Goel says:
"Remember, that you need to publish video as H264 codec and audio as AAC/HEAAC for it to work with ios player.. Flash player may not yet support this combination.."
It is not possible for me to use Flash Media Live Encoder.
In this article:
http://sonnati.wordpress.com/2011/05/14/a-dream-comes-true-h-264-encoding-into-flash-playe r-11/
it says:
"A dream comes true: H.264 encoding into Flash Player 11"
and its done by this actionscript 3 code:
var h264Settings:H264VideoStreamSettings = new H264VideoStreamSettings();
h264Settings.setProfileLevel(H264Profile.BASELINE, H264Level.LEVEL_2);
stream.videoStreamSettings = h264Settings;
which will be connected to the NetStream like this i guess:
var h264Settings:H264VideoStreamSettings = new H264VideoStreamSettings();
h264Settings.setProfileLevel(H264Profile.BASELINE, H264Level.LEVEL_2);
nc = new NetConnection();
nc.connect("rtmp://example.com/livepkgr");
ns = new NetStream(nc);
ns.videoStreamSettings = h264Settings;
ns.publish("livestream?adbe-live-event=liveevent");
Is this what need to be done in order to publish video as H264 from webcam and Flash Player 11 in user browser?
And what about publish audio as AAC/HEAAC?
Another important thing:
I am writing my actionscript 3 code in my flash CS4 and publish it to swf.
With this code obviously it give me error...
do i need to upgrade to CS5 or CS5.5 and then i will be able to publish with this code?
again big thanks for any help
cheinan

Similar Messages

  • AS3 - 1) Countdown Timer to the designated Date for Live-Broadcast-Event yet to take place; 2) Detect End of Live Stream on FLVPlayback with FLVPlayback.isLive = true - ActionScript 3 - flash cs3 cs4

    Hi folks,
    Ronny's here again on forums, having particularly 2 (two) questions/problems to resolve:
    1) Countdown Timer to the designated Date for Live-Broadcast-Event yet to take place
    2) Detect End of Live Stream on FLVPlayback with FLVPlayback.isLive = true
    attached is the .zip file (as3_Countdown Timer_ver 1.0.1_by Ronny Depp.zip) with all flash source files containing:
    a) The FLash Source (file: timer_module.fla) - (FLA flash source file - Flash CS3 Professional, Flash Player 9, actionscript 3.0)
    b) com.othenticmedia.utils.dateAndTimeManagement package including 2 .as actionscript 3.0 Class files.
           i) com.othenticmedia.utils.dateAndTimeManagement.DateAndTimeManager Class in the said package. (file: DateAndTimeManager.as)
           ii) com.othenticmedia.utils.dateAndTimeManagement.CountdownTimer Class in the package. (file: CountdownTimer.as)
    c) The compiled SWF file version of this Application's blueprint. (file: timer_module.swf).
    What i need to confirm is: ........................................................ see the next post of mine. (for Problems  need to be Resolved)

    Problems to Resolve:
    Problem#1) - Countdown Timer to the designated Date for Live-Broadcast-Event yet to take place.
    Problem#2) - Detect End of Live Stream on FLVPlayback with FLVPlayback.isLive = true;
    Problem#1 Description:
    I need to pinpoint the Logical TimeSync Exception, i am still unable to figure out. That is I'm using a webservice in my Application to Synchronize the Time with the actual ET (eastern time) with accomodation of auto-adjustment for EDT GMT-4 (eastern daylight time) & EST GMT-5 (eastern standard time), times. I am using the zipcode: "10012" to pass it to the Web Service in urlRequest object, to retrieve the Current ET eastern time according to EDT & EST time settings for Manhattan/Brooklyn areas or others within  New York, NY 10012.
    Currently the Web Service is returning accurate date/time based on local EDT GMT-4 daylight time.
    Is there some defined set of dates for EDT & EST times for New York region that I can check for to ensure the correct Dates/Times for Eastern Time in New York area ??? I am using NY zipcodes because i am sure to get correct ET values.
    The Major Problem Part: is I need to correct the time by 2 seconds or approx. 2 secs, some millisecs.
    When I retrieve the Time Value from WebService, it lags behind for 2 seconds as compared to DateObj i create using computer's local time, on my Windows XP Service Pack 2 with Automatic Updates turned-on. And I'm sure about my Windows will be having latest updates for Time Management already installed. I also added the 2 secs. to the TimeSync(ed) Date to make correction to this Date obj.
    I call my custom fucntion addSeconds(dateObj:Date, secs:int) to add 2 seconds to the Date by Converting Seconds to Milliseconds.
    Please comb through the as code in files attached and Help Me Out !!!
    Problem#2 Description:
    Secondly I need to Detect the End of Stream state while using FLVPlayback component, an rtmp:// live Stream from FLASH MEDIA SERVER.
    I need to Play a YuMe Post-Roll Ad when Steam Finishes/Ends.
    Live Broadcast Stream Event starts every night on Wednesdays & Saturdays exactly  at 10:59 PM EDT GMT-4.
    Live Events only Streams/Broadcasts the stream for 50secs. exactly. When [playback stopped] it plays a PostRoll Ad and after the CountdownTimer again comes back to life. The Next upcoming Event is calculated & the Countdown begins until Next Event's time/date is reached.
    Here is the  code on the frame 1 on the MainTimeline: (rest of the params like source, volume, skinAutoHide are Set using Property Inspector for FLVPlayback instance on Stage)
    //myStream instance of FLVPlayback is on the Stage
    myStream.isLive = true;// Frame 1 Actions in the FLA
    myStream.addEventListener(VideoEvent.COMPLETE, onEndOfStream);
    myStream.addEventListener(VideoEvent.STATE_CHANGE, onState);
    myStream.addEventListener(String(VideoError.NO_CONNECTION), onStreamError);
    myStream.addEventListener(NetStatusEvent.NET_STATUS, netStatusHandler);
    /*if(myStream.stopped){
         trace("tracy: "+myStream.state);
    } else if(myStream.state == VideoState.STOPPED){
         trace("tracy: "+myStream.state);
    function onStreamError(event:VideoError) {
         trace(event.code + "\n\t" + event);
    function onState(event:VideoEvent) {
         trace(event.state + "\n\t" + event.toString());
    function onEndOfStream(event:VideoEvent) {
         trace(event.state + "\n\t" + event.toString());
    function netStatusHandler(event:NetStatusEvent):void {
         switch (event.info.code) {
              case "NetConnection.Connect.Success":
                   //connectStream();
                   break;
              case "NetStream.Play.StreamNotFound":
                   trace("Stream not found: "/* + myStream.source*/);
                   break;

  • Live Stream Green Screen with Adobe Premiere

    So at my school I am head tech for the morning news which is a live broadcast that is sent throughout the entire school. Right now we have a blank green background and my Teacher asked me to setup a way for us to broadcast the news with a green screen live. After about a month of attempts none have proven worthy for the situation is not the best:
    - The output signal that is used to broadcast to the school is old analog: it runs on a coaxial plug. However, we have a modulator that takes in s-video and RCA and converts that to coaxial
    - We have adobe Premiere on new computers but they don't have a suitable coaxial output, I could possibly install a graphics card we have but since they are new and are school PCs, I cant really do that
    - The camera is nothing fancy, put does have a fire wire and RCA output, but the computers do not have none of those inputs, however we do have a converter from fire wire to USB
    - The new computers have no RCA or s-video outputs(like I said, they are new and the broadcasting setup is old) but we still have one old laptop that we can clone the screen to the broadcast for it has an s-video output.(Possibly we could stream from the new PC to the old one and clone it to the TV?)
    - The new computers do have extra VGA outputs, so could you possibly convert that to s-video? (Cus they are both analog?)
    Like I said, this is a very sticky situation. My closest idea is that we could some how stream the green screen video from the new PC to the old one that would then broadcast it, but I do not know how to do that and the camera will not register with Premiere even when plugged in via converter... I believe it is because we do not have a capture card, the only capture card I could even use though would have to be external for I cannot open the new PCs under the school policies. If anyone has any suggestions or ideas on how I could solve this enigma, that would mean a million.

    Green screen (99.9% of the time) is caused by outdated or corrupted video drivers: http://forums.adobe.com/thread/945765

  • Re-Start a Live Stream using AMS 5 through Cloudfront

    We have AMS 5 running on an Amazon EC2 instance. We send a Live Stream from an FMLE encoder using H.264/AAC.
    If we do an Origin Pull from that instance running the AMS 5 and Live steam through one of our commercial CDN accounts, we can stop and re-start our Live stream without issue. It stops and re-starts as it should. When we re-connect to the AMS using FMLE, using the Origin Pull, the stream re-starts in either the OSMF or JW Player 6 in under 30 seconds.
    If we use that same AMS 5 on the same EC2 instance and Live stream through an Amazon Cloudfront distribution, if we stop our Live stream, when we re-start the stream it takes an unacceptably long time for the stream to refresh - if ever. Basically, after the re-start of the Live stream through FMLE, the Live stream at the player plays the first 10 or 15 seconds of the intial Live stream and then stops. If you refresh the player, it will play that same 10 or 15 seconds again - and stop again.
    It would seem that the AMS 5 is setup correctly. It works through our commercial CDN.
    It also would seem that the issue is with the cache on Amazon Cloudfront. Our assumption is that the Cloudfront cache is not properly updating. But we have tried about every variable in every setting for Cloudfront that we can find. We can't get past this.
    Any suggestions?
    Thanks
    As a followup to this discussion, you can follow the issue through the AWS.Amazon forums at https://forums.aws.amazon.com/thread.jspa?threadID=120448&tstart=0

    We also have other threads going.
    http://forums.adobe.com/thread/1180721?tstart=0
    https://forums.aws.amazon.com/thread.jspa?threadID=120448&tstart=0
    There appears to be 2 different issues - one issue with Amazon Cloudfront and a second issue with Adobe Media Server. Unfortunately, they are tied together in the real world application of these programs.
    In AMS. when we use the adbe-record-mode=record, then the AMS works fine because the AMS is clearing the streams on each re-publish. Unfortunately, this causes a status code error on Amazon Cloudfront. And that particular status code error from a custom origin server causes Cloudfront to "override the minimum TTL settings for applicable cache settings". As the programmers explained, basically what happens is that Cloudfront overrides our 30 second cache settings and goes into a standard cache setting of 5 minutes. And waiting 5 minutes to re-start a Live stream just doesn't work in the real world for real customers.
    We can fix that Cloudfront status code error issue by using adbe-record-mode=append on the AMS. If we use that AMS setting, then we can stop a Live stream, re-start the stream, the Cloudfront cache clears and everything works - unless you also change the stream settings in your FMLE. If you stop the stream and change the video input or output size in the FMLE, then the AMS server gets stuck and can't/won't append to the f4f record file.
    Does anyone have any ideas about how we can adjust all of this so that it works together?
    Thanks

  • User Auth. for Live Streaming with Media Encoder 2.5 and FMSS

    Hi,
    i am currently thinking of buying the Flash Media STREAMING
    Server (~995 $) and in order to see if this is the right choice, i
    also take a look at the Wowza Media Server for comparison.
    I want to support Live Streaming from Events with Flash Media
    Encoder 2.5 and for this i'll need a user authentification for
    giving access to special publishing Users only. Wowza Media Server
    supports a User Authentification by parsing the connecting URL,
    witch is delivered by the Media Encoder on starting a Connection,
    and therefor grants or denies access to the publishing function.
    Only Users with the right Access Persmissions, which is written
    trough a password and the onPublish Comand in the url, can publish
    Live Streams. Everyone can watch the Live Stream, if no restriction
    to the onConnect Command was made.
    The Problem in this case is, i will need seperate Application
    folders for every user who should be able to start a live stream.
    This would be a horrible effort...
    more info ==>
    http://www.wowzamedia.com/forums/showthread.php?t=1281
    Is there a similar way to support authorisation for starting
    publishing a LiveStream to the Flash Media Streaming Server, or do
    i need the Flash Media Interactive Server?
    How can i restrict access to users, who should not be able to
    start a live stream on my FMSS?
    Thx for your help,
    Tobi

    quote:
    Is there a similar way to support authorisation for starting
    publishing a LiveStream to the Flash Media Streaming Server, or do
    i need the Flash Media Interactive Server?
    Unfortunately no, there isn't. I struggled with same issue
    earlier, and we were forced to purcahse the Interactive edition. As
    the 'Streaming' version only runs Adobe-signed apps, there is no
    way to change things that happen within the onConnect handler.
    Furthermore, guys at Adobe haven't provided any possibility for
    even simple access control through configuration files for built-in
    apps that come along with the Streaming Server edition.
    - Jakki

  • Can storing a live stream using actionscript fail by the cache filling up with overhead?

    Hi,
    Lately we have been seeing a problem with the archives of live streams we create using FMS. We use FMS for live streaming and concurrently store the stream in a file using ActionScript. We use the following code to record:
    var s2 = Stream.get('mp4:' + mp4name);
    application.publishedStreamMP4= s2;
    application.publishedStreamMP4.record();
    application.publishedStreamMP4.play(application.publishedStream.name,-1,-1);
    (some lines have been removed that are used for logging, etc).
    Sometimes some of these functions fail and return false. In these cases FMS's core log shows that the cache is full:
    2013-06-11 11:45:55        13863   (w)2611372      The FLV segment cache is full.  -
    In investigating this issue I have not yet been able to recreate this exact situation. By lowering the cache to 1MB I have however been able to create a situation where storing a stream can stop because the cache is full. The situation occurs as follows:
    * The server is restarted, the cache is empty.
    * A live stream is started, which is also recorded.
    * Via the Administration API the cache values <bytes> and <bytes_inuse> show to be exactly the same as the <overhead> of the object that relates to the file being saved. The <bytes> and <bytes_inuse> values of the object are 0.
    * This continues in the same way untill the cache is full.
    * When the limit of the cache is reached the message
    2013-06-11 12:07:35        13863   (w)2611372      The FLV segment cache is full.  -
    is shown in the core log and storing of the file stops. The instance log also show status changes:
    2013-06-11 12:07:35        13863   (s)2641173      MP4 recording status for livestream.mp4: Code: NetStream.Record.NoAccess Level: status Details:         -
    2013-06-11 12:07:35        13863   (s)2641173      MP4 recording status for livestream.mp4: Code: NetStream.Record.Stop Level: status Details:     -
    In the filesystem I can confirm that the last change of the file is on this moment (in this case 12:07). The live stream continues without problems.
    I have reproduced this several times. Though I can understand that caches can fill up and this can cause trouble I feel like this situation is a bug in FMS. The cache fills up with overhead, which is apparently reserved untill writing the file ends.
    I hope someone here can help out. Has anyone seen a situation like this and is there any remedy for it? Or even a workaround where this overhead in the cache can be released so the cache does not fill up?
    We use FMS version 4.5.1.

    You can use an XML socket, but the FMS application will need to initiate the connection, as FMS has no support for listening for anything other than RTMP and HTTP requests.
    Stream switching can happen on the FMS side. In your FMS application, you'll create a server side stream, and use the Stream.play method for playing other sources (live streams or recorded flv/h.264 files) over that stream. Your subscribers will connect to the server side stream
    See the FMS docs for the Stream class and the XMLSocket class.

  • Pls playlist of mp3 live stream doesn't work properly, what to do?

    Can others confirm the following issue with a live mp3 stream on version 7.0.2.16 on windows xp (sp2):
    Go to wnyc.org and click on the AM or FM live stream (32k) links at the top left of the page. They should link to either of the following urls:
    http://www.wnyc.org/stream/fm.pls
    http://www.wnyc.org/stream/am.pls
    Open these playlists with the current iTunes for windows. You should hear a 15 second sponsor pre-roll for this NPR station, and the live stream should start after that. I only hear the pre-roll, and then iTunes skips to the next item in the library.
    Do you only hear the sponsor pre-roll, and then nothing OR does it work properly for you? If not, what to do?

    Reboot your computer in Safe Mode and then try to reset all Skype settings.
    Go to Windows Start and in the Search/Run box type %appdata% and then press Enter or click the OK button. The Windows File Explorer will pop up. There locate a folder named “Skype”. Rename this folder to something different, e.g. Skype_old.
    Next go to Windows Start and in the Search/Run box type %temp%\skype and then press Enter or click the OK button. Delete the DbTemp folder.
    Reboot in normal mode and try to start Skype.
    N.B. If needed, you will still be able to re-establish your call and chat history. All data is still saved in the Skype_old folder.

  • Trying to find lowest latency setting for FMS 3.0 live streaming app

    Hi,
    I work for a technology company in the agricultural industry,
    and we recently bought Adobe FMS 3.0 for live streaming of an
    online auction application. Initially the tests were ok, but the
    stream latency is becoming a real issue. It starts at around 1s
    delay, which is barely acceptable, but an hour into the sale it
    stretches to 2-3s, and more. This is becoming a real deal breaker
    for the Flash solution that looked very promising at the beginning.
    We need to find the absolute lowest latency setting for the
    live streaming application supplied with FMS3.0.
    Low latency (1s max) and stable latency has highest priority,
    more important than picture quality or uninterrupted streaming.
    How can we achieve this? Are there any settings where I can
    tweak the stream latency?
    I am not a Flash expert, but I can find and modify the
    configuration files on the FMS Windows 2003 server.
    Any help would be immensely appreciated.

    On server side:
    FMS installs itself with default configurations optimized for
    Max CPU efficiency. Reducing/shutting-off FMS publisher queues
    using Application.xml config should help in removing the delay that
    you experienced.
    <Application>
    <StreamManager>
    <Live>
    <Queue enabled="true">
    </Queue>
    </Live>
    </StreamManager>
    </Application>
    Changing default setting to <Queue enabled="false">
    disables live publisher queues & aggregate messages ad
    therefore "the delay". (the delay should be reduced to <100 ms
    for a perfect kind of network or LAN)
    --> On client side: Do NOT use NetStream.setBufferTime API
    at all ! i.e. do not set any buffer on client side.; Flash player
    (upto 10, the current version) would try its best to keep the
    buffer around 0.1 seconds, which is the default value.
    OmicronPersei8,
    Which publisher are you using to publish the live stream?
    Sometimes , it is possible that a some fault at stream publisher
    may make buffer at flash player grow.
    I hope this would help !

  • Live stream on FMS issues when reaching 1600 concurrent users

    Hey all,
    We have an FMS 4.5.2 installation with 1x Origin and 2x Edge setup only for Live streaming. We are using FMLE latest version to stream to the Origin server and Edge servers are connecting to the Origin server. We are running four applications stream1, stream2, stream3 and stream4 all for live streaming (copied live) application. All servers have default configuration except edge/origin setup.
    All servers have 4Gbit connections in bonding and network traffic is nicely distributed across all bonds and uplink is 10Gbit to the ISP.
    Server is a 2 CPU / Quad Core HP DL 380 with 64Gbit of memory running Ubuntu 10.04 LTS
    Now to the problem.
    We are streaming live stream from FMLE with 700kbit/s to for example stream1 application and when we hit around 1600 concurrent users with 50% on edge1 and 50% on edge2 (800 concurrent users on each edge) in the player buffer drops and all people experience buffering issues at approx 30secs - 2minutes intervals. (network is not congested because if at that time do an SCP from the server outside network it copies 1Gbyte file within seconds.
    Funny thing is that If I start a second FMLE and stream live stream to for example stream2 application at same time and open up second player on client the video runs great without any buffer issues from the same server at the same time.
    edge1 server:
    application: stream1 with 800 concurrent users, player has buffer issues
    application: stream2 with 4 concurrent users, player has no issues
    edge2 server
    same as above
    So my current conclusion is that it has to be something per application since other application does not have this issues when running simultaneously? We tried changing fmscore number settings and some buffer settings but nothing helped.
    at the time when we get buffering issues this are the only problematic things that get in the log and its in edge.00.log on both EDGE servers but not at same times:
    Edge1:
    2012-04-23    12:50:21    29270    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    12:55:30    29270    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    12:56:42    29270    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    12:56:42    29270    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    13:14:40    29270    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    13:20:30    29270    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    Edge2:
    2012-04-23    12:56:32    9625    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    13:02:23    9625    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    2012-04-23    13:08:03    9625    (e)2661034    Connect failed ( , 8134 ) : Connect failed: Connection refused (111)    -
    there is no packet loss between edge and origin servers and latency is at 0.2ms and nothing in the logs of the origin server
    We even tried to deploy Wowza Medis Servers with edge / origin setup and were able to handle around 4000 (2000 on one and 2000 on second edge) concurrent users without any issues.
    Anyone has any ideas or at least what are our next options to do and what settings to change on the FMS? or how to debug and what to check when buffering issue appears? any more debugging we can enable on the FMSs? too se if we hit some kind of limit somewhere?
    thanks

    hey, i got the similar problem, logging like this
    2012-11-12
    18:50:12
    23434
    (e)2661034
    Connect failed ( , 5779212 ) : Connect failed: Connection refused (111)
    2012-11-12
    18:50:54
    23434
    (e)2661034
    Connect failed ( , 5779212 ) : Connect failed: Connection refused (111)
    2012-11-12
    18:51:36
    23434
    (e)2661034
    Connect failed ( , 1166880400 ) : Connect failed: Connection refused (111)
    2012-11-12
    18:54:14
    23434
    (e)2661034
    Connect failed ( , 1175301776 ) : Connect failed: Connection refused (111)
    2012-11-12
    18:54:55
    23434
    (e)2661034
    Connect failed ( , 1164775056 ) : Connect failed: Connection refused (111)
    2012-11-12
    18:55:37
    23434
    (e)2661034
    Connect failed ( , 16 ) : Connect failed: Connection refused (111)
    2012-11-12
    19:13:08
    23434
    (e)2661034
    Connect failed ( , 1158459024 ) : Connect failed: Connection refused (111)
    it seems that the port number is invalid, but we never use such ports.

  • Live stream to android devices? shouldn't be this difficult.

    Why doesn't Adobe post a simple tutorial on how to stream live video from FMS 4.5 to Android device? The developer center has a tutorial series that hasn't been updated in years. The Part 8: Streaming to Android Devices (to Come) hasn't been updated in 1.5 years. The server is useless if we can't watch video on Android devices. Surely someone out there has successfully streamed live video to an Android device and can explain how to do it?
    Do we need to create an Android Air App to accomplish this?
    Thanks for any help!
    Dave

    As an avid Wowza developer/integrator/consultant (as well as the occasional AMS work), I concur with JayCharles on all of his assertions. The transcoder is completely optional, and you can actually make your own Wowza transcoder module with FFmpeg if the Transcoder AddOn costs are really too steep for deployment.
    Having said that, my current solution for live streaming to Android with an off-the-shelf (OTS) player is JW Player, and setting its fallback property to false. It looks something like this:
    <div id="videoPlayer">
         <!-- Code for non-supported JW Player case. JW Player 6 will not show HLS source on Android. -->
         <video src="http://media_server/app_name/_definst_/mp4:stream_ref/playlist.m3u8" controls ></video>
         <p>Video not playing? <a href="rtsp://media_server/app_name/_definst_/mp4:stream_ref">Click this link to play in an external media player.</a></p>
    </div>
    <script type="text/javascript">
    jwplayer("videoPlayer").setup({
                                                                playlist: [
                                                                                    sources: [
                                                                                                        file: "rtmp://media_server/appname/_definst_/mp4:streamname"
                                                                                                        file: "http://media_server/appname/_definst_/mp4:streamname/playlist.m3u8"
                                                                                    title: "My Sample Live Stream"
                                                                height: 180,
                                                                primary: "html5",
                                                                autostart: false,
                                                                width: 320,
                                                                fallback: false
    </script>
    HTH.

  • My IPad cannot download live Tv from skygo. I have no problem with my lap top so I assume the broad band is okay. Can anybody suggest why the live streaming will not work on my iPad one.

    My IPad cannot download live Tv from skygo. I have no problem with my lap top so I assume the broad band is okay. Can anybody suggest why the live streaming will not work on my iPad .

    Are you using the Sky Go app to try and watch it ? If so are you logged in with your Sky account ?
    If you are using the app then you could try closing the app completely and see if it works when you re-open it : from the home screen (i.e. not with Sky Go 'open' on-screen) double-click the home button to bring up the taskbar, then press and hold any of the apps on the taskbar for a couple of seconds or so until they start shaking, then press the '-' in the top left of the Sky Go app to close it, and touch any part of the screen above the taskbar so as to stop the shaking and close the taskbar.
    If that doesn't work then you could try a reset : press and hold both the sleep and home buttons for about 10 to 15 seconds (ignore the red slider), after which the Apple logo should appear - you won't lose any content, it's the iPad equivalent of a reboot.

  • Problem of republishing remote rtmp live stream with AMF

    Hi, Guys
    I am trying recieving one live stream from a remote rtmp server and republishing this stream in AMF.
    a remote rtmp (rtmp://192.168.1.100/live/remotestream)
    AMF instance(codes in application/app/main.asc) recieve and republish to another AMF instance.
    I can watch with url(rtmp://192.168.1.100/live/remotestream ) in a player (ex. OSMF player).
    But,no data from remote rtmp streaming can be recieved in AMF.
    Here is the log:
    NetConnection.Connect.Success
    Sending error message: Method not found (onBWDone).
    mystream.onStatus: NetStream.Publish.Start
    mystream.onStatus: NetStream.Play.Reset
    My codes in application/app/main.asc
    application.onAppStart = function()
              mystream = Stream.get("myvideo");
              myRemoteConn = new NetConnection();
              myRemoteConn.connect("rtmp://192.168.1.100/live");
              myRemoteConn.onStatus =function(info){
               trace(info.code);
               if(info.code == "NetConnection.Connect.Success"){
                    mystream.play("remotestream", -1, -1, true, myRemoteConn);
                    //mystream.play("sample", 0, -1, true);<-----if play a local flv video file ,it works.republish fine.
              nc = new NetConnection();
              nc.connect("rtmp://localhost/demo");
              ns = new NetStream(nc); 
              ns.setBufferTime(2);
        mystream.onStatus = function(sinfo)
            trace("mystream.onStatus: "+sinfo.code);
            if(sinfo.code == "NetStream.Publish.Start")
                                  attach_retVal = ns.attach(mystream);
                if(attach_retVal==true){
                        ns.publish(mystream.name,"live");

    On the Macintosh side there is the Camtwist app which allows you to do this and it works smoothly with FMLE. On the PC side I´m not sure but there just got to be more apps capable of this than on the mac side.

  • Start/stop/restart live stream - do stream files need to be removed first?

    When streaming a live event if something happens and we need to stop our encoder and then restart it again.
    Do we need to delete the stream files created in the 'application/event/' directory and also delete the f4f stream,bootstrap, etc... files too?
    If we don't delete the stream files before restarting the live stream. It will only play about 10 seconds of the stream then just hang buffering.
    What is causing this?
    My application log shows no access when I restarted the stream.
    2013-09-04
    09:02:21
    3960
    (s)2641173
    onPublish : cfhsstream4
    2013-09-04
    09:02:21
    3960
    (s)2641173
    Stream name is: cfhsstream4 and live event is: cfhs
    2013-09-04
    09:02:21
    3960
    (i)2611173
    Event level recording configuration: General[flags: 3, keyframe interval: 60000, max size: -1, max length: -1, io buffer size: 0], Fragment[frag duration: 4000, seg duration: 0, duration precision: 0, allowed drift: 1000, frame rate: 0.000000, frames per interval: 0, intervals per frag: 1, frame precision: 1, disk management duration: 0.000000], ContentProtection[enabled=false(overridable)]
    2013-09-04
    09:02:21
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Publish.Start
    2013-09-04
    09:02:21
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Play.Reset
    2013-09-04
    09:02:21
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Data.Start
    2013-09-04
    09:02:22
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Record.Start
    2013-09-04
    09:02:23
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Record.NoAccess
    2013-09-04
    09:02:23
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Record.Stop
    2013-09-04
    09:02:23
    3960
    (s)2641173
    f4f:cfhsstream4 - NetStream.Record.NoAccess
    Thanks,
    Dave

    I have a similar issue like Mr Dave explained. If need to stop, start, stop and restart again, the encoder is sending, but in player not shows nothing. Then we need to stop and disconnect in encoder, and remove the stream file in /event and the cache files in livestream folder for after then restart the streaming. Only after this action, the streaming works fine and the streaming shows again in player.
    Have other way to solve this issue without need all that actions? Note that I use the Adobe FME to live streaming and encoding in HDS http streaming
    I hope for a help
    Thanks
    Pedro
    Message was edited by: asaweb2013

  • Live stream HDS and HLS with one webpage.

    Hello,
    First off, I am not a web developer, but I can put together a page if I have to. I have been tasked with setting up our Adobe Media Server 5 for live streaming and have successfully gotten the HDS and HLS streams running. I have created the two pages on our webserver to link to the HDS and HLS streams and I wanted to be able to integrate them into one page so the users don't have to choose the appropriate stream. Can anyone point me to a site or maybe an adobe help on how to do this simply and effectively? Does adobe have some prewritten code for this somewhere?

    Hi,
    One way you can do this is to use php script to identify the user agent of the client (http://php.about.com/od/learnphp/p/http_user_agent.htm) and then generate dynamic webpages based on the client (HLS for iOS based clients and HDS for flash based clients). As Apache can only serve static webpages, you'll need to use CQ/MySQL/something else. I've personally not attempted to do so, but this may be a good starting point : http://www.ibm.com/developerworks/linux/tutorials/l-php/
    Hope this helps.
    Thanks,
    Apurva

  • Live stream problem with FMS 4.5 + FMLE

    Hi,
    I've got a live stream (and DVR) setup using FMS 4.5 and FMLE.
    The hardware for the FMS server we use is 2x Xeon E5645 2.4GHz, 32GB RAM, OCZ RevoDrive R3 X2 PCI-e SSD. Installed OS is CentOS 5.8.
    We have 6 workstation PCs, each equiped with 4 BlackMagic Intensity PRO capture cards and FMLE 3.2. Using these PCs we stream Live video to FMS server.
    Video stream is grabbed from HDMI cable, to BlackMagic Cards, which are encoding by FMLE using H.264 format, 700kbps bitrate per stream.
    Video stream is published to server through RTMP stream, which is later saved for DVR in RAW format. We only keep 3 days of recording and delete the old ones.
    Server and Encoder PCs are in the same network, connected by gigabit managed switch.
    The problem I'm having is that after 10-15 hours FMLE starts to drop frames, because video buffer is increased. What I observed is that this happens immediately when Server CPU load increases above 60%.
    Based on the above observation I decreased the number of channels streamed by the server to 10, which reduced CPU load. But the problem still persists.
    Whenever I restart FMS and delete all DVR data, the CPU load (when streaming 10 live channels) is only 1%, but after 2 days CPU load increases to 50-60%.
    Whenever I restart FMS and don't delete DVR data, the CPU load is 5-10%, and after 2 days it still increases to 60-70%.
    Another thing I observed is that there is only single fmscore process running, but it has lots of threads which are switched on and off in split seconds. These threads are launched on different CPU cores, but at any given point in time the distribution of the load isn't equal among CPU cores. This leads certain CPU cores being loaded by more than 60% and frame drops start to occur.
    For the moment there are just 10 users watching this service, so I don't think this load accounts for the problem.
    Has anybody had similar problem, or know how can I optimize or finetune the system to run without problem? I would appreciate any suggestions.
    Another thing I noted through last couple of days:
    When I was restarting FMS the CPU load was reduced to 30%, but after 1 week past when I restart it CPU load only goes down to 75%. Everything is the same, nothing has been changed and there is no disk IO issues involved.
    P.S. I've modified application.xml using these values:
    <Scope>vhost</Scope>
    <Distribute numprocs="5">app</Distribute>
    <LifeTime>
    <RollOver></RollOver>
    <MaxCores></MaxCores>
    </LifeTime>

    Hi
    How many channels are you publishing?
    If there are too many channels, it is recommended to have one  FMSCore process start for each of them. To do so, you will have to change the scope to app.
    ---snippet---
    <Application>
    <Process>
        <Scope>app</Scope>
       <Distribute numprocs="3">inst</Distribute>
    </Process>
    </Application>
    Also, to delete older content, you will have to enable disk management. Refer http://help.adobe.com/en_US/flashmediaserver/devguide/WSeb6b7485f9649bf23d103e5512e08f3a33 8-8000.html#WSec225f632fa00875-23954b6f1300b641158-8000 for more info.

Maybe you are looking for