Codec used in Live streaming

Dear all
I work with live streaming for a few weeks, using flash.media.Camera and flash.net.NetStream class. I have not found any place where the codec is specified, that is used to encode the live stream.
Anyone knows which codec is used and if there is a possiblitiy to use H264 codec for live streaming? I found that live streaming with h.264 codec seems to be possible with the flashmedia encoder (http://www.adobe.com/products/flashmediaserver/flashmediaencoder/). Is this possible with flash player alone?
Thanks for your help,
Marc

After doing some research I found this thread http://balazs.sebesteny.com/implement-h264-live-streaming-with-adobe-air-20/. It uses FMLE but other open source tools perhaps would be better because of licensing issues.
It would be really nice if encoding with h264 would be enabled with flash player. Now that decoding works with Hardware acceleration, perhaps encoding will come as well. Any plans exist?
Thanks,
Marc

Similar Messages

  • Using http live streaming

    Hello.  For a project I am working on, I need to design a secure streaming system for audio files such that they can be streamed but not downloaded (due to copy right issues).  I was looking into http live streaming as a possible solution.  However, I am new to developing web applications and the like and was hoping for some help, maybe links to sample code, how to use http live streaming for my server/in my code, etc.  In particular, I was hopping to make the streaming available to firefox, safari, and apple mobile devices.  I am hoping to develop this using html 5.
    Thanks.

    I have same issue with 64bit Safari 5.1 and 10.6.8.
    Video plays if not protected. If Protected says "Loading" and never plays.
    With 32bit Safari looks the same but...
    There is a login window behind the Safari window. If I find it and login, mp4 plays.
    Is quite annoying. Any fixes?

  • Using http live streaming to stream data from a Webcam

    JavaFX 2.2 includes support for http live streaming: http://en.wikipedia.org/wiki/HTTP_Live_Streaming
    How could one create a http live stream from a Webcam and thereby enable a JavaFX application to view and hear the Webcam video in realtime?

    Bom Dia Saille,
    Eu acredito que o que você tenha hoje seja isso aqui (Desenho bem simplificado):
    Hoje o que você tem é o Aplicativo do Medidor + USB Driver controlando o seu medidor. Basicamente, sua aplciação está dividida em três camadas:
    Aplicação - Onde estão as funcionalidades principais do programa (interface com o usuário, apresentação de dados, geração de arquivos, etc.)
    VISA - Uma arquitetura de Software para controle de instrumentos. Basicamente, ele se comunica com o Driver USB para poder enviar e receber pacotes de dados através do barramento USB.
    Device Driver - Intruções de SW de baixo nível para controlar um periférico através de um barramento.
    No desenho, eu aponto duas camadas onde você pode tentar atuar para automatizar suas medições:
    Aplicação - Você pode usar o VI Server para controlar o Aplicativo do Medidor (Se este aplicativo tiver sido desenvolvido em LabVIEW, o que eu suponho ser verdade). No entanto é necessário saber se o desenvolvedor do produto permitiu isso (Veja Using VI Server to Pass Data Between a VI and a LabVIEW Executable). Outra alternativa é conversar com o fabricante para ver se ele desenvolveu alguma API (Application Programming Interface) para que você consiga controlar o instrumento diretamente.
    VISA - O LabVIEW Fornece uma API chamada NI-VISA para poder enviar e receber informações através de diversos barramentos (Veja Serial Instrument Control Tutorial)
    É possível acessar o Driver diretamente também, mas eu não recomendo. O resultado não compensa o esfroço!
    Espero ter esclarecido suas dúvidas e espero que você tenha sucesso na sua aplicação!!
    Atenciosamente.
    Felipe Flores
    Engenharia de Aplicações
    National Instruments Brasil

  • How to use Youtube Live Streaming API ??

    I want to make a WPF application that gets the video from my ip camera and sends it to my youtube channel live. I look around all of the websites but there's no example how can i live stream a video to Youtube with c#. There are examples in google's website
    but they were written with PHP, Java and Phyton but I don't know this programming languages so i couldn't use the API.
    I tried to write a little bit but it didn't work. Here's my code that i wrote looking through the Java example.
    string devkey = "AIzaSyCbxm6g9orAw9PF3MkzTb_0PGbpD3Xo1Qg";
    string username = "MyYoutubeChannelEmailAdress";
    string password = "MyPassword";
    YouTubeRequestSettings youtubereqsetting = new YouTubeRequestSettings("API Project", devkey, username, password);
    YouTubeRequest youtubereq = new YouTubeRequest(youtubereqsetting);
    LiveBroadcastSnippet broadcastSnippet = new LiveBroadcastSnippet();
    broadcastSnippet.Title = "Test Live Stream";
    broadcastSnippet.ScheduledStartTime = new DateTime(2015, 3, 12, 19, 00, 00);
    broadcastSnippet.ScheduledEndTime = new DateTime(2015, 3, 12, 20, 00, 00);
    LiveBroadcastStatus status = new LiveBroadcastStatus();
    status.PrivacyStatus = "Private";
    LiveBroadcast broadcast = new LiveBroadcast();
    broadcast.Kind = "youtube#liveBroadcast";
    broadcast.Snippet = broadcastSnippet;
    broadcast.Status = status;
    Google.Apis.YouTube.v3.LiveBroadcastsResource.InsertRequest liveBroadcastInsert = new Google.Apis.YouTube.v3.LiveBroadcastsResource.InsertRequest(service, broadcast, "");
    LiveBroadcast returnLiveBroadcast = liveBroadcastInsert.Execute();

    This article should get you going:
    https://developers.google.com/youtube/v3/code_samples/dotnet
    Also, questions on YouTube API issues should get the best service here:
    https://groups.google.com/forum/#!forum/google-api-dotnet-client

  • Http Live Streaming

    Hi *,
    i'm trying to accomplish in my local network a live streaming using HTTP LIVE STREAMING. I'm already able to do streaming video with a file stored on the disc, and now I want to capture the video stream from the webcam and send it. Can I use Quicktime Broadcaster ... and if so, which settings? My problem is that the mediastreamsegmenter accepts only a particular encoding....

    As this is an OS feature it is a question that only Microsoft can answer. All I know about the iPlayer is on your other post.
    Ask Microsoft here:
    http://answers.microsoft.com/en-us/winphone

  • Need license for using H.264 Video codec for live stream with FMS 3.5 series

    Hi,
        I am creating a live audio-Video chat application  in which I use H264 video codec for live streaming.  I am using Flash Media Server version 3.5.7. I read some where that H.264 Video streaming using FMS required seperate licencse. So I want to confirm , It is true that H264 video streaming required seperate license? If yes then what is the procedure to get license.
    Please let me know its urgent.
    Regrads,

    Hi Again,
    Thanks for your reply.
    Here are some articles/posts which says about Royalties to use h264:
    http://www.mpegla.com/Lists/MPEG%20LA%20News%20List/Attachments/226/n-10-02-02.pdf 15:27
    http://www.mpegla.com/main/programs/AVC/Documents/AVC_TermsSummary.pdf
    http://www.streamingmedia.com/Articles/ReadArticle.aspx?ArticleID=65403&PageNum=3
    http://www.mpegla.com/main/programs/avc/Documents/AVC_TermsSummary.pdf
    According to these, any  product/service provider, need to pay one time fee based on # of subscribers if we are using h264 for encoding and is charging fee from its users.
    About our business: We are web-conferencing solution tool provider and are using Flash Media Server version 3.5.7. I for audio/video. Please help me out if h264 is really free for us or we need to buy license to use that.
    Looking forward to hear from you soon.
    Regards,

  • Live stream and codecs

    Hello.
    Here is described how to use different codecs for different
    clients (Sorenson Spark for Flash Player 7- and On2 VP6 for Flash
    Player 8+).
    http://livedocs.adobe.com/fms/2/docs/00000081.html
    But as I can see, it is only suitable for recorded .flv
    files.
    Is there any possibilities to customize stream delivery or
    live streams?
    Can FMS2 transcode streams on the fly?
    Thank you.

    FMS can't transcode or do any sort of recompression. It's
    bits in, bits out.
    If you publish through the Flashplayer, you will alwaus have
    sorenson spark encoded video (the flashplayer only has the decoder
    half of the VP6 codec). If you want to publish live VP6 video, look
    into the Adobe Flash Video Encoder, or On2 Flix Live.

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder which does not have firewire out/input? it comes only with a component video output, USB, HDMI and composite RCA output?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • How do I use multiple cameras to live stream through FME?

    I am looking to live stream pool tournaments from multiple angles but don't know what software or hardware I might need. Anybody have any good how to advice or links they might share? I stream through Ustream.tv if that makes a difference. Should I look for something different? Thanks

    I am working on getting just the counter working by using
    the program posted previously, and I am running into issues. Periodically I get
    the error:
    Error -200141
    occurred at DAQmx Read (Counter DBL 1Chan 1Samp).vi
    Possible reason(s):
    Data was overwritten
    before it could be read by the system.
    If Data Transfer
    Mechanism is Interrupts, try using DMA. Otherwise, divide the input signal
    before taking the measurement.
    It seems to work better if I use cascaded counters, but I need timer 0 for
    analog channels when I run this code along with the program for the other
    measurements.
    I have tried averaging, and selecting different values for the millisecond
    timer, and these did not seem to have an effect.
    I tried different DAQms configurations and "Counter DBL 1Samp" seemed
    to work the best.
    The program will work for a while and then it will give me the above error
    message.
    If I use counter 0 as a cascaded counter input, the program runs fine. If I run
    this with other analog channels, it errors out because the analog channels use
    counter 0.
    If I use counter 1 as a cascaded counter input, it seems to work better than a
    single channel, but it will still error out with the above error.
    If I use only counter 1, I get the error above even faster.
    Also, none of the
    configurations give measurements outside the While Loop.
    The only place I can add a speed dial for the front panel is within the While
    Loop.
    Is there someway to get the signal to continuously send out of the while loop?
    I thought if I could get the signal out of the while loop, I could condition it
    anyway I wanted without the program erroring out.
    Any suggestions would be much appreciated.
    Thank you.
    Attachments:
    Counter_error.jpg ‏45 KB

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder ..that doesn't have firewire out/input?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • How to use Phone camera for live streaming rather than using WireCast camera in Live streaming using Azure Media Services

    Hi,
    am planning to develop a live streaming app using C# and xaml.
    I have gone through the below sample and am able to stream the live content using wireacst camera.
    Azure Media Services Live Streaming Sample
    But i want to use Phone camera to stream the live video rather using wirecast camera.
    How can i use that? How to get the published URL from the C# code, is there any API to get that?
    Anybody please help me
    Regards,
    Santhosh

    Hello,
    There are no in box components that will allow you to transmit captured video from a Windows Phone device to Azure Media Services. You will need to create you own implementation. You can do this by creating a custom Media Foundation sync that can plug into
    the MediaCapture element's pipeline. You will then need to modify the captured and compressed video to align to one of the two protocols that Azure Media services can use. At this time Azure Media services can use either Smooth Streaming or RTMP for ingestion.
    You will need to create your own packager (or find compatible packager code from a 3rd party) to package the MP4 encoded data into the necessary format.
    Here is an example of how to create a custom Media Foundation sync (please note you must write your Media Foundation components using C++ MoCOM. We do not support Media Foundation components written in managed languages):
    Real-time communication sample
    I hope this helps,
    James
    Windows SDK Technologies - Microsoft Developer Services - http://blogs.msdn.com/mediasdkstuff/

  • X264vfw.dll - using x264 in live streaming...

    I heard that it is possible to use x264vfw.dll for encoding live video & that it contains an exportable function called "Configure" which can
    be used for configuring the encoder.
    Is it possible to use this dll in Flex applications? I have never tried hooking any dll files with Flex but heard that some wrappers allow u to
    do that. Is this option feasible? Are there any sites where such Wrappers are available for download?
    Please help me...

    Hello,
    There are no in box components that will allow you to transmit captured video from a Windows Phone device to Azure Media Services. You will need to create you own implementation. You can do this by creating a custom Media Foundation sync that can plug into
    the MediaCapture element's pipeline. You will then need to modify the captured and compressed video to align to one of the two protocols that Azure Media services can use. At this time Azure Media services can use either Smooth Streaming or RTMP for ingestion.
    You will need to create your own packager (or find compatible packager code from a 3rd party) to package the MP4 encoded data into the necessary format.
    Here is an example of how to create a custom Media Foundation sync (please note you must write your Media Foundation components using C++ MoCOM. We do not support Media Foundation components written in managed languages):
    Real-time communication sample
    I hope this helps,
    James
    Windows SDK Technologies - Microsoft Developer Services - http://blogs.msdn.com/mediasdkstuff/

  • Live Streaming Using my Webcam As a secruity camera

    I'm lost here someone please help, : \ im trying to use my isight as a security camera over the internet so i can watch my dog i just got her.
    So the question is what software out there i can use to put my webcam as a camera ( again my powerbook is kinda old im still running on panther so i have 10.3.9 not 10.4 can i work anyway around this) where i can watch live.
    Can someone help me i know i need my ip address i have that but when i hook everythign else up where do i go when im somewhere else when i want to look at my live streaming from work to my house? please can someone help me thank you if you can. : )

    Hello, keith28
    Until we hear back from juanorfrankie, here is my suggestion for your question:
     I would like to access the isight using my iPhone.
     Can anyone make some suggestions?
    If iPhone will can support viewing a streaming webcam page, use EvoCam with your iSight to publish a webcam with streaming video. That would probably be the simplest way to do what you want, but, because I do not have iPhone, so I cannot test this.
    See the instructions and FAQs at EvoCam for system requirements and details on how to set up your webcam page.
    EZ Jim
    PowerBook 1.67 GHz   Mac OS X (10.4.10)    G5 DP 1.8  External iSight

  • Live streaming not working using REOPS

    I made live streaming work using Live Encoder and a flash player I created with FlvPlayback component. It works fine. Now I like to try live streaming using REOPS. I created a config xml file using similar settings in src/assets/data/Playerconfig_LiveStreaming.xml:
    isLive = true
    <streamtype>live</streamtype>
    part of my xml file is like this:
    <?xml version="1.0" encoding="UTF-8"?>
    <player
        width="640" height="480"
        scaleMode="stretch" isLive="true" autoPlay="true" updateInterval="250"
        hasCaptions="false">
        <mediaElement>
            <mediaElement>
                <id>seeker</id>
                <mimeType>video/x-flv</mimeType>
                <streamType>live</streamType>
                <deliveryType>streaming</deliveryType>
                <media url="rtmp://<myMediaStreamingServerName>:1935/<myapplicationName>/live/livestream" />
            </mediaElement>
        </mediaElement>
    </player>
    in adobe live encoder, I have setting like this:
    URL: rtmp://<myMediaStreamingServerName>:1935/<myapplicationName>/live
    stream:  livestream
    but no matter how hard I tried, it never works for me. I used the same settings for my own player which works fine with live streaming.
    Anyone has any ideas on how to make live streaming work with REOPS?
    Thanks a lot.

    Is the XML a copy/paste? If so you may have a typo in your media url. There
    is a space in "liv e". That could be the issue.
    hth!

  • Problem relaying live stream using different urls

    Hi,
    I have setup QT Broadcaster to announce a stream to my qtss.
    I announced the stream to a directory live/stream.sdp
    This works without a clinch when using the same URL for viewing, e.g.
    rtsp://rtsp.myqttss.com/live/stream.sdp
    If I create a symlink (or simply copy the sdp file) from live/stream.sdp to test/stream.sdp and try to access the stream through
    rtsp://rtsp.myqttss.com/test/stream.sdp
    it does not work anymore. QT Viewer will connect, but does not play the stream.
    I am wondering why this is. Is the mountpoint relevant for accessing the stream? All neccessary information should be contained in the sdp file.
    The reason I would like to do that is because we have a payment module for QTSS that will use a 'virtual' address to dispatch a stream to the client. This works perfectly for non live content.
    Thanks
    NikWest

    Ok.
    No problem now.
    Thanks.

Maybe you are looking for