Videographer Wants To Live Stream - Lost At Amazon

I am a freelance videographer and I am researching methods of live video streaming. I want to offer my clients, some of whom are entertainers and event promoters, the ability to stream events, and on-demand recordings, and also to control access to them. In other words, I need the flexibility to stream on my own Livestream platform or some other platform and have it be protected. The basic Livestream platform only supports fully public streams and recordings.
I have a computer fully equipped to perform live streaming, but I have not set up the services associated with it (except for the Basic Livestream platform)
I have an Adobe Flash Live Media Encoder installed as part of the Livestream Studio desktop application for multi-camera productions. I also have a Teradek Vidiu - a camera-top unit that allows me to stream via WIFI or hard-wired internet on remote shoots without a computer. Finally, I also have a subscription to Adobe Creative Suites 2014.
Two days ago, I streamed an event live for the first time using Livestream, and it was a success! However, it was mainly an experiment to test my equipment and the quality of the video stream. And because the stream was public, there was no way to charge the viewers for watching. Livestream does offer the services that I think I need to control access to my video streams - password protection, viewer login, pay-per-view, etc., but they charge a very high Enterprise level fee that is inaccessible to a lone operator like myself. So I am looking for alternatives.
I also came across something called WOWZA, to which I have a trial account, but I know nothing about how to use.
The Adobe Live Flash website led me to Amazon, and here I am. I already had an Amazon.com account, so it was easy to create an AWS account. But I have no idea what I am doing here. I see all sorts of Amazon web and cloud services, but I seem to have lost any links or references to live media streaming, which I originally set out to do. 
Replies: 0    |  Pages: 1

I may as well update my own thread because nobody out there seems to see it. I have upgraded my Livestream account to the Premiere level of their platform, which gives me some nice capabilities. But, because everything is done in house at Livestream, I am not really learning the nuts and bolts of how anything works.
I know nothing about Adobe Flash Live Media Encoder, which happened to be installed along as part of the Livestream Studio application. I am trying to understand it apart from LS, and how it might integrate with some of the other CDNs. I'm not learning much here on this forum about Adobe's own application. I though this might be the place to go for answers

Similar Messages

  • Re-Start a Live Stream using AMS 5 through Cloudfront

    We have AMS 5 running on an Amazon EC2 instance. We send a Live Stream from an FMLE encoder using H.264/AAC.
    If we do an Origin Pull from that instance running the AMS 5 and Live steam through one of our commercial CDN accounts, we can stop and re-start our Live stream without issue. It stops and re-starts as it should. When we re-connect to the AMS using FMLE, using the Origin Pull, the stream re-starts in either the OSMF or JW Player 6 in under 30 seconds.
    If we use that same AMS 5 on the same EC2 instance and Live stream through an Amazon Cloudfront distribution, if we stop our Live stream, when we re-start the stream it takes an unacceptably long time for the stream to refresh - if ever. Basically, after the re-start of the Live stream through FMLE, the Live stream at the player plays the first 10 or 15 seconds of the intial Live stream and then stops. If you refresh the player, it will play that same 10 or 15 seconds again - and stop again.
    It would seem that the AMS 5 is setup correctly. It works through our commercial CDN.
    It also would seem that the issue is with the cache on Amazon Cloudfront. Our assumption is that the Cloudfront cache is not properly updating. But we have tried about every variable in every setting for Cloudfront that we can find. We can't get past this.
    Any suggestions?
    Thanks
    As a followup to this discussion, you can follow the issue through the AWS.Amazon forums at https://forums.aws.amazon.com/thread.jspa?threadID=120448&tstart=0

    We also have other threads going.
    http://forums.adobe.com/thread/1180721?tstart=0
    https://forums.aws.amazon.com/thread.jspa?threadID=120448&tstart=0
    There appears to be 2 different issues - one issue with Amazon Cloudfront and a second issue with Adobe Media Server. Unfortunately, they are tied together in the real world application of these programs.
    In AMS. when we use the adbe-record-mode=record, then the AMS works fine because the AMS is clearing the streams on each re-publish. Unfortunately, this causes a status code error on Amazon Cloudfront. And that particular status code error from a custom origin server causes Cloudfront to "override the minimum TTL settings for applicable cache settings". As the programmers explained, basically what happens is that Cloudfront overrides our 30 second cache settings and goes into a standard cache setting of 5 minutes. And waiting 5 minutes to re-start a Live stream just doesn't work in the real world for real customers.
    We can fix that Cloudfront status code error issue by using adbe-record-mode=append on the AMS. If we use that AMS setting, then we can stop a Live stream, re-start the stream, the Cloudfront cache clears and everything works - unless you also change the stream settings in your FMLE. If you stop the stream and change the video input or output size in the FMLE, then the AMS server gets stuck and can't/won't append to the f4f record file.
    Does anyone have any ideas about how we can adjust all of this so that it works together?
    Thanks

  • Rtmp live streaming of videos!

    I am loving Azure Platform. I am having certain problems, i hope you can help me resolve them.
    Problem : I want to live stream videos using rtmp. The input for rtmp will be using ffmpeg running in my azure VM. I am new to using rtmp. Any help is welcome. Thank you.

    hi i met a problem
    ffmpeg -v verbose -y -i sample.mp4  -strict -2 -c:a aac -b:a 16k -ar 24000 -r 30 -g 60 -b:v 2m -c:v libx264 -f flv rtmp://livetest1-tracyone.channel.mediaservices.chinacloudapi.cn:1935/live/1ea048c582e34115b9d76cb98c6e9fb6
    Parsing...
    Parsed protocol: 0
    Parsed host    : livetest1-tracyone.channel.mediaservices.chinacloudapi.cn
    Parsed app     : live
    RTMP_Connect1, ... connected, handshaking
    RTMP_Connect1, handshake failed.
    rtmp://livetest1-tracyone.channel.mediaservices.chinacloudapi.cn:1935/live/1ea048c582e34115b9d76cb98c6e9fb6: Unknown error occurred

  • When I want to watch live stream, the buffer comes on all the time. I have 32bit, server is wildblue, highest option

    32 bit laptop, wildblue, highest speed. Live stream is choppy. it doesn't run smoothly

    I had this exact same problem- Found a suggestion on a simular thread that fixed it. Right Click anywhere on your screen to get to Display Properties. Find out the make and model of your video card(in my case it was Nvidia). Go to their website and to the driver download link. Download the updated driver for your card and follow the instructions to install.

  • Debugging live streaming ingest protocol

    Hi everyone,
    I posted a separate question last week about live streaming protocols. I had been searching for documentation on fmp4 live streaming ingestion and found nothing, but afterward I found the Smooth Streaming protocol specification (MS-SSTR-- the forum won't
    let me link to it directly), which seems to apply to my use case. Now I've spent a few days trying to make it work and so far haven't been able to do so.
    I'm not looking for a canned solution; I'm looking for resources to help me debug the problem.
    According to the specification, I should be able to make a long-lived, chunked http POST request to my ingest endpoint, sending the fragmented mp4 data as the request body. I've tried to implement this without success-- I open the request stream and
    write 1000 byte chunks at a time, but the connection is reset after only one or two chunks. I can't tell if the problem is the data or the chunk size or something else.
    And if I mimic Expression Encoder 4, which seems to diverge from the spec and not send its data chunked at all, the connection isn't reset no matter how much data I send, but the video never makes it to the preview stream on the Azure portal.
    So the question is, how can I debug this? Is there any error log or diagnostic output I can access? Each time I try something new, it doesn't work, but I have no way of knowing whether I am getting closer to or further from a working solution. I'm a bit
    lost here, so help would be greatly appreciated.

    Hi Jason, and thanks for getting back to me.
    The IP security thing is a good thought, but I think I have that part taken care of already. I saw the option to restrict ingest to the IP I used to configure the channel from in the Portal, and I made sure to set it to Off. Also, I can stream successfully
    from Expression, which would have the same source IP.
    As for why I want to use Smooth Streaming, basically there are two reasons. First, it seems like the simpler protocol conceptually. I'm already generating fragmented mp4 files (though I think they might be packaged wrong; I can't really tell which part is
    causing problems), and it seems simple enough that I could implement it without turning to a third-party library (the part that's feeding the stream is written in C#, though I'd be willing to rewrite in C++ if necessary, and there's no RTMP client
    that seems trustworthy-- FluorineFX, for instance, seems to have ceased development years ago now).
    Second, while reading the RTMP spec, I realized that even if I had such a library it isn't clear how I would need to use it with Azure. It allows creating objects or invoking methods on the server, but I couldn't find any guide to which objects and methods
    might be supported. And the spec also doesn't define specifically what video format, codec options, or packaging method I need to use; essentially I would seem to have all the same problems I'm having with Smooth, only with the added difficulty of needing
    a third-party library to implement the spec properly.
    I guess the main problem is that I don't really know how close or far I am from success. Like I said, I think I am doing something wrong in how the video is encoded, but for all I know I'm being dropped because I'm feeding the data too quickly
    or slowly. (I can't find any documentation on how much data Azure can buffer or how it handles underruns when the video to be streamed live is generated on the fly and not tied to strictly realtime output). If there's more documentation on specifically how
    Azure uses or expects data from these protocols, it'd be very helpful.

  • Live Streaming Using my Webcam As a secruity camera

    I'm lost here someone please help, : \ im trying to use my isight as a security camera over the internet so i can watch my dog i just got her.
    So the question is what software out there i can use to put my webcam as a camera ( again my powerbook is kinda old im still running on panther so i have 10.3.9 not 10.4 can i work anyway around this) where i can watch live.
    Can someone help me i know i need my ip address i have that but when i hook everythign else up where do i go when im somewhere else when i want to look at my live streaming from work to my house? please can someone help me thank you if you can. : )

    Hello, keith28
    Until we hear back from juanorfrankie, here is my suggestion for your question:
     I would like to access the isight using my iPhone.
     Can anyone make some suggestions?
    If iPhone will can support viewing a streaming webcam page, use EvoCam with your iSight to publish a webcam with streaming video. That would probably be the simplest way to do what you want, but, because I do not have iPhone, so I cannot test this.
    See the instructions and FAQs at EvoCam for system requirements and details on how to set up your webcam page.
    EZ Jim
    PowerBook 1.67 GHz   Mac OS X (10.4.10)    G5 DP 1.8  External iSight

  • Live Streaming from static video file

    Hi everybody,
    I'm using Amazon Cloud & Flash MMedia Server 4.5.
    I would like to take one of my exits videos (a static file), and turn it into a live stream. People can view this video (synchronous). I'm using Flash Media Live Encoder 3.2, but it only capture video from devices as camera...
    How I can do it?
    Cheers.

    Hi Huy,
    Please find the zip attached where I have written the scripts for you to deploy on your local FMS server, in context of option 2:
    You may like to follow these steps:
    1. unzip the folder.
    2. deploy FLVpublishonLoad (in case your recorded file is a FLV use case) on your lcoal server
    3. deploy toPublish on your remote aws server
    4. make sure to correct the path for aws in FLVPublishonLoad application @ line#19:
    // Please mention your aws instance hostname instead of localhost , and application name ....
    nc.connect("rtmp://localhost/toPublish");
    5. go to admin console of your local development server and load the instance of FLVPublishonLoad as shown in the howtoLoadfromAdminConsole.png image attached for refrence purpose.
    6. it will automatically connect to the "toPublish" app on your remote server and start publishing your local vod file as a live stream to your aws instance
    7. start your subscriber app and subscribe to the stream you have used in FLVPublishonLoad app for Stream.get() method, see line#29 and the subscriberScreenExample.png file attached for the illustration purpose.
    few points to note here:
    1. I have removed the sample.flv and "sample1_1500kbps.f4v" for keeping the zip size lower.
    2. You need to put your stream name at line#55 where you call mystream.play()
    3. The stream name that you specify in Stream.get() will be used by subscribers.
    4. if your use case is mp4, then please use "MP4PublishonLoad" instead of "FLVPublishonLoad" application.
    =============
    As I am unable to attach the files here therefore copy-pasting the code for you and others to be re-used:
    =======
    main.asc code for FLVPublishonLoad:
    var nc;
    var ns;
    application.onAppStart = function()
        trace("hello client: ");
        publishIt();
    function publishIt()
        trace("publishing");
        nc = new NetConnection();
        // Please mention your aws instance hostname instead of localhost , and application name ....
        nc.connect("rtmp://localhost/toPublish");
        nc.onStatus = function(info)
            trace(info.code);
        ns = new NetStream(nc);
        // Exact stream name available for subscribers .....
        mystream = Stream.get("myvodfile");
        mystream.onStatus = function(sinfo)
            trace("mystream.onStatus: "+sinfo.code);
            if(sinfo.code == "NetStream.Publish.Start")
                attach_retVal = ns.attach(mystream);
                if(attach_retVal==true)
                        trace("stream attach was successful ...");
                        startPublish();
                else
                    trace("The attempt to attach stream source to NetStream failed");
        // Please put the stream name here inside double-quotes that you want to publish and is available there in streams/_definst_ folder ......
        mystream.play("sample",0,-1,true);
    function startPublish()
        ns.publish(mystream.name,"live");
    main.asc code for MP4PublishonLoad:
    var nc;
    var ns;
    application.onAppStart = function()
        trace("hello client ");
        publishIt();
    function publishIt()
        trace("publishing");
        nc = new NetConnection();
        // Please mention your aws instance hostname instead of localhost , and application name ....
        nc.connect("rtmp://localhost/toPublish");
        nc.onStatus = function(info)
            trace(info.code);
        ns = new NetStream(nc);
        // Exact stream name available for subscribers .....
        mystream = Stream.get("mp4:myvodfile.f4v");
        mystream.onStatus = function(sinfo)
            trace("mystream.onStatus: "+sinfo.code);
            if(sinfo.code == "NetStream.Publish.Start")
                attach_retVal = ns.attach(mystream);
                if(attach_retVal==true)
                        trace("stream attach was successful ...");
                        startPublish();
                else
                    trace("The attempt to attach stream source to NetStream failed");
        // Please put the stream name that you want to publish and is available there in streams/_definst_ folder ......
        mystream.play("mp4:sample1_1500kbps.f4v",0,-1,true);
    function startPublish()
        trace("#### " + mystream.name);
        ns.publish("mp4:" + mystream.name,"live");
    main.asc code for "toPublish" app
    =====================
    application.onPublish = function(clientObj, streamObj)
        trace("published: " + streamObj.name);
    =====================
    Please revert back to me in case of further query.
    Regards,
    Shiraz Anwar

  • Live Stream connection limit per server

    I Need to Broadcast 20,000 to 40,000 Live Streams.! (Bitrate : 128Kbps / Viewers Per Stream Apprx-10 Server : Adobe FMS Server Amazon EC2 Instance - Bandwidth Apprx 100mbps)
    My Question is
    1. How Many Concurrent connections Possible for one Adobe FMS Server? (Amazon EC2 Instance)
    2. is this possible to use RTMFP for this process? (20,000 Connections Via Amazon Adobe FMS Server EC2 Instance )
    3. if Possible Over RTMFP then How Many Developer Keys Need for this Operation. (or) Pricing Details for the Developer Key. (or) Any Separate Key[s] Available For Amazon EC2 Instance?
    Thanks
    Sarma

    1.  Looking at your stream specs - its like  128 X 10 ~ around 1 Mbps for one stream considering 10 clients. And considering bandwidth of 100 mbps for instance you can support around 1000 clients. But few things i want to know before we can conlude on figures - what is NIC capacity of the instance.
    2. Yes if you using FMS on AWS offering - you can use RTMFP
    3. You don't need developer keys if you are using FMS on AWS offering to use RTMFP

  • HTTP Live Streaming and 3rd party servers.

    I have a working HTTP Live Streaming (HLS) setup on my Snow Leopard server, where I have the live stream embedded in a Safari page via HTML5. I want to setup a very basic kind of CDN with something like a Microsoft 2003 Server, where the encoding and segmenting of the stream is handled by the Snow Leopard Server  but the hosting / distribution of the stream / site is handled by the Microsoft server.
    I'm looking for documentation on how to go about doing this. I know that it's possible to configure HLS to work with commercial CDNs like Amazon Cloudfront and Akamai, but my needs are much smaller as I'm looking to serve very few clients.
    Thanks

    https://support.mozilla.org/en-US/kb/troubleshoot-firefox-issues-caused-malware

  • Multiple Live Streaming Video in flash?

    Hello Everyone,
    I have this project in mind which I want to create a video
    montage of different live streaming videos from a few different
    users with their webcam in different locations. There will be the
    website which the participants will be able to view the live video
    montage in real time. However, I am unsure of where to start with
    (live video & server). Therefore I will be really thrilled if
    someone knows an example similar to this or someone who can offer
    me a lead in this topic!!
    I have knowledge in flash and abit of actionscripting.
    Much THANKS!!
    jen

    The 2-way video chat module example in the sample chapter
    from "Learning Flash Media Server 3" should give you some ideas.
    Obviously, you only want the 1-way (user to server) part of
    that, as you'll be displaying the video montage I guess in a web
    page as a set of regular flash streams.
    Main page
    http://www.adobe.com/devnet/flashmediaserver/articles/ora_learning_fms3.html
    Sample chapter (PDF) :
    http://www.adobe.com/devnet/flashmediaserver/articles/ora_learning_fms3/learning_fms3_ch05 .pdf
    It looks to be an O'Reilly book, so you may be able to get
    the whole thing online, or get it delivered from O'Reilly or Amazon
    in a few days.

  • Doubt regarding HTTP Live Streaming for windows phone 8.1 silverlight app?

    Hi,
    I am developing a Windows phone 8.1 silverlight app.
    I want to add live streaming youtube channel , thats why I used  https://phonesm.codeplex.com/ this link as a reference , and I am able to play the links given in the samples, but I am unable to play my link https://www.youtube.com/watch?v=p2GF-QQS6n0
    I am getting 3108 unable to play media error with my link,
    My question Is there any need of streaming server ?
    If yes can anyone help how to  create the streaming server.
    Thanks....
    Suresh.M

    Hi Suresh.M
    As I can see a similar question on the Windows Phone Streaming Media discussion:
    https://phonesm.codeplex.com/discussions/545737
    You may need check into the call stack and find out what is the root cause, for instance if there any chance that your phone cannot reach to the video address, and did you enable some network capabilities, etc.
    --James
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder which does not have firewire out/input? it comes only with a component video output, USB, HDMI and composite RCA output?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • How do I use multiple cameras to live stream through FME?

    I am looking to live stream pool tournaments from multiple angles but don't know what software or hardware I might need. Anybody have any good how to advice or links they might share? I stream through Ustream.tv if that makes a difference. Should I look for something different? Thanks

    I am working on getting just the counter working by using
    the program posted previously, and I am running into issues. Periodically I get
    the error:
    Error -200141
    occurred at DAQmx Read (Counter DBL 1Chan 1Samp).vi
    Possible reason(s):
    Data was overwritten
    before it could be read by the system.
    If Data Transfer
    Mechanism is Interrupts, try using DMA. Otherwise, divide the input signal
    before taking the measurement.
    It seems to work better if I use cascaded counters, but I need timer 0 for
    analog channels when I run this code along with the program for the other
    measurements.
    I have tried averaging, and selecting different values for the millisecond
    timer, and these did not seem to have an effect.
    I tried different DAQms configurations and "Counter DBL 1Samp" seemed
    to work the best.
    The program will work for a while and then it will give me the above error
    message.
    If I use counter 0 as a cascaded counter input, the program runs fine. If I run
    this with other analog channels, it errors out because the analog channels use
    counter 0.
    If I use counter 1 as a cascaded counter input, it seems to work better than a
    single channel, but it will still error out with the above error.
    If I use only counter 1, I get the error above even faster.
    Also, none of the
    configurations give measurements outside the While Loop.
    The only place I can add a speed dial for the front panel is within the While
    Loop.
    Is there someway to get the signal to continuously send out of the while loop?
    I thought if I could get the signal out of the while loop, I could condition it
    anyway I wanted without the program erroring out.
    Any suggestions would be much appreciated.
    Thank you.
    Attachments:
    Counter_error.jpg ‏45 KB

  • How can I do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder ..that doesn't have firewire out/input?

    I need to do live streaming with a Mac Book Pro 2011, using a new model of Sony HD camcorder (http://store.sony.co...ber=HDRAX2000/H) ..this camcorder model does not have firewire out/input ..it comes only with a component video output, USB, HDMI and composite A/V video output..
    I wonder how can I plug this camcorder to the firewire port of my laptop? Browsing on internet I found that Grass Valley Company produces this converter http://www.amazon.co...=A17MC6HOH9AVE6 ..but I am not sure -not even after checking the amazon reviews- if this device will send the video signal through firewire to my laptop, in order to live streaming properly? ..anyone in this forum could help me please?
    Thanx

    I can't broadcast with the built in iSight webcam... how would I zoom in or zoom out? or how would I pan? I've seem people doing it walking with their laptops but that's not an option for me... there's nothing wrong with my USB ports but that's neither an option to stream video because as far as I know through USB you can't connect video in apple operating systems ..you can for sure plug any video cam or photo camera through usb but as a drive to transfer data not as a live video camera...  is by firewire an old interface developed by apple that you can connect all sorts of cameras to apple computers... unfortunately my new sony HDR-AX2000 camcorder doesn't have firewire output...
    thanx

  • How do I create a variable video delay of a live stream...

    I have a live video feed from my webcam in my Flash Application. I have a second video window next to it that I want to place a variable delayed video of the live stream. Therefore, I need to save the video stream in memory/disk and create this variable delay, say 5-60 seconds. As well as, continue to capture the live stream. The camera I am using suppots H.264 encoding.
    How do I create a variable length queue or buffer to hold the video stream coming into the flash application. Do I create a memory variable or write this to disk ? I have been looking over the ActionScript 3.0 documentation and I can not figure out how to code this either way ( as a memory variable or write to disk queue ).
    I want to be able to change the delay, view the delay stream in slow motion and scrub through the delayed stream.
    I would like to do this with out having to use Flash Media Server.
    Thanks,
    Bob

    I am sure it is practically impossible to accomplish that on the client side. Although theoretically it is conceivable to use NetStream.appendBytes() but it requires an extremely complex implementation.
    I suggest you look into FMS DVR capacities:
    http://help.adobe.com/en_US/flashmediaserver/devguide/WS236AE81A-5319-4327-9E44-310A93CA09 C6Dev.html

Maybe you are looking for

  • How do I get old emails from Time Machine into a new Mail?

    I have (I hope) email data in Time machine as of April 8  which I can't figure out how to get into my newly reinstalled version of Mail.  I took over my brother-in-law's iMac on April 8, and closed his User ID, saving it in a .dmg file.  Now that he

  • Photoshop CS4 Activation Error

    Hello, I've been using Adobe products for a while now. However, I'm experiencing a problem with Photoshop CS4 - I have been for a while but I thout it'd eventually go away. I have 2 computers. I have CS4 installed on both. However, a while back It wa

  • FB60-No display of special gl transaction

    Hi, in the T.code FB60 special g/l transaction symbols are not displayed. According to my client he want to post special g/l transactions in FB60 only. he dont want F-43 and anyother. When i want to in FB60. to choose a special gl indicator the syste

  • Sales Document Mapping Error

    Tcode SMW01 in SAP CRM Bdoc Error Bdoc State: E08 Bdoc Description: Mapping error Bdoc Type: BUS_TRANS_MSG Bdoc Detail Error Description: Business transaction type ZZZZ does not exist The ZZZZ is a quotation transaction type maintained in ECC. But we

  • Problem opening downloaded PDF files

    When I click on a PDF link in Safari, I would normally expect the document to open with Acrobat 7.0. What is happening now, is that the screen changes to 'browse' option asking me to identify which application will open the PDF. When I click on Acrob