Large scale Connect Session, 3 Continents, 1000s of participants, tips?

I am doing a large scale webinar using Adobe Connect with 20 different presenters in three different continents.  I will be the host, but will give up host duties to others who will act as technical team members in the three regions (Europe, Americas, Pacific).  This sounds like alot of moving parts, and it is my first show with this software.
Does anyone have any tips on this level of participation?  Any information would be helpful...
Thanks

Congrats. Lots of customers are doing that today. How many participants? 1000's?
And you are running this on your own server or Adobe Hosted?
You using VOIP or using an integrated audio provider?
Suggestions:
1. Hosts should avoid using wireless connections as they are always risky for interference. Best practice is to use wired ethernet.
2. Keep the use of webcams to the start but after the meeting gets going reduce the use of webcam video as sometimes it can be distracting....all depends on the focus of your meeting
3. Mute everyone but the hosts to keeo audio interference and noise to a minimum
4. Suggest that VOIP users have proper headsets designed for VOIP. This is s best practice.
5. Make sure you add a nice background image to the room using preferences...personalize it
6. Use the Q&A Pod to field questions.....download the multi-language chat pod from the Adobe Connect Exchange....trnaslates chat into various languages on the fly....very cool with the U.N.

Similar Messages

  • Very-large-scale searching in J2EE

    I'm looking to solve a very-large-scale searching problem. I am creating a site
    where users can search a table with five million records, filtering and sorting
    independantly on ten different columns. For example, the table might be five million
    customers, and the user might choose "S*" for the last name, and sort ascending
    on street name.
    I have read up on a number of patterns to solve this problem, but anticipate some
    performance issues. I'll explain below:
    1) "Page-by-Page Iterator" or "Value List Handler"
    In this pattern, it appears that all records that match the search criteria are
    retrieved from the database and cached on the application server. The client (JSP)
    can then access small pieces of the cached results at a time. Issues with this
    include:
    - If the customer record is 1KB, then wide search criteria (i.e. last name =
    S*) will cause 1 GB transfer from the database server to app server, and then
    1GB being stored on the app server, cached, waiting for the user (each user!)
    to ask for the next 10 or 100 records. This is inefficient use of network and
    memory resources.
    - 99% of the data transfered from the database server will not by used ... most
    users flip through a couple of pages and then choose a record or start a new search
    2) Requery the database each time and ask for a subset
    I haven't seen this formalized into a pattern yet, but the basic idea is this:
    If a clients asks for records 1-100 first (i.e. page 1), only fetch that many
    records from the db. If the user asks for the next page, requery the database
    and use the JDBC API's ResultSet.absolute(int row) to start at record 101. Issue:
    The query is re-performed, causing the Oracle server to do another costly "execute"
    (bad on 5M records with sorting).
    To solve this, I've beed trying to enhance the second strategy above by caching
    the ResultSet object in a stateful session bean. Unfortunately, this causes a
    "ResultSet already closed" SQLException, although I ensure that the Connection,
    PreparedStatement, and ResultSet are all stored in the EJB and not closed. I've
    seen this on newsgroups ... it appears that WebLogic is forcing the Connection
    closed. If this is how J2EE and pooled connections work, then that's fine ...
    there's nothing I can really do about it.
    Another idea is to use "explicit cursors" in Oracle. I haven't fully explored
    it yet, but it wouldn't be a great solution as it would be using Oracle-specific
    functionality (we are trying to be db-agnostic).
    More information:
    - BEA WebLogic Server 8.1
    - JDBC: Oracle's thin driver provided with WLS 8.1
    - Platform: Sun Solaris 5.8
    - Oracle 9i
    Any other ideas on how I can solve this issue?

    Michael McNeil wrote:
    I'm looking to solve a very-large-scale searching problem. I am creating a site
    where users can search a table with five million records, filtering and sorting
    independantly on ten different columns. For example, the table might be five million
    customers, and the user might choose "S*" for the last name, and sort ascending
    on street name.
    I have read up on a number of patterns to solve this problem, but anticipate some
    performance issues. I'll explain below:
    1) "Page-by-Page Iterator" or "Value List Handler"
    In this pattern, it appears that all records that match the search criteria are
    retrieved from the database and cached on the application server. The client (JSP)
    can then access small pieces of the cached results at a time. Issues with this
    include:
    - If the customer record is 1KB, then wide search criteria (i.e. last name =
    S*) will cause 1 GB transfer from the database server to app server, and then
    1GB being stored on the app server, cached, waiting for the user (each user!)
    to ask for the next 10 or 100 records. This is inefficient use of network and
    memory resources.
    - 99% of the data transfered from the database server will not by used ... most
    users flip through a couple of pages and then choose a record or start a new search
    2) Requery the database each time and ask for a subset
    I haven't seen this formalized into a pattern yet, but the basic idea is this:
    If a clients asks for records 1-100 first (i.e. page 1), only fetch that many
    records from the db. If the user asks for the next page, requery the database
    and use the JDBC API's ResultSet.absolute(int row) to start at record 101. Issue:
    The query is re-performed, causing the Oracle server to do another costly "execute"
    (bad on 5M records with sorting).
    To solve this, I've beed trying to enhance the second strategy above by caching
    the ResultSet object in a stateful session bean. Unfortunately, this causes a
    "ResultSet already closed" SQLException, although I ensure that the Connection,
    PreparedStatement, and ResultSet are all stored in the EJB and not closed. I've
    seen this on newsgroups ... it appears that WebLogic is forcing the Connection
    closed. If this is how J2EE and pooled connections work, then that's fine ...
    there's nothing I can really do about it.
    Another idea is to use "explicit cursors" in Oracle. I haven't fully explored
    it yet, but it wouldn't be a great solution as it would be using Oracle-specific
    functionality (we are trying to be db-agnostic).
    More information:
    - BEA WebLogic Server 8.1
    - JDBC: Oracle's thin driver provided with WLS 8.1
    - Platform: Sun Solaris 5.8
    - Oracle 9i
    Any other ideas on how I can solve this issue? Hi. Fancy SQL to the rescue! If the table has a unique key, you can simply send a
    query per page, with iterative SQL that selects the next N rows beyond what was
    selected last time. Eg:
    Let variable X be the highest key value you've seen so far. Initially it would
    be the lowest possible value.
    select * from mytable M
    where ... -- application-specific qualifications...
    and M.key >= X
    and (100 <= select count(*) from mytable MM where MM.key > X and MM.key < M.key and ...)
    In English, this says, select all the qualifying rows higher than what I last saw, but
    only those that have fewer than 100 qualifying rows between the last I saw and them (ie:
    the next 100).
    When processing this query, remember the highest key value you see, and use it for the
    next query.
    Joe

  • ACS issues in large scale network with Prime Infra and WAAS express

    Hi,
    I wonder if there is a common practice or a recommended way for deploying large scale network where there are Prime Infrastructure (PI) and WAAS Central manager keep logging into routers (scale of 1000 or more) to collect statistics. The way PI and WAAS CM collect stats from the routers (besides using SNMP) is that they log in (authenticate) themselves with there usernames and password and issue multiple show and config commands on the routers. Imagin this routine happens every 5 - 10 minutes with all 1000+ routers at the same time and the impact to the ACS server in terms of authentication requests and AAA logs. Appreciate if somebody could recommend a solution where these elements can work together in a large scale network.
    Thanks,
    Tos

    The AEBS is connected to the TC via an ethernet run from the basement to the main floor... its not connected wirelessly.
    The "extend" feature is intended for wireless, not wired connections. Since you have the base stations connected by Ethernet, the downstream router just need to be reconfigured as a bridge. The bridged router would then perform as a combination Wireless Access Point and Ethernet switch. Neither base station should be configured for "extending."
    Basically, you will want both to be configured for a "roaming" network.
    o Setup the base station connected to the Internet to "Share a public IP address."
    Internet > Internet Connection > Connection Sharing: Share a public IP address
    o Setup the remaining base station(s), as a bridge.
    Internet > Internet Connection > Connection Sharing: Off (Bridge Mode)
    For each base station in the roaming network:
    o Connect to the same subnet of the Ethernet network
    o Provide a unique Base Station Name
    o The Network Name should be identical
    o If using security, use the same encryption type (WEP, WPA, etc.) and password.
    o Make sure that the channel is set at least three channels apart from the next base station.
    while the TC is running at 2.4ghz since my MBP is connected at speeds around 240 to the AEBS at the same time that my ipod is connected to the TC at speeds of only 54 max.
    The iPod is a 802.11b/g wireless device. It cannot connect at greater than the maximum bandwidth for that mode ... which is 54 Mbps, regardless of the bandwidth available.

  • Flex deployed on a large scale?

    We plan on developing a new product and Flex popped in my
    mind as a development platform. I know a good deal of Flex 1.5, but
    only used it for personal sites.
    My question is how well Flex behaves in a large scale
    environment to those who have deployed it in such. Server load will
    be at least thousand / day.
    Thanks!

    Hmm... I made one medium sized application in 1.5 (approx 10
    screens, user access <1000 times per day) and it seems to be
    working alright for the client.
    Now I am working on a major application (over 20 main
    screens, and definately access>1000 times per day) and it is not
    going well. I am really worried about the bugs and memory issues of
    Flex 2.0. I have also not found a sure-fire way to address these
    potential issues. I can say this: for the size of application we
    are making, Flex and Flash Player just aren't up to the job.
    Compiled and executed as a single .swf application results in 755MB
    ram usage and for some reason a constant CPU access of 60% (Pentium
    4 proc.) after accessing every screen. And this is just FlashPlayer
    doing what it is supposed to. Me, not being a computer engineer,
    can't really address these problems. Flex and AS aren't C. I can't
    control memory usage with my code. By breaking up the huge
    application into smaller ones and then loading those via an
    SWFLoader I may be able to avoid this rampant resource hogging but
    it's sort of illogical from an application architecture standpoint
    because this is ONE application.
    As a developer, I can see plenty of places to streamline the
    application but this simply isn't possible when dealing with the
    client. They want this screen to look and act this way and that
    screen to look and act the other way. I can talk about how if both
    screens use the same layout and logic they can both use the same
    template class, share static resources, blah blah until I am blue
    in the face but it won't matter because they are the client and
    they decide how the application is going to look--at the expense of
    streamlining. That's just the real world. Then I have to somehow
    make it work.
    By the way... before you think "just use view states!", I do
    use those--and bitwise logic flags for more complicated
    configurations--it's still not enough, although it did cut approx
    160 screens in documentation form down to just 20 in
    implementation.
    In worst case scenarios, I have to deny the client what they
    want and if they ask me why, I have to reply "it can't be done with
    Flex". Then their satisfaction in the product drops. Flex suddenly
    isn't as incredible as it seemed at first. Doesn't matter how
    pretty and animated the screens are when if you run them over an
    hour your computer slows to a halt or .ttc fonts stop loading (HUGE
    issue here in Japan).
    I have yet to see a sample application that comes close to
    the scale of our current project: A library book browser? neat but
    that's just square one; A Commodore 64 emulator? cute. no place in
    business; A real-estate browswer? in our project that would be the
    equivalent of ONE SCREEN out of the entire application.
    I like Flex. It's fun--on a small scale. But I never want to
    develop a real world business application using it again. There are
    way more (and way more skilled) Java, JSP, PHP, etc. etc.
    developers out there than Flex developers who can make much more
    robust applications. It's a shame the client got caught up in the
    hype of Flex RIA before the technology was ready for the task.
    Very long story short: Beware using Flex for an involved
    application.
    It's going to require exponentially more time than a smaller,
    less ambitious project--especially if you don't purchase FDS. And
    oh my god implementing a Flex application on a legacy Struts
    framework... kill me now! As much as I hate "page-refresh"
    applications, Flex (both 1.5 and 2.0) has not proven to be the
    god-send that I had hoped and dreamed it would be as a developer.
    What can you expect, though? It's only been out a few years.... And
    as far as clients' perspectives go, the price for FDS also
    certainly doesn't help make it appealing. That is why it is so
    embarrassing to tell them their dream application is quickly
    becoming an egregious memory hog.
    Anyway, good luck if you take on your project with flex. Just
    be careful!

  • Typical/Common large-scale ACE deployment or designs?

    I am deploying several ACE devices and GSS devices to facilitate redundancy and site load balancing at a couple of data centers.  Now that I have a bunch of experience with the ACE and GSS, are there typical or common ACE deployment methods?  Are there reference designs?  I have been looking, and haven't really found any.
    Even if they are not Cisco 'official' methods, I'm wondering how most people, particularily those who deploy a lot of these or deploy them with large-scale systems, typically do it.
    I'm using routed mode (not one-arm mode) and I'm wondering if most people use real server (in my case, web servers) with dual-NICs to support connectivity to back-end systems?  Or do people commonly just route it all through the ACE?
    Also, how many VIPs and real servers have been deployed in a single ACE 4710 device?  I'm trying to deploy about 700 VIPs with about 1800 Real Servers providing content to those VIPs.
    How do people configure VIPs, farms, and sticky?  I'm looking for how someone who wants to put a large ammount of VIPs and real servers into the ACE would succeed at doing it.  I have attempted to add a large number in the 'global' policy-map, but that uses too many PANs.
    I have tried a few methods myself, and have run into the limit on Policy Action Nodes (PANs) in the ACE device.  Has anyone else hit this issue?  Any tips or tricks on how to use PANs more conservitively?
    Any insight you can share would be appreciated.
    - Erik

    As far as i can see from your requirements i suggest you create 1 ear file for your portal and 1 ear file per module.
    The ear file from your portal is the main application and the ear files of your modules are shared libraries that contain the taskflows. These taskflows can be consumed in the portal application.
    This way, you can easily deploy 1 module without needing to deploy the main application or the other application.
    It also let you devide your team of developer so everybody can work on a sepperate module without interfering.
    On a sidenote: when you have deployed your main application, and later you create a new module, than you have to register that module to your application so then you will need to redeploy your portal but if you update an existing module, you won't need to redeploy your portal.
    As for the security, all your modules will inherit the security model of your portal application.

  • Kernel parameter for large concurrent connections database.

    Hi,
    Does any one have suggestion about kernel parameters we have to modify for large concurrent connections database (about 20000 concurrent connections) ? We would like to use Sun M8000 as our database server. Oracle installation guide indicated many parameters and values, but it seems ok for general database, not the one we need. Please somebody help us. Thanks !

    I certainly don't have any idea where the 150G memory usage for 20000 concurrent users come about. That's 7.5MB per user session.
    The session memory usage is depends on Oracle version, OS , Database configuration and application behavior etc..
    From following query we can tell, average session memory usage is around 1.5M in normal usage on 10g linux platform.
    It could go up to 20-30M if user session doing large data manipulation.
    SYS@test >  l
      1   select value, n.name|| '('||s.statistic#||')' , sid
      2    from v$sesstat s , v$statname n
      3    where s.statistic# = n.statistic#
      4    and n.name like '%ga memory%'
      5*   order by sid
    SYS@azerity > /
         VALUE N.NAME||'('||S.STATISTIC#||')'        SID
        486792 session pga memory max(26)             66
        486792 session pga memory(25)                 66
        225184 session uga memory(20)                 66
        225184 session uga memory max(21)             66
        225184 session uga memory max(21)             70
        486792 session pga memory max(26)             70
        486792 session pga memory(25)                 70
        225184 session uga memory(20)                 70
        225184 session uga memory(20)                 72
        486792 session pga memory max(26)             72
        225184 session uga memory max(21)             72
        486792 session pga memory(25)                 72
        486792 session pga memory max(26)             74
        225184 session uga memory(20)                 74
        486792 session pga memory(25)                 74
        225184 session uga memory max(21)             74
    ......Aggregate by SID
      1   select sum(value),sid
      2    from v$sesstat s , v$statname n
      3    where s.statistic# = n.statistic#
      4    and n.name like '%ga memory%'
      5    group by sid
      6*   order by sid
    SYS@test > /
    SUM(VALUE)        SID
       1423952         66
       1423952         70
       1423952         72
       1423952         74
       1423952         75
       1423952         77
       1423952         87
       1423952        101
       2733648        104
       1423952        207
      23723248        209
      23723248        210
       1293136        215
       7370320        216
    .........

  • Difference between connection, session and process

    Hi all,
    Can anyone please update me on the difference between connection,session and process.
    Thanks in advance,
    - Sri

    I got this useful note by googled in net. It describes session,connection,process gracefully.
    A connection is a physical circuit between you and the database.A connection
    might be one of many types -- most popular begin DEDICATED server and SHARED
    server. Zero, one or more sessions may be established over a given connection
    to the database as show above with sqlplus. A process will be used by a session
    to execute statements. Sometimes there is a one to one relationship between
    CONNECTION->SESSION->PROCESS (eg: a normal dedicated server connection).
    Sometimes there is a one to many from connection to sessions (eg: like
    autotrace, one connection, two sessions, one process). A process does not have
    to be dedicated to a specific connection or session however, for example when
    using shared server (MTS), your SESSION will grab a process from a pool of
    processes in order to execute a statement. When the call is over, that process
    is released back to the pool of processes.

  • How to increase the Max. No. of connections(Sessions) & Processes in DB?

    Hi,
    What are all the options to increase the Max. No. of connections(Sessions) & Processes in Database?
    Also, I would like to know How we have to validate the count to fix the No. of connections based on the Database Health?
    Kindly clarify me on this.
    Thanks,
    Orahar.

    Orahar wrote:
    Hi,
    What are all the options to increase the Max. No. of connections(Sessions) & Processes in Database?Set PROCESSES parameter. SESSIONS is a derived parameter.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/initparams169.htm#i1132608
    Also, I would like to know How we have to validate the count to fix the No. of connections based on the Database Health?Check v$license.SESSIONS_HIGHWATER value. That would give you an indication of highest possible load on the database. In case of a brand new setup (with no information about the number of concurrent users), you need to start with a reasonable value and revise if needed.

  • Hi, My printing has suddenly changed in adobe to a large scale, as in, what should be one page of print comes out as 24 pages?   I havent changed anything, its happening on more than one document also, I have to stop my printer before all the pages spew o

    Hi, My printing has suddenly changed in adobe to a large scale, as in, what should be one page of print comes out as 24 pages?   I havent changed anything, its happening on more than one document also, I have to stop my printer before all the pages spew out. I have tried printing 'one single page' and it does exactly the same? Help?

    Is the Poster Print feature turned ON?

  • When printing from an online PDF, the page prints in extra large scale. How do I fix this?

    when printing from an online PDF, the page prints in extra large scale. How do I fix this?

    This can happen when Firefox has misread the paper size from the information supplied by Windows. Clearing it can involve finding some obscure settings, but here goes:
    (1) In a new tab, type or paste '''about:config''' in the address bar and press Enter. Click the button promising to be careful.
    (2) In the search box above the list, type or paste '''print''' and pause while the list is filtered
    (3) For each setting for the problem printer, right-click and Reset it. The fastest way is to right-click with the mouse and then press the r key on the keyboard with your other hand.
    Note: In a couple other threads involving Brother printers, the preference '''printer_printer_name.print_paper_data''' was set to 256 and when the user edited it to 1 that fixed the paper size problem. If you see a 256 there, you can edit the value by doubling-clicking it or using right-click>Modify.

  • Grant debug connect session to user

    I understand that if DBA hasn't granted following option to "user" following way:
    grant debug connect session to user;, then "user" wil lget following error when she tries to debug:
    ORA-0131:Insufficient Priviledges
    Debugging requires Debug connect session system priviledgesIs there any explanation why DBA decides to not grant such privilege? Maybe such privilege slows down something? Maybe it can create security "hole"? Maybe there is another good explanation?
    The "user" in this example was Schema/Account ment for Development personell who developes functionality, but it doesn't matter i think what kind the "user" is.
    so why you think i don't have such privilege, what bad could such privilege when granted give?

    Its just that You have not been Granted that Privilege. Mostly in what ever shop i have worked Such grants are given on request Base.
    Debugging is a Developers Right and no one can denie it.
    But i never use any debugging tool. I only relay on my Instrumented Code.

  • Line appears when applying drop shadow on large scale

    Hello!
    Some weeks ago I had to make a large scale graphic (800mmx2000mm) for a roll-up banner. I wanted to apply a drop shadow to a rounded shape, and ugly lines came up. Since it was a bit urgent, I decided not to use it.
    But now I'm curious, so I quickly made an ellipse and added a shadow, so you know what I mean. This also happens when I save it as pdf or image.
    Perhaps someday I will have to use a drop shadow on large scale. So, does anybody knows how to fix this or what could I do in case I need to use this effect in these conditions? I use Illustrator CS6 in Mac with Mavericks.
    Thanks in advance.

    Mike Gondek wrote:
    I was able to create an ellipse to your dimension and got a good drop shadow. What happens if you manually make a drop shadow using appearance?
    FYI I tried making the same using drop shadow filter in CS5 and got this error.
    Incase your file was created in CS5 and opened in 6, I would recreate the drop shadow in CS6. I know they redid the gaussian blur in CS6, but not sure if that affected drop shadow.
    CS6 is better on raster effects at large sizes.
    My file was created and opened in CS6.
    My ellipse is around 175cmx50cm. I tried it manually like you said and got the same results:
    So, I guess I'm alone with such a problem. No idea what is wrong :/

  • Granting system privileges DEBUG ANY PROCEDURE and CONNECT SESSION in 10gXE

    I am using Oracle DB 10g Express Edition and trying out SQL Developer. Whenever I want to debug a procedure I got the error:
    This session requires DEBUG CONNECT SESSION and DEBUG ANY PROCEDURE user privileges.
    But I don't see these privileges to grant in the XE GUI for users.
    Does anybody know if this is supported in XE.
    Thanks.

    Hi
    Use SQL*Plus (or the apex interface), logon as dba and type this sql statements:
    grant DEBUG CONNECT SESSION to <username>;
    grant CONNECT SESSION to <username>;

  • Tweaking product prices on a large scale - how?

    My Client has a software store on BC. His supplier is constantly changing their prices and my client wants to be able to quickly review prices, make changes to reflect supplier prices every few days
    If I export the Product List the Excel export is unusable with it full of HTML markup from the product descriptions.
    Apart from opening each product individually to check and tweak prices how is everyone ammending prices on a large scale.... My client only has 60 products at the moment but this is soon to quadruple and I have prospective clients looking at BC for their ecommerce solution and they have thousands of items.
    Regards
    Richard

    If its just prices you want to input, see if you can just eliminate all the other columns that are not needed and only import the price column with its product identifier ofcourse, and see if it will just update the price and not have to deal with the descriptions...Just a thought...

  • OraOLEDB.Oracle provider create connections (sessions).

    Hi all,
         I have Link server between oracle and MSSQL with following versions,
    Oracle – 10G
    MSSQL – SQL Server 2008 R2
    Link server provider=N'OraOLEDB.Oracle'
    After created link server then without actually doing any queries or anything oracle side create 100 or 120 connections (sessions) from this SQL server to oracle server. Any idea on this?
    Thanks
    Tharindu Dhaneenja

    Alright, so it's not that. It's possible the OLE DB provider isn't installed, depending on how you installed the Oracle client.
    This link has a little windows script that can list all the known OLE DB providers. That should tell you if its installed and SSIS isn't finding it, or if it's not installed at all. http://www.motobit.com/help/regedit/sa117.htm

Maybe you are looking for

  • Problems with Adobe Acrobat - last resort!

    I've tried at the Adobe Macintosh forum, but haven't had any advice that's helped, so thought I'd try here!... I have Adobe CS3 Design Standard Suite, Educational version, running on OS X 10.5.5 on my MacBook Pro. 1 - I cannot update Acrobat from 8.0

  • User decision history report, workflow summary log

    Hello SAP Workflow community, In our project there is a task to implement User Decision history/log management. I want to get programmatically (via ABAP, in order to be able to store this data in DB too) the history all user decisions (statuses of al

  • Unicode in XI

    Hi, Our scenario is like this...We send an XML file with some unicode characters via HTTP adapter to XI and an RFC FM receives the same and sends the same text as response. But in this entire process we are losing all Unicode characters and some junk

  • Installing Adobe Extension Manager and get the error message

    I am trying to install Adobe Extension Manager and get the error message ' Requires Photoshop 32 version in range of inclusively 13.0 and 13.9'  Please help me solve this, not sure how to proceede after getting a message that there is four updates av

  • Time Machine and Upgrade to Lion

    At our office we have a Mac mini Model A1347. It was running Snow Leopard. We went to do some updates and the computer started to lag really bad. We ended up booting in Safe Mode only to find that Disk Utility said the ServerHD drive was failing. So