Enable lastlogin timestamp cause big performance hit

We have installed the ODSEE 11g in our environment and we want to enable the lastlogin timestamp but base on the slamd testing result, the performance got hit almost 50% for user authentication. Is this normal? I created SR with oracle and was told it is normal. is anybody using this function as well? if so any tricks or workaround to increase the perfromance?
Thanks

It's basically adding an update for every BIND operation. If the lastlogin attribute is indexed, there is even more write load. The tuning you can do should start with the kind of tuning you'd do for a high volume of updates to the Directory. You should instrument the host, paying particular attention to iostat data like service times on disks and how much CPU time is spent in IOWait.

Similar Messages

  • Joining two queries causes big performance hit

    There are two queries: one with a sub query and one without
    although they both return the same data, one uses a full table scan of the indexed 'trade_index' table
    and thus is much slower
    any Ideas why it does this?:
    SQL> select count(*) from trade_index ti, counterparty_xref cx
    2 where ti.CPY_SID = cx.CPY_SID and ti.TRADE_OR_ARCHIVE = 'A'
    3 and cx.XREF_VALUE in (select ml.VALUE from v_mis_lists ml where ml.LIST_NAME = 'Deutsche Bank DB')
    4 and cx.XREF_TYPE = 'Client RXM'
    5 and ti.START_DATE < to_date('01/01/2004','DD/MM/YYYY')
    6 and (ti.END_DATE >= to_date('01/01/2003','DD/MM/YYYY') or ti.END_DATE = '01-Jan-1900') ;
    COUNT(*)
    6195
    Elapsed: 00:00:15.08
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=2572 Card=1 Bytes=57)
    1 0 SORT (AGGREGATE)
    2 1 HASH JOIN (Cost=2572 Card=26581 Bytes=1515117)
    3 2 HASH JOIN (Cost=460 Card=5619 Bytes=219141)
    4 3 VIEW OF 'VW_NSO_1' (Cost=8 Card=16 Bytes=272)
    5 4 SORT (UNIQUE) (Cost=8 Card=16 Bytes=960)
    6 5 HASH JOIN (Cost=4 Card=16 Bytes=960)
    7 6 TABLE ACCESS (BY INDEX ROWID) OF 'MIS_LISTS' (Cost=1 Card=2 Bytes=60)
    8 7 INDEX (RANGE SCAN) OF 'MISL_NAT_UK' (UNIQUE) (Cost=2 Card=2)
    9 6 TABLE ACCESS (FULL) OF 'MIS_LIST_VALUES' (Cost=2 Card=817 Bytes=24510)
    10 3 TABLE ACCESS (FULL) OF 'COUNTERPARTY_XREF' (Cost=451 Card=35116 Bytes=772552)
    11 2 TABLE ACCESS (FULL) OF 'TRADE_INDEX' (Cost=2056 Card=286928 Bytes=5164704)
    SQL> select ml.VALUE from v_mis_lists ml where ml.LIST_NAME = 'Deutsche Bank DB';
    VALUE
    B49540
    B60010
    Elapsed: 00:00:00.00
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE
    1 0 NESTED LOOPS
    2 1 TABLE ACCESS (BY INDEX ROWID) OF 'MIS_LISTS'
    3 2 INDEX (RANGE SCAN) OF 'MISL_NAT_UK' (UNIQUE)
    4 1 INDEX (RANGE SCAN) OF 'MISLV_PK' (UNIQUE)
    SQL> select count(*) from trade_index ti, counterparty_xref cx
    2 where ti.CPY_SID = cx.CPY_SID and ti.TRADE_OR_ARCHIVE = 'A'
    3 and cx.XREF_VALUE in ('B49540','B60010')
    4 and cx.XREF_TYPE = 'Client RXM'
    5 and ti.START_DATE < to_date('01/01/2004','DD/MM/YYYY')
    6 and (ti.END_DATE >= to_date('01/01/2003','DD/MM/YYYY') or ti.END_DATE = '01-Jan-1900') ;
    COUNT(*)
    6195
    Elapsed: 00:00:02.06
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=532 Card=1 Bytes=40)
    1 0 SORT (AGGREGATE)
    2 1 NESTED LOOPS (Cost=532 Card=41 Bytes=1640)
    3 2 TABLE ACCESS (FULL) OF 'COUNTERPARTY_XREF' (Cost=451 Card=1 Bytes=22)
    4 2 TABLE ACCESS (BY INDEX ROWID) OF 'TRADE_INDEX' (Cost=81 Card=286928 Bytes=5164704)
    5 4 INDEX (RANGE SCAN) OF 'CPY_SID_IDX' (NON-UNIQUE) (Cost=2 Card=286928)
    What I dont understand is that:
    cx.XREF_VALUE in (select ml.VALUE from v_mis_lists ml where ml.LIST_NAME = 'Deutsche Bank DB')
    and
    cx.XREF_VALUE in ('B49540','B60010')
    are essentially the same thing
    (ie the subquery returns the values I have used)
    but the optimizer uses a totally different (slower) method to execute the rest of the former query?
    Many thanks
    Ian

    Hi,
    Here is with join and not using the view: still have full table scan of large table trade_index
    SQL> select count(*) from mis_lists ml, mis_list_values mlv, counterparty_xref cx, trade_index ti
    2 where ti.CPY_SID = cx.CPY_SID and ti.TRADE_OR_ARCHIVE = 'A'
    3 and cx.XREF_VALUE = mlv.VALUE
    4 and mlv.MISL_SID = ml.MISL_SID
    5 and ml.LIST_NAME = 'Deutsche Bank DB'
    6 and cx.XREF_TYPE = 'Client RXM'
    7 and ti.START_DATE < to_date('01/01/2004','DD/MM/YYYY')
    8 and (ti.END_DATE >= to_date('01/01/2003','DD/MM/YYYY') or ti.END_DATE = '01-Jan-1900') ;
    COUNT(*)
    6195
    Elapsed: 00:00:12.08
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=2602 Card=1 Bytes=10
    0)
    1 0 SORT (AGGREGATE)
    2 1 HASH JOIN (Cost=2602 Card=26581 Bytes=2658100)
    3 2 HASH JOIN (Cost=456 Card=5619 Bytes=460758)
    4 3 HASH JOIN (Cost=4 Card=16 Bytes=960)
    5 4 TABLE ACCESS (BY INDEX ROWID) OF 'MIS_LISTS' (Cost1 Card=2 Bytes=60)
    6 5 INDEX (RANGE SCAN) OF 'MISL_NAT_UK' (UNIQUE) (Cost=2 Card=2)
    7 4 TABLE ACCESS (FULL) OF 'MIS_LIST_VALUES' (Cost=2 Card=817 Bytes=24510)
    8 3 TABLE ACCESS (FULL) OF 'COUNTERPARTY_XREF' (Cost=451Card=35116 Bytes=772552)
    9 2 TABLE ACCESS (FULL) OF 'TRADE_INDEX' (Cost=2056 Card=286928 Bytes=5164704)
    Statistics
    0 recursive calls
    47 db block gets
    16519 consistent gets
    4981 physical reads
    0 redo size
    205 bytes sent via SQL*Net to client
    622 bytes received via SQL*Net from client
    3 SQL*Net roundtrips to/from client
    0 sorts (memory)
    0 sorts (disk)
    1 rows processed
    Thanks for you help,
    Ian

  • When Coherence security is enabled there is a significant performance hit

    When Coherence security is enabled, every call to do the cache operation do a security check. This is negatively impacting application's performance. I'm using the built in DefaultController & java keystore login. Is there any workaround so that the performance can be improved.

    Hi user9980369
    It depends how you have implemented security. What version of Coherence are you using?
    JK

  • Just upgraded - and took a big performance hit

    As prompted, I upgraded Safari and Leopard just an hour ago, and my computer has been lagging ever since. My word processor (Mellel) labors over every letter, web sites groan to display themselves, and even typing this brief message is a chore.
    Is this just me?

    Is this just me?
    Yes, it's just you. Unfortunately, it's hard to say what it may take to get you back to normal. These sorts of things happen when you upgrade over an already corrupted system. Since you didn't provide any information on exactly what you did, how you did it, in what order, or over what system level I can only stab at possible solutions.
    First, try the following:
    Repairing the Hard Drive and Permissions
    Boot from your OS X Installer disc. After the installer loads select your language and click on the Continue button. When the menu bar appears select Disk Utility from the Installer menu (Utilities menu for Tiger and Leopard.) After DU loads select your hard drive entry (mfgr.'s ID and drive size) from the the left side list. In the DU status area you will see an entry for the S.M.A.R.T. status of the hard drive. If it does not say "Verified" then the hard drive is failing or failed. (SMART status is not reported on external Firewire or USB drives.) If the drive is "Verified" then select your OS X volume from the list on the left (sub-entry below the drive entry,) click on the First Aid tab, then click on the Repair Disk button. If DU reports any errors that have been fixed, then re-run Repair Disk until no errors are reported. If no errors are reported click on the Repair Permissions button. Wait until the operation completes, then quit DU and return to the installer. Now shutdown the computer for a couple of minutes and then restart normally.
    If DU reports errors it cannot fix, then you will need Disk Warrior (4.0 for Tiger, and 4.1 for Leopard) and/or TechTool Pro (4.6.1 for Leopard) to repair the drive. If you don't have either of them or if neither of them can fix the drive, then you will need to reformat the drive and reinstall OS X.
    Afterwards I would suggest downloading the standalone Safari 3.1 upgrade and the Leopard Combo upgrade from Apple's Download Site. Reinstall the upgrades. Be sure when you do the upgrades that no other applications are running and that you do not use the computer in any way until the upgrades are completed.
    Second, do some maintenance:
    Kappy's Personal Suggestions for OS X Maintenance
    For disk repairs use Disk Utility. For situations DU cannot handle the best third-party utilities are: Disk Warrior; DW only fixes problems with the disk directory, but most disk problems are caused by directory corruption; Disk Warrior 4.0 is now Intel Mac compatible. TechTool Pro provides additional repair options including file repair and recovery, system diagnostics, and disk defragmentation. TechTool Pro 4.6.1 is Leopard compatible; Drive Genius is similar to TechTool Pro in terms of the various repair services provided. The current version, 1.5.1, is Intel Mac compatible.
    OS X performs certain maintenance functions that are scheduled to occur on a daily, weekly, or monthly period. The maintenance scripts run in the early AM only if the computer is turned on 24/7 (no sleep.) If this isn't the case, then an excellent solution is to download and install a shareware utility such as Macaroni, JAW PseudoAnacron, or Anacron that will automate the maintenance activity regardless of whether the computer is turned off or asleep.
    OS X automatically defrags files less than 20 MBs in size, so unless you have a disk full of very large files there's little need for defragmenting the hard drive. As for virus protection there are few if any such animals affecting OS X. You can protect the computer easily using the freeware Open Source virus protection software ClamXAV. Personally I would avoid most commercial anti-virus software because of their potential for causing problems.
    I would also recommend downloading the shareware utility TinkerTool System that you can use for periodic maintenance such as removing old logfiles and archives, clearing caches, etc.
    For emergency repairs install the freeware utility Applejack. If you cannot start up in OS X, you may be able to start in single-user mode from which you can run Applejack to do a whole set of repair and maintenance routines from the commandline. Note that presently AppleJack is not compatible with Leopard.
    When you install any new system software or updates be sure to repair the hard drive and permissions beforehand. I also recommend booting into safe mode before doing system software updates.
    Get an external Firewire drive at least equal in size to the internal hard drive and make (and maintain) a bootable clone/backup. You can make a bootable clone using the Restore option of Disk Utility. You can also make and maintain clones with good backup software. My personal recommendations are (order is not significant):
    1. Retrospect Desktop (Commercial - not yet universal binary)
    2. Synchronize! Pro X (Commercial)
    3. Synk (Backup, Standard, or Pro)
    4. Deja Vu (Shareware)
    5. PsynchX 2.1.1 and RsyncX 2.1 (Freeware)
    6. Carbon Copy Cloner (Freeware - 3.0 is a Universal Binary)
    7. SuperDuper! (Commercial)
    8. Data Backup (Commercial)
    The following utilities can also be used for backup, but cannot create bootable clones:
    1. Backup (requires a .Mac account with Apple both to get the software and to use it.)
    2. Toast
    3. Impression
    4. arRSync
    Apple's Backup is a full backup tool capable of also backing up across multiple media such as CD/DVD. However, it cannot create bootable backups. It is primarily an "archiving" utility as are the other two.
    Impression and Toast are disk image based backups, only. Particularly useful if you need to backup to CD/DVD across multiple media.
    Visit The XLab FAQs and read the FAQs on maintenance, optimization, virus protection, and backup and restore.
    Additional suggestions will be found in Mac Maintenance Quick Assist.
    Referenced software can be found at www.versiontracker.com and www.macupdate.com.
    Third, try updating pre-binding:
    1. Open the Terminal application in your Utilities folder.
    2. At the prompt enter or paste the following command:
    sudo update_prebinding -root / -force
    Press RETURN. You will be prompted for your password which will not be echoed. Wait until the prompt reappears then quit the Terminal and restart your computer.

  • I was wondering can they fix iphone 5  cause today i was on the ferry and a big wave hit the ferry and my whole clothes got wet now the iphone wont charge or no sound at all

    anyone can help with this  i was wondering if apple care can fix the iphone 5 cause today wasnt my day wen i was on ferry back to the city a big wave hit the ferry and i got all wet and now the phone wont turn on or charge with no sound  i wanna know how much im looking for it to be fixed

    Could be new or refurbished, there's no way to tell...they all come in a plain white box. Doesn't really matter though, as all refurbished iPhones get new batteries & cases.
    You get a 90 day warranty or the balance of your original warranty, whichever is greater.

  • Performance hit using "where" clause in the query

    Hi All,
    I am facing a huge performance hit in the java code when using "where" clause in queries. Following are the details:
    1. SELECT * FROM Employee
    2. SELECT * FROM Employee where employeeid in (26,200,330,571,618,945)
    There is no difference in Query Execution Time for both queries.
    Business Logic Time is huge in second case as compared to first one (ratio - 1:20).
    Rows returned are more in first case as compared to second case.(ratio - 1:4)
    Business Logic is same for both the cases where I iterate through the ResultSet, get the objects and set them in a data structure.
    Does anybody know the reason of unexpected time difference for the business logic in the second case?

    Since you're mentioning clustering your index, I'll assume you are using Oracle. Knowing what database you are using makes it a lot easier to suggest things.
    Since you are using Oracle, you can get the database to tell you what execution plan it is using for each of the 2 SQL statements, and figure out why they have similar times (if they do).
    First, you need to be able to run SQL*Plus; that comes as part of a standard database installation and as part of the Oracle client installation - getting it set up and running is outside the scope of this forum.
    Second, you may need your DBA to enable autotracing, if it's not already:
    http://asktom.oracle.com/~tkyte/article1/autotrace.html
    http://www.samoratech.com/tips/swenableautotrace.htm
    Once it's all set up, you can log in to your database using sql*plus, issue "SET AUTOTRACE ON", issue queries and get execution plan information back.
    For example:
    SQL> set autotrace on
    SQL> select count(*) from it.ticket where ticket_number between 10 and 20;
      COUNT(*)
            11
    Execution Plan
    Plan hash value: 2983758974
    | Id  | Operation         | Name       | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT  |            |     1 |     4 |     1   (0)| 00:00:01 |
    |   1 |  SORT AGGREGATE   |            |     1 |     4 |            |          |
    |*  2 |   INDEX RANGE SCAN| TICKET_N10 |    12 |    48 |     1   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       2 - access("TICKET_NUMBER">=10 AND "TICKET_NUMBER"<=20)
    Statistics
              0  recursive calls
              0  db block gets
              1  consistent gets
              0  physical reads
              0  redo size
            515  bytes sent via SQL*Net to client
            469  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
              1  rows processed
    SQL> This tells me that this query used an INDEX RANGE SCAN on index TICKET_N1; the query can't do much better than that logically... In fact, the statistic "1 consistent gets" tells me that Oracle had to examine only one data block to get the answer, also can't do better than that. the statistic, "0 physical reads" tells me that the 1 data block used was already cached in Oracle's memory.
    the above is from Oracle 10g; autotrace is available back to at least 8i, but they've been adding information to the output with each release.
    If you have questions about sql_plus, check the forums at asktom.oracle.com or http://forums.oracle.com/forums/category.jspa?categoryID=18
    since sql*plus is not a JDBC thing...
    Oh, and sql*plus can also give you easier access to timing information, with "set timing on".

  • Performance Hit After Oracle Database Upgrade to 10.2.0.4

    We have a couple dozen workbooks that took this performance hit after the upgrade of the database/migration to a new server. Worksheets that executed in the ten second range are now running for hours or simply not finishing. We took the new server factor out of the equation by rolling back the database to 10.2.0.3 where a test EUL resides and the problem was resolved. Has anyone seen this issue? Does anyone have an suggestions? An early reply would be greatly appreciated.
    Thanks,
    Jerre

    Rod,
    Thanks for the quick reply. We are looking at the different plans and modifying the optimizer settings, switching back and forth, as we speak. We are now starting with the hints. Currently our Server 'optimizer_mode' parameter is ALL_ROWS. We are planning to change the to 'Choose' and see what happens. The workbooks that are impacted are on our oldest business areas of Finance and HR. The former setup was borrowed from another school for a quick, low cost start up. The latter was thrown together by novices. Our true datamarts developed by knowledgeable personnel with star schemas are not impacted. Of course we are planning on redoing the older business areas but time, personnel and money matters slow things down. It is these workbooks on the older business areas that are greatly affected by the migrations and upgrade. We eventually get things to settle down but past actions do not always have the same resolution with newer and better servers and upgrades.
    Thanks,
    Jerre

  • How can I run two independant LabView applications from the same computer, without taking a performance hit?

    I have two identical, but independant test stations, both feeding data back to a Data Acquisition Computer running LabView 6.1. Everything is duplicated at the computer as well, with two E-series multifunction I/O cards (one for each test station) and two instances of the same LabView program for acquiring and analysing the data. The DAQ computer has a Celeron processor w/ 850Mhz clock and 512MB memory, and is running on Windows NT.
    I have noticed that when I run both the applications simultaneously, I take a substantial performance hit in terms of processing speed (as opposed to running just one program). Why does this happen and how can I prevent it? (In t
    his particular case, it may be possible to combine both the tests into one program since they are identical, but independant, simultaneous control of two different LabView programs is a concept I need to prove out).
    Thanks in advance for any tips, hints and spoon feedings (!)....

    Depending on your application, you may or may not be able to improve things.
    Firstly, each task requires CPU time, so a certain performance difference is guaranteed. Making sure you have a "wait until ms" in every while loop helps in all but the most CPU intensive programs.
    Secondly, if you are
    1) streaming data to disk
    2) Acquiring lots of data over the PCI bus
    3) Sending lots of data o ver the network
    you can have bottlenecks elsewhere than in your program (limited Disk, PCI or Network bandwidth).
    Avoid also displaying data which doesn`t need to be displayed. An array indicator which only shows one element still needs a lot of processing time if the array itself is large.... Best is to set the indicator invisible if this is the case.
    I think
    it would be best if you could give some more information about the amount of data being acquired, processed and sent. Then maybe it will be more obvious where you can optimise things. If you are running W2000, try activating the task manager while the program(s) are sunning to see where the bottleneck is.
    Shane
    Using LV 6.1 and 8.2.1 on W2k (SP4) and WXP (SP2)

  • Bug report: Energy Manager causes slow performance...

    Very slow Windows Startup..
    Energy Manager's plans and the problem occured using any plan as I remember, and it wasn't an issue with the maximum processor state. I had heavy disk usage, especially access to the Page File, which is what the person in the second link I posted also described. CPU usage was low. The disk usage was seemingly not related to Energy Manager (Chrome was usually the worst), but removing Energy Manager completely resolved the problem. If you would find it helpful. I'm quite confident that even with Maximum Processor state at %100, CPU usage was low when it was slow, and that disk access was the culprit.
    previously have use 1.0.0.31 or 32 not sure.
    Windows 8.1
    Im using Lenovo Energy Management Driver
    1.0.0.28
    4.6.2014

    I actually have the same problem.  Sorry to work on a old thread, but...
    I've got my Lenovo B50-70 less than 1 month ago, and had switched it to SSD and had set everything up and is working perfectly.  However there are a bunch of lenovo programs that came pre-installed.  This computer is used mainly for work and I really don't like programs that suddenly pop-up and ask you to do stuff, as it is quite distracting, and with Lenovo Energy Manager being one of them so I removed it.  (I did do some research before I remove...)
    Everything was fine, until when I try to switch on my wifi (I usually have it off, as I use LAN most of the time)...  For the B50-70 there is a airplane button on F7, which allow the user to switch wifi and bluetooth on/off.  That button was unresponsive...  After some digging it is actually related to the power energy manager being removed.  So I've downloaded the latest version from Lenovo and it was working again...
    However though, whenever I start my computer it is taking so long to start...  And I tried uninstalling the power energy manager and it solved the problem...  And I am stuck with either having no wifi on/off ability or a really slow boot, which loses the whole purpose of the SSD.  After you enter your username and password it originally takes around 5~10 seconds to login, now it takes over 1 minute.
    I have found this other link, which says Lenovo Power Manager having the same problem.
    https://forums.lenovo.com/t5/ThinkVantage-Technologies/Bug-report-Power-Manager-causes-slow-performa...
    There is a solution, which you can change the registry to switch a function off.  I've tried it and it didn't work, which is not surprising as it's a different program..  But I am hoping that Lenovo can provide a similar solution for Lenovo Energy Manager...

  • Performance hit when using a FirePro GPU?

    Hey there!
    I'm interested in purchasing a cheap ATI FirePro V4900 or similar GPU in order to take advantage of my 10 bit TFT when using Photoshop. I was wondering if I have to expect a performance hit when using PS 6.0 with such a card as opposed to a normal "gamer" GPU like the NVidia 670 GTX when :
    - performing normal file handling, opening PSD files, paning, zooming, brushing, etc.
    - applying some of the newer GPU-enhanced filters like Liquify, Oil Paint, Iris Blur, 3d enhancements, etc.
    - actually applying/rendering a more demanding fliter such as Iris Blur?
    Does anyone know about this? I'm afraid I could not find any benchmarks at all except for the one on Tomshardware regarding OCL, but that one does not include professional GPUs...
    Thanks for any info in advance!

    There will not be synchronization when the method of A
    is being called. The method 1) certainly saves memory
    space, but will the performance be hurt since there
    will only be one object accessed by multiple threads?
    Or maybe it doesn't matter?If there is no synchronization, it will not matter. Threads execute methods. Methods do not run on Objects. The Object is just data that is implicitly linked to the method.
    Just make sure it's safe to keep the method unsynchronized.

  • Which more performance hit: trip to DB, traverse XML DOM in memory?

    Which is more of a performance hit:
         1. Make trip to DB, get Data, come back
              a. Includes Make trip to DB, alter smth, come back
         2. Use DOM Parser to get data from XML (XML is in Memory)
              a. Includes altering the DOM Document that is in memory

    The trip to the DB depends on network latency and sometimes (but not often) network speed and of course the complexity of doing the update on the DB (how many indices chang, what trigggers fire, what integrity constraints must be validated, etc.).
    The XML parsing is generally done once and from then on you manipulate the tree directly - that latter part is efficient. Of course the cost depends on your technique, the hit taken to parse the XML (which parser? validating?).
    The first part will probably consume less CPU on the host running the Java program (assuming DB is remote) but take more time total.
    The latter may (or may not) consume more local CPU but should be faster.
    Your mileage may vary. Test it.
    Chuck.

  • How bad is the performance hit with RTMPT?

    In a conversation with engineers recently at a CDN it was suggested to me that streaming all video over RTMPT only was a viable solution to the barriers posed by firewalls and proxies blocking port 1935.  They indicated that they had seen no signifigant performance degradation with tunneling and that many of their clients were making the switch from "rollover" type connection models to connection via RTMPT only.
    This ran counter to my notion of the process.  I have always thought the packet overhead was signifigant.
    Is it?  How bad is the performance hit for streaming live h.264 video?

    Hmmm...  That's quite a hit.
    I'm trying to determine the best strategy for reaching the most people with the least performance hit.  A guy lays out his strategy here:
    http://www.kensodev.com/2010/02/19/rtmp-being-blocked-by-firewalls-flash-media-server/
    He basically says ditch 1935.  Never use it.  Always use 80.  Like this:
    rtmp://your_ip_address:80/app_name
    If that fails, do this:
    rtmpt://your_ip_address:80/app_name
    Does that seem valid?  Does that first option avoid the perfomance hit of tunnelling while getting you more connections?  If so, it makes me think there is no benefit at all to connecting via 1935.

  • RLDP enabled on WLC causes client drops?

    The release notes for WLC code 4.0.206.0 states, "Enabling RLDP may cause access points connected to the controller to lose connectivity with their
    clients for up to 30 seconds." Does anyone have more information about this? I don't like the word "may" in there. I need to enable RLDP but I can't afford to have clients dropping. Is this something that will work differently in a new release?

    There's no may about it. If clients are on an AP that decides to go rogue-hunting they will be told to "get out of the pool" (de-authenticated) to facilitate their going elsewhere. The same applies to DCA - Dynamic Channel Allocation - part of Auto-RF. If you don't have AP density, clients that use fast roaming, or tolerant apps, you would be best served to turn these features off (or On-Damand).

  • BPC 7.5 AfterRange Causes Extreme Performance Issue

    In a report when I apply second AfterRange the run time goes from under 2 minutes to 15 minutes. I know there are performance penalties for using Before/AfterRange but I have never seen the discrepancy to this proportion.
    Report contains 1 Column (Time) and 4 Rows in Expansion. The first range which causes no issues for Cost Centers is defined Q73:V74, the range which causes the performance issue and applies for each Employee is defined as Q71:V72
    To test this I have two versions of essentially the same report, one with only Cost Center AfterRange and this version runs in less than 2 minutes, then I have the other version in which I have added one more AfterRange to another column and that's when the hell breaks loose and this version of the report runs in 15 minutes. The number of records is roughly 2000, and I realize that the range causing the issue is applied on pretty much every rcord in result set but still I'd not expect the run time to increase by this much.
    BPC Experts, ideas on what could be causing this issue, any help would be highly appreciated. thanks!

    Hi
    1> Please try to restart your Gernal Server
    2> Check after optimization that helps
    Regards
    Vinoo

  • Is there a big performance difference between HD's

    Hi All
      Is there a big performance difference between a 5400 & 7200 Rpm hardrive? I'm asking because I want to pick up a new Imac and am stuck between choosing the higher end 21.5" and lower end 27" and am trying to determine if the difference in price is worth the extra sheckles.

    I was wanting to know the same question (5400rpm vs. 7200rpm) in a 15" standard MBP, deciding on whether to get a 1TB serial ATA drive @ 5400rpm vs. a 750GB serial ATA drive @ 7200rpm. (Sorry to jump in)
    For the most part, I'm a general user - web surfing/research, Word processing/Excel/Powerpoint (pretty basic), etc. BUT I do like to take alot of photos and plan on doing some editing on them (nothing advanced) and creating slideshows with my photos/music (ex. my Europe trip of photos or a slideshow of the grandchildren as a gift to my parents, etc.)
    Some folks mention "video editing" in reference to gonig with the faster speed (7200rpm) if that's what you plan on doing. But, what do they exactly mean by "video editing"? Is slideshow creation the same? 
    Just wondering for my needs, whether I should go with the 750GB serial ATA drive @ 7200 rpm or the 1TB serial ATA drive @ 5400rpm ($50 more yet with more storage space which would help with my increasing photo files every year).
    Thanks

Maybe you are looking for

  • How to find the user who did the personalization?

    Hi All, Is there a way to find out which user has done personalizations in a given OA page? For e.g. if a user has done personalization for hiding a text bean in a page, I want to find out that user's name by passing the page name to a script? Even i

  • White lines appear in PDF

    I have a user that is having white lines appear around different objects in PDFs when she attaches them into Mail. Anyone else having this problem or have a solution. I can't find anything searching it.

  • Praviliuon dv7 hard drive; need to transfer entire contents to new hard drive

    My hard drive is on it last leg and I would like to transfer the entire contents to a new drive to invlude my operating system.

  • How to bind register in Popup/Dialog to be filtered by ID

    Hello all, we use Oracle JDeveloper 11.1.2.3.0 and I'm relatively new to this development tool. We currently open a dialog from a table (bound to DataControl). The dialog is opened with a common ShowPopupBehavior on a commandImageLink and already sho

  • Dashes convert to single line

    Hello, On my iMac OS X 10.9.4, i am unable to create in TextEdit or Pages guitar tabs in the Courier font, like i did for years, by simply keeping my finger pressed on the dash key: all my dashes attach themselves together to form a continuous line a