Performance tuning in CRM server

Hi everyone,
   My CRM server working very slow. Idont know the exact reason. My server configuration IBM Xeon processor 3 Ghz 3GB RAM.There are about 25 users. Is this problem with SAP or Hardware.If any one have performance tuning document plz post it.
Thanks,
Murali

Chech the following:-
Processor usage
Ram usage
background jobs
how many user at a time
what kind of job they do
Any Network Problem
Increse Work Processor
increase Virtual Ram
And for 25 users, You need System with 2 CPU and 4 Gb Ram
Regards

Similar Messages

  • Performance Tuning in CRM

    Hi All,
    Can any one help me to important OSS Notes in CRM performance tuning, CRMIC webclient performance, web client channel performance.
    Thanks & Regards,
    Sandeep

    Hi Sandeep,
    Please take a look at the following SAP NOTES for performance tuning in CRM 2007 Webclient:
    Note 1048388 - General Performance improvements of BSP transactions
    Note 1228076 - CRM Web UI: Frontend performance DDLB and input changes
    Note 1246144 - Advanced Search right side hidden when resolution: 1024X768
    Note 1242599 - IE7: Message # shines through the "More" dropbox
    Note 1240769 - Native DDLB value not set when its disabled
    Note 1237437 - Config Mode: FireFox support and Visual enhancements
    Note 1230443 - Scrollbar missing or not showing up on some CRM views
    Best Regards,
    Gabriel

  • Need clear steps for doing performance tuning on SQL Server 2008 R2 (DB Engine, Reporting Services and Integration Services)

    We have to inverstigate about a reporting solution where things are getting slow (may be material, database design, network matters).
    I have red a lot in MSDN and some books about performance tuning on SQL Server 2008 R2 (or other) but frankly, I feel a little lost in all that stuff
    I'am looking for practical steps in order to do the tuning. Someone had like a recipe for that : a success story...
    My (brain storm) Methodology should follow these steps:
     Resource bottlenecks: CPU, memory, and I/O bottlenecks
     tempdb bottlenecks
     A slow-running user query : Missing indexes, statistics,...
     Use performance counters : there are many, can one give us the list of the most important
    how to do fine tuning about SQL Server configuration
    SSRS, SSIS configuration ? 
    And do the recommandations.
    Thanks
    "there is no Royal Road to Mathematics, in other words, that I have only a very small head and must live with it..."
    Edsger W. Dijkstra

    Hello,
    There is no clear defined step which can be categorized as step by step to performance tuning.Your first goal is to find out cause or drill down to factor causing slowness of SQL server it can be poorly written query ,missing indexes,outdated stats.RAM crunch
    CPU crunch so on and so forth.
    I generally refer to below doc for SQL server tuning
    http://technet.microsoft.com/en-us/library/dd672789(v=sql.100).aspx
    For SSIS tuning i refer below doc.
    http://technet.microsoft.com/library/Cc966529#ECAA
    http://msdn.microsoft.com/en-us/library/ms137622(v=sql.105).aspx
    When I face issue i generally look at wait stats ,wait stats give you idea about on what resource query was waiting.
    --By Jonathan KehayiasSELECT TOP 10
    wait_type ,
    max_wait_time_ms wait_time_ms ,
    signal_wait_time_ms ,
    wait_time_ms - signal_wait_time_ms AS resource_wait_time_ms ,
    100.0 * wait_time_ms / SUM(wait_time_ms) OVER ( )
    AS percent_total_waits ,
    100.0 * signal_wait_time_ms / SUM(signal_wait_time_ms) OVER ( )
    AS percent_total_signal_waits ,
    100.0 * ( wait_time_ms - signal_wait_time_ms )
    / SUM(wait_time_ms) OVER ( ) AS percent_total_resource_waits
    FROM sys.dm_os_wait_stats
    WHERE wait_time_ms > 0 -- remove zero wait_time
    AND wait_type NOT IN -- filter out additional irrelevant waits
    ( 'SLEEP_TASK', 'BROKER_TASK_STOP', 'BROKER_TO_FLUSH',
    'SQLTRACE_BUFFER_FLUSH','CLR_AUTO_EVENT', 'CLR_MANUAL_EVENT',
    'LAZYWRITER_SLEEP', 'SLEEP_SYSTEMTASK', 'SLEEP_BPOOL_FLUSH',
    'BROKER_EVENTHANDLER', 'XE_DISPATCHER_WAIT', 'FT_IFTSHC_MUTEX',
    'CHECKPOINT_QUEUE', 'FT_IFTS_SCHEDULER_IDLE_WAIT',
    'BROKER_TRANSMITTER', 'FT_IFTSHC_MUTEX', 'KSOURCE_WAKEUP',
    'LAZYWRITER_SLEEP', 'LOGMGR_QUEUE', 'ONDEMAND_TASK_QUEUE',
    'REQUEST_FOR_DEADLOCK_SEARCH', 'XE_TIMER_EVENT', 'BAD_PAGE_PROCESS',
    'DBMIRROR_EVENTS_QUEUE', 'BROKER_RECEIVE_WAITFOR',
    'PREEMPTIVE_OS_GETPROCADDRESS', 'PREEMPTIVE_OS_AUTHENTICATIONOPS',
    'WAITFOR', 'DISPATCHER_QUEUE_SEMAPHORE', 'XE_DISPATCHER_JOIN',
    'RESOURCE_QUEUE' )
    ORDER BY wait_time_ms DESC
    use below link to analyze wait stats
    http://www.sqlskills.com/blogs/paul/wait-statistics-or-please-tell-me-where-it-hurts/
    HTH
    PS: for reporting services you can post in SSRS forum
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • How to performance tuninig  in weblogic server and AM module.

    Hi,
    I am using Jdev11.1.1.2.0 and weblogic server 10.3.0.My Application is working fine but i want to better performance in my application.
    Single user accessing its fast but more than 4-5 user accesing my application its very slow .I have increase the memory in setDomainEnv.cmd.
    My operaitng system is windows server2003 and RAM is 5GB .
    So where i should to increased the performance tuning in weblogic server or In AM Module any configuration is needed .
    Please help me.
    Thanks
    Anup

    hi Anup, May be helpful for you.
    ADF BC Tuning
    http://andrejusb.blogspot.com/2011/11/stress-testing-oracle-adf-bc_16.html
    http://andrejusb.blogspot.com/2011/10/adf-bc-tuning-with-do-connection.html
    http://www.avromroyfaderman.com/2008/11/adf-bc-tuning-iii-view-objects-part-1/
    http://andrejusb.blogspot.com/2010/02/optimizing-oracle-adf-application-pool.html
    VO tunning
    www.gebs.ro/blog/oracle/adf-view-object-performance-tuning-analysis
    http://www.sagecomputing.com.au/presentations_sage_computing_services.html // pls Download power point title JDeveloper ADF and the oracle database - Friends not Foes
    Weblogic
    http://developmentips.blogspot.com/2009/04/avoid-java-out-of-memory-with-weblogic.html
    Regards
    KT

  • SAP CRM Performance tuning

    From where can I get information or documents how to do performance tuning in SAP CRM?

    Hi, Performance Tuning a SAP system is unique in every environment, also there are many area's in SAP to tune even area's outside SAP, eg... your Operating System/Network/Client Machines/Virtual Landscape/Cloud.
    I recommend that you engage in a performance testing execise which will identify what area of your SAP Landscape needs tuning, first you need to identify the bottleneck before you can tune it.
    The way the question was posed is exactly the same as wanting to fix a car but you have no idea which part of the car to fix...
    *LoadRunner is an excellent tool to simulate virtual work load on an SAP environment*
    Regards
    Sodick25
    http://www.2createawebsite.co.za
    Edited by: Sodick25 on Aug 15, 2011 11:58 AM
    Edited by: Sodick25 on Aug 15, 2011 11:59 AM

  • Performance tuning content server

    Hi,
    I would like to profile my UCM based application. I am using Oracle content server 10g version and has custom java components for my business functionality. I would like to profile this application to identify performance bottlenecks. Is there any profiling tool in market that would get plugged /integrated with oracle content server (UCM 10g)?
    What are the general tips for performance tuning custom components?
    Thanks in advance
    Siva

    Some information can be found in the following document on Metalink :
    Use VisualVM Tools to Troubleshoot UCM and Observe Performance Problems Occurring in Java Virtual Machine (Doc ID 950621.1)

  • Performance Tuning Guidelines for Windows Server 2008 R2 mistake?

    Hi all,
    I'm reading the "Performance Tuning Guidelines for Windows Server 2008 R2" and I think there is same kind of error. At page 53 there is a sub chapter "I/O Priorities" that explain how to handle I/O priorities.
    In the guide it is said that in the registry I have to have a key:
    HKEY_LOCAL_MACHINE\System\CurrentControlSet\Control\DeviceClasses\{Device GUID}\DeviceParameters\Classpnp\
    Where {Device GUID} I think it is the value I get in device manager under one of my disks -> properties -> details -> "Device Class GUID"
    the trouble is that under the key HKEY_LOCAL_MACHINE\System\CurrentControlSet\Control\DeviceClasses\ I've no subkey matching my disk's device class guid! Maybe I'm drunk but I've checked every sub key under HKEY_LOCAL_MACHINE\System\CurrentControlSet\Control\DeviceClasses\
    but neither of them as a subkey named "DeviceParameter"!!
    So am I wrong? 

    Hi Andrea,
    To find the "DeviceParameter", please try to follow the steps below:
    1.Find this registry key and note the DeviceInstance value:
    HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\DeviceClasses\
    2.Find the device instance registry key under "HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Enum" which contains information about the devices on the system and get the device interface GUID:
    HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Enum\USB\<hardware id>\<instance id>\Device Parameters
    I hope this helps.

  • Performance tuning for siebel CRM application on oracle database

    Hi,
    Please send me the link for Performance tuning for siebel CRM application on oracle database. If there are any white papers please send me the link.
    Thanks,
    Rajesh

    Hi,
    This metalink document is very useful, if you have any other documents or links please inform me.
    Thanks once again
    Rajesh

  • IIR 2.8.0.7 performance tuning suggestions / documents for Oracle 11?

    Would we have any hints or white papers that would support a customer in IIR matching server tuning for initial load performance,
    beyond the Siebel specific
    Performance Tuning Guidelines for Siebel CRM Application on Oracle Database (Doc ID 781927.1)
    which does NOT generate any statistics on the Informatica Schema?
    Customer is starting production data load into Siebel UCM of over 5 million customer records . Their current bottle neck seems to be IIR queries and high IIR host resources usage.
    This would be for Siebel 8.1.1.4 [21225] on Oracle 11.1.0.6; I currently do not know if ACR 475 with its EBC access is used or not; I'd be looking for any performance tuning suggestions on the Informatica & database side - I have not found anything helpful in the informatica knowledgebase and in the 2.8.0.7 docs the only performance tuning suggestions seem to be on IDT loads.
    Obviously performance can be influenced by the matching fields used, but this customer will be looking for performance tuning suggestions keeping the matching fields and thresholds they got approved from the business side.
    Any documents we could share?

    Hi Varad,
    Well, Oracle Metalink actually got it fixed for me! Got to give them credit.
    I ran the 1st bit of SQL they sent me and it fixed it so I think it was only cosmetic. I've enclosed all of their response in case others may find themselves in similar situation.
    Thanks again for YOUR help too.
    Perry
    ==========================================================
    1. Do you see any invalid objects in the database?
    2. Please update the SR with the output of running the commands below
    1.Connect as SYS.
    set serveroutput on (This will let you see if any check fails)
    exec validate_apex; (If your version is 2.0 or above)
    If all are fine, the registry should be changed to VALID for APEX and you should see the following
    message:
    PL/SQL procedure successfully completed.
    Note : The procedure validate_apex should always complete successfully.
    With serveroutput on, it should show why Application Express is listed as invalid.It should show output like:
    FAILED CHECK FOR <<obect_name>>
    Or
    FAILED EXISTENCE CHECK FOR <<object_name>>

  • Data source migration in parallel to CRM server upgradation

    Hi Experts,
    We are working on BI 7.0 system which pulls data from CRM 2004. Recently we upgraded BI environment from BW3.5 (to BI.70 SP -16).
    Now we are re-modeling the existing BI design due to performance issues & to use BI functionalities. We would like to migrate the data sources in BI also to BI version so that we can get advantage of using DTP & transformations.
    But the issue here is now we started to up grade our source system also from CRM 2004 --> CRM 2007. I beleive that the data source version will be maintained in source system (CRM) tables also.
    So if we migrate the data source now (before CRM upgrade), will it cause any issues in BI once CRM server upgrade done?
    Note: We can implement all of our new designs with out data source migration.
    Please advice us to proceed. Let me know if you need more details.
    Thanks in advance.
    Regards,
    Arun Thangaraj.

    Hi reddy,
    1). Why Migration Process?
    Migration process is to convert BW 3.x objects ( such as Data source, Trasfer rules and update rules etc) into BI 7.0 objests.
    2). Why we are Migrating BW 3.X objects into 7.0 objects?
    We are Migrating 3.x data source into 7.0 inorder to avail three functions, Those are
    a). Direct data acquisition
    b). Realtime Data Acquisition (RDA)
    c). BI Accelarator.
    3) .Regarding With export and Withour export options ?
    These two options come in the process of Migration step. Regarding these options clearly described in the fourth point as bellow.
    4). What will happen in this Migration process?
    ( Remember at the time Migration Transfer structure and Transfer rlues deleted, keep in mind go throgh bellow steps).
    What ever may be the option selected (i.e, With Export or Without export ) data source will migrate to 7.0.
    But if you choose With export option deleted Transfer Structure and Trasfer Rules will store in the table " RSDSEXPORT". If you select other option (i.e, Without export ) the deleted Trasfer structure and trasfer rules perminently not avaliable.
    Thats why always go with With Export option only. This will help to revert back 7.0 data source into 3.5 data source when ever is there any requirement in future.
    Please go look into bellow link.
    Re: Migration of datasource from 3.5 to 7.0
    if u r satisfed about my answer please assign points.
    Thanks & Regards,
    Kotesh.

  • Performance tuning in PL/SQL code

    Hi,
    I am working on already existing PL/SQL code which is written by someone else on validation and conversion of data from a temporary table to base table. It usually has 3.5 million rows. and the procedure takes arount 2.5 - 3 hrs to complete.
    Can I enhance the PL/SQL code for better performance ? or, is this OK to take so long to process these many rows?
    Thanks!
    Yogini

    Can I enhance the PL/SQL code for better performance ? Probably you can enhance it.
    or, is this OK to take so long to process these many rows? It should take a few minutes, not several hours.
    But please provide some more details like your database version etc.
    I suggest to TRACE the session that executes the PL/SQL code, with WAIT events, so you'll see where and on what time is spent, you'll identify your 'problem statements very quickly' (after you or your DBA have TKPROF'ed the trace file).
    SQL> alter session set events '10046 trace name context forever, level 12';
    SQL> execute your PL/SQL code here
    SQL> exitWill give you a .trc file in your udump directory on the server.
    http://www.oracle-base.com/articles/10g/SQLTrace10046TrcsessAndTkprof10g.php
    Also this informative thread can give you more ideas:
    HOW TO: Post a SQL statement tuning request - template posting
    as well as doing a search on 10046 at AskTom, http://asktom.oracle.com will give you more examples.
    and reading Oracle's Performance Tuning Guide: http://www.oracle.com/pls/db102/to_toc?pathname=server.102%2Fb14211%2Ftoc.htm&remark=portal+%28Getting+Started%29

  • Performance tuning in EP6 SP14

    Hi,
    We just migrated our development EP5 SP6 portal to NW04 EP6 SP14 and are seeing some performance problems with a limited number of users (about 3). 
    Please point me to some good clear documentation related to performance tuning.  Better yet, please tell me some things you have done in your EP6 portal to improve performance.
    Any help is greatly appreciated.
    Regards,
    Rick

    Hi
    In general we have experienced that NW04 is faster than earlier versions.
    One of the big questions is if the response time is SERVER time (used for processing on the J2EE server), NETWORK time (latency and many roundtrips), CLIENT time (rendering, virus-scanning etc) or a combination.
    1) Is the capacity on the server ok?  CPU utilization and queue lenght high?  Memory swapping?
    2) A quick optimazation server-side is logging: Plese verify that that log and trace levels are ERROR /FATAL or NONE on J2EE and also avoid logging on IIS and IIS proxy is used
    3) Using a HTTP trace you can see if the NW portal for some reason generetes more roundtrips (more GETS) that the old portal for the same content - what is the network latency?  Try to do a "ping <server>" from a client pc and see the time (latency) - if it is below 20 ms the network should do a big difference.
    4) On the client try to call the portal with anti-virus de-activated if the delta/difference in response times are HUGE your client could be to old? (don't underestimate the client). Maybee the compression should be set different to avoid compres (server) and uncompres (client) - this is a tradeoff with network latency however.
    Also client-side caching (in browser) is important.
    These are just QUICK point to consider - tuning is complex
    BR
    Tom Bo

  • Performance Tuning in IR

    Hello All,
    We have created some reports using Interactive Reporting Studio. The volume of data in that Oracle database are huge and in some tables of the relational database are having above 3-4 crores rows individually. We have created the .oce connection file using the 'Oracle Net' option. Oracle client ver is 10g. We earlier created pivot, chart and report in those .bqy files but had to delete those where-ever possible to decrease the processing time for getting those report generated.
    But deleting those from the file and retaining just the result section (the bare minimum part of the file) even not yet helped us out solving the performance issue fully. Still now, in some reports, system gives error message 'Out of Memory' at the time of processing those reports. The memory of the client PCs,wherefrom the reports are being generated are 1 - 1.5 GB. For some reports, even it takes 1-2 hours for saving the results after process. In some cases, the PCs gets hanged at the time of processing. When we extract the query of those reports in sql and run them in TOAD/SQL PLUS, they take not so much time like IR.
    Would you please help us out in the aforesaid issue ASAP? Please share your views/tips/suggestions etc in respect of performance tuning for IR. All reply would be highly appreciated.
    Regards,
    Raj

    SQL + & Toad are tools that send SQL and spool results; IR is a tool that sends a request to the database to run SQL and then fiddles with the results before the user is even told data has been received. You need to minimize the time spent by IR manipulating results into objects the user isn't even asking for.
    When a request is made to the database, Hyperion will wait until all of the results have been received. Once ALL of the results have been received, then IR will make multiple passes to apply sorts, filters and computed items existing in the results section. For some unknown reason, those three steps are performed more inefficiently then they would be performed in a table section. Only after all of the computed items have been calculated, all filters applied and all sorts sorted, then IR will start to calculate any reports, charts and pivots. After all that is done, the report stops processing and the data has been "returned"
    To increase performance, you need to fine tune your IR Services and your BQY docs. Replicate your DAS on your server - it can only transfer 2g before it dies, restarts and your requested document hangs. You can replicated the DAS multiple times and should do so to make sure there are enough resources available for any concurrent users to make necessary requests and have data delivered to them.
    To tune your bqy documents...
    1) Your Results section MUST be free of any sorts, filters, or computed items. Create a staging table and put any sorts or local filters there. Move as many of your computed items to your database request line and ask the database to make the calculation (either directly or through stored procedures) so you are not at the mercy of the client machine. Any computed items that cannot be moved to the request line, need to be put on your new staging table.
    2) Ask the users to choose filters. Programmatically build dynamic filters based on what the user is looking for. The goal is to cast a net only as big as the user needs so you are not bringing back unnecessary data. Otherwise, you will bring your server and client machines to a grinding halt.
    3) Halt any report pagination. Built your reports from their own tables and put a dummy filter on the table that forces 0 rows in the table until the report is invoked. Hyperion will paginate every report BEFORE it even tells the user it has results so this will prevent the user from waiting an hour while 1000s of pages are paginated across multiple reports
    4) Halt any object rendering until request. Same as above - create a system programmically for the user to tell the bqy what they want so they are not waiting forever for a pivot and 2 reports to compile and paginate when they want just a chart.
    5) Saved compressed documents
    6) Unless this document can be run as a job, there should be NO results stored with the document but if you do save results with the document, store the calculations too so you at least don't have to wait for them to pass again.
    7) Remove all duplicate images and keep the image file size small.
    Hope this helps!
    PS: I forgot to mention - aside from results sections, in documents where the results are NOT saved, additional table sections take up very, very, very small bits of file size and, as long as there are not excessively larger images the same is true for Reports, Pivots and Charts. Additionally, the impact of file size only matters when the user is requesting the document. The file size is never an issue when the user is processing the report because it has already been delivered to them and cached (in workspace and in the web client)
    Edited by: user10899957 on Feb 10, 2009 6:07 AM

  • Performance tuning in progress

    Due to performance issues caused by high traffic, if you see
    a blank page, just go to Create to start using kuler. Performance
    tuning is in progress.

    Hello, I'm the lead developer on kuler. Over the weekend we
    added a new database server as well as made many performance
    enhancements to the site. The kuler site is a lot more stable now
    and you should not experience the slowdowns previously encountered.
    Thanks,
    Tim Strickland

  • Performance tuning in ODI

    Hi,
    Can any help how to do performance tuning in ODI. What are the procedure for performance tuning. What needs to checked for performace tuning. Is there any document for ODI performace tuning.
    Thanks,
    Gnanamanju

    Hi Gnanamanju,
    Let me try to contribute a little,
    ODI performance strategy tells,
    1. Its recommended to run the AGENT in the target host rather than source. In case if your target is a remote host.
    2.Selecting appropriate KMs to load data, for example, if its a simple file to oracle transfer, its recommended to use native DB utilities like SQLLDR, External Table and so on.
    3. In case, if your staging and target are on the same server then make sure in Physical architecture your selected the correct schema (Data Schema and Work Schema)
    4. Right area of execution of joins and filters. If your source has more records then use the filter condition to be executed on SOURCE rather staging.
    5. Increase the Array Fetch Size and Batch Fectch Size of your Data server ( so that agent can accordingly fetch/insert the data)
    And more....
    Thanks,
    Guru

Maybe you are looking for