Improve Portal performance with BI-Java by adding additional server node.

Hi SAP Expert
I wonder anyone might have attempted this before,
In an environment with BI-Java and Portal Java stack installed on the same server, is it possible to improve the system performance by adding additional portal server node? or is there any improvement seen by adding the extra server node?
I am after suggestion and experience on how the portal performance can be improve, with consideration of BI-Java sharing Portal resources.
Any comment will be most appreciated.

Hi Jim,
We've this configuration at our site, Portal and BI running together (not federated).  Recommendations would be to:
1. Set a sensibly large max heap size on each server node of Portal (at least 2GB if not larger) and implement several nodes at least.  For example we have 4 x physical nodes, each running 3 server nodes of 2GB max heap size apiece so a total of 12 nodes.
2. Implement the BI Safety Belt for large results sets (1127156)
3. Implement the latest JVM / JDK you can on your environment.  We found much improved performance after implementing JDK SR10 with the J9/2.3 options enabled e.g.,  -Xjvm:j9vm23
-Xsoftrefthreshold0
4. Patch your BI components on the Portal (BIBASES, BIWEBAPP etc) to the latest available patch level for your Portal SP level
5. Make sure all your RFC connections are load balanced to the backend and tuned for the kind of load you expect on your reports, and the BI is sized appropriately in terms of app servers and dialog work processes, RFC-enabled login groups in SMLG etc.
6. Consult notes 1048691 (especially useful, opions &PROFILING=X&TRACE=X added to reports for example for performance tracing), 937697, 1021921, 948158 for information about problem analysis for this scenario
7. Implement SAP Web Dispatcher for Portal to reduce load on static mime files
8. Tune the Portal application.  There's plenty information on SDN related to Portal performance tuning.
I hope this helps!
Cheers,
Marc

Similar Messages

  • Adding additional j2ee node

    Hello,
    we are running BI 7 - abap+java stacks with 16 GB ram (recently added)
    i would like to know what is the thumb rule for adding additional j2ee node?
    when is it neccassary to add additional j2ee node?
    Regards,
    Moshe

    Hi,
    The JLaunch starts a Java program. It loads the JVM into its own address space and then represents the required cluster element. The program can receive notification from the JControl process via named pipes to stop the cluster element, and terminates, if the JControl stops running (fork emulation under Windows).
    So Jlaunch consuming high CPU time means that server, dispatcher or bootstrap process consume it. If there are not enough system or application threads avaliable means that the engine must buffering requests for web application or system processes. Things to check in this case are:
    SAP NOTE 1016241 - J2EE Engine hangs because of unsufficient system threads
    SAP note: 716604  and upgrade JDK.
    SAP note: 723909 -> set correct VM parameters
    and if issue persists, then you will need to analyse the application causing teh leak...
    SAP note: 1259465: Memory analyser.
    Regards,
    Hemanth

  • Improving Portal Performance

    Hi,
    We are looking at improving our Portal's performance.  At the moment, it seems like we can achieve this by adding either Java Instance or Server Instance.  I have read that if there are frequent access to the Portal, it would be best that additional server instances/processes can be added.  I have followed the guidelines in help.sap.com to add server instances.  However, I am not able to find any documentation which recommends/advices on how many server instances a server needs in order to achieve maximum performance.  Would it be as simple as 1 server process per processor?  Hence, assuming that I have a 4-CPU server, I should at least have 4 server processes?
    Aside from looking at adding Java instances and Server instances, we are also increasing the heap size of each local disks. 
    Any other pointers to which we can further improve the performance of the server are welcomed.  Please excuse my misuse of terms and poor-understanding of concepts.  Please feel free to correct as needed.  If there is any other posts or boards I can refer to, please feel free to redirect.
    Thank You.

    Hi,
    Note that a java process cannot as far as I know use more than 1 CPU. Therefore, if you have 4 CPUs and only two java process, you basically have to idle CPUs.
    Memory is also important. Normally, java cannot address more than 2 GB of memory, and you don't really want it to, because the full garbage collection that will eventually occur, will block the entire JVM for many seconds (the more memory, the more time).
    Regards
    Dagfinn

  • Improve Indesign performance with a huge number of links?

    Hi all,
    I am working on a poster infographic with a huge number of links, specifically around 4500. I am looking to use indesign over illustrator for the object styles (vs graphic styles in illustrator) and for the interactive capacity.
    The issue I am having is indesign's performance with this many links. My computer is not maxed out on resources when indesign is going full power, but indesign is still very slow.
    So far, here are the things I have tried:
    Display performance to fast
    Switching from link AI files to SVGs
    Turning off preflight
    Turning off live-draw
    Turning off save preview
    Please let me know if you have any suggestions on how to speed up indesign! See below system specs
    Lenovo w520
    8GB DDR3 @1333mhz
    nVidia 2000M, 2GB GDDR
    Intel Core i7 2760QM @ 2.4GHz
    240GB Samsung 840 SSD
    Adobe CS6
    Windows 8
    The only other thing I can think to try is to break up the poster into multiple pages/docs and then combine it later, but this is not ideal. Thank you all for your time.
    Cheers,
    Dylan Halpern

    I am not a systems expert, but I wonder if you were to hide the links, and keep InDesign from accessing them that it might help. Truly just guessing.
    Package the file so all the graphics are in a single folder. Then set the File Handling Preferences so InDesign doesn't yell at you when it can't find the links.
    Then quit InDesign and move the folder to a new place. Then reopen InDesign. The preview of the graphics will suck, but it might help. And one more thing, close the Links panel.

  • How to improve Query performance on large table in MS SQL Server 2008 R2

    I have a table with 20 million records. What is the best option to improve query performance on this table. Is partitioning the table into filegroups  is a best option or splitting the table into multiple smaller tables? 

    Hi bala197164,
    First, I want to inform that both to partition the table into filegroups and split the table into multiple smaller tables can improve the table query performance, and they are fit for different situation. For example, our table have one hundred columns and
    some columns are not related to this table object directly (for example, there is a table named userinfo to store user information, it has columns address_street, address_zip,address_ province columns, at this time, we can create a new table named as Address,
    and add a foreign key in userinfo table references Address table), under this situation, by splitting a large table into smaller, individual tables, queries that access only a fraction of the data can run faster because there is less data to scan. Another
    situation is our table records can be grouped easily, for example, there is a column named year to store information about product release date, at this time, we can partition the table into filegroups to improve the query performance. Usually, we perform
    both of methods together. Additionally, we can add index to table to improve the query performance. For more detail information, please refer to the following document:
    Partitioning:
    http://msdn.microsoft.com/en-us/library/ms178148.aspx
    CREATE INDEX (Transact-SQL):
    http://msdn.microsoft.com/en-us/library/ms188783.aspx
    TechNet
    Subscriber Support 
    If you are
    TechNet Subscription user and have any feedback on our support quality, please send your feedback
    here.
    Allen Li
    TechNet Community Support

  • How can we improve query performance with out indexes?

    Hello Experts,
    I have a problem with table(calc) which contain 3 crore records and table doesn't have index, on that table one of the view(View_A) created.
    When i use the view in  below query SELECT count(*)
    FROM
      Table A INNER JOIN Table B ON (A.a=B.b)
       LEFT OUTER JOIN View_A ON ( Table A.a=View_A.a)
       LEFT OUTER JOIN View_B ON (Table A.a=View_B.a)
    In above query View_A is causing the problem, view_A is created on Calc table. One more thing when i execute select statement on the view it was running fine.
    With out View_A query fetching data fine. Update stats of the table also fine. When i run cost plan for scanning only cost % is 90.
    Can any help me please?.
    Thank you all.
    Regards,
    Jason.

    Jason,
    Not sure what you are trying to do. But outer joins are bad for performance. So try to avoid them.
    You also say that you have a view on a calc table. What are you calculating? Are you using user defined functions maybe?
    Regards,
    Nico

  • Improving SSD performance with this??

    OK ..in researching SSD for my Mac Pro that is on its way ...I found this:
    http://www.silverstonetek.com/products/p_spec.php?pno=HDDBOOST&area=usa
    Has anyone tried one of these things with a Mac Pro ? And even if it has not been tried ....what do you think of the concept ....would it really be of advantage?

    Hi,
    I didn't try this one but I've install OCZ SSD in my MAC PRO 8
    and what I can say is simply monster.
    System load less than 8 sec.
    Cheers.

  • Adding new server node in EP

    HI All,
    We need to add additional one server node to our existing EP server. Server is having total of 4GB ram and CPU ENT0.50
    vmstat 5 5
    System configuration: lcpu=4 mem=4096MB ent=0.50
    Is there any calculation /thumb role  of how much resource  required in order to add a new server node.
    Thanks
    -Tom

    Hi,
    Check the below links as well... hope fully u will get the good information there.
    How many server nodes can you create per instance
    http://santy.wdfiles.com/local--files/web-dynpro/logon_page_branding.pdf
    -Srini

  • Performance with LinkedList in java

    Hello All,
    Please suggest me any solution to further improve the performance with List.
    The problem description is as follows:
    I have a huge number of objects, lets say 10,000 , and need to store the objects in such a collection so that if i want to store an object at some particular index i , I get the best performance.
    I suppose as I need indexed based access, using List makes the best sense as Lists are ordered.
    Using LinkedList over ArrayList gives the better performance in the aforementioned context.
    Is there are way I can further improve the performance of LinkedList while solving this particular problem
    OR
    Is there any other index based collection using that i get better performance than LinkedList?
    Thanks in advance

    The trouble with a LinkedList as implemented in the Java libraries is that if you want to insert at index 100, it has no option but to step through the first 100 links of the list to find the insert point. Likewise is you retrieve by index. The strength of the linked list approach is lost if you work by index and the List interface gives no other way to insert in the middle of the list. The natural interface for a linked list would include an extended Iterator with methods for insert and replace. Of course LinkedLists are fine when you insert first or last.
    My guess would be that if your habitual insertion point was half way or more through the list then ArrayList would serve you better especially if you leave ample room for growth. Moving array elements up is probably not much more expensive, per element, than walking the linked list. Maybe 150% or thereabouts.
    Much depends on how you retrieve, and how volatile the list is. Certainly if you are a read-mostly situation and cannot use an iterator then a LinkedList won't suit.

  • Report  performance with Hierarchies

    Hi
    How to improve query performance with hierarchies. We have to do lot of navigation's in the query and the volume of data size very big.
    Thanks
    P G

    HI,
    chk this:
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/4c0ab590-0201-0010-bd9a-8332d8b4f09c
    Query Performance – Is "Aggregates" the way out for me?
    /people/vikash.agrawal/blog/2006/04/17/query-performance-150-is-aggregates-the-way-out-for-me
    ° the OLAP cache is architected to store query result sets and to give all users access to those result sets.
    If a user executes a query, the result set for that query’s request can be stored in the OLAP cache; if that same query (or a derivative) is then executed by another user, the subsequent query request can be filled by accessing the result set already stored in the OLAP cache.
    In this way, a query request filled from the OLAP cache is significantly faster than queries that receive their result set from database access
    ° The indexes that are created in the fact table for each dimension allow you to easily find and select the data
    see http://help.sap.com/saphelp_nw04/helpdata/en/80/1a6473e07211d2acb80000e829fbfe/content.htm
    ° when you load data into the InfoCube, each request has its own request ID, which is included in the fact table in the packet dimension.
    This (besides giving the possibility to manage/delete single request) increases the volume of data, and reduces performance in reporting, as the system has to aggregate with the request ID every time you execute a query. Using compressing, you can eliminate these disadvantages, and bring data from different requests together into one single request (request ID 0).
    This function is critical, as the compressed data can no longer be deleted from the InfoCube using its request IDs and, logically, you must be absolutely certain that the data loaded into the InfoCube is correct.
    see http://help.sap.com/saphelp_nw04/helpdata/en/ca/aa6437e7a4080ee10000009b38f842/content.htm
    ° by using partitioning you can split up the whole dataset for an InfoCube into several, smaller, physically independent and redundancy-free units. Thanks to this separation, performance is increased when reporting, or also when deleting data from the InfoCube.
    see http://help.sap.com/saphelp_nw04/helpdata/en/33/dc2038aa3bcd23e10000009b38f8cf/content.htm
    Hope it helps!
    tHAK YOU,
    dst

  • Oracle: slow performance with SELECT using ojdbc14 and connection pooling

    Hello,
    i'm working hard the last days to solve a performance problem with our customer using a oracle 10g database. For testing I used our oracle 9.2.0.1.0 database which shows the same symptoms. All doing solved nothing: the performance while using this oracle is much slower than other databases. This result I cannot trust and so I need some advice. What is missing to improve the performance on the java side?
    The webapplication I use runs fast on MySQL 4.x and SQLServer 2000, but on the above mentioned Oracle it was always 4 times slower. The webapplication uses a lot of simple SELECT-Statements without complicated joins and so on (because it should run on many different databases). Doing some days of creating tests within this webapplication, I was not able to find any entrance point for a change. All databases server I'm using, having only the default configurations after a common installation.
    To reduce the complexity I wrote a simple java application with connection pooling using only the latest libraries from apache-commons(dbcp, pool), and the latest ojdbc14 for oracle 9.2.
    First the results than the code: MySQL needed less than 1000 millisecond, SQLServer around 1000 milliseconds and Oracle over 2000 milliseconds. I stopped pooling and the results are for Oracle even worse: over 18000 milliseconds (mysql:2500, sqlserver:4100).
    I changed the classes for Oracle and used the class oracle.jdbc.pool.OracleConnectionCacheImpl from the ojdbc14-library. No difference (around 100 milliseconds more or less).
    The only Select-Statement works on this table, which has one index on HICTGID.
    It contains 259 entrances.:
    CREATE TABLE HIERARCHYCATEGORY (
      HICTGID                 NUMBER (19)   NOT NULL,
      HICTGLEVEL              NUMBER (10)   NOT NULL,
      HICTGEXTID              NUMBER (19)   NOT NULL,
      HICTGEXTPARENTID        NUMBER (19)   NOT NULL,
      HICTGNAME               VARCHAR2(255) NOT NULL
    );The application simply loops through this table using
    SELECT Hictgid, Hictgname FROM HIERARCHYCATEGORY WHERE HICTGID = ?, but I always open a connection before this query and closes this connection afterwards. So I use the pooling as much as possible. That's all SQL I'm using.
        protected static DataSource setupDataSource(String sDriver, String sUrl, String sUser, String sPwd) throws SQLException {
            BasicDataSource ds = new BasicDataSource();
            ds.setDriverClassName(sDriver);
            ds.setUsername(sUser);
            ds.setPassword(sPwd);
            ds.setUrl(sUrl);
            // The maximum number of active connections:
            ds.setMaxActive(3);
            // The maximum number of active connections that can remain idle in the pool,
            // without extra ones being released, or zero for no limit:
            ds.setMaxIdle(3);
            // The maximum number of milliseconds that the pool will wait (when there are no available connections)
            // for a connection to be returned before throwing an exception, or -1 to wait indefinitely:
            ds.setMaxWait(3000);    
            return ds;
        }I can switch by using external properties between three databases (oracle, mysql and sqlserver) and if I want I can switch pooling off. And all actions I'm interested are logged by Log4J.
        public static Connection getConnection() throws SQLException {
            Connection result = null;
            String sJdbcDriver = m_oJbProp.getString("jdbcDriver");
            String sJdbcUrl = m_oJbProp.getString("databaseConnection");
            String sJdbcUser = m_oJbProp.getString("dbUsername");
            String sJdbcPwd = m_oJbProp.getString("dbPassword");
                try {
                    if (m_oJbProp.getString("useConnectionPooling").equals("true")) {
                         if (log.isDebugEnabled()) {
                              log.debug("ConnectionPooling true");
                        if(null == m_ds) {
                            m_ds = setupDataSource(sJdbcDriver,sJdbcUrl,sJdbcUser,sJdbcPwd);
                              if (log.isDebugEnabled()) {
                                   log.debug("DataSource created");
                        result = m_ds.getConnection();
                    } else {
                        // No connection pooling:
                         if (log.isDebugEnabled()) {
                              log.debug("ConnectionPooling false");
                        try {
                            Class.forName(sJdbcDriver);
                            result = DriverManager.getConnection(sJdbcUrl, sJdbcUser, sJdbcPwd);
                        } catch (ClassNotFoundException cnf) {
                            log.error("Exception: Class Not Found. ", cnf);
                            System.exit(0);
    (.. ErrorHandling ...)Here is the code fragment which is doing the work:
                     StringBuffer sb = new StringBuffer();
                while (lNextBottom <= lNextCeiling) {
                     con = getConnection();
                     innerSelStmt = con.prepareStatement("SELECT Hictgid, Hictgname FROM HIERARCHYCATEGORY WHERE HICTGID = ?");
                     innerSelStmt.setLong(1, lNextBottom);
                     rsInner = innerSelStmt.executeQuery();
                     if ((rsInner != null) && (rsInner.next())) {
                         sb.append(rsInner.getLong(1) + ", " + rsInner.getString(2) + "\r");
                          if (log.isDebugEnabled()) {
                               log.debug("Inner Statement: " + rsInner.getLong(1) + "\r");
                     rsInner.close();
                     con.close();
                     lNextBottom++;
                 if (log.isInfoEnabled()) {
                      log.info("\rResult values: Hictgid, Hictgname \r");
                      log.info(sb.toString());
                 }and the main method:
        public static void main(String[] args) {
            try {
                 long lStartTime = System.currentTimeMillis();
                 JdbcBasic oJb = new JdbcBasic();
                 boolean bSuccess = false;
                 bSuccess = oJb.getHierarchycategories();
                 if (log.isInfoEnabled()) {
                      log.info("Running time: " + (System.currentTimeMillis() - lStartTime));
                 if (null != m_ds) {
                     printDataSourceStats(m_ds);
                      shutdownDataSource(m_ds);
                      if (log.isInfoEnabled()) {
                           log.info("Datasource closed.");
             } catch (SQLException sqe) {
                  log.error("SQLException within  main-method", sqe);
        }My database values are
    databaseConnection=jdbc:oracle:thin:@SERVERDB:1521:ora
    jdbcDriver=oracle.jdbc.driver.OracleDriver
    databaseConnection=jdbc:jtds:sqlserver://SERVERDB:1433/testdb
    jdbcDriver=net.sourceforge.jtds.jdbc.Driver
    databaseConnection=jdbc:mysql://localhost/testdb
    jdbcDriver=com.mysql.jdbc.Driver
    dbUsername=testusr
    dbPassword=testpwdThanks for your reading and maybe for your help.

    A few comments.
    There is of course another difference between your test cases then just the database. There is also the driver. And I suspect that in at least the case with the jtds driver it is helping you along where you are doing something silly and the Oracle driver is not.
    Before I explain the next part I would say the speed differences between MS-SQL and MySQL look about right I think you are aiming here for MS-SQL level performance not MySQL. (For a bunch of reasons MySQL is inherently faster but there are MANY drawbacks as well which have been well discussed on previous threads)
    Here is where I believe your problem lies
    while (lNextBottom <= lNextCeiling) {
                     con = getConnection();
                     innerSelStmt = con.prepareStatement("SELECT Hictgid, Hictgname FROM HIERARCHYCATEGORY WHERE HICTGID = ?");
                     innerSelStmt.setLong(1, lNextBottom);
                     rsInner = innerSelStmt.executeQuery();
                     if ((rsInner != null) && (rsInner.next())) {
                         sb.append(rsInner.getLong(1) + ", " + rsInner.getString(2) + "\r");
                          if (log.isDebugEnabled()) {
                               log.debug("Inner Statement: " + rsInner.getLong(1) + "\r");
                     rsInner.close();
                     con.close();
                     lNextBottom++;
                 }There at least four things that are wrong with above.
    1) Why are you preparing the statement INSIDE the loop. Let us for a moment say that the loop will spin 100 times. That means that you are preparing the same statement 100 times. This is bad. It is also very relevant because for example the Jtds driver is going to be caching the prepared statements you make so that actually while you try and prepare it 100 times it only actually does it once... but in Oracle I don't know what it is doing for sure but if it is preparing on each pass well than that bit of it is going take 100 times longer then it should.
    2) You are opening and closing the connection on each pass through the loop... also a terrible idea. You need to fix this first so that you can repeatedly use the same prepared statement.
    3) Why are you looping in the first place? More on this later.
    4) Where do you close the PreparedStatement? It doesn't look like you do.
    Okay so for starters your loop should look a lot more like this...
    code]
    con = getConnection();
    innerSelStmt = con.prepareStatement("SELECT Hictgid, Hictgname FROM HIERARCHYCATEGORY WHERE HICTGID = ?");
    while (lNextBottom <= lNextCeiling) {
    innerSelStmt.setLong(1, lNextBottom);
    rsInner = innerSelStmt.executeQuery();
    if ((rsInner != null) && (rsInner.next())) {
    sb.append(rsInner.getLong(1) + ", " + rsInner.getString(2) + "\r");
    rsInner.close();
    lNextBottom++;
    innerSelStmt.close();
    con.close();
    I think the code above (and you can put your debug stuff back if you want) which uses ONE connection and ONE prepared Statement will improve your performance dramatically.
    The other question though I would as is why in the hell you are doing 100 or whatever number of queries anyway. This can be done all in ONE query which again will improve performance.
    Your query and such should look like this I think.
    String sql = "SELECT Hictgid, Hictgname FROM HIERARCHYCATEGORY WHERE HICTGID >=? AND HICTGID<=?";
    PreparedStatement ps = conn.prepareStatement(sql);
    ps.setLong(1,lNextBottom );
    ps.setLong(2,lNextCeiling);
    ResultSet rs = ps.executeQuery();
    while(rs.next()){
      // your appending to string buffer code goes here
    }and I can't understand why you're not doing that in the first place.

  • SAP ECC 6.0 EHP4 installation with ABAP+JAVA stack

    Hi,
    I need to install SAP ECC6.0 EHP4 with ABAP+JAVA stack on windiws server 2008 with oracle.
    I have read that for this, we need to install ECC 6.0 EHP4 ready and then upgrade to EHP4 using ehpi installer.
    I have found the documents seperately for ABAP and JAVA stack and not for ABAP+JAVA.
    I need clarifications on these.
    1.Do we need to install ABAP and JAVA instances seperately with EHP4 and then upgrade(In that case I have to install both instances on single host)?
    2.Should we give seperate <SID> for both instances and create seperate databases?
    3.Hard disk requirement for this set up.(We have 150 GB including C drive)
    4.Steps to perform after installing ECC6.0 EHP4 ready
    5.Please send me the doc. if anyone has
    Anyhelp would be highly appreciated.
    Regards,
    Surya

    > 1.In the installation document for EHP4 ready for SAP ERP 6.0 ABAP.It is mentioned that this installation contains only basic EHP4 for ERP 6.0 and we need to install additional EHP4  functionality in the SAP ERP 6.0 -EHP4 ready system.
    > It states that atleast Support pacakge stack (SPS) 1 of SHP4 needs to be installed.So If I install this stack using EHPi installer in the post installation activities my SAP ERP 6.0 EHP4 will be in place?Please confirm.
    Yes..
    >
    > 2.Can I install SAP SR3 with dual stack and upgaradte it directly to EHP4 at a time?
    Yes you can do it.....as SAP still supports dual stack ERP systems to EHP4
    >
    > Any solutions or suggestions from others also welcome.
    Why do u want a dual stack ECC ystem?? any specific requirement???
    Regards,
    Nirmal.K
    Edited by: nirmal konchada on Feb 2, 2011 12:16 PM

  • Query Performance with Exception aggregation

    Hello,
    My Query Keyfigures has exception aggregation on order line level as per requirement.
    Currently cube holds 5M of records, when we run query its running more than 30min.
    We cont remove exception aggregation.
    Cube is alredy modeled correctly and we dont want to use the cache.
    Does anybody can please advice if there is any other better approach to improve query performance with exception agg?
    Thanks

    Hi,
    We have the same problem and raised an OSS ticket. They replied us with the note 1257455 which offers all ways of improving performance in such cases. I guess there s nothing else to do, but to precalculate this exception aggregated formula in data model via transformations or ABAP.
    By the way, cache can not help you in this case since exc. agg. is calculated after cache retrieval.
    Hope this helps,
    Sunil

  • Does anyone have a Portal working with 50 or more Users concurrently ?

    I WOULD LIKE TO KNOW IF ANYBODY HAS A PRODUCTION PORTAL WITH 50 OR MORE USERS WORKING CONCURRENTLY.
    I AM INTERESTED TO KNOW HOW PORTAL PERFORMS WITH MANY USERS.
    ANY ADVICE OR SUGGESTIONS WILL BE VERY WELLCOME.
    TKS!

    Without knowing your hardware setup, your functional requirements, and the user base, and can't answer that question with 100% certainty. But, I can tell you that a multiple server configuration will definitely provide you with more performance and scalability (assuming proper configuration).
    As far as using Standard Edition vs Enterprise Edition, there will be no differences from a performance/scalability perspective. With Enterprise edition you get some tools you don't get with Standard Edition such as WebCache, Forms, Reports, Discoverer, 8i Cache (database cache), and iFS. I'm not certain if you get OID with both editions, but I think you might. If you dont have OID you wont be able to use the LDAP authentication features of Portal (unless you already have your own LDAP directory service set up).
    The product lists are from my memory (not always dependable), so you should verify those through your sales representative before making any decisions.
    Hope this helps...
    null

  • SSL impact on Portal performance.

    I was wondering if anyone knows of any benchmarks of Portal performance with SSL enabled versus a non SSL enabled Portal.
    I realize that SSL does create some additional overhead on the server during the encyrpt/decrypt process. I'm trying to determine if a dedicated reverse proxy may be more beneficial to support approximately 1000 concurrent connections.
    Thanks in advance, and as always points awarded.

    I haven't seen any official figures, but you also have to consider the applications running in the portal. For example, if you plan to encrypt SAPGUI for HTML, that will have an imapact etc.
    Reallistically you need to do your own tests, as relying on someone else's figures based on their specific portal applications may lead you to problems later on.

Maybe you are looking for