Best Practice SAP connection

Hello All
I am just starting creating reports with CR2008 connected to SAP ECC 6.0.  I am suffering some performance issues when trying to create a report on tables VBAK, VBAP and LIPS. Report performance is extremely slow. I have a few questions on this:
1. Can anyone point me to some documentation that would help me set up SAP and CR2008 for improved performance when reporting directly on tables. 
2. Performance wise Is it better to report on an Classic Infoset rather than directly on a table in CR2008
Any other tips would be greatly appreciated.
Thanks
Phillip

RFC Statement                                                                               
Function module name                                                   
  Source IP Address         192.168.3.220                                
  Source Server             cham-vm1_CV1_00                              
  Destination IP Address                                                 
  Destination Server                                                     
  Client/Server             Client                                       
  Conversation ID                                                        
  RFC Trace Rec. Status     2                                            
  Sent Bytes                0                                            
  Received Bytes            0                                            
  Total Sent Bytes                                                       
  Total Received Bytes                                                   
  ABAP program name         %_T002S0                                     
  RFC Time                           0                          
  Function module name      /CRYSTAL/OSQL_EXECUTE_INTERNAL                 
  Source IP Address                                                        
  Source Server             cham-vm1_CV1_00                                
  Destination IP Address    192.168.3.220                                  
  Destination Server                                                       
  Client/Server                                                            
  Conversation ID                                                          
  RFC Trace Rec. Status     5                                              
  Sent Bytes                0                                              
  Received Bytes            12.166                                         
  Total Sent Bytes                                                         
  Total Received Bytes                                                     
  ABAP program name                                                        
  RFC Time                  364328.935

Similar Messages

  • What are the best practices to connect 30-40 iPads to Wi-Fi in a single room?

    What are the best practices to connect 30-40 iPads to Wi-Fi in a single room?

    I don't use it but it does say this in the help section...

  • Best practices for connecting to DB

    Hi,
    I am having 3 different java classes which will contact the DB for getting the data from the table. I wrote a separate java class for DB connection like this :
    public class DBConnection {
    private Connection con;
    public Connection getConnection() {
    try {
    Class.forName("org.apache.derby.jdbc.ClientDriver");
    con = DriverManager.getConnection("jdbc:derby://localhost:1527/StrutsDB", "username", "password");
    } catch (SQLException ex) {
    Logger.getLogger(DBConnection.class.getName()).log(Level.SEVERE, null, ex);
    } catch (ClassNotFoundException ex) {
    Logger.getLogger(DBConnection.class.getName()).log(Level.SEVERE, null, ex);
    return con;
    How to use this single connection object in all of the 3 classes? What are all the ways of doing so?
    Is there any best practices for connecting to DB?
    Thanks in advance,

    The problem with "best practice" is it really depends on your situation. If you are creating a single user, desktop application then what you are doing will work but would be more efficient if the connection was declared as static and you used the singleton pattern.
    public class DBConnection {
        private static final Connection con;
        private DBConnection() {
            try {
                Class.forName("org.apache.derby.jdbc.ClientDriver");
                con = DriverManager.getConnection("jdbc:derby://localhost:1527/StrutsDB", "username", "password");
            } catch (SQLException ex) {
                Logger.getLogger(DBConnection.class.getName()).log(Level.SEVERE, null, ex);
            } catch (ClassNotFoundException ex) {
                Logger.getLogger(DBConnection.class.getName()).log(Level.SEVERE, null, ex);
        public static Connection getConnection() {
            return con.
    }The private constructor guarantees only one instance of the class will be created (since you can't use 'new' to create one) and initialized the database connection. Then any other object that require a connection simply call DBConnection.getConnection() and they will get the same database connection each time.
    Note that this is a little simplistic and is not thread safe. If your classes will be executing on different threads you will need a more sophisticated approach. You will also need to make sure you commit or rollback any transactions when done or the next time you get the connection you may be in the middle of an existing transaction.

  • What is the best Practice JCO Connection Settings for DC  Project

    When multiple users are using the system data is missing from Web Dynpro Screens.  This seems to be due to running out of connections to pull data.
    I have a WebDynpro Project based on component development using DC's.  I have one main DC which uses other DC's as Lookup Windows.  All DC's have their Own Apps.  Also inside the main DC screen, the data is populated from multiple function modules.
    There are about 7 lookup DC Apps accessed by the user
    I have created JCO destinations with following settigns
    Max Pool Size 20
    Max Number of Connections 200
    Before I moved to DC project it was regular Web Dynpro Project with one Application and all lookup windows were inside the same Project.  I never had the issue with the same settings.
    Now may be becuase of DC usage and increase in applications I am running out of connections.
    Has any one faced the problem.  Can anyone suggest the best practice of how to size JCO connections.
    It does not make any sense that just with 15-20 concurrent users I am seeing this issue.
    All lookup components are destroyed after its use and is created manually as needed.  What else can I do to manage connections
    Any advise is greatly appreciated.
    Thanks

    Hi Ravi,
    Try to go through this Blog its very helpful.
    [Web Dynpro Best Practices: How to Configure the JCo Destination Settings|http://www.sdn.sap.com/irj/scn/weblogs;jsessionid=(J2EE3417600)ID2054522350DB01207252403570931395End?blog=/pub/wlg/1216]
    Hope It will help.
    Regards
    Jeetendra

  • Best practice SAP landscape

    Hello all,
    I would like to know if there is some kind of best practice regarding SAP landscape in a big company.
    For example is it recommended to have in the landscape a SAP Quality Assurance System open for customizing (transaction SCC4) so that quick customizing tests are performed at any moment, instead of customizing in Development system and then transports in QaS. (this can be very frustrating because for solving and testing an issue it's possible that numerous customizing tasks and reset of customzing is neccessary) ?
    How SAP compliant would this solution be?
    Thank you very much for your help!
    Daniel Nicula

    Hmmm, I do not know exactly if the question can be posed here in GRC related threads.
    But it seemed to me that it is somehow connected.
    Anyway, I agree with you that final customizing should be done in DEV and then transported in QAS.
    What i am not sure is if it is against SAP recommendations to have a QAS opened for customizing and try all the solutions for an issue. And in the end when you are sure of what you want to do and to obtain, then you do the customizing also in DEV and follow the normal transport route.
    Which can be the risks in case you have a QAS opened for customizing?
    Thank you.

  • Best Practice - WAP connecting switchport configuration.

    Is there a best practice for deploying the WAP's in a WAP/WLC infrastructure?  Should the connecting switchport be an Access port or a Trunk port?  I've seen this implemented in both fashions and wasn't sure if one was a better choice than the order.  What is the difference?
    My other question is regarding applying additional switchport configurations.  Is there anything wrong with applying either spanning-tree portfast, spanning-tree bpdguard, or switchport port-security. 

    Hi Ken,
    Access port all the time, everywhere, UNLESS the AP is configured for HREAP/FLEX then trunk. Or if you deploy a AP in monitor mode then TRUNK.
    QOS -- if its access port trust dscp. If you truck trust cos.
    No you are fine. Portfast is highly recommended.
    "Satisfaction does not come from knowing the solution, it comes from knowing why." - Rosalind Franklin
    ‎"I'm in a serious relationship with my Wi-Fi. You could say we have a connection."

  • Best Practice SAP HealthCare Portal

    Hi SAP Team,
    I have a following questions about the SAP HealthCare integration with SAP Portal, What is the best practice for implementation in these case? Its necessary to develop the application using webdynpro or Now I available the SAP Portal Business Package for these solution?
    Best Regards
    Roberto

    I  am also search on net & found some links & files about it, it found this PPT file on this topic, hope this will helps you.
    http://net.educause.edu/ir/library/powerpoint/EDU03146.pps
    [healthcare technology|http://www.nx3corp.com/]

  • Best Practices for Connecting to WebHelp via an application?

    Greetings,
    My first post on these forums, so I appologize if this has already been covered (I've done some limited searching w/o success).  I'm developing a .Net application which is accessing my orginazation's RoboHelp-generated webhelp.  My organization's RoboHelp documentation team is still new with the software and so it's been up to me to chart the course for establishing the workflow for connecting to the help from the application.  I've read up on Peter Grange's 'calling webhelp' section off his blog, but I'm still a bit unclear about what might be the best practices approach for connecting to webhelp.
    To date, my org. has been delayed in letting me know their TopicIDs or MapIDs for their various documented topics.  However, I have been able to acquire the relative paths to those topics (I achieved this by manually browsing their online help and extracting out the paths).  And I've been able to use the strategy of creating the link via constructing a URL (following the strategy of using the following syntax: "<root URL>?#<relative URI path>" alternating with "<root URL>??#<relative URI path>").  It strikes me, however, that this approach is somewhat of a hack - since RoboHelp provides other approaches to linking to their documentation via TopicID and MapID.
    What is the recommended/best-practices approach here?  Are they all equally valid or are there pitfalls I'm missing.  I'm inclined to use the URI methodology that I've established above since it works for my needs so far, but I'm worried that I'm not seeing the forest for the trees...
    Regards,
    Brett
    contractor to the USGS
    Lakewood, CO
    PS: we're using RoboHelp 9.0

    I've been giving this some thought over the weekend and this is the best answer I've come up with from a developer's perspective:
    (1) Connecting via URL is convenient if (#1) you have an established naming convention that works for everyone (as Peter mentioned in his reply above)
    (2) Connecting via URL has the disadvantage that changes to the file names and/or folder structure by the author will break connectivity
    (3) Connecting via TopicID/MapID has the advantage that if there is no naming convention or if it's fluid or under construction, the author can maintain that ID after making changes to his/her file or folder structure and still maintain the application connectivity.  Another approach to solving this problem if you're working with URLs would be to set up a web service that would match file addresses to some identifier utilized by the developer (basically a TopicID/MapID coming from the other direction).
    (4) Connecting via TopicID has an aesthetic appeal in the code since it's easy to provide a more english-readable identifier.  As a .Net developer, I find it easy and convenient to construct an enum that matches my TopicIDs and to utilize that enum to construct my identifier when it comes time to make the documentation call.
    (5) Connecting via URL is more convenient for the author, since he/she doesn't have to worry about maintaining IDs
    (6) Connecting via TopicIDs/MapIDs forces the author to maintain those IDs and allows the documentation to be more easily used into the future by other applications worked by developers who might have their own preference in one direction or another as to how they make their connection.
    Hope that helps for posterity.  I'd be interested if anyone else had thoughts to add.
    -Brett

  • What is the best practice for connecting to different schemas?

    Hi all,
    We are porting an application from SQL Server to oracle and would like to know what the best practices are in oracle for user connections to an Oracle instance.
    More or less the question could be put like this:
    1) The equivalent of a SQL Server Database in Oracle is a Schema. (more or less)
    2) A specific application has it's own schema where it keeps all related objects (Tables, etc)
    3) In SQL Server you grant access to the Database and its objects (Tables, etc) to all users of the application.
    4) In Oracle do you grant access to the Schema and its objects (Tables, etc) to all users of the application also? Or do all users log
    in as the schema owner?
    So in Oracle if there existed [SchemaApplication].[table1], how would [userChris] and [userDave] query [SchemaApplication].[table1]?
    Would Chris and Dave log in as [userChris] and [userDave], or would they normally log in as [userApplication]?
    finally, is it good practice to log in as a unique user eg [userChris] and then issue the
    alter session set current_schema = shemaApplication;
    command to change the way references to tables are interpreted?

    We are porting an application from SQL Server to oracle and would like to know what the best practices are in oracle for user connections to an Oracle instance.
    More or less the question could be put like this:
    1) The equivalent of a SQL Server Database in Oracle is a Schema. (more or less)
    2) A specific application has it's own schema where it keeps all related objects (Tables, etc)
    3) In SQL Server you grant access to the Database and its objects (Tables, etc) to all users of the application.
    4) In Oracle do you grant access to the Schema and its objects (Tables, etc) to all users of the application also? Or do all users log
    in as the schema owner?There are ways to implement the same.
    Case 1.
    Create different roles, such as APP_ROLE, READONLY_ROLE. Create public synonym for objects in SchemaApplication user. Grant these role to single user say appUser this is different from you SchemaApplication user. Use appUser to connect to application and for different user like userChris, userDave provide another layer of security. Say userDave is allowed only to deal with cash related transaction, so allow him to open only those screens which are related to cash transaction only.
    Case 2.
    Create public synonym and grant privilege on tables from SchemaApplication to different users (say userChris, userDave).
    So in Oracle if there existed [SchemaApplication].[table1], how would [userChris] and [userDave] query [SchemaApplication].[table1]?This is resolved by public synonym. There are private synonym as well, you can create this also but in this case you have to create private synonym for each of the users.
    Would Chris and Dave log in as [userChris] and [userDave], or would they normally log in as [userApplication]? I would suggest you to connect either using a new user(Case 1) or the user itself has account in the database(Case2).
    finally, is it good practice to log in as a unique user eg [userChris] and then issue the
    alter session set current_schema = shemaApplication;
    No. It is not a good practice to allow the user to login to database using the application owner.
    command to change the way references to tables are interpreted?The public/private synonym can be used to resolve the schema.object value. For example, if SchemaApplication has table T, then you can create public synonym as 'CREATE PUBLIC SYNONYM T FOR SchemaApplication.T'; and now you can refer this table as T from any other schema(user).
    HTH
    Virendra

  • What is the best practice to connect 2 sites and replicate ADDS between 2 different sites using windows server 2012 r2?

    I found that there is new features remote access inside windows server 2012 r2 which is much more easier. Is there anybody can suggest me how to connect 2 different sites and it will be the tunnel to replicate ADDS (RW) between both sites.

    See this
    https://social.technet.microsoft.com/Forums/windowsserver/en-US/133e7780-6a59-4bd7-906b-70830bea48d0/sites-connections-best-practice?forum=winserverDS
    Regards,
    Biswajit
    MCTS, MCP 2003,MCSA 2003, MCSA:M 2003, CCNA, Enterprise Admin, ITIL F 2011
    Blog:
      Script Gallary:
      LinkedIn:
    Note: Disclaimer: This posting is provided & with no warranties or guarantees and confers no rights..

  • What are best practices for connecting asa to nexus 5000

    just trying to get a feel for the best way to connect redundant asa to redundant nexus 5000
    using a vpc vlan is fine, but then running a routing protocol isn't supported, so putting static routes on 5000 works, but it doesn't support ip sla yet so you cant really stop distributing the default if your internet goes down. just looking for what was recommended.

    you want to test RAC upgrade on NON RAC database. If you ask me that is a risk but it depends on may things
    Application configuration - If your application is configured for RAC, FAN etc. you cannot test it on non RAC systems
    Cluster upgrade - If your standalone database is RAC one node you can probably test your cluster upgrade there. If you have non RAC database then you will not be able to test cluster upgrade or CRS
    Database upgrade - There are differences when you upgrade RAC vs non RAC database which you will not be able to test
    I think the best way for you is to convert your standalone database to RAC one node database and test it. that will take you close to multi node RAC

  • Best practice for connecting to SAP backend from JSPDynPage

    Hi,
    Can anyone help clear up some confusion I have over connecting to SAP from Java?
    We have a number of JSPDynPage portal applications on EP7 Ehp1 that connect to SAP ECC6.
    We currently use 2 methods to call remote functions on the ECC system.
    1) Enterprise connector and JCo Client Service
    This method has the advantage of automatically generated proxy classes and typed access to function module parameters. It is also more closely aligned to Web Dynpro from a design time point of view
    However it also has a number of disadvantages in that it does not support table APPEND structures and requires regeneration for every change to the ABAP structures involved even if the field that has changed is not being referred to. In addition the use of the JCo client service means that the connection pooling must be handled ourselves explicitly
    2) Connector Framework - SAP System Connector
    This method has the advantage that connection management is handled by the framework. It also has the advantage that fields are referred to by name so if an ABAP structure changes then the remote call will not fall over as long as the named fields are still available.
    However this method is more cumbersome because no proxy classes are generated and function module parameters are referred to generically so errors are not picked up by the complier.
    Given the limitations of the above 2 methods, what is the recommended approach?
    In the book u2018Inside Web Dynpro for Javau2019 Chris Whealy says the following about the Adaptive RFC Layer:
    u2018Any Java program running in the AS Java u2013 not just Web Dynpro u2013 can make use of the Adaptive RFC layeru2019
    Has anyone been able to use the Adaptive RFC layer from a NON-WebDynpro java application?
    Any thoughts on the above would be most welcome.
    Thanks,
    Denis.

    Hello,
    The recommended way is with JCA Connector.
    Adaptive RFC is included from NW7.1,
    See "Creating an Adaptive RFC2 Sample Application without using Web Dynpro"
    in the SAP HELP of NWCE7.1 or NWCE7.2.
    Seems to be based in JCo 3.0 and Java5 so it will not work in NW70.
    For older Versions like EP 5 there is also a JCo ClientService available.
    Regards
    Johannes

  • Best Practiced SAP Netweaver Portal

    Hi All,
    We are trying to determine what should be the breaking point between what the portal dev team handles and what becomes SAP Securities responsibility. We are thinking that security should take it over when assigning IViews to Roles and then Roles to Groups and lastly groups to users. This is a very high view and welcome as much detailed information as you can provide.
    Thanks,
    Mary Sims

    Hi Mary,
    I fully agree with John. Portal roles are used to build up navigation. Therefor they define, what a user would see (if he/she would have read permission).
    The mapping between user groups and portal roles, the assignment of users to roles and the setting of read permissions on certain iViews (pages) or folders in KM is a different thing. This is a security issue and should be done by security department.
    We face this responsibility discussion in all our portal projects. You just have to rethink the word 'role', which normally isn't used for structuring content. This helps clearing up things. But at the end of the day, its an organizational discussion which depends on your internal organization.
    At least, I made the experience, that it's best if you assign portal roles to portal groups but use the groups for setting the access control list - not the roles.
    Maybe another cent,
    Carsten

  • Best Practice SAP HANA Unify tables

    I have 2 tables from differents systems With The same structure . We need unify. It´s possible Do It with Calculate View, and after the start this analytics and anothers Calculate View ?.
    Same Name, and different Schema
    Thanks

    I have 2 tables from differents systems With The same structure . We need unify. It´s possible Do It with Calculate View, and after the start this analytics and anothers Calculate View ?.
    Same Name, and different Schema
    Thanks

  • NetApp direct connect to UCS best practices

    Folks,
    I have installed many FlexPods this year but all involved either Nexus 5Ks or 7Ks with vPC. This protects the NFS LUN connections from both network outages or UCSM FIM outages. What is not clear to me is the best practice on connecting a NetApp directly to the UCS via applaince ports. Appliance ports seem like a great idea but they seem to ass issues to designs, both in VMware and the network.
    Does anyone have a configuration example on both the NetApp, UCS & Vmware side?
    I thought I would ask the group their opinion.
    Cheers,
    David Jarzynka

    Hi David
    Can you clarify your last posting a little bit.
    I never installed direct attached NetApp Storage to the UCS as well.
    One drawback i see is, if you have your NetApp System in active/standby mode connected to the FI, then for example all Servers connect through FI A to the NetApp System (NIC with VMKernel for all Server active on FI A). If the NetApp Link fails and switches over to standby link on Fabric B all the Traffic will go form Server to FI A to Uplink Switch to FI B and then to the NetApp System - because the Server is not aware that the NetApp System did a failover. So not all companies will have 10G Uplink Switches which will cause a bottleneck in that case.
    What other things do you see? I agree completely with you - all say is a wonderful feature on the slides - but i don't think it's that smart in practise?
    Thanks for a short replay.
    Cheers
    Patrick

Maybe you are looking for