Accrual Reconciliation Load Run accoutning data

Hi,
Can anyone let me know from which table the Accrual Reconciliation Load Run picks up the acocunting CCIDs?
Thanks and Regards,
MPH

We experienced similar issue. As the runtime of this report depends on the input parameters. Remember, your first run of this report going to take significant amount of time and the subsequent runs will be much shorter.
But w had to apply the patches referred in the MOS article to resolve the performance issue.
Accrual Reconciliation Load Run Has Slow Performance [ID 1490578.1]
Thanks,
Sunthar....

Similar Messages

  • Accrual Reconciliation Load Run  report performance issue.

    we have significant performance issues when running accrual reconciliation load run report. we had to cancel after have it run for a day. any idea on how to resolve it?

    We experienced similar issue. As the runtime of this report depends on the input parameters. Remember, your first run of this report going to take significant amount of time and the subsequent runs will be much shorter.
    But w had to apply the patches referred in the MOS article to resolve the performance issue.
    Accrual Reconciliation Load Run Has Slow Performance [ID 1490578.1]
    Thanks,
    Sunthar....

  • Accounts Payable and PO Accrual Reconciliation Summary Report?

    Dear all,
    We are using R12.1.3 and doing the reconciliation AP Vs PO Accrual. However, I run the report "Accounts Payable and PO Accrual Reconciliation Summary Report" and found that one record is strange, like the PO Accrual balance is something -10,000 and the Account Payable balanced is 0.00.
    What does it mean for my case? Please advice.
    Thanks in advance.

    Dear all,
    I rerun " Accrual Reconciliation Load Run" program with data range. After that I re-run "Accounts Payable and PO Accrual Reconciliation Summary Report" and found that everything is fine now.
    I just wondering am I need to run "Accrual Reconciliation Load Run" every month end closing or just run once for up-grade POs?
    Please advice.
    Thanks.

  • R12: How to filter AP and PO Accrual Reconciliation Report based on GL Date

    Hi,
    Anybody know how to filter AP and PO Accrual Reconciliation Report based on GL Date from and GL Date to in R12?
    Since we just upgraded from 11.5.10 to 12.1.3 and found we cannot filter those report for specific date. It shown all data included the old data from 8 years ago also.
    We need to run the report only for specific date only. Please share with me if anyone know about this.
    Thanks.

    In 12 you can rebuild and run the report for a particular period, but not a single day if that's what's required

  • PO issue (Accrual Reconciliation Report )

    Hi,
    when run the accrual reconciliation report for PO,I find the quantity *Unit price is not equal Amount when A/P Receipt Match,the details as follows:
    when find the original PO line ,find that the PO line's status is cancelled ,the price have changed from 6 to 45 , from 10 to 50 before cancelled this PO lines, how can do now let them equal ?
    and check the invoice for this po ,
    20081022     1     Bar     20     6     120.00
    CH0000059 1     Bar     -20     6     (120.00)
    20081022     2     Bar     20 10     200.00
    CH0000059 2 Bar     -20 10 (200.00)
    CHSZ-Compart Hi-Precision Tech                         Accrual Reconciliation Report                          Report Date:      02-12-2008 20:00     
    Account:     24.00000.317010.0000.0000.000                                                                 
    PO Number     Item+      Transaction     Src     Org      Date     Document Number     Line      Unit     Quantity      Unit Price      Amount      Variance      Balance     
    Vendor Name:     : ShenzhenBaoan                                                       
    240803218     Pdtn consumables     Receive     PO     CHS     2008-10-29     30509262     1     Bar     -20     6     (120.00)          (120.00)     
         Pdtn consumables     Return to Supplier     PO     CHS     2008-11-5     30509262     1     Bar     20     6     120.00           120.00      
         Pdtn consumables     A/P Receipt Match      AP     CHS     2008-10-29     20081022     1     Bar     *20     45     120.00           120.00*
         Pdtn consumables     A/P Receipt Match      AP     CHS     2008-11-5     CH0000059     1     Bar     -20     45     (900.00)          (900.00)     
         Pdtn consumables     Receive     PO     CHS     2008-10-29     30509262     2     Bar     -20     10     (200.00)          (200.00)     
         Pdtn consumables     Return to Supplier     PO     CHS     2008-11-5     30509262     2     Bar     20     10     200.00           200.00      
         Pdtn consumables     A/P Receipt Match      AP     CHS     2008-10-29     20081022     2     Bar     *20*     *50*     *200.00*           *200.00*
    Pdtn consumables     A/P Receipt Match      AP     CHS     2008-11-5     CH0000059     2     Bar     -20     50     (1,000.00)          (1,000.00)     
                                  Vendor          Subtotal               (1,580.00)     0     (1,580.00)     
                                                           ---------------- -     ---------------     ----------------     GL
                                       Account Subtotal:                     (67,550.27)     0     (67,550.27)     (67,550.27)
                                       Report Total:                     (67,460.15)     -90     (67,550.15)     
                                                           ================ =     ===============     ================
    Edited by: user638343 on Dec 28, 2008 10:56 PM
    Edited by: user638343 on Dec 28, 2008 10:57 PM
    Edited by: user638343 on Dec 28, 2008 10:58 PM
    Edited by: user638343 on Dec 28, 2008 10:59 PM
    Edited by: user638343 on Dec 28, 2008 11:00 PM
    Edited by: user638343 on Dec 28, 2008 11:07 PM

    have anybody can give advise ?
    Edited by: user638343 on Jan 8, 2009 10:24 PM

  • PowerPivot Data Refresh Fails - We cannot locate a server to load the workbook Data Model

    Hello,
    I have installed sql01\PowerPivot instance of Analysis services on my existing SQL Server and configured it per Configure Power Pivot for SharePoint 2013
    http://technet.microsoft.com/en-us/library/dn456880(v=office.15).aspx . I did not install a second Database Engine on the server.
    I created an Excel 2013 workbook using the Excel 2013 PowerPivot add-in and saved it to my SharePoint BI Center Document library.
    When I try to refresh my workbook, I get 2 errors in the SharePoint log.
    http://support.microsoft.com/kb/2756665?wa=wsignin1.0 indicates the sproc is deprecated, but that does not help me fix the issue.
    Log errors:
    Following error occured while trying to execute a sql query: System.Data.SqlClient.SqlException (0x80131904):
    Could not find stored procedure 'DataRefresh.GetSchedule'.
    EXCEPTION: NoAvailableStreamingServerException: We cannot locate a server to load the workbook Data Model.
    macrel

    Hi Marcel,
    Based on my research, it seems that you have Microsoft SQL Server 2012 PowerPivot for SharePoint 2013 add-in installed on a computer. When you perform a data refresh operation on some PowerPivot workbooks in Microsoft SharePoint 2013, you receive two error
    message.
    As to the first error, this issue occurs because the Custom Properties on the Excel workbook are not correctly cleared before you set a new serialized schedule. This causes an extra custom property to remain after you set the new schedule. Therefore, the
    new schedule fails. To fix this issue, please
    install Cumulative Update 9 for SQL Server 2012 SP1.
    As to the second error, please
    running a repair using PowerPivot configuration wizard for SharePoint 2013. For more information about Troubleshooting PowerPivot Data Refresh, please see: 
    http://social.technet.microsoft.com/wiki/contents/articles/troubleshooting-powerpivot-data-refresh.aspx
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • XEM - Unable to load the initial data or the variances(delta) data into sys

    I am installing xEM 2.0 SP 10 (SAP xApp Emissions Management) in a windows environment with SQL 5000.  I installed xEM on NW 2004, usage types AS Java and EP 6.
    I am attempting to load the initial data or the variances (delta) data into the system.  Instruction is on page 15 in the install guide.
    I am supposed to enter the following in the command line:
    java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=[JDBC Driver];[JDBCUrl];[User];[Password]
    Example command for import into SQL Server:
    java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=com.ddtek.jdbc.sqlserver.SQLServerDriver; jdbc:datadirect:sqlserver://vma03:1433;SAPC11DB;password
    The customer I am with is running the xEM database on a different instance.  This is where I run into a problem.  I am not sure how to specify the instance in the script.  This is what  I have attempted so far:
    C:\>cd temp\load
    C:\Temp\load>java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=com.ddte
    k.jdbc.sqlserver.SQLServerDriver;jdbc:datadirect:sqlserver://PRODSQL43:SQL3:1534;SAPPEMDB;password
    java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:datadirect:sqlserver:/
    /PRODSQL43:SQL3:1534 as user SAPPEMDB (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver]Unable to connect.  Inva
    lid URL.): [DataDirect][SQLServer JDBC Driver]Unable to connect.  Invalid URL.
            at com.sap.sdm.util.dbaccess.DBTask.dbImport(DBTask.java:356)
            at com.sap.sdm.util.dbaccess.SapTransTask.perform_import(SapTransTask.java:293)
            at com.sap.sdm.util.dbaccess.SapTransTask.execute(SapTransTask.java:51)
            at com.sap.sdm.util.dbaccess.SapTrans.main(SapTrans.java:21)
    import aborted with java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:da
    tadirect:sqlserver://PRODSQL43:SQL3:1534 as user SAPPEMDB (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver]Unab
    le to connect.  Invalid URL.): [DataDirect][SQLServer JDBC Driver]Unable to connect.  Invalid URL.
    C:\Temp\load>java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=com.ddte
    k.jdbc.sqlserver.SQLServerDriver;jdbc:datadirect:sqlserver://PRODSQL43;SQL3:1534;SAPPEMDB;password
    java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:datadirect:sqlserver:/
    /PRODSQL43 as user SQL3:1534 (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver]Error establishing socket. Connec
    tion refused: connect): [DataDirect][SQLServer JDBC Driver]Error establishing socket. Connection refused: connect
            at com.sap.sdm.util.dbaccess.DBTask.dbImport(DBTask.java:356)
            at com.sap.sdm.util.dbaccess.SapTransTask.perform_import(SapTransTask.java:293)
            at com.sap.sdm.util.dbaccess.SapTransTask.execute(SapTransTask.java:51)
            at com.sap.sdm.util.dbaccess.SapTrans.main(SapTrans.java:21)
    import aborted with java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:da
    tadirect:sqlserver://PRODSQL43 as user SQL3:1534 (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver]Error establi
    shing socket. Connection refused: connect): [DataDirect][SQLServer JDBC Driver]Error establishing socket. Connection ref
    used: connect
    C:\Temp\load>java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=com.ddte
    k.jdbc.sqlserver.SQLServerDriver;jdbc:datadirect:sqlserver://PRODSQL43:1534;SQL3;SAPPEMDB;password
    java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:datadirect:sqlserver:/
    /PRODSQL43:1534 as user SQL3 (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver][SQLServer]Login failed for user
    'SQL3'.): [DataDirect][SQLServer JDBC Driver][SQLServer]Login failed for user 'SQL3'.
            at com.sap.sdm.util.dbaccess.DBTask.dbImport(DBTask.java:356)
            at com.sap.sdm.util.dbaccess.SapTransTask.perform_import(SapTransTask.java:293)
            at com.sap.sdm.util.dbaccess.SapTransTask.execute(SapTransTask.java:51)
            at com.sap.sdm.util.dbaccess.SapTrans.main(SapTrans.java:21)
    import aborted with java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:da
    tadirect:sqlserver://PRODSQL43:1534 as user SQL3 (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver][SQLServer]Lo
    gin failed for user 'SQL3'.): [DataDirect][SQLServer JDBC Driver][SQLServer]Login failed for user 'SQL3'.
    C:\Temp\load>java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=com.ddte
    k.jdbc.sqlserver.SQLServerDriver;jdbc:datadirect:sqlserver://PRODSQL43:1534:SQL3;SAPPEMDB;password
    java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:datadirect:sqlserver:/
    /PRODSQL43:1534:SQL3 as user SAPPEMDB (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver]Unable to connect.  Inva
    lid URL.): [DataDirect][SQLServer JDBC Driver]Unable to connect.  Invalid URL.
            at com.sap.sdm.util.dbaccess.DBTask.dbImport(DBTask.java:356)
            at com.sap.sdm.util.dbaccess.SapTransTask.perform_import(SapTransTask.java:293)
            at com.sap.sdm.util.dbaccess.SapTransTask.execute(SapTransTask.java:51)
            at com.sap.sdm.util.dbaccess.SapTrans.main(SapTrans.java:21)
    import aborted with java.lang.Exception: ERROR: Cannot connect to the database:ERROR: Connect failed to database jdbc:da
    tadirect:sqlserver://PRODSQL43:1534:SQL3 as user SAPPEMDB (java.sql.SQLException/[DataDirect][SQLServer JDBC Driver]Unab
    le to connect.  Invalid URL.): [DataDirect][SQLServer JDBC Driver]Unable to connect.  Invalid URL.
    C:\Temp\load>
    My last attempt was a command using colons and semicolons with the following results.  The closest (there was a significant delay before the error or failure) appears to have been //PRODSQL43;SQL3:1534; (second attempt).  The error listed from this attempt is   "Error establishing socket. Connection refused: connect".
    I also checked the default database that user SAPPEMDB has in place and it is assign the correct database.
    Please help.
    Message was edited by: Owner
            Mike Smayrabunya

    Hey,
    It looks like one of the following:
    1. The DB is down,
    2. The user SAPPEMDB does not have the right authorization.
    3. The password of the user SAPPEMDB is not password
    4. The syntax is incorrect
    in order to find what is the problem,
    please:
    1. Login in the the DB PRODSQL43:1534 with the user "SAPPEMDB" and the password "password",
    this will eliminate the options 1 - DB down, 2 -SAPPEMDB does not have authorization and 3 - password of the user SAPPEMDB is not password.
    2. If the login failed, than please run sql trace with security elements (in the client there is a tool called "SQL Profiler"
    3. If the login is correct, than you check the syntax of the command:
    "java -Djava.ext.dirs=. -jar SAPtrans.jar import logfile=import.log datafile=init.dat connectstring=com.ddtek.jdbc.sqlserver.SQLServerDriver; jdbc:datadirect:sqlserver://vma03:1433;SAPC11DB;password"
    According to the error message "Error establishing socket. Connection refused"
    it looks like The DB is down or syntax is incorrect.

  • External Data Refresh Failed. We cannot locate a server to load the workbook Data Model. ThisWorkBookDataModel

    Hi All,
    I have been trying to fix this for days now. I have tried solutions in many articles but to no avail. So while the error message is something you may have seen may times, I just can't find a solution in my case.
    This is the error:
    And in text just in case the image isn't viewable:
    "External Data Refresh Failed. We cannot locate a server to load the workbook Data Model. We were unable to refresh one or more data connections in this workbook. The following connections failed to refresh: ThisWorkBookDataModel."
    What is worse is I have checked the ULS (SharePoint Trace Logs), the Event Viewer Logs and the OWA Logs and I cannot find a specific error that would help pin point the problem.
    Excel Workbook
    So what am I doing? I have an Excel 2013 workbook and I create an "SQL Server" connection to the AdventureWorksDW database, add a pivot table and a pivot chart, test in in Excel and all works fine.
    I save the Excel workbook to SharePoint 2013 and then select "Data" then "Refresh All Connections" and then I get the error in the picture above.
    Even more puzzling is I have another Excel workbook that also has pivot tables and pivot charts in the AdventureWorksDW2012Multidimensional cube database in "SQL Analysis Services" and this works fine. Hmmm.
    My Environment
    My environment is Windows 2008 R2 Server, SharePoint 2013 with the April Service Pack1 and a separate server with OWA2013 SP1. It has an SQL Server 2008 R2 database which has been upgraded to SQL Server 2012.
    Data Model Settings
    In Excel Services this is set to my server name which is "server-name". As I do not have instances all I can enter is the server name. As this works everywhere else including the workbook outside of SharePoint I do not think this is the problem.
    But I could be wrong.
    Unattended Account
    I have set this up for the PowerPivot Services App and Excel Services App.
    ODC Connections in Excel
    I have tried all 3 authentication modes, Windows, Secure Store ID and "None" which is the unattended account. I have not tried the other connection types, should I?
    Not in WOPI
    I am not in WOPI mode.
    AD Accounts
    I have added permissions in the SharePoint Services and SQL Server, and as they work in Excel outside of SharePoint, I do not think it is a permissions issue. I could be wrong of course, but the problem is in one of SharePoint, OWA, AD,
    SQL Server, Excel, and Windows Server.
    Isolate the Error
    Below is a list of errors I think are relevant but they do not tell me much. The SharePoint logs are not really giving me an error that tells me what to do and where to do it, or even why it cannot refresh, (perhaps not noticeable by the untrained eye).
    Problem with SQL Server Not Analysis Services
    So my cube database in analysis services works fine in SharePoint/OWA but not the databases in sql server. This is my best clue but I have no idea what it means. Why would it work with an Analysis Services connection but not an "SQL Server" connection?
    It Works Outside of SharePoint
    If I run the excel worksheet outside of SharePoint all works fine. When inside OWA this is where the refresh error occurs.
    Errors from Event Viewer on SharePoint Server using ULS Viewer
    "Failed to create an external connection or execute a query. Provider message: There are no servers available or actively being initialized., ConnectionName: , Workbook:"
    "Refresh failed for 'ThisWorkbookDataModel' in the workbook 'http://server...'. [Session: 1.V22.26itT0lx8piNFeqtuGVhN214.5.en-US5.en-US36.98c0e158-9113-46e9-850e-edda81d9ed1c1.A1.N User: 0#.w|ad\testuser1]"
    And an error in the ULS under the "Data Model" category:
    "--> Check Deployment Mode (server-name): Fail (Expected: SharePoint, Actual: Multidimensional)."
    This last error, as it turned out, defined the problem concisely, although I was yet to work out what it meant in some detail.

    I finally solved this myself (or should I say with the help of several key articles).
    The refresh did not work because the database was not in "SharePoint Mode". Yes, SQL Server has modes, 3 of them in fact.
    If you installed SharePoint to the default SQL instance which would be called <servername> then you cannot use this default instance for Excel 2013 workbooks in OWA 2013 because the refresh only works if the database is in SharePoint mode.
    So what are these 3 modes? The Deployment Mode property in the msmdsrv.ini file has them as:
    0 = Multidimensional mode (the default whenever you install SQL Server normally)
    1 = PowerPivot for SharePoint mode
    2 = Tabular mode mode
    How do you know what mode you are in? That's easy, open SQL Studio Manager and connect to all your SQL database engine instances (ignore Analysis Services or SSRS as they are not database engines). If you only have the default instance then that is almost
    definately in Multidimensional mode which is the default and what SharePoint installs its databases to.
    You must have an instance called <servername>\POWERPIVOT. This instance is the "sharepoint mode" needed, and the default instance name when you install an SQL instance in this mode.
    If you don't see <servername>\POWERPIVOT in SQL server then you are not in "sharepoint mode". It is more accurate to say, you do not have an instance that is in sharepoint mode. This is because you cannot simply switch modes on an SQL server.
    You have to install a new instance in the required mode, thats the only way.
    That's easy enough. Load up the SQL Server setup CD and run setup. Install a brand new instance and select "SQL Server PowerPivot for SharePoint" when you get there in the wizard.
    Now you will have the default instance that stores all the SharePoint databases and that is in mode 0, and a new instance called <servername>\POWERPIVOT that is in mode 1. The "<servername>\POWERPIVOT" instance connection is what you
    will use for Excel 2013 when rendering in OWA 2013.
    You also need to ensure OWA 2013 is not in WOPI mode for Excel worksheets. See the last link below for more information about WOPI.
    Next you should go to the Excel Service App in CA and click Data Model Settings and add the <servername>\POWERPIVOT instance.
    Then you have to either turn off the firewall on the SQL server machine, or create an inbound rule on the Windows firewall to open the TCP port for the <servername>\POWERPIVOT instance:
    1. Start Task Manager and then click Services to get the PID of the MSOLAP$InstanceName.
    2. Run netstat –ao –p TCP from the command line to view the TCP port information for that PID.
    Finally, you can now create Excel 2013 workbooks that run in OWA without refresh errors, as long as you are connecting to the <servername>\POWERPIVOT instance. Hooray.
    REFERENCES
    Look for the string "There are no servers available or actively being initialized" in this article:
    http://blogs.msdn.com/b/analysisservices/archive/2012/08/02/verifying-the-excel-services-configuration-for-powerpivot-in-sharepoint-2013.aspx
    Determine the server mode:
    http://msdn.microsoft.com/en-au/library/gg471594(v=sql.110).aspx
    Install the SharePoint PowerPivot instance (aka SharePoint mode)
    http://msdn.microsoft.com/en-au/library/eec38696-5e26-46fa-bc83-aa776f470ce8(v=sql.110)
    Open the port for the new SQL instance:
    http://msdn.microsoft.com/en-us/library/ms174937(v=sql.110).aspx
    Turn Off WOPI for Excel OWA
    http://blogs.technet.com/b/excel_services__powerpivot_for_sharepoint_support_blog/archive/2013/01/31/powerpivot-for-sharepoint-browser-refresh-fails-data-refresh-not-supported-in-office-web-apps.aspx

  • Accrual Reconciliation Rebuild - No Report/Accrual Reconciliation Report

    Hi Experts,
    It is well-known that Accrual Reconciliation Report shows out the data from po_accrual_reconcile_temp_all i.e. only that data which resides in po_accrual_reconcile_temp_all table will be picked and showed on the report based on net_po_line_amount and net_po_line_quantity (i.e. depends on the tolerance levels given). But if I want to show the data from backend thru query, i will use this table and give my conditions for net_po_line_amount and net_po_line_quantity. As mentioned in Note:433683.1 in Metalink, it is mentioned that:
    The following two columns in the PO_ACCRUAL_RECONCILE_TEMP_ALL table
    determines whether a Transaction appears on the Accrual Reconciliation
    Report.
    Net_po_line_quantity
    Net_po_line_amount
    For a Transaction not to appear on the Accrual Reconciliation Report
    net transaction quantity AND the net transaction amount for each PO Line must be
    less than or equal to the values entered in the tolerance parameters (or <= 0,
    which is the default value).
    This means I will make my query which should match with the report data:
    select * from po_accrual_reconcile_temp_all
    where net_po_line_amount>0 and net_po_line_quanity>0;
    But I have seen some cases where both these are zero and still shown on the report, i have seen few more cases where these are null and still shown on the report. May be we need to add more conditions or somewhere I am doing wrong! Please suggest what all conditions we should consider while incorporating with this type of requirement.
    Any suggestions on this will be highly appreciated.
    Thanks in advance!
    Regards,
    Sumir Chawla

    Hi,
    Pleaser refer the below documentation ...
    Accrual Reconciliation Report Errored Out With APP-FND-01564: ORACLE error 942 in FDFGDF [ID 1440284.1]
    Regards,
    Ivruksha

  • Error while loading  the master data

    Hi Masters,
              When I tried to load the master data, data was not loaded. Monitor tells, in YELLOW color (0 records from 0). Status telling following things
    Diagnosis
    No errors could be found. The current process has probably not finished yet.
    System response
    The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
    and/or
    the maximum wait time for this request has not yet run out
    and/or
    the batch job in the source system has not yet ended.
    Current status
    No Idocs arrived from the source system.
    Could you please help me...
    Thanks & Regards
    Raja.S

    Hi,
    Check the TRFC log.
    You can get to this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Removing errors:
    If the TRFC is incorrect, check whether the source system is completely connected to the SAP BW. Check especially the authorizations of the background user in the source system.
    Resolutions:
    Reload the Master Data manually again from Info-package at RSA1.

  • Performance problem in loading the Mater data attributes 0Equipment_attr

    Hi Experts,
    We have a Performance problem in loading the Mater data attributes 0Equipment_attr.It is running with psuedo delta(full update) the same infopakage runs with diffrent selections.The problme we are facing is the load is running 2 to 4 hrs in the US morning times and when coming to US night times it is loading for 12-22 hrs and gettin sucessfulluy finished. Even it pulls (less records which are ok )
    when i checked the R/3 side job log(SM37) the job is running late too. it shows the first and second i- docs coming in less time and the next 3and 4 i- docs comes after 5-7 hrs gap to BW and saving in to PSA and then going to info object.
    we have userexits for the data source and abap routines but thay are running fine in less time and the code is not much complex too.
    can you please explain and suggest the steps in r/3 side and bw side. how can i can fix this peformance issue
    Thanks,
    dp

    Hi,
    check this link for data load performance. Under "Extraction Performance" you will find many useful hints.
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3a699d90-0201-0010-bc99-d5c0e3a2c87b
    Regards
    Andreas

  • SQL Loader - Field in data file exceeds maximum length

    Dear All,
    I have a file which has more than 4000 characters in a field and I wish to load the data in a table with field length = 4000. but I receive error as
    Field in data file exceeds maximum lengthThe below given are the scripts and ctl file
    Table creation script:
    CREATE TABLE "TEST_TAB"
        "STR"  VARCHAR2(4000 BYTE),
        "STR2" VARCHAR2(4000 BYTE),
        "STR3" VARCHAR2(4000 BYTE)
      );Control file:
    LOAD DATA
    INFILE 'C:\table_export.txt'
    APPEND INTO TABLE TEST_TAB
    FIELDS TERMINATED BY '|'
    TRAILING NULLCOLS
    ( STR CHAR(4000) "SUBSTR(:STR,1,4000)" ,
    STR2 CHAR(4000) "SUBSTR(:STR2,1,4000)" ,
    STR3 CHAR(4000) "SUBSTR(:STR3,1,4000)"
    )Log:
    SQL*Loader: Release 10.2.0.1.0 - Production on Mon Jul 26 16:06:25 2010
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Control File:   C:\TEST_TAB.CTL
    Data File:      C:\table_export.txt
      Bad File:     C:\TEST_TAB.BAD
      Discard File:  none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 0
    Bind array:     64 rows, maximum of 256000 bytes
    Continuation:    none specified
    Path used:      Conventional
    Table TEST_TAB, loaded from every logical record.
    Insert option in effect for this table: APPEND
    TRAILING NULLCOLS option in effect
       Column Name                  Position   Len  Term Encl Datatype
    STR                                 FIRST  4000   |       CHARACTER           
        SQL string for column : "SUBSTR(:STR,1,4000)"
    STR2                                 NEXT  4000   |       CHARACTER           
        SQL string for column : "SUBSTR(:STR2,1,4000)"
    STR3                                 NEXT  4000   |       CHARACTER           
        SQL string for column : "SUBSTR(:STR3,1,4000)"
    value used for ROWS parameter changed from 64 to 21
    Record 1: Rejected - Error on table TEST_TAB, column STR.
    Field in data file exceeds maximum length
    MAXIMUM ERROR COUNT EXCEEDED - Above statistics reflect partial run.
    Table TEST_TAB:
      0 Rows successfully loaded.
      1 Row not loaded due to data errors.
      0 Rows not loaded because all WHEN clauses were failed.
      0 Rows not loaded because all fields were null.
    Space allocated for bind array:                 252126 bytes(21 rows)
    Read   buffer bytes: 1048576
    Total logical records skipped:          0
    Total logical records read:             1
    Total logical records rejected:         1
    Total logical records discarded:        0
    Run began on Mon Jul 26 16:06:25 2010
    Run ended on Mon Jul 26 16:06:25 2010
    Elapsed time was:     00:00:00.22
    CPU time was:         00:00:00.15Please suggest a way to get it done.
    Thanks for reading the post!
    *009*

    Hi Toni,
    Thanks for the reply.
    Do you mean this?
    CREATE TABLE "TEST"."TEST_TAB"
        "STR"  VARCHAR2(4001),
        "STR2" VARCHAR2(4001),
        "STR3" VARCHAR2(4001)
      );However this does not work as the error would be:
    Error at Command Line:8 Column:20
    Error report:
    SQL Error: ORA-00910: specified length too long for its datatype
    00910. 00000 -  "specified length too long for its datatype"
    *Cause:    for datatypes CHAR and RAW, the length specified was > 2000;
               otherwise, the length specified was > 4000.
    *Action:   use a shorter length or switch to a datatype permitting a
               longer length such as a VARCHAR2, LONG CHAR, or LONG RAW*009*
    Edited by: 009 on Jul 28, 2010 6:15 AM

  • SQL LOADER Problem when data is loaded but not come in standard formate

    Hi guys,
    I got problem when sql loader run data loaded successfully in table but UOM data not come in standard formate.
    UOM table column contains the Unit of measure data but in my excel sheet it's look like :
    EXCEl SHEET DATA:
    1541GAFB07080          0     Metres
    1541GAFE10040          109.6     Metres
    1541GAFE10050          594.2     Metres
    1541GAFE10070          126.26     Metres
    1541GAFE14040          6.12     Metres
    1541GAFE14050          0     Metres
    1541SAFA05210          0     Metres
    1541SAFA07100          0     Metres
    1551EKDA05210          0     Nos
    1551EKDA07100          0     Nos
    1551EKDA07120          0     Nos
    1551EKDA07140          0     Nos
    1551EKDA07200          0     Nos.
    1551EKDA08160          0     Nos.
    1551EKDA08180          0     Nos.
    1551EKDA08200          0     Nos.
    1551EKDA10080          41     Nos.
    1551EKDA10140          85     Nos.
    .ctl file :
    OPTIONS (silent=(header,feedback,discards))
    LOAD DATA
    INFILE *
    APPEND
    INTO TABLE XXPL_PO_REQUISITION_STG
    FIELDS TERMINATED BY ','
    OPTIONALLY ENCLOSED BY'"'
    TRAILING NULLCOLS
    ( ITEM_CODE CHAR,
    ITEM_DESCRIPTION CHAR "TRIM(:ITEM_DESCRIPTION)",
    QUANTITY,
    UOM,
    NEED_BY_DATE,
    PROJECT,
    TASK_NAME,
    BUYER,
    REQ_TYPE,
    STATUS,
    ORGANIZATION_CODE,
    LOCATION,
    SUBINVENTORY,
    LINE_NO,
    REQ_NUMBER,
    LOADED_FLAG CONSTANT 'N',
    SERIAL_NO "XXPL_PRREQ_SEQ.NEXTVAL",
    CREATED_BY,
    CREATION_DATE SYSDATE,
    LAST_UPDATED_BY,
    LAST_UPDATED_DATE,
    LAST_UPDATED_LOGIN
    Some output came in table like:
    W541WDCA05260 0 Metres|
    W541WDCA05290 3 Metres|
    W541WDCA05264 4 Metres|
    W541WDCA05280 8 Metres|
    1551EADA04240 0 Nos|
    1551EADA07100 0 Nos|
    1551EKDA10080 0 Nos.|
    1551EKDA10080 41 Nos.|
    proble in | delimiter...how to remove it ' | ' from my table when sqlloader program runnig ...... where i can change in .ctl file or excel file....it's urgent guys olz help me ..
    thanks

    Hi,
    How are you extracting the data to Excel sheet ?
    Please check the format type of the column in Excel sheet for UOM.
    There is no issue in the SQL loader control file, but issue is there in your source excel file. (Try using a different method to extract the data to Excel sheet.)
    Regards,
    Yuvaraj.C

  • Most efficient way to load XML file data into tables

    I have a complex XML file running into MBs. I want to load it's data into 7-8 tables.
    Which way will be better:
    1) Use SQL Loader to actually load directly into the 7-8 tables directly by modifying the control card.
    Is this really possible and feasible? I am not even sure about it
    2) Load data as XML Type in a table and register it. Then extract from there to load into various tables.
    Please help. I have to find the most efficient way of doing it.
    Regards,
    Sudhir

    Yes it is possible to use SQL*Loader to parse and load XML, but that is not what it was designed for and so is not recommended. You also don't need to register a schema, just to load/store/parse XML in the DB either.
    So where does that leave you?
    Some options
    {thread:id=410714} (see page 2)
    {thread:id=1090681}
    {thread:id=1070213}
    Those talk some about storage options and reading in XML from disk and parsing XML. They should also give you options to consider. Without knowing more about your requirements for the effort, it is difficult to give specific advice. Maybe your 7-8 tables don't exist and so using Object Relational Storage for the XML would be the best solution as you can query/update tables that Oracle creates based off the schema associated to the XML. Maybe an External Table definition works better for reading the XML into the system because this process will happen just once. Maybe using WebDAV makes more sense for loading XML to be parsed (I don't have much experience with this, just know it is possible from what I've read on the forums). Also, your version makes a difference as you have different options available depending upon the version of Oracle.
    Hope all that helps as a starter.
    Edited by: A_Non on Jul 8, 2010 4:31 PM
    A great example, see the answers by mdrake in {thread:id=1096784}

  • Impact on transactional Loads if Master data is not loaded

    Scenario for your reference:
    The loads to ZPO_LINE from PC3 was failing for past 20 days from The fix was applied on April 21. 
    I need help to decide the effects of not loading ZPO_LINE for past 20 days. And create a detailed plan for data loads required.
    If the master data is not loaded for 20 days will it affect the transaction loads happened during those days?
    And how can i find out the impact and correct the transaction data if it does?
    Can any 1 help me with this?

    Hi,
    If i understand your scenario, you have a scenario where the Master data load has not been updated for the last 20days but, the transaction data was loaded without any interruption.
    In such a scenario, the Transactional loads will only be affected if there is some field which undergoes transformation after looking up on the master data.
    So, first load the master data and run the attrib change run process.
    After this, if the transaction data is in full update mode, then you dont need to do anything as data will be refreshed with correct values in the next load.
    If delta loads are there, you might need to performa  full repair.
    Regards,
    Rahul

Maybe you are looking for