Data Loads Stuck in RSSELDONE

Hi All,
We have a scenario in our Q system where we cannot see the data load request in RSMO. When we click on a request (which does not have a status light), it doesn't show anything on the right side pane and will eventually timeout after a while. Analyzing the short dump we found that the RSSELDONE table has a lot of entries and the selections in Infopackage is about 80000.
RSREQDONE doesnot have any status for these requests. The requests were created a week back and couple of days back the system was also bounced. But still we cannot see the request in RSMO to turn it to red. Further loads from the same data source also doesn't start or doesn't show "the load is requested" message.
We are considering the RSREQARCH_WRITE. But not sure if this would resolve the issue. Anyone faced a similar situation?
All helpful suggestions welcome.
Thank you,
Vijay

Hi,
Probably Ask your Basis to do Database statistics refresh with 80% of table RSSELDONE. This wiould enhance reading performance and reduce the time required to generate new request number.
- Your Basis team needs to look at this problem and search for some related OSS notes. They might need to manipulate some system parameters.
You can also look for the note 620361, which indicates :
Cause and preconditions
You have a data target with a large number of requests.
Since certain information of all requests is required in a data target
with the above-described actions, a relatively large number of entries
must be read and processed in different tables.
As a result, the processing times become longer and longer.
=========================
It also provides a solution to improve the data load performance.
you can delete db statistics of table RSSELDONE, RSLDTDONE, RSREQDONE  and try and load, you might be lucky.
Regards,
Bhargava

Similar Messages

  • Data load stuck from DSO to Master data Infoobject

    Hello Experts,
    We have this issue where data load is stuck between a DSO and master data infoobject
    Data uploads from DSO( std) to master data infoobject.
    This Infoobject has display and nav attributes in it which are mapped from DSO to Infoobject.
    Now we have added a new infoobject as attribute to the master data infoobject and made it as NAV attri.
    Now when we are doing full load via DTP the load is stuck and is not processing.
    Earlier it took only 5 mns of time to complete the full load.
    Please advise what could be the reason and cause behind this.
    Regards,
    santhosh.

    Hello guys,
    Thanks for the quick response.
    But its nothing proceeding further.
    The request is still running.
    earlier this same data is loaded in 5 mns.
    Please find the screen shot.
    master data for the infoobjects are loaded as well.
    I can see in SM50 the process at P table of the infoobject the process is.
    Please advise.
    Please find the detials
    Updating attributes for InfoObject YCVGUID
    Start of Master Data Update
    Check Duplicate Key Values
    Check Data Values
    Process time dependent attributes- green.
    No Message: Process Time-Dependent Attributes- yellow
    No Message: Generates Navigation Data- yellow
    No Message: Update Master Data Attributes - yellow
    No Message: End of Master Data Update - yellow
    and nothing is going further in Sm37
    Thanks,
    Santhosh.

  • IDOC RSRQST is getting stuck frequently, it leads to delay in data loading

    hi,
    we are using SAP R/3 4.7 E and SAP Netweaver 2004s as BI server .
    On daily basis we are loading the data from R/3 to BI system . we are facing problems in data loading
    IDOC RSRQST is getting stuck frequently, which leads to delay in data loading..
    Please guide me to resolve this issue .
    Thanks in advance .

    Thanks for your reply kumarsen   . I have gone through the referred forum .
    In my case , I have checked the TRFC,  conncetivity between R/3 and BW  and Job scheduled as priority C .
    Still many Idocs are got stuck in R/3 [approx every 5 min once] . we are releasing the Idoc's through BD87 only .
    Please suggest for some permanent solution .

  • Partial data loading from dso to cube

    Dear All,
    I am loading the data from dso to the cube through DTP. The problem is that some records are getting added  through data
    packet 1 to the cube around 50 crore records while through data packet2 records are not getting added.
    It is full load from dso to cube.
    I have tried deleting the request and executing DTP but again the same no of records get added to the cube through datapacket 1   and after that records are not added to the cube through datapacket 1  ;request remains in yellow state only.
    Please suggest .

    Nidhuk,
    Data load transfers package by package. Your story sounds like it got stuck in second package or something. Suggest you check package size, try to increase to a higher number to see if anything  changes ? 50 records per package kind of low, your load should not spread out into too many packages.
    Regards. Jen
    Edited by: Jen Yakimoto on Oct 8, 2010 1:45 AM

  • Data Load : Number of records count

    Hi Experts,
              I want to document number of records transferred to BW during an infopackage execution.
              I want to automate the process by running a report in background which will fetch a data from SAP tables about number of records been transfered by all my InfoPackage .
    I would like to know how should I proceed with.
             I want to know some System tables which contains same data as that of RSMO transaction displays to us.
    Kindly help with valuable replies.

    HI,
    inorder to get the record counts report you need to create a report based on below tables
    rsseldone, rsreqdone, rsldpiot, rsmonfact
    Check the below link which explain in detail with the report code as well.
    [Data load Quick Stats|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/90215bba-9a46-2a10-07a7-c14e97bdb764]
    This doc also explains how to trigger a mail with the details to all.
    Regards
    KP

  • How to identify the maximum data loads done for a particular day?

    Hi all,
    There is huge volume of data been loaded on a last monday, and then data was deleted from the cubes on the same day. And hence i needs to see which are all the cubes which was loaded wtih lot of records for a particular day,
    i happened to look at rsmo, i am unable to see the ods nor the cube data loads,
    were do i seet it?
    Thanks

    See if the table RSSELDONE helps. This will give you the recent data load details. Based on those loads , you can search the targets.
    And also check table TBTCO  which will give the latest job details. You will have to analyze the same jobs to know what loads were done . Give a selection for date.

  • ERPI data load error

    Hello,
    We are loading one month data from peoplesoft to planning application using ERPI 11.1.2.1. It takes about 1 hrs to load one month data.Is it normal? When looking at ODI PS_GL_LOAD_BALANCES_DATA is always successful but COMM_LOAD_BALANCES fails at the end with this error.Did anyone have this error before. Message in HyS9aifWeb-sysout file is:
    FROM AIF_HS_BALANCES
    WHERE LOADID = 132
    [EssbaseRuleFile] Locking rules file AIFData
    [EssbaseRuleFile] Successfully locked rules file AIFData
    [EssbaseRuleFile] Copying rules file OWBPLAND for data load as AIFData
    [EssbaseRuleFile] Unlocking rules file AIFData
    [EssbaseRuleFile] Successfully unlocked rules file AIFData
    [EssbaseRuleFile] The data rules file has been created successfully.
    [EssbaseRuleFile] Locking rules file AIFData
    [EssbaseRuleFile] Successfully locked rules file AIFData
    [EssbaseRuleFile] Load data into the cube by launching rules file...
    <Mar 1, 2012 1:09:30 PM PST> <Error> <WebLogicServer> <BEA-000337> <[STUCK] ExecuteThread: '5' for queue: 'weblogic.kernel.Default (self-tuning)' has been busy for "625" seconds working on the request "weblogic.servlet.internal.ServletRequestImpl@3bc2189[
    POST /aif/services/HPLService HTTP/1.1
    Content-type: text/xml;charset="utf-8"
    Accept: text/xml, multipart/related, text/html, image/gif, image/jpeg, *; q=.2, */*; q=.2
    Soapaction: "http://DEVAPPHPN001.indymacdev.biz:19000/aif/services/HPLServiceexecuteDataLoad"
    User-Agent: JAX-WS RI 2.1.7-b01-
    Content-Length: 638
    Connection: Keep-Alive
    Proxy-Client-IP: 10.205.101.108
    X-Forwarded-For: 10.205.101.108
    X-WebLogic-KeepAliveSecs: 30
    X-WebLogic-Force-JVMID: 2124105652
    ]", which is more than the configured time (StuckThreadMaxTime) of "600" seconds. Stack trace:
    Thread-176 "[STUCK] ExecuteThread: '5' for queue: 'weblogic.kernel.Default (self-tuning)'" <alive, suspended, priority=1, DAEMON> {
    jrockit.net.SocketNativeIO.readBytesPinned(SocketNativeIO.java:???)
    jrockit.net.SocketNativeIO.socketRead(SocketNativeIO.java:24)
    java.net.SocketInputStream.socketRead0(SocketInputStream.java:???)
    java.net.SocketInputStream.read(SocketInputStream.java:107)
    com.essbase.services.olap.main.main_direct.EssNetClient.readFrom(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssNetClient.readPackage(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssNetClient.receiveResponse2(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssNetClient.adNetReceiveResponse(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssAPIData._adImport(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssAPIData.ImportToEssbaseASO(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssMAPIDir.BeginDataload(Unknown Source)
    com.essbase.server.framework.EssOlapMainService.BeginDataload(Unknown Source)
    com.essbase.server.framework.EssOrbPluginDirect.essMainBeginDataload(Unknown Source)
    com.essbase.api.session.EssOrbPlugin._invokeMainMethod(Unknown Source)
    com.essbase.api.session.EssOrbPlugin._invokeMethod2(Unknown Source)
    com.essbase.api.session.EssOrbPlugin._invokeMethod(Unknown Source)
    com.essbase.server.framework.EssOrbPluginDirect._invokeProtected(Unknown Source)
    com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
    com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
    com.essbase.api.session.EssOrbPlugin.essMainBeginDataload(Unknown Source)
    com.essbase.api.datasource.EssCube.beginDataload(Unknown Source)
    com.hyperion.aif.essbase.EssbaseRuleFile.executeDataRuleFile(EssbaseRuleFile.java:314)
    com.hyperion.aif.webservices.HPLService.executeDataLoad(HPLService.java:167)
    sun.reflect.GeneratedMethodAccessor7310.invoke(Unknown Source)
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    java.lang.reflect.Method.invoke(Method.java:575)
    org.apache.axis.providers.java.RPCProvider.invokeMethod(RPCProvider.java:397)
    org.apache.axis.providers.java.RPCProvider.processMessage(RPCProvider.java:71)
    org.apache.axis.providers.java.JavaProvider.invoke(JavaProvider.java:265)
    org.apache.axis.strategies.InvocationStrategy.visit(InvocationStrategy.java:32)
    org.apache.axis.SimpleChain.doVisiting(SimpleChain.java:113)
    org.apache.axis.SimpleChain.invoke(SimpleChain.java:78)
    org.apache.axis.handlers.soap.SOAPService.invoke(SOAPService.java:435)
    org.apache.axis.server.AxisServer.invoke(AxisServer.java:132)
    org.apache.axis.transport.http.AxisServlet.doPost(AxisServlet.java:586)
    javax.servlet.http.HttpServlet.service(HttpServlet.java:700)
    org.apache.axis.transport.http.AxisServletBase.service(AxisServletBase.java:325)
    javax.servlet.http.HttpServlet.service(HttpServlet.java:815)
    weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:224)
    weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:108)
    weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:206)
    weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:111)
    oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:299)
    oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:405)
    oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:94)
    oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:98)
    oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:70)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:86)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:25)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3687)
    weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3681)
    weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:308)
    weblogic.security.service.SecurityManager.runAs(SecurityManager.java:116)
    weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2213)
    weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2135)
    weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1420)
    weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    weblogic.work.ExecuteThread.run(ExecuteThread.java:168)
    Any feedback is appreciated.Thank you for taking ime to read through.
    Regards.

    Hello,
    Part of the message you are getting is because the WebLogic process is a bit intuitive. It knows when something is running and waits. It also assumes anything running for a certain period of time is considered "stalled/stuck" and reports it. The time it waits is dependant on the Weblogic configuration.
    As for the time it takes to load data... that is very vague. Without knowing more about your environment that is very hard to state. We know nothing about the hardware, platform, dataset size, etc. You would need to review ODI to find out what is taking the long amount of time and try and replicate it outside of ERPi/ODI to see if you get the same results.
    Thank you,

  • What is the Short Key,For checking the check box in Data Loader ?

    When i am Transfering the data from a legacy to Oracle Apps DB, Using Data Loader tool.I stuck with an issue his i need to enable the check in a particular form using the data loader,Is there a short key Like Enter( ENT), Save and Proceed(*SP) .. Needed Badly ...
    Thank you ..

    You can be able to enable a check box or disable a check box by using the space bar, you make check box item by navigating using tab and use Space bar *SB.
    try this out
    Regards
    Ramesh Kumar S

  • SPM: Data Load Program / API Codes

    Hi Experts,
    In our Project,
    In Order to analyze an SPM Data Load Issue, we need to analyze Program / API Codes that Load Data into SPM Inbound Layer.
    How / where can we find the details of those Programs / API Codes ?
    Regards
    Pankaj

    Hi Neil,
    Thanks for your response. This is the Issue we are stuck currently, and need urgent help.
    Can you let me know which program loads the Data from Flat file into the Inbound DSO, so that we can debug the same. We are facing the issue with currency conversion. Overall there are 10 Currencies, but for EUR Currency, conversion is not happening, but its happening for others. so we need to debug the program to find the root cause. We have done all basic checks, like TCURR Table, OB08 etc.
    Really appreciate your help.
    Regards
    Pankaj

  • Error is Master Data load.

    Hi Experts,
    From last 1 month the ( 0MAT_UNIT ) master data load is getting failed. Its a full load. So the the target is not updated from last one month( August)
    Now the problem is , when I am checking a report. The materials which are new or changed are not getting the data properly. It is because of the master data got stucked at PSA level.And I cant rectify at PSA level because, There is 673 errorneous record in 1 package.
    I tried to find the invalid character , but didn't get anything.
    Please suggest.
    Thanks,
    TapuKumar

    Hi Anand,
            This is the long text msg
    Data records in the PSA were marked for data package 1 . Here there were
    673 errors. The system wrote information or warnings for the remaining 
    data records.
                                                                                    System response                                                                               
    The data package was not updated.                                                                               
    Procedure                                                                               
    Correct the incorrect data records of the data package (for example by   
         manually editing them in the PSA maintenance). You can find the error    
         message for each of the records in the PSA by double-clicking on the     
         status of the record.                                                                               
    Note                                                                               
    Note that you are only allowed to edit and manually update the request   
         if it is contained in the request list in the data target. Where         
         necessary, first delete it from the data target.                         
    And it is a master data load( Full ), The data target is a INFO OBJECT.
    Please suggest, I can't rectify each and every reecord. Because the number of record id very high..
    Thanks,
    Tapu Kumar.
    Edited by: NewToBW on Sep 2, 2010 1:42 PM

  • Data Load - ALE inbox of SAP BW is identical to ALE outbox of the SS

    Hello,
    I am doing a init for a cube, and the load is still running without ending.
    no errors given but when I look at the step by stpe analysis it gives the following message.
    Diagnosis
        No errors could be found. The current process has probably not finished
        yet.
    System response
        The ALE inbox of the SAP BW is identical to the ALE outbox of the source
        system
        and/or
        the maximum wait time for this request has not yet run out
        and/or
        the batch job in the source system has not yet ended.
    Thanks,
    VJ

    Hi Vijay,
    The message you are seeing is a standard message that is shown while the data load is in running state in the monitor.
    Periodically refresh the monitor screen and check whether records are getting extracted.Example if records getting extracted really slow,you will see the figure 'XYZ(records) from XYZ' slowly increasing.
    Also keep checking detail tab to see if steps are slowly turning GREEN or not.
    If load still stuck in yellow,maybe u need to re-run it and look into performance tuning options.
    cheers,
    Vishvesh

  • Data load problem - BW and Source System on the same AS

    Hi experts,
    I’m starting with BW (7.0) in a sandbox environment where BW and the source system are installed on the same server (same AS). The source system is the SRM (Supplier Relationship Management) 5.0.
    BW is working on client 001 while SRM is on client 100 and I want to load data from the SRM into BW.
    I’ve configured the RFC connections and the BWREMOTE users with their corresponding profiles in both clients, added a SAP source system (named SRMCLNT100), installed SRM Business Content, replicated the data sources from this source system and everything worked fine.
    Now I want to load data from SRM (client 100) into BW (client 001) using standard data sources and extractors. To do this, I’ve created an  InfoPackage in one standard metadata data source (with data, checked through RSA3 on client 100 – source system). I’ve started the data load process, but the monitor says that “no Idocs arrived from the source system” and keeps the status yellow forever.
    Additional information:
    <u><b>BW Monitor Status:</b></u>
    Request still running
    Diagnosis
    No errors could be found. The current process has probably not finished yet.
    System Response
    The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
    and/or
    the maximum wait time for this request has not yet run out
    and/or
    the batch job in the source system has not yet ended.
    Current status
    No Idocs arrived from the source system.
    <b><u>BW Monitor Details:</u></b>
    0 from 0 records
    – but there are 2 records on RSA3 for this data source
    Overall status: Missing messages or warnings
    -     Requests (messages): Everything OK
    o     Data request arranged
    o     Confirmed with: OK
    -     Extraction (messages): Missing messages
    o     Missing message: Request received
    o     Missing message: Number of sent records
    o     Missing message: Selection completed
    -     Transfer (IDocs and TRFC): Missing messages or warnings
    o     Request IDoc: sent, not arrived ; Data passed to port OK
    -     Processing (data packet): No data
    <b><u>Transactional RFC (sm58):</u></b>
    Function Module: IDOC_INBOUND_ASYNCHRONOUS
    Target System: SRMCLNT100
    Date Time: 08.03.2006 14:55:56
    Status text: No service for system SAPSRM, client 001 in Integration Directory
    Transaction ID: C8C415C718DC440F1AAC064E
    Host: srm
    Program: SAPMSSY1
    Client: 001
    Rpts: 0000
    <b><u>System Log (sm21):</u></b>
    14:55:56 DIA  000 100 BWREMOTE  D0  1 Transaction Canceled IDOC_ADAPTER 601 ( SAPSRM 001 )
    Documentation for system log message D0 1 :
    The transaction has been terminated.  This may be caused by a termination message from the application (MESSAGE Axxx) or by an error detected by the SAP System due to which it makes no sense to proceed with the transaction.  The actual reason for the termination is indicated by the T100 message and the parameters.
    Additional documentation for message IDOC_ADAPTER        601 No service for system &1, client &2 in Integration Directory No documentation exists for message ID601
    <b><u>RFC Destinations (sm59):</u></b>
    Both RFC destinations look fine, with connection and authorization tests successful.
    <b><u>RFC Users (su01):</u></b>
    BW: BWREMOTE with profile S_BI-WHM_RFC (plus SAP_ALL and SAP_NEW temporarily)
    Source System: BWREMOTE with profile S_BI-WX_RFCA (plus SAP_ALL and SAP_NEW temporarily)
    Someone could help ?
    Thanks,
    Guilherme

    Guilherme
    I didn't see any reason why it's not bringing. Are you doing full extraction or Delta. If delta extraction please check the extractor is delta enabled or not. Some times this may cause problems.
    Also check this weblog on data Load errors basic checks. it may help
    /people/siegfried.szameitat/blog/2005/07/28/data-load-errors--basic-checks
    Thanks
    Sat

  • BI 7.0 data load issue: InfoPackage can only load data to PSA?

    BI 7.0 backend extraction gurus,
    We created a generic datasource on R3 and replicated it to our BI system, created an InfoSource, the Transformation from the datasource to the InfoSource, an ODS, the transformation from the InfoSource to the ODS. 
    After the transformation creation between the InfoSource and the ODS is done on this BI system, a new folder called "Data Transfer Process" is also created under this ODS in the InfoProvider view.  In the Data Transfer Process, in the Extraction tab, picks 'Full' in the field Extraction Mode, in the Execute tab, there is a button 'Execute', click this button (note: so far we have not created InfoPackage yet) which sounds like to conduct the data load, but find there is no data available even if all the status show green (we do have a couple of records in the R3 table). 
    Then we tried to create an InfoPackage, in the Processing tab, find 'Only PSA' radio button is checked and all others like 'PSA and then into Data Targets (Package by Package)" are dimmed!  In the Data Target tab, find the ODS as a target can't be selected!  Also there are some new columns in this tab, 'Maintain Old Update Rule' is marked with red color 'X', under another column 'DTP(S) are active and load to this target', there is an inactive picture icon, that's weird since we have already activated the Data Transfer Process!  Anyway, we started the data load in the InfoPackage, and the monitor shows the records are brought in, but since in the Process tab in the InfoPackage, 'Only PSA' radio button is checked with all others dimmed that there is no any data going to this ODS!  Why in BI 7.0, 'Only PSA' radio button can be checked with others all dimmed?
    Many new features with BI 7.0!  Any one's idea/experience is greatly appreciate on how to load data in BI 7.0!

    You dont have to select anything..
    Once loaded to PSA in DTP you have the option of FULL or DELTA ,full loads all the data from PSA and DELTA loads only the last load of PSA.
    Go through the links for Lucid explainations
    Infopackage -
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/03808225cf5167e10000000a1553f6/content.htm
    DTP
    http://help.sap.com/saphelp_nw2004s/helpdata/en/42/f98e07cc483255e10000000a1553f7/frameset.htm
    Creating DTP
    http://help.sap.com/saphelp_nw2004s/helpdata/en/42/fa50e40f501a77e10000000a422035/content.htm
    <b>Pre-requisite-</b>
    You have used transformations to define the data flow between the source and target object.
    Creating transformations-
    http://help.sap.com/saphelp_nw2004s/helpdata/en/f8/7913426e48db2ce10000000a1550b0/content.htm
    Hope it Helps
    Chetan
    @CP..

  • Open HUB ( SAP BW ) to SAP HANA through DB Connection data loading , Delete data from table option is not working Please help any one from this forum

    Issue:
    I have SAP BW system and SAP HANA System
    SAP BW to SAP HANA connecting through a DB Connection (named HANA)
    Whenever I created any Open Hub as Destination like DB Table with the help of DB Connection, table will be created at HANA Schema level ( L_F50800_D )
    Executed the Open Hub service without checking DELETING Data from table option
    Data loaded with 16 Records from BW to HANA same
    Second time again executed from BW to HANA now 32 records came ( it is going to append )
    Executed the Open Hub service with checking DELETING Data from table option
    Now am getting short Dump DBIF_RSQL_TABLE_KNOWN getting
    If checking in SAP BW system tio SAP BW system it is working fine ..
    will this option supports through DB Connection or not ?
    Please follow the attachemnet along with this discussion and help me to resolve how ?
    From
    Santhosh Kumar

    Hi Ramanjaneyulu ,
    First of all thanks for the reply ,
    Here the issue is At OH level ( Definition Level - DESTINATION TAB and FIELD DEFINITION )
    in that there is check box i have selected already that is what my issue even though selected also
    not performing the deletion from target level .
    SAP BW - to SAP HANA via DBC connection
    1. first time from BW suppose 16 records - Dtp Executed -loaded up to HANA - 16 same
    2. second time again executed from BW - now hana side appaended means 16+16 = 32
    3. so that i used to select the check box at OH level like Deleting data from table
    4. Now excuted the DTP it throws an Short Dump - DBIF_RSQL_TABLE_KNOWN
    Now please tell me how to resolve this ? will this option is applicable for HANA mean to say like , deleting data from table option ...
    Thanks
    Santhosh Kumar

  • Data load times

    Hi,
    I have a question regarding data loads. We have a process cahin which includes 3 ods and cube.
    Basically ODS A gets loaded from R/3 and the from ODS A it then loads into 2 other ODS B, ODS C and CUBE A.
    So when I went to monitor screen of this load ODS A-> ODS B,ODS C,CUBE A. The total time shows as 24 minutes.
    We have some other steps in process chain where ODS B-> ODS C, ODS C- CUBE 1.
    When I go the monitor screen of these data loads the total time the tortal time  for data loads show as 40 minutes.
    I *am suprised because the total run time for the chain itself is 40 minutes where in the chain it incclude data extraction form R/3 and ODS's Activations...indexex....
    *Can anybody throw me some light?
    Thank you all
    Edited by: amrutha pal on Sep 30, 2008 4:23 PM

    Hi All,
    I am not asking like which steps needed to be included in which chain.
    My question is when look at the process chain run time it says the total time is equal to 40 minutes and when you go RSMO to check the time taken for data load from ODS----> 3 other data targets it is showing 40 minutes.
    The process chain also includes ods activation buliding indexex,extracting data from R/3.
    So what are times we see when you click on a step in process chain and displaying messages and what is time you see in RSMO.
    Let's take a example:
    In Process chain A- there is step LOAD DATA- from ODS A----> ODS B,ODS C,Cube A.
    When I right click on the display messages for successful load it shows all the messages like
    Job started.....
    Job ended.....
    The total time here it shows 15 minutes.
    when I go to RSMO for same step it shows 30 mintues....
    I am confused....
    Please help me???

Maybe you are looking for

  • How to unlock my iphone if i forgot my pin

    I managed to activate the phone locking feature and forgot the pin I used. Now the phone is locked and I don't know what to do. Please help

  • Program not working since 8.0.8 update

    The application has not worked (connected my mac to windows 7 PC) since the 8.0.8 update. 8.0.9 has since been released and has still not fixed the problem. Any suggestions as to what is wrong? I see this is a common issue with other users.

  • 4G future devices

    I was just wondering and please excuse me if this has been addressed. I know that the smart phones are going to eventually all be 4G. Will eventually all Verizon be 4G. Or is it only smart phones that will be 4G. I have a smart phone right now and li

  • Changing ORMI port !!!

    Hi All, I had OBIEE and Oracle warehouse builder both running on the same server as both were using OC4J service on port 23791, I changed the ORMI port in OWB_HOME\owb\jrt\config\rmi.xml configuration file from 23791 to 24791. I started OC4j for OBIE

  • Cancellation of my account

    I wish to cancell skype services. My name is [removed to maintain privacy], my skype name is [removed to maintain privacy]. I can not get suffient information on my account with skype or a satisfied answer to my questions. Thank you