Push from ODS to cube taking a long time

Hi All,
I've created a new ODS that pushes data to a new cube.  I am trying to load 3000 records and its taking over 30 minutes, and I can never get it to complete successfully.  When I look in sm50, there is nothing running.  The are no short dumps in st22. But when I look in sm21 I get the following:
Documentation for system log message R6 8 :                                                
An error has causes an SAP rollback.  All database updates are reset.                                                                               
Technical details                                                                          
File................ 009581                                                                
Position............ 0000072900                                                            
Entry type.......... m      ( Error (Function,Module,Row)    )                             
Message ID.......... R6 8                                                                  
Variable parts......                                       ThIRollroll bathxxhead1248                                                                               
My update rules between the ODS and cube are very straightforward.  All fields are a direct mapping to the ODS, no routines.  I have a start routine, but when I comment out the entire start routine and run the update between ODS and cube, I still get the same issue.  We are on version 3.5.
Any suggestions?
Thanks
Charla

Hi,
Try "generate export data source" from ODS and load the data. Sometimes there is problem in DDIC tables and structures are not generated properly.
Regards,
Kams

Similar Messages

  • Transferring to my new iMac from my MacBook is taking a long time. Is this normal?

    Transferring from my old MacBook to my me iMac is taking a long time. Is this normal?

    See Pondini's Setup New Mac guide for starters.

  • Error when trying to load data from ODS to CUBE

    hi,
      Iam getting a short dump  when trying to load data from ODS to CUBE. The Run time error is 'TYPELOAD_NEW_VERSION' and the short text is 'A newer version of data type "/BIC/AZODS_CA00" was found than one required.please help me out.

    Hi,
    Check this thread.........Ajeet Singh  has given a good solution here.........
    Re: Error With Data Load-Getting Canceled Right Away
    Also check SAP note: 382480..................for ur reference............
    Symptom
    A DART extraction job terminates with runtime error TYPELOAD_NEW_VERSION and error message:
    Data type "TXW_INDEX" was found in a newer version than required.
    The termination occurs in the ABAP/4 program "SAPLTXW2 " in "TXW_SEGMENT_RECORD_EXPORT".
    Additional key words
    RTXWCF01, LTXW2U01, TXW_INDEX
    Cause and prerequisites
    This problem seems to happen when several DART extraction jobs are running in parallel, and both jobs access table TXW_INDEX.
    Solution
    If possible, avoid running DART extractions in parallel.
    If you do plan to run such jobs in parallel, please consider the following points:
    In the DART Extract configuration, increase the value of the parameter "Maximum memory allocation for index (MB)" if possible. You can estimate reasonable values with the "File size worksheet" utility.
    Run parallel DART jobs on different application servers.
    As an alternative, please apply note 400195.
    It may help u.........
    Regards,
    Debjani.......

  • SM59 - Remote Logon taking very long time - RFC is receiver

    Hi Friends,
    We have ABAP connection (SM59) to R/3 system from PI system.  The connection test is fine. But, when we click the Remote Logon from SM59, it is taking very long time. It is unable to login.
    Due to this, messages are in scheduled state in the receiver CC monitoring. (SOAP to RFC - Async ). Receiver RFC CC also are green.
    We checked with R/3 team, the remote user which is used in SM59 is not locked in R/3 system and password also not changed.
    Friends, kindly clarify why it is taking very long time from PI in SM59 when the connection is fine.
    Kind regards,
    Jegathees P.

    Hi Sabarish,
    I checked in SM58. There are no entries.  SM59 is for other interfaces (IDoc).  In the Message Monitoring all messages showing 'To be Delivered' and 'Delivering' from yesterday onwards. In the CC monitoring, it shows 'Message Processing Started' from yesterday onwards. But not completed.
    I checked in SM21 take log of type 'RD'.  I doubt is this the correct log for this problem.
    The specified operating system call was returned with an error.        |
    For communication calls (receive, send, etc) often the cause of errors
    are network problems.
    It could also be a configuration problem at operating system level.
    (file cannot be opened, no space in the file system etc.).
    Additional specifications for error number 146
    Name for errno number ECONNREFUSED
    Interprocess communication (e.g. TCP/IP) connection refused by partner.
    |This usually means that the necessary receiver program is not running.
    Kindly clarify.
    Kind regards,
    Jegathees P.

  • How to delete init request when load is taking place from ods to Cube

    Dear All,
    I have initiated the init load from ODS to Cube, but the init request was not successfull... Now on clicking the init infopkg i get the short dump error....
    I am trying to acess this infopkg so that i can delete the old init and reschedule it again...
    Now in order to avoid this error i need to delete the init request somehow..... Since data load is happening in BW system only there is no way to go to delta queue and delete the init...
    How can i delete the init request in this case or how do i avoid this dumup error???
    Thanks & Regards,
    Anup

    From se11, go to the table, see the content... from this window mark the check box for relevant entries in the first column and click on Display button (F7)
    enter "/h" in the command prompt and press enter
    Click the top most green button to execute.
    From the debugging screen, double click on variable "CODE" change the value to "DELE" (from "SHOW") and save (by pressing enter in bi7 and by clicking pencil icon in bw 3.x ).
    Press F8
    In the result screen you will find the delete entry in the application bar
    Delete one by one after ensuring the record content.
    Better you can ask any abaper to delete the table entries in the debug mode coz its a production server..else you can call me 9500066350(I prefer you to go along with abaper if you have no option dont wait for calling).

  • Index for loads from ODS To Cube

    The load from ODS to cube is taking a long time - In the start routine another ODS is being looked up - The keys for look up is say X and Y
    There is already an index existing on keys X , Y & Z -
    Will this index be used while doing the select on that ODS or I need to create a new index with only X and Y keys ?
    Thnx

    When you are running the start routine - run an SQL trace - ST05 - that will tell you if the index is being used.
    Arun

  • Rollup in cube is taking very long time (BW 3.5)

    Hi,
    Rollup in one of our cube is taking a long time from what it is expected.
    Previously this rollup was taking 2-3 hrs but now it is taking more than 8 hrs.
    To solve this I even tried compressing the cube till last month but of no use. (this worked last year when we faced a similar issue but not sure why it wont help this time)
    Through this rollup 2-3 active aggregates are filled up.
    Is there any way to solve this issue.
    I am using BW 3.5 version.
    (PS - Can't use compress after rollup option.)

    Dear,
    When ever you do a rollup of data to its aggregates, the system automatically deletes and rebuilds the indexes of the aggregates.
    Now when you run the program SAP_RSADMIN_MAINTAIN with the above mentioned parameters the system skips this process of deleting and rebuilding the index of aggregates .
    Hence you dont have to do anything manually except to rebuild the indexes of the aggregates(on the "Performance" tab when you "Administrate data targets" in the Admininistrator Workbench) once your process chain has run successfully .
    So the steps should be as follows :
    1) Run the program SAP_RSADMIN_MAINTAIN (One time activity)
    2) Delete index (for infocube in process chain)
    3) Load data into infocube (in process chain)
    4) Create index (for infocube in process chain) .
    5)Rollup (in process chain )
    6)Rebuild the indexes of the aggregates(on the "Performance" tab when you "Administrate data targets" in the Admininistrator Workbench . (Manually) .
    Note : You can always revert back the setting what you are maintaining through program SAP_RSADMIN_MAINTAIN by just filling the same parameters and slecting the Delete radio button.
    Hope this helps .
    Regards .

  • Data loading from ODS to CUBE

    Hi All,
    I have loaded data from ODS to CUBE. now i have requirement to add some fields in the standard cube. so, for testing purpose i have created copy of the original and created transformation . now when i try to load data from ODS it shows me no more data available . while data is already there in ODS.
    now what should i do ? i don't want to delete data from original cube. is there any other way to load data through transformation ?
    Regards,
    Komik Shah

    Hi,
    Check the DTP of old cube n see whether its Delta. If yes then check whether any one of the foll is check:
    1) get delta only once
    2) get data by request.
    If 1 is checked then delta wont come for the second cube as it says to get delta once and delta is already in one of the cube.
    Generally both should be unchecked but can vary as per requirements.
    Now for your new DTP, i dont think it will aloow you to change to FULL.
    If its allowing you to select FULL, then select it and select from acive table.
    try to load and see.
    regds,
    Shashank

  • Error in updating data from ODS to CUBE.

    Hi,
    I am tryin to load data manually from ODS to CUBE in NW2004s.
    This is a flat file load from the datasource to the ODS and then from the ODS to the CUBE.
    In the CUBE, I am trying to populate fields by using the ODS fields.
    For eg.
    In the ODS, a CHAR Infoobject has the data in the timestamp format(i.e. mm/dd/yyyy hh:mm ). I need to split this data and assign them to the two individual DATE and TIME Infoobject in the CUBE.
    For this, I have done the coding in the Transfer Structure in the Rule Group.
    The time field is gettin populated , but the date field is not getting populated.
    I get an error as Eg:
    <b>Value '04052007' for CHAR 0DATE is not plausible</b>
    Due to this, the corresponding records is not getting displayed
    Also, the records where the time id displayed, the date is not getting displayed inspite of the date being correct.
    Please help me with a solution for this.
    <b><u><i>REMOVED</i></u></b>
    Thanks In Advance.
    Hitesh Shetty

    Hello Hitesh
    SAP accepts the date format in YYYYMMDD, so in the routine where you have concatenate the day month year...just do it in reverse order.....
    Thanks
    Tripple k

  • Error while uploading data from ODS to Cube

    Hi All,
    Will you please help out this issue.As this is a priority high issue,please reply if you know the answers.
    I am facing an error while loading the data from ODS to CUBE,the error
    is
    <b>1.Name is not in the namespace for generated BW Metaobjects
    2.Error 18 in the update</b>
    And this error is occuring only in quality server.
    Thanks,
    Ram.

    HI RAM SIVA,
    make sure that the Data source is replicated and tranfer rules r active.
    And also check whether the all transport requests r imported properly.
    hope it helps
    bhaskar

  • Error while loading data from ODS to CUBE.

    Hi friends,
    When l am loading data from ODS to Cube with help of data mart, I am getting error in QA system , IN  DM system ,every thing went well.IF i see the detail tab in monitor under Processing .
    Its is showing like this .
    Transfer Rules : Missing Massage.
    Update PSA : missing massage.
    Processing end : missing message.
    I have checked the coding in update rules, everything is ok.
    Plz any inputs.
    hari
    Message was edited by:
            hari reddy

    Might means that IDocs flow is not defined properly in your QA system for myself-SourceSystem.
    Regards,
    Vitaliy

  • Transformation from ods to cube IN BI 7.0

    Hi there, I will appreciate if anybody can help, I want to know if I want to assign the constant value to a field in transformation from ods to cube what is the way out.as i can choose the constant in update in BW3.5 but what is the step in BI 7.0
    thanks
    Soniya

    Go to rule type column right clicj and select rule details. Under rule type u can fuind the option of constant.
    http://help.sap.com/saphelp_nw04s/helpdata/en/e5/f913426908ca7ee10000000a1550b0/content.htm
    KJ!!!

  • Select query based on joining of tables from different database is taking too long

    Hi Team,
    Select query on table with millions of records is taking very long time. It took 50mins, the below query is joining on multiple table from two databases DB1.dbo.Table1 contains 100 million records and also Table3 and Table4 (of different database) might
    contain close to 1million records.
    Select T1.*
    From DB1.dbo.Table1 T1
    Join DB1.dbo.Table2 T2 on RTRIM(T1.Col3) = RTRIM(T2.Col3)
    Join Table4 CA on RTRIM(T1.Col1) + T2.Col2 = CA.Col1
    Join Table3 U on CA.Col2 = U.Col2 AND RTRIM(T2.Col2) = U.Col3
    Where U.Col4 NOT IN ('A1', 'A2', 'A3', 'A4', 'A5', 'A6','A7','A8','A9')
    And (T1.flg IS NULL OR T1.flg = 'N')
    And LTRIM(RTRIM(T2.Col2)) NOT IN ('B1','B2')
    How can i improve the performance of this query. Actual thing is update the data in Db1.dbo.table1 based on the conditions but if the select is taking close to 1hr then update will take hours together. Indexes already implemented on all the tables.
    Thanks,
    Eshwar.
    Please don't forget to Marked as Answer if my post solved your problem and use Vote As Helpful if a post was useful. It will helpful to other users.

    Thanks.
    Join Table4 CA on RTRIM(T1.Col1) + T2.Col2 = CA.Col1
    This join is not working with spaces if Rtrim is not used.
    Thanks! Eshwar.
    Please don't forget to Marked as Answer if my post solved your problem and use Vote As Helpful if a post was useful. It will helpful to other users.

  • Automatic loading from ODS to Cube in 3.5

    Hi All
    I was under the impression that in version 3.5 in order to load delta from ODS to Cube you had to run the 8 series Ipak.
    However I have recently noticed that this ipak is running automatically after a delta load into the ODS even when the load is not via a process chain.
    Can somebody where and how this setting is maintained.
    Regards
    A

    Hi,
    Go to ODS display mode and check if "Update Data Automatically" is ticked in Settings.
    Regards,
    Kams

  • Error when loading from ODS to Cube

    Hello Friends,
    I am having trouble loading data from ODS to Infocube in my 2004S system.When loading data i get this message
    07/03/2007     13:10:25     Data target 'ODSXYZ ' removed from list of loadable targets; not loadable.
    07/03/2007     13:28:42     Data target 'ODSXYZ ' is not active or is incorrect; no loading allowed     
    I checked for ODSXYZ in my data target but there is nothing by that name.Even the infopackage doesnt have it.What needs to be done.Please help.
    Thanks.

    Its a expected behavior. When you migrate ur DS, the infopacks associated with it, will grey out all the data targets that they were feeding before, that applies to any infopack you create even after the migration. You womt be able to delete it.
    Having said this, this shouldnt impact ur loads from ODS to Cube. As this shoudl be taken care of by ur DTP's rather than ur infopacks.
    A few questions:
    How are you loading ur cube?
    Did the data get thru fine to PSA with the infopack in question?
    How did you load ur DSO(assuming the load was successful)?
    Message was edited by:
            voodi

Maybe you are looking for