DUmp Erro ASSIGN_TYPE_CONFLICT while loading from DSO to INFOCUBE

HI ALL,
Here i had rquirement of modifying the Tranformation rules and i have done some routine coding for a particular infoobject
and deleted the data from DOS  & Cube , then i started Loading to DSO, here i got the sucess for DSO, now i have got  the modified data of the particular Infoobject in the DSO,
But i am facing a Dump Error while loading data from DSO to Cube thorugh the DTP .
HEre is the Error "ASSIGN_TYPE_CONFLICT" which i see in ST22.
Plz,
Thanks in Advance,
sravan

HI,
When i started the load for the first time i got he same error, so i have activated the DTP and laoded it again,
then i faced the same problem,
where the Modified data is already in DSO and  i have validated the Data also,  that is ok ...
so i need to Delete or to need to create another DTP to load the data from DSA to CUBE.
and even i have cheked all the tansformation rules they are all fine  the DSO structure and Infocube Structure is ok,
Please  suggest,
Thanks LAX,

Similar Messages

  • Error while loading from DSO to DSO in BI

    Hi All,
    While we are loaind the data from base DSO to Target DSO. Using DTP we are getting the following error. Could any one please provide me the solution for this.
    "The argument 'B' cannot be interpreted as a number"
    i am getting the above message. The lload is successfull from R3 to base DSO, but the same data is ffailed in target DSO.
    Helpful answers will be rewared with the points.
    Many thanks,
    Khaja

    Hi Siva,
    thanks for ur reply. Mapping in transformation is Ok. And we are using the same info objects which we have used in base DSO. The load is successfull to the base DSO it got failed in target DSO.
    How to know to which info object it is throwing the error. we have 80 info objects. even in log also it is not giving the info object name.
    Thanks
    Khaja

  • Data loading from DSO to Cube

    Hi,
    I have a question,
    In book TBW10 i read about the data load from DSO to InfoCube
    " We feed the change log data to the InfoCube, 10, -10, and 30 add to the correct 30 value"
    My question is cube already have 10 value, if we are sending 10, -10 and 30 Values(delta), the total should be 40 instead of 30.
    Please some one explaine me.
    Thanks

    No, it will not be 40.
    It ll be 30 only.
    Since cube already has 10, so before image ll nullify it by sending -10 and then the correct value in after immage ll be added as 30.
    so it ll be like this 10-10+30 = 30.
    Thank-You.
    Regards,
    Vinod

  • Partial data loading from dso to cube

    Dear All,
    I am loading the data from dso to the cube through DTP. The problem is that some records are getting added  through data
    packet 1 to the cube around 50 crore records while through data packet2 records are not getting added.
    It is full load from dso to cube.
    I have tried deleting the request and executing DTP but again the same no of records get added to the cube through datapacket 1   and after that records are not added to the cube through datapacket 1  ;request remains in yellow state only.
    Please suggest .

    Nidhuk,
    Data load transfers package by package. Your story sounds like it got stuck in second package or something. Suggest you check package size, try to increase to a higher number to see if anything  changes ? 50 records per package kind of low, your load should not spread out into too many packages.
    Regards. Jen
    Edited by: Jen Yakimoto on Oct 8, 2010 1:45 AM

  • Data flow tasks faills while loading from database to excel

    Hello,
    I am getting error while loading from oledb source to excel and the error as shown below.
    Error: 0xC0202009 at DFT - Company EX, OLE DB Destination [198]: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80004005.
    Error: 0xC0209029 at DFT - Company EX, OLE DB Destination [198]: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR.  The "input "OLE DB Destination Input" (211)" failed because error code 0xC020907B occurred, and the error row
    disposition on "input "OLE DB Destination Input" (211)" specifies failure on error. An error occurred on the specified object of the specified component.  There may be error messages posted before this with more information about the
    failure.
    Error: 0xC0047022 at DFT - Company EX: SSIS Error Code DTS_E_PROCESSINPUTFAILED.  The ProcessInput method on component "OLE DB Destination" (198) failed with error code 0xC0209029. The identified component returned an error from the ProcessInput
    method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.  There may be error messages posted before this with more information about the failure.
    Error: 0xC02020C4 at DFT - Company EX, OLE DB Source 1 [1]: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.
    Error: 0xC0047021 at DFT - Company EX: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread0" has exited with error code 0xC0209029.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047038 at DFT - Company EX: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "OLE DB Source 1" (1) returned error code 0xC02020C4.  The component returned a failure code when the pipeline engine
    called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    Error: 0xC0047021 at DFT - Company EX: SSIS Error Code DTS_E_THREADFAILED.  Thread "SourceThread0" has exited with error code 0xC0047038.  There may be error messages posted before this with more information on why the thread has exited.
    Any help would be appreciated ASAP.
    Thanks,
    Vinay s

    You can use this code to import from SQL Server to Excel . . .
    Sub ADOExcelSQLServer()
    ' Carl SQL Server Connection
    ' FOR THIS CODE TO WORK
    ' In VBE you need to go Tools References and check Microsoft Active X Data Objects 2.x library
    Dim Cn As ADODB.Connection
    Dim Server_Name As String
    Dim Database_Name As String
    Dim User_ID As String
    Dim Password As String
    Dim SQLStr As String
    Dim rs As ADODB.Recordset
    Set rs = New ADODB.Recordset
    Server_Name = "EXCEL-PC\EXCELDEVELOPER" ' Enter your server name here
    Database_Name = "AdventureWorksLT2012" ' Enter your database name here
    User_ID = "" ' enter your user ID here
    Password = "" ' Enter your password here
    SQLStr = "SELECT * FROM [SalesLT].[Customer]" ' Enter your SQL here
    Set Cn = New ADODB.Connection
    Cn.Open "Driver={SQL Server};Server=" & Server_Name & ";Database=" & Database_Name & _
    ";Uid=" & User_ID & ";Pwd=" & Password & ";"
    rs.Open SQLStr, Cn, adOpenStatic
    ' Dump to spreadsheet
    With Worksheets("sheet1").Range("a1:z500") ' Enter your sheet name and range here
    .ClearContents
    .CopyFromRecordset rs
    End With
    ' Tidy up
    rs.Close
    Set rs = Nothing
    Cn.Close
    Set Cn = Nothing
    End Sub
    Also, check this out . . .
    Sub ADOExcelSQLServer()
    Dim Cn As ADODB.Connection
    Dim Server_Name As String
    Dim Database_Name As String
    Dim User_ID As String
    Dim Password As String
    Dim SQLStr As String
    Dim rs As ADODB.Recordset
    Set rs = New ADODB.Recordset
    Server_Name = "LAPTOP\SQL_EXPRESS" ' Enter your server name here
    Database_Name = "Northwind" ' Enter your database name here
    User_ID = "" ' enter your user ID here
    Password = "" ' Enter your password here
    SQLStr = "SELECT * FROM Orders" ' Enter your SQL here
    Set Cn = New ADODB.Connection
    Cn.Open "Driver={SQL Server};Server=" & Server_Name & ";Database=" & Database_Name & _
    ";Uid=" & User_ID & ";Pwd=" & Password & ";"
    rs.Open SQLStr, Cn, adOpenStatic
    With Worksheets("Sheet1").Range("A2:Z500")
    .ClearContents
    .CopyFromRecordset rs
    End With
    rs.Close
    Set rs = Nothing
    Cn.Close
    Set Cn = Nothing
    End Sub
    Finally, if you want to incorporate a Where clause . . .
    Sub ImportFromSQLServer()
    Dim Cn As ADODB.Connection
    Dim Server_Name As String
    Dim Database_Name As String
    Dim User_ID As String
    Dim Password As String
    Dim SQLStr As String
    Dim RS As ADODB.Recordset
    Set RS = New ADODB.Recordset
    Server_Name = "Excel-PC\SQLEXPRESS"
    Database_Name = "Northwind"
    'User_ID = "******"
    'Password = "****"
    SQLStr = "select * from dbo.TBL where EMPID = '2'" 'and PostingDate = '2006-06-08'"
    Set Cn = New ADODB.Connection
    Cn.Open "Driver={SQL Server};Server=" & Server_Name & ";Database=" & Database_Name & ";"
    '& ";Uid=" & User_ID & ";Pwd=" & Password & ";"
    RS.Open SQLStr, Cn, adOpenStatic
    With Worksheets("Sheet1").Range("A1")
    .ClearContents
    .CopyFromRecordset RS
    End With
    RS.Close
    Set RS = Nothing
    Cn.Close
    Set Cn = Nothing
    End Sub
    Knowledge is the only thing that I can give you, and still retain, and we are both better off for it.

  • Data not loading for new Fields InfoObjects from DSO to InfoCube

    Hi Gurus -
                   I have a DataSource that is providing data to existing DSO then to Infocube. My client asked me to added couple fields to DataSource and get the data to DSO and then DSO to InfoCube.
    Here is the Old scenarion:   DataSource -> DSO -> InfoCube.
    Here is the New Scenation:
         DataSource ( Added new fields) -> DSO (added new InfoObjects for Corresponding DataSource FIelds) -> InfoCube( Added new InfoObjects to mapped from DSO)
    I added the new fields to DataSource, added the corresponding InfoObjects to DSO and InfoCube.
    I successfully loaded data from DataSource to DSO. Data is populating for the new Fields/InfoObjects in DSO.
    But when I load data from DSO to InfoCube, I don't see any data for New Fields InfoObjects in the InfoCube.
    Data from DSO to InfoCube is loading fine for the Old InfoObjects Fields but not for the New InfoObjects I added in InfoCube.
    -Sonali

    Hi,
    Why dont u debug the load through DTP debugging and check what happens to the source field and target field once it passes through the transformation. You can easily trace back where the fields are becoming blank.
    The loads which you have mentioned earlier did it have values in Added Records/Transferred Records column for the cube.
    Regards,
    Mani

  • Delta records are not loading from DSO to info cube

    My query is about delta loading from DSO to info cube. (Filter used in selection)
    Delta records are not loading from DSO to Info cube. I have tried all options available in DTP but no luck.
    Selected "Change log" and "Get one request only" and run the DTP, but 0 records got updated in info cube
    Selected "Change log" and "Get all new data request by request", but again 0 records got updated
    Selected "Change log" and "Only get the delta once", in that case all delta records loaded to info cube as it was in DSO and  gave error message "Lock Table Overflow" .
    When I run full load using same filter, data is loading from DSO to info cube.
    Can anyone please help me on this to get delta records from DSO to info cube?
    Thanks,
    Shamma

    Data is loading in case of full load with the same filter, so I don't think filter is an issue.
    When I follow below sequence, I get lock table overflow error;
    1. Full load with active table with or without archive
    2. Then with the same setting if I run init, the final status remains yellow and when I change the status to green manually, it gives lock table overflow error.
    When I chnage the settings of DTP to init run;
    1. Select change log and get only one request, and run the init, It is successfully completed with green status
    2. But when I run the same DTP for delta records, it does not load any data.
    Please help me to resolve this issue.

  • Data load from DSO to cube fails

    Hi Gurus,
    The data loads have failed last night and when I try to dig inside the process chains I find out that the cube which should be loaded from DSO is not getting loaded.
    The DSO has been loaded without errors from ECC.
    Error Message say u201DThe unit/currency u2018source currency 0currencyu2019 with the value u2018spaceu2019 is assigned to the keyfigure u2018source keyfigure.ZARAMTu2019 u201D
    I looked in the  PSA has about 50, 000 records .
    and all data packages have green light and all amounts have 0currency assigned
    I went in to the DTP and looked at the error stack it has nothing in it then I  changed the error handling option from u2018 no update,no reportingu2019 to u2018valid records update, no reporting (resquest red)u2019 and executed then the error stactk showed 101 records
    The ZARAMT filed has 0currency blank for all these records
    I tried to assign USD to them and  the changes were saved and I tried to execute again but then the message says that the request ID should be repaired or deleted before the execution. I tried to repair it says can not be repaired so I deleted it and executed . it fails and the error stack still show 101 records.when I look at the records the changes made do not exist anymore.
    If I delete the request ID before changing and try to save the changes then they donu2019t get saved.
    What should I do to resolve the issue.
    thanks
    Prasad

    Hi Prasad....
    Error Stack is request specific.....Once you delete the request from the target the data in error stach will also get deleted....
    Actually....in this case what you are suppose to do is :
    1) Change the error handling option to u2018valid records update, no reporting (resquest red)u2019 (As you have already done....) and execute the DTP......all the erroneous records will get sccumulated in the error stack...
    2) Then correct the erroneos records in the Error Stack..
    3) Then in the DTP----in the "Update tab"....you will get the option "Error DTP".......if it is already not created you will get the option as "Create Error DT".............click there and execute the Error DTP..........The Error DTP will fetch the records from the Error Stack and will create a new request in the target....
    4) Then manually change the status of the original request to Green....
    But did you check why the value of this field is blank....if these records again come as apart of delta or full then your load will again fail.....chcek in the source system and fix it...for permanent solution..
    Regards,
    Debjani......

  • Data load from DSO to Cube in BI7?

    Hi All,
    We just migrated a dataflow from 3.5 to 7 in development and moved to production. So till now in production, the dataloads happend using the infopackages.
    a. Infopackage1 from datasource to ODS and
    b. Infopackage2 from ODS to the CUBE.
    Now after we transported the migrated dataflow to production, to load the same infoproviders I use
    1. Infopackage to load PSA.
    2. DTP1 to load from PSA to DSO.
    3. DTP2 to load from DSO to CUBE.
    step1 and step2 works fine but when I run the DTP2 it is getting terminated. But now when I tried the step b (above), it loads the CUBE fine using the infopackage. So I am unable to understand why the DTP failed and why the infopackage load is successful. In order to use the DTP do we need to do any cleanup when using it for first time? Please let me know if you have any suggestions.
    Please note that the DSO already has data loaded using infopackage.  (Is this causing the problem?)
    Thanks,
    Sirish.

    Hi Naveen,
    Thanks for the Reply. The creation of DTP is not possible without a transformation.
    The transformation has been moved to production successfully.

  • Error while creating transformation from DSO to InfoCube

    Hi Gurus,
       When i am creating tranformation from DSO to InfoCube getting an error,        " The unit/currency 'Source Currency 0CURRENCY ' with the value 'space' is  assigned to the key figure 'Source Key Fig. ZAMT_PLC ' with the value   
    '8.50 ' "
    ZAMT_PLC is created with 0CURRENCY.
    On checking we found that there is no value stored for Currency field in table level itself. Now in this case how do we avoid this error.
    I want to know why is this we need to have value for 0Currency field???
    Thanks & Regards,
    Anup

    Hello Anup,
    If you load a key figure which has a unit assigned to it,
    then for every records where the key figure has a valid value
    (value 0) the corresponding unit has to be delivered as well.
    Additionally please check note 1366474 regarding the RSTRAN 312
    error message:
    "The corrections contained in Note 1165168 are incomplete.
    Depending on the sequence during loading, the system may not issue error
    message RSTRAN 312 even though data regarding the assigned unit or
    currency "space" is incorrect."
    This means actually, that the system should have issued the RSTRAN312
    error message even before the upgrade, but for some reason this
    check found no issues in some constellations. This means furthermore
    that your system behaved incorrectly in the past, that this
    error message was not triggered if you loaded key figures without
    any unit.
    I'm sorry to tell you, but the system behaves and works as designed
    as it should be even before the correction from note 1366474.
    Please change your transformations for key figures which
    have a unit assigned, that the unit field is filled correctly.
    with the corresponding units, otherwise the system will abort
    the processing of the data with the above mentioned error
    message.
    Best Regards,
    Des

  • Error while loading from PSA to Cube. RSM2-704 & RSAR -119

    Dear Gurus,
    I have to extract about 1.13 million records from set up tables to the MM cube 0IC_C03. While doing, the following two errors occured while loading from the psa to the cube:
    Error ID - RSM2 & error no. 704: Data records in the PSA were marked for data package 39 . Here there were 2 errors. The system wrote information or warnings for the remaining data records.
    Error ID- RSAR & No. 119: The update delivered the error code 4 .
    (Data records for package 39 selected in PSA - 2 error(s). Record 5129 :No SID found for value 'EMN ' of characteristic 0BASE_UOM . Record 5132 :No SID found for value 'EMN ' of characteristic 0BASE_UOM )
    I tried to change the defective records in psa by deleting the erraneous field value EMN and tried to load but failed.
    Now, my questions are:
    How can I resolve the issue ?
    (ii) How to rectify the erraneous record ? should we only delete the erraneous field value or delete the entire record from the psa.
    (iii) How to delete a record from psa.
    Thanks & regards,
    Sheeja.

    Hi,
    Data records for package 39 selected in PSA - 2 error(s). Record 5129 :No SID found for value 'EMN ' of characteristic 0BASE_UOM . Record 5132 :No SID found for value 'EMN ' of characteristic 0BASE_UOM
    The issue with record no. 5129 and 5132.
    In PSA check errorneous records and it will display only the error records, you just edit as per the require and try reloading into cube.
    Deleting single record is not possible.
    Let us know if you still have any issues.
    Reg
    Pra

  • Data Records are missing in between while loading from R/3 (ECC) to BI.

    Dear Experts,
    I have created a custom DataSource on Custom Function Module.  This datasource contains 600 fields. (I know its a monster and splitting options are thinner).
    1) Validate the data using RSA3 in R/3 and showed the correct record count.
    2) Validate the data by debugging the FM, still showed the correct record count.
    But while loading from R/3 to BI, records are missing.
    Various Scenarios load from R/3 to BI:
    1a) Loaded full load (78000 records) with all default data transfer settings.  PSA showed up with 72000 records (missing 6000) only.  Compared the Idocs vs data packets, both reconciled.
    1b) Loaded full load (78000) with modified settings (15000 KB / data packet).  PSA showed up with 74000 records (missing 4000) only.
    2a) Loaded with selection parameters (took a small chunk) (7000 records) with default data transfer settings.  PSA showed up only 5000 records (missing 2000).
    2b) Loaded with selection parameters (7000 records) with modified settings (15000 KB / data packet).  PSA showed up all 7000 records.
    3a) Loaded with selection parameters (took further small chunk) (4000 records).  PSA showed up all records regardless data transfer settings.
    Also please look at this piece of code from the function module,
    IF l_wa_interface-isource = 'ZBI_ARD_TRANS'.
          l_package_size = l_wa_interface-maxsize DIV 60.
        ENDIF.
    I really appreciate your advise or help in this regard.
    Thanks much,
    Anil

    Hi,
    Which module u want?
    if its SD(for example)
    steps>>
    1>In AWB goto "business content"
    2> goto "Info provider"
    3>Under infoarea select SD cubes
    4> Drag related cubes and ODS to right panel
    5> Set the grouping option "In Data flow before&afterwards"
    6>Install the collected objects
    Go to R/3
    7> Use Tcode RSA5 Transfer all regarding SD module datasources
    Goto BW
    8> Right click on the source system "Replicate datasources"
    [DataSources|http://help.sap.com/saphelp_nw70/helpdata/en/3c/7b88408bc0bb4de10000000a1550b0/frameset.htm]
    Edited by: Obily on Jul 10, 2008 8:36 AM

  • Load from ODS into InfoCube gives TIME-OUT runtime error after 10 minutes ?

    Hi all,
       We have a full load from ODS into InfoCube and it was working fine till the last week upto with 50,000 records. Now, we have around 70,000+ records and started failing with TIME_OUT runtime error.
       The following is from the Short Dump (ST22):
       The system profile "rdisp/max_wprun_time" contains the maximum runtime of a
    program. The current setting is 600 seconds. Once this time limit has been exceeded, the system tries to terminate any SQL statements that are currently being executed and tells the ABAP processor to terminate the current program.
      The following are from ROIDOCPRMS table:
       MAXSIZE (in KB) : 20,000
       Frequency       :  10
       Max Processes : 3
      When I check the Data Packages under 'Details' tab in Monitor, there are four Data Packages and the first three are with 24,450 records.  I will right click on each Data Package and select 'Manual Update' to load from PSA. When this Manual Update takes more than 10 minutes it is failing with TIME_OUT again.
      How could I fix this problem, PLEASE ??
    Thanks,
    Venkat.

    Hello A.H.P,
    The following is the Start Routine:
    PROGRAM UPDATE_ROUTINE.
    $$ begin of global - insert your declaration only below this line  -
    TABLES: /BIC/AZCPR_O0400, /BIC/AZCPR_O0100, /BIC/AZCPR_O0200.
    DATA: material(18), plant(4).
    DATA: role_assignment like /BIC/AZCPR_O0100-CPR_ROLE, resource like
    /BIC/AZCPR_O0200-CPR_BPARTN.
    $$ end of global - insert your declaration only before this line   -
    The follow definition is new in the BW3.x
    TYPES:
      BEGIN OF DATA_PACKAGE_STRUCTURE.
         INCLUDE STRUCTURE /BIC/CS8ZCPR_O03.
    TYPES:
         RECNO   LIKE sy-tabix,
      END OF DATA_PACKAGE_STRUCTURE.
    DATA:
      DATA_PACKAGE TYPE STANDARD TABLE OF DATA_PACKAGE_STRUCTURE
           WITH HEADER LINE
           WITH NON-UNIQUE DEFAULT KEY INITIAL SIZE 0.
    FORM startup
      TABLES   MONITOR STRUCTURE RSMONITOR "user defined monitoring
               MONITOR_RECNO STRUCTURE RSMONITORS " monitoring with record n
               DATA_PACKAGE STRUCTURE DATA_PACKAGE
      USING    RECORD_ALL LIKE SY-TABIX
               SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS
      CHANGING ABORT LIKE SY-SUBRC. "set ABORT <> 0 to cancel update
    $$ begin of routine - insert your code only below this line        -
    fill the internal tables "MONITOR" and/or "MONITOR_RECNO",
    to make monitor entries
       clear DATA_PACKAGE.
       loop at DATA_PACKAGE.
          select single /BIC/ZMATERIAL PLANT
             into (material, plant)
             from /BIC/AZCPR_O0400
             where CPR_EXT_ID = DATA_PACKAGE-CPR_EXT_ID
             and ( MATL_TYPE = 'ZKIT' OR MATL_TYPE = 'ZSVK' ).
           if sy-subrc = 0.
              DATA_PACKAGE-/BIC/ZMATERIAL = material.
              DATA_PACKAGE-plant = plant.
              modify DATA_PACKAGE.
              commit work.
           endif.
           select single CPR_ROLE into (role_assignment)
                         from /BIC/AZCPR_O0100
                         where CPR_GUID = DATA_PACKAGE-CPR_GUID.
            if sy-subrc = 0.
              select single CPR_BPARTN into (resource)
                         from /BIC/AZCPR_O0200
                         where CPR_ROLE = role_assignment
                         and CPR_EXT_ID = DATA_PACKAGE-CPR_EXT_ID.
                   if sy-subrc = 0.
                      DATA_PACKAGE-CPR_ROLE = role_assignment.
                      DATA_PACKAGE-/BIC/ZRESOURCE = resource.
                      modify DATA_PACKAGE.
                      commit work.
                   endif.
              endif.
           clear DATA_PACKAGE.
           endloop.
    if abort is not equal zero, the update process will be canceled
      ABORT = 0.
    $$ end of routine - insert your code only before this line         -
    Thanks,
    Venkat.

  • Problem in transfering data from DSO to InfoCube

    Hi,
    I am trying to transfer data from DSO to Infocube. But it is giving the following message.
    Last delta update is not yet completed
    Therefore, no new delta update is possible.
    You can start the request again
    if the last delta request is red or green in the monitor (QM activity)
    It is not allowing to load data to InfoCube from DSO. I checked the request number. But there is no request number in InfoCube. If request no is there I can delete the request from InfoCube and delete the data mart status in DSO and reload again.But there is no request id in IC. So plz guide me how I can tranfer data to InfoCube overcoming this problem.
    Actually further updata was running from DSO to IC. But it failed. Executing Function Module : RSPC_PROCESS_FINISH, I set the chain  green before loading
    data to InfoCube. So it starts executing the next chain. So there is no request id. So plz guide me how I can retain this request id.
    Edited by: Purushotham Reddy on Jan 28, 2008 11:19 AM

    Hi Reddy,
    Make the QM Status Red of the failed request even if it is red.
    and then run the same infopackage used in the chain. This will ask for repeat of last delta.
    If this process is understood to u then its fine otherwise read below:
    the other option is check if this failed request is present in PSA. If it is present in PSA then update the failed request from PSA, here u wont need to do a repeat of last delta. This Option is easier. this u can do from the ,manage tab of the data source I think this u might be surely knowing....
    Assign Points if it helps u.....
    Thanks,
    Nagesh.

  • Error in uploading data from DSO to Infocube in 3.5(rsa1old)

    While uploading data from DSO to Infocube, I selected Initial update option
    In schedule tab when I clicked start, an error occured it says:
    Delete init request
    REQU_D4YTAGX8PEOUQJLSKOEPH9EV before running
    init. again with same selection.
    Can anyone let me know how should I delete that request.
    Thanks,
    Soujanya.

    Hi,
    Go to RSRQ and enter the request then execute. This should bring you to Monitor u2013 Administration Workbench. From there you will see the InfoSource (usually has u201C8u201D as a prefix). Get the InfoSource name. Go to RSA1 to find the InfoSource. From here it gets a bit tricky. Go to the datasource and drill down to the related InfoPackage. Go to Infopackage scheduling. Then go  to Scheduler Menu, the Initialization Options for Source System should be open. You can see from there the request with error. Let me know if this answers your question.
    Regards,
    JD Datuin

Maybe you are looking for