Updating to InfoCube IC_TEST : 2 - 0 Data Records

Hi friends i have to create dtp with flat file and loading data some errors are there
plz give me solution.
Request 25269
Process Request
Data Package 1: Errors During Processing
Updating to InfoCube IC_TEST : 2 -> 0 Data Records
Messages for data records saved; request is red acc. to configuration

Hi Prasana,
Please provide the error message?
check in the monitor tab for the error message.
Regards,
Ravi Kanth

Similar Messages

  • How to UPDATE MASTER DATA RECORD PA0377 using FM HR_INFOTYPE_OPERATION

    How to UPDATE MASTER DATA RECORD IN INFOTYPE 0377 DIRECTLY using function module HR_INFOTYPE_OPERATION ? When i use operation = 'MOD'  im getting an ERROR as NO DATA STORED FOR 0377 IN SELECTION PERIOD.
           Can anyone please help me how to use HR_INFOTYPE_OPERATION to UPDATE MASTER DATA RECORD.
    Thanks,
    Karthi.

    Hi,
    please check this code
    it is very useful
    infotypes: 0105.
    parameters: p_pernr type p0105-pernr,
                p_subty type p0105-subty,
                p_begda type p0105-begda,
                p_endda type p0105-endda,
                p_opera type pspar-actio,
                p_usrid type p0105-usrid,
                p_commt as checkbox.    "Useful only when NO_COMMIT is 'X'
    data: w_return type bapireturn1,    "Error handling data
          w_key type bapipakey,         "If record has been created or changed the created/ changed
                                        "record's key (PSKEY) will be populated
          v_message type string.        "For printing returned messages
    call function 'BAPI_EMPLOYEE_ENQUEUE'
      exporting
        number = p_pernr
      importing
        return = w_return.
    if w_return is initial. "Employee is not locked
      p0105-pernr = p_pernr.
      p0105-subty =
      p0105-usrty = p_subty.
      p0105-begda = p_begda.
      p0105-endda = p_endda.
      p0105-usrid = p_usrid.
    *Calls Internally HR_MAINTAIN_MASTERDATA
      call function 'HR_INFOTYPE_OPERATION'
        exporting
          infty         = '0105'
          number        = p_pernr
          subtype       = p_subty
          validityend   = p_endda
          validitybegin = p_begda
          record        = p0105
          operation     = p_opera
          dialog_mode   = '0'
          nocommit      = 'X'
        importing
          return        = w_return
          key           = w_key.
    endif.
    if w_return is not initial.
      message id w_return-id type w_return-type number w_return-number
              with w_return-message_v1 w_return-message_v2 w_return-message_v3
              w_return-message_v4 into v_message.
      case w_return-type.
        when 'A' or 'E'.
          format color col_negative. write:/ v_message.
        when others.
          format color col_heading. write:/ v_message.
      endcase.
      call function 'BAPI_EMPLOYEE_DEQUEUE'
        exporting
          number = p_pernr.
    else.
      format color col_positive. write:/ 'Record Updated Successfully.'.
      if w_key is not initial.
        write:/ 'Key contains',
              / 'Personnel Number:', w_key-employeeno,
              / 'Subtype (absence type):', w_key-subtype,
              / 'Start date of Absence:', w_key-validbegin,
              / 'End date of Absence:', w_key-validend.
      endif.
      commit work.
    endif.
    Thanks and regards
    durga.K

  • IDOC:Master Data Records are not going to be updated after posting the IDOC

    I need to create employee Master data records in R/3 .I am using the IDOC HRMD_A07, though the IDOC posting is successful no tables, corresponding to employee master Data are getting  updated,
    Can anybody plz Let me know the possible reasons and solutions for this?
    Thanks in Advance,
    Sandhya

    hi i hecked the job log entries in SM37 in source system, the idocs were returned with '0' records. i took the Idoc number and checked in BD87 it displays some logs in that one node shows the red status with the detail as follow  IDoc entries in tRFC queue      3383
    the next one is green                Data passed to port OK                    1
    i selected the red staus node and clicked 'process'  button but it displays as
    The operation cannot be carried out with this node type
    wht should i do now, wht may be the reason for the datasource not pulling the data from R/3 to BI?

  • Check which data records have been updated or created by the business rule.

    Dear all,
       Is there any way to check which data records have been updated or created by the business rule?
       For ex:
       When I run a data package which execute some standard store procedure and the log shows :
    Executing SPRUNCONVERSION  'LEGALAPP', [ACTUAL], [GGGG],'GLOBAL','SPSCOPE_42412','SPLOG_125272'
    SPRunConversion Version 2.06
    20071200 - 210 Rows Calulated
    20071200 - 84 Rows Updated
    Anyway to see the data content of "84 Rows Updated "?
    Best Regards,
    Jeff

    Yes each aplication has 3 fact tables:
    FACTWB:Real time storage, when you send data it is stored here and this is linked to the FACTWB partition in OLAP which is a ROLAP partition to enable real time updates of the data.
    FAC2: Short term storage. If too many records are in the FACTWB the aplication becomes slow because these records need to be queried by the OLAP engine everytime new records are send to the database. To clean up te real time storage you can do a lite optimize, which will move the FACTWB records to the FAC2 partition. this lite optimize can be done even when users are in the system.
    FACT: this is the long term storage. When your reporting is finished you can run a full optimize to clean up both FACTWB and FAC2 tables and store it in the FACT table. This contains all your historical records.
    Evertime you run a reort, SAP BPC is looking at all 3 partitions and combines the data of the 3 partitions to present the correct value in your report. Remember that BPC is posting differences in the WB and not the final values. So if you have for example a record with 10 in the FAC2 partition and you send 12 to the database for that intersection, BPC will post only 2. When running a report it combines the 10 and the 2 which makes 12.
    Hope this makes it more clear.
    Joost
    Edited by: Joost Hoppenbrouwers on May 6, 2008 12:19 PM : Seems that James and I where writing pretty much the same post at the same time

  • Unique data record means you can't  update a record from ECC with same key.

    Unique data record means you can't  update a record from ECC with same key fileds right?
    Details: For example i have two requests Req1 and Req2 in DSO with unique data record setting checked. on day one Req1 has a filed quantity with value 10 in Active data table. On day two Req1 can not be overwitten from ECC with Req2 with the same key fields but different value 20 in the filed quantity because of the Unique data record settings. finally the delta load fails from ECC going to DSO because of this setting. is it right?
    I think we can only use unique record setting going from DSO to cube right?
    Please give me a simple scenario in which we can use this setting.
    I already search the threads and will assign points only to valuable information.
    Thanks in advance.
    York

    Hi Les,
    Unique Data Records:
    With the Unique Data Records indicator, you determine whether only unique data records are to be updated to the ODS object. This means that you cannot load a data record into the ODS object the key combination for which already exists in the system – otherwise a termination occurs. Only use this setting when you are sure that only unique data records are to be loaded into the ODS object (for example, single documents). A typical application of this is in the loading of mass data. It improves the load performance.
    Hope it Helps
    Srini

  • Error 4 in the update: Time Overlapping of data records

    Hi Experts,
    I am loading 0EQUIPMENT master data load, after scheduling i'm getting Time overlapping of the data records. can anybody give the inputs for this...
    Regards
    SK

    Hi,
    Take a look at: Master data load failure
    Also maybe OSS note 314964 will help.
    Hope it helps,
    Gili

  • Infocube/DSO/Process Chain - data load statistics

    Hi,
    I'm trying to look for the total data records extracted  per day and extraction time from ECC into a particular DSO or infocube.
    Also the start and end times of the process chain which executes this infocube.
    We are trying to do some performance analysis for the volume testing.So ,we can compare the data volume loaded vs the execution times.
    I was able to look under the infocube/dso -> right click -> MANAGE -> transfered records/update records , request time,update time
    For this, do I need to look at the full update details or the delta update details?
    Also under a particular process chain,-> right click on First step ->Display messages to get the start time of the job
    Right Click on last Step ->Display mseeages to get the job finished time.
    I just wanted to get your thoughts on this, if this is the correct way to get the total records extracted and extraction times along with process chain times.
    Thank you.

    Hi,
    Based on the time char like 0calmonth, 0calday you will check the data.
    check the jobs in SM37 at ECC and BI system.
    select the specific jobs it will give clear picture. how much taking time form source to target.
    you use the DB02 and STO5 and RSTT you can do performance analysis.
    manage screen you will get either full or delta and which one source system and which type of data source.
    Process chains have it will automate the PRD system
    Thanks,
    Phani.

  • ERROR WHILE UPDATING TO INFOCUBE

    Hi All,
    I am using a flat file to load data from datasource into infocubes.The data reaches PSA tables with no errors but after the execution of DTP , the monitors shows up Error Updating to infocube, intitally I thought there could be some problem in the data format of certain fields in the flat file but everything looks to be correct, there are no errors in the transformation rules.
    Here's the ST22 analysis :
    what happened?                                                          
    The current application program detected a situation which really   
    should not occur. Therefore, a termination with a short dump was    
    triggered on purpose by the key word MESSAGE (type X).              
    Termination occurred in the ABAP program "CL_RSAWBN_NAV_TRACE===========CP" -
      in "DUMP".                                                                  
    The main program was "RSAWBN_START ".                                                                               
    In the source code you have the termination point in line 42                 
    of the (Include) program "CL_RSAWBN_NAV_TRACE===========CM004".       
    I have been deleting requests from the infocube , datasources and deleting entries from PSA but always endup getting the "Error while updating to Infocube"
    Have to figure out the solution as the planning team is waiting for me to copy the actual data to the plan cube
    Thanks

    Hi,
    You should have the following format in the flatfile :
    Considering today's date(18 Apr 2008) for example:
    0CALDAY            20080418
    0CALMONTH       200804
    0FISCPER           0042008   
    0FISCPER3         0042008 (I think this can have additional periods as well, not sure though)
    0FISCVARNT       No idea about this.
    Reward with points if helpful.
    Regards
    Hemant Khemani

  • DTP Error: Duplicate data record detected

    Hi experts,
    I have a problem with loading data from DataSource to standart DSO.
    In DS there are master data attr. which have a key  containing id_field.
    In End routine I make some operations which multiple lines in result package and fill new date field - defined in DSO ( and also in result_package definition )
    I.E.
    Result_package before End routine:
    __ Id_field ____ attra1 ____  attr_b  ...___   attr_x ____ date_field
       ____1________ a1______ b1_________ x1         
       ____2________ a2______ b2_________ x2       
    Result_package after End routine:
    __ Id_field ____ attra1 ____  attr_b  ..___   attr_x ____ date_field
       ____1________ a1______ b1_________ x1______d1         
       ____2________ a1______ b1_________ x1______d2    
       ____3________ a2______ b2_________ x2______d1         
       ____4________ a2______ b2_________ x2______d2   
    The  date_field (date type)  is in a key fields in DSO
    When I execute DTP I have an error in section Update to DataStore Object: "Duplicate data record detected "
    "During loading, there was a key violation. You tried to save more than one data record with the same semantic key."
    As I know the result_package key contains all fields except fields type i, p, f.
    In simulate mode (debuging) everything is correct and the status is green.
    In DSO I have uncheched checkbox "Unique Data Records"
    Any ideas?
    Thanks in advance.
    MG

    Hi,
          In the end routine, try giving
    DELETE ADJACENT DUPLICATES FROM RESULT_PACKAGE COMPARING  XXX  YYY.
    Here XXX and YYY are keys so that you can eliminate the extra duplicate record.
    Or you can even try giving
        SORT itab_XXX BY field1 field2  field3 ASCENDING.
        DELETE ADJACENT DUPLICATES FROM itab_XXX COMPARING field1 field2  field3.
    this can be given before you loop your internal table (in case you are using internal table and loops)  itab_xxx is the internal table.
    field1, field2 and field 3 may vary depending on your requirement.
    By using the above lines, you can get rid of duplicates coming through the end routine.
    Regards
    Sunil
    Edited by: Sunny84 on Aug 7, 2009 1:13 PM

  • Duplciuate Data record error while reloading a DTP

    I have a DTP created to load etxt datasource and put in the process chain. I have set "delta mode" so that it picks up if additional records are available.
    But, I noiticc that when there are no additional records (for eg: earlier load had 269 records and current load has same 269 records), I am gettign the error - "Duplciate data record" and status is set as "Red". I do not want to see error in this case as it did not find any new records and want the status to be green if there are no new records.
    Could you please suggest what settings will do that.
    Regards
    Raj

    Delta DTP will fetch only unloaded requests ( requests which donot exist in target ) and not additional records.
    Is the text datasource delta enabled ?
    Do you have Infopackage of Update type Delta setup?
    Did you run a Full DTP before Delta DTP?
    Assuming Full Infopackage has loaded 269 records to PSA and the same will be loaded ( if no additional records in source system ) again.
    Req 1 - 269 - Yesterday
    Req 2 - 269  - Today
    Full DTP ran yesterday will load 269 records to target.
    Delta DTP ran today will load Req1 & Req 2 to target ( master data will be overwritten ) - Reason Delta DTP is acting like Init with data transfer.
    You can start off with a Delta DTP instead Full, if Full DTP is ran before a Delta DTP maje sure you delete requests loaded by Full DTP.
    This can be ignored as this is Master Data which wil be overwritten.
    To get rid of error jst Check " Handle Duplicate Records " in  Update Tab of DTP.

  • Error on activation of data records from ODS object

    hi,
    Kindly help with my BW Problem, the error occured on the
    activation of ODS. The error says "<i>Request
    REQU_43MNOPW29W4F5M037J7OAFP52 in ODS ZPUR_O01 must have QM
    status green before it is activated</i>" and "<i>Activation of data
    records from ODS object ZPUR_O01 terminated</i>"
    I tried to check the data in the ODS Object by choosing
    <LS>Administrator Workbench > data targets > context menu on ODS
    Object ZPUR_O01 but the status are green and when i tried to
    activate it the logs says "<i>ODS object ZPUR_O01 was built
    incorrectly. Cannot update request
    REQU_43MODQ3GF69X17B71ZKQMN85I(20,154)
    ODS object ZPUR_O01 was built incorrectly. Cannot update request
    REQU_43MQ8A8F0L1US5COQTUUWX3BA(20,155)
    ODS object ZPUR_O01 was built incorrectly. Cannot update request
    REQU_43MQ8OQQTY21YQ37RMA1WMO06(20,156)
    Data to be activated successfully checked against archiving
    objects                                   
    Activation of data records from ODS object ZPUR_O01 terminated</i>"
    Please help me with this problem.

    Hi Jay
    U need to update these requests to ODS
    REQU_43MODQ3GF69X17B71ZKQMN85I(20,154)
    REQU_43MQ8A8F0L1US5COQTUUWX3BA(20,155)
    REQU_43MQ8OQQTY21YQ37RMA1WMO06(20,156)
    1) delete the current request
    2)Reconstruct the requests above three request
    3)and then Load the current reqest (REQU_43MNOPW29W4F5M037J7OAFP52)
    Then only u will be able to activate the ODS....
    This is an serialization issue...
    hope it helps
    regrds
    AK
    assign points if usefull

  • Getting Duplicate data Records error while loading the Master data.

    Hi All,
    We are getting Duplicate data Records error while loading the Profit centre Master data. Master data contains time dependent attributes.
    the load is direct update. So i made it red and tried to reloaded from PSA even though it is throwing same error.
    I checked in PSA. Showing red which records have same Profit centre.
    Could any one give us any suggestions to resolve the issues please.
    Thanks & Regards,
    Raju

    Hi Raju,
            I assume there are no routines written in the update rules and also you ae directly loading the data from R/3 ( not from any ODS). If that is the case then it could be that, the data maintained in R/3 has overlapping time intervals (since  time dependency of attributes is involved). Check your PSA if the same profit center has time intervals which over lap. In that case, you need to get this fixed in R/3. If there are no overlapping time intervals, you can simply increase the error tolerance limit in your info-package and repeat the load.
    Hope this helps you.
    Thanks & Regards,
    Nithin Reddy.

  • Issue in update of BANK INFORMATION in PERSONAL DATA INFORMATION(ESS)

    Hi Experts,
    i have an issue while i am trying to update the BANK INFORMATION in PERSONAL INFORMATION, i am getting the fallowing error:
    Data record 0000001300090      2010033120090401000 has grouping value "" instead of "40"
    Data record 0000001300090      9999123120100401000 has grouping value "" instead of "40"
    please help me out

    Dear Siva,
    You can do this using the personalization
    1. The pernr should be assigned to user id you are using having Content Administrator rights.
    2. Preview the iview for Bank Inforamtion and ctrl + right click on the "Edit Button".
    3. Select "Yes" for the Invisible property of button Edit.
    4. Click Apply and Ok.
    The button will not be visible anymore.
    Regards,
    Samir

  • Purpose of Update/Overwrite Hierarchy in Import master data package

    Hi All,
    Please let me know what is the difference in update and overwrite option in import master data package. Here when i tried running the package, both the option are altering the hierarchy of existing member.
    My requirement is just to add a new member ID, if it already exist it should not affect those records and only update the new master data records.
    Thanks & Regards,
    Ramanathan

    Hi Ramanathan,
    You can run Import Master Data DM package in BPC to perform delta master data load. While you run DM package you get option to load Master Data from BW/BI where you can select:
    1. MERGE: this will keep old data intact and also load new master data, and
    2. COPY and REPLACE: This will copy new data and delete old master data.
    As per your requirement, you need to select the First option. If you select Update option then it will not delete old master data, instead update the members and hierarchy according to the parent-children design which you would have given in the dimension membersheet.
    Hope this corresponds to your requirement.
    Rgds,
    Poonam

  • Can we update BW infocube from SAP using Custom ABAP program ?

    Hi,
    My requiremnet is to update the infocube using data from R/3 .
    I have a custom report program and users want this data in BW.
    Let me if there are any FM? from which we can call via RFC or any other alternatives?
    Rgds
    Praveen

    report must be an ABAP program in ECC I guess...Why dont you look forward for generic extraction...
    Check this one:
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/a0f46157-e1c4-2910-27aa-e3f4a9c8df33

Maybe you are looking for

  • Analysis Authorization Issue

    Hi: I created an analysis authorization ZCO_CODE to trstrict it by a company code. I added following objects in authorization with values. 0COMP_CODE = 1000 0TCAACTVT = 03 0TCAIFAREA = * 0TCAIPROV = * 0TCAVALID = * Then I created a role Z:00:BW_REPOR

  • Crystal Report: Multiple Join in Table

    Hi experts, I have a problem adding a table in a report. The layout I'm modifing is the Purchase Order (CR). The report contains the necessary tables linked via the Database Expert to OPOR master table. I need to add the Vendor Item Code in the OSCN

  • How can i retrieve saved ppt downloaded by firefox, not active to open from downloads?

    i downloaded a ppt 2013 document from my email using firefox, worked on it, was hitting save button several times. when done, i couldn't retrieve it. I see inactive ppt doc on the downloads section of mozilla on the right upper area that can't be ope

  • Help! trying to start HP Pavillion G7 Windows 8.1 laptop in safe mode

    on friday last week my boss' laptop had an issue with outlook, and we had to do a live chat with microsoft and allow remote connection to get that resolved. the issue was resolved, and he took the laptop home. however, when he attempted to boot up ag

  • Issue with Attribute - BUK

    Hi , Company code attribute is not picking when i tried to raise the shopping cart, saying that no values found. I have already assigned the Company code attribute(BUK) in the Org structure. Also i have assigned the Location in the extended attribute