Archive only masked records error

Hi SDNers,
I want to Archive repository only with 100 records.
I have tried Masking those records and on Archiving used the option to select only that Mask.
But it is resulting in Archive fail and the report says:
4952  2009/10/13 15:53:20.744 Report Info  XCatDBCopy.cpp     Table 55-HCLIQ00CUSTOMERLTM_m000-A2i_48_Tu_114
4952  2009/10/13 15:53:20.744 Report Error  XCatDBCopy.cpp       Error accessing table A2i_48_Tu_114
4952  2009/10/13 15:53:20.744 Report Info  XCatDBCopy.cpp   --~ Total KBytes: 34048, Partition#0(Main) KBytes: 33803
4952  2009/10/13 15:53:20.837 Report Warning  XCatDBCopy.cpp   $$$ Operation ended in ERROR : 80000001H : Illegal value for parameter
4952  2009/10/13 15:53:20.868 Report Warning  XCatDBCopy.cpp   Archiving operation for HCLI Q00 CUSTOMER LT Masking1 ended in error
4952  2009/10/13 15:53:20.868 Report Info  XCatDBCopy.cpp       0min 10sec
4952  2009/10/13 15:53:20.868 Report Info UNC TextReportLog.cpp   Report Duration: 0 day(s), 00:00:10
Any idea?
Thanks,
Priti

Hi Priti,
I came across this note and thought of sharing it with you
"Note 1424126 - Archiving a repository, that includes Tuples, via Mask"
Check if this is the cause of your problem.
Best Regards,
Shiv

Similar Messages

  • Archiving of Info Records

    Hi,
    I have 22 Materials and now user wants me to archive all the INFO Records related to these 22 materials?
    Please guide me what are the prerequisite needed to be done by me before archiving all the INFO Records.
    Note: We had errors in Production system, which we want to reproduce in Quality system.
    Cheers,
    Kumar.S

    Hi
    Follow the below steps:
    1. Flag Info-Records for deletion (ME15)
    This will Set the Deletion Flag only in Purchase Info Record.
    Once Deletion Flag is set you need to Archive them Using SARA with Archiving Object MM_EINA which will delete the Info Records from the Database.
    2. Run the archiving program - RM06IW30 (Menu path -> Logistics - Materials Management u2013Purchasing - Master Data - Info Record - Follow on Functions - Archive)
    Create Archive File: Info Record:
    a) Select Action: Archive and enter a new Variant, for example: Z_EINA_ARCH_ 01, press 'Maintain'
    b) On selection screen enter the data range (Vendor, Material, etc.) you want to archive.
    c) Deselect the 'Test' flag if you don't want to test first.
    d) Press green back-arrow and enter the description of this new variantt on the screen which follows.
    e) Save the variantt which brings you again to the selection screen. Press green back arrow again.
    f) To start archiving process (batch-job), press the 'Start Date' button and select the time when you want to start this process. Select 'Immediate' for instant processing and press the 'Save' button on the bottom of the Start Time' window.
    g) Select the 'Spool Parameter' button and save entries. Eventually enter a valid printer to have the result outputted.
    h) You are ready now to start the process. Press the 'Start' button and monitor the success with the 'Job Overview' button You can also go the 'fast path' by using transaction SE38, program RM06IW30 to archive info records. For large data archiving, use the background jobs and run those during off-peak times. If you run the program online, you will see a confirmation on the status bar telling 'New Archive file created:.... ' 
    Delete Archived Records: Info Record
    a) Follow the menu path: Tools - Administration - Administration - Archiving
    b) Select the Object Name MM_EINA for info records
    c) Select the menu button 'Delete'
    d) Select the menu button: 'Archive Selection'
    e) Click the archive created in previous step
    f) Select Start Date for process and Spool Parameters for output
    g) Submit selection.
    h) Check status by pressing the Job Overview button 
    regards
    Antony

  • Archiving only compressed InfoCubes?

    Hello Colleges,
    We have an InfoCube with 5 Mill Records. If I try to select  a daterange for Archiving, I get allways the error the request is not compressed. I could not find in any document any record size when it has to be compressed. Is it a generall restriction, which is never mentioned in a document. Does anybody has experince with archiving and had also this problem?
    Thanks for the help
    Henning

    Hi,
    It is recommended to archive only fully compressed InfoCubes as you want to ensure that you do not archive data from requests while they are being loaded.
    In addition, if you InfoCube has a non-cumulative key figure, there is a hard check to ensure all requests are compressed first before an archive job can execute.
    I hope this helps,
    Mike

  • Servererror Trigger to mask ORA error

    When I place one tablespace offline, I get ORA-00376 and ORA-01110 errors. Ok, must return these errors same.
    The problem is that the ORA-01110 error shows the way of the archive of tablespace that is offline.
    I would like to mask this error through a servererror trigger.
    But errors 376 and 1110 together continue being shown with the message that I configured.
    Below it is my trigger.
    CREATE OR REPLACE TRIGGER check_tbs_status
    AFTER SERVERERROR ON DATABASE
    BEGIN
    if is_servererror(376) and is_servererror(1110) then
    raise_application_error(-20002,'TABLESPACE OFFLINE',true);
    end if;
    end;
    Suggestions?

    This was my last test, but if I to use only one of the errors, the same thing happens.
    Examples:
    CREATE OR REPLACE TRIGGER check_tbs_status
    AFTER SERVERERROR ON DATABASE
    BEGIN
    if is_servererror(1110) then
    raise_application_error(-20002,'TABLESPACE OFFLINE',true);
    end if;
    end;
    OR
    CREATE OR REPLACE TRIGGER check_tbs_status
    AFTER SERVERERROR ON DATABASE
    BEGIN
    if is_servererror(376)then
    raise_application_error(-20002,'TABLESPACE OFFLINE',true);
    end if;
    end;
    OR
    CREATE OR REPLACE TRIGGER check_tbs_status
    AFTER SERVERERROR ON DATABASE
    BEGIN
    if is_servererror(376) OR is_servererror(1110) then
    raise_application_error(-20002,'TABLESPACE OFFLINE',true);
    end if;
    end;
    OR with <raise_application_error(-20002,'TABLESPACE OFFLINE',FALSE);>
    In all cases the same thing happens.

  • BI 7.0 - Duplicate Record Error while loading master data

    I am working on BI 7.0 and I am trying to load master data to an info object.
    I created an Infopackage and loaded into PSA.
    I created transformation and DTP and I get an error after I execute the DTP about duplicate records.
    I have read all previous threads about duplicate record error while loading master data and most of the them suggested to check 'Ignore duplicate records' option in the infopackage. But in 7.0, I can only load to PSA with infopackage and it doesn't have any option for me to ignore duplicate records.
    My data is getting loaded to PSA fine and I get this error while loading to info object using DTP.
    I would appreciate your help to resolve this issue.
    Regards,
    Ram.

    Hi,
    Refer:
    http://help.sap.com/saphelp_nw2004s/helpdata/en/45/2a8430131e03c3e10000000a1553f6/frameset.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/d0/538f3b294a7f2de10000000a11402f/frameset.htm
    With rgds,
    Anil Kumar Sharma .P

  • Bulk collect limit 1000 is looping only 1000 records out of 35000 records

    In below code I have to loop around 35000 records for every month of the year starting from Aug-2010 to Aug-2011.
    I am using bulk collect with limit clause but the problem is:
    a: Limit clause is returning only 1000 records.
    b: It is taking too much time to process.
    CREATE OR REPLACE PACKAGE BODY UDBFINV AS
    F UTL_FILE.FILE_TYPE;
    PV_SEQ_NO NUMBER(7);
    PV_REC_CNT NUMBER(7) := 0;
    PV_CRLF VARCHAR2(2) := CHR(13) || CHR(10);
    TYPE REC_PART IS RECORD(
    PART_NUM PM_PART_HARSH.PART_NUM%TYPE,
    ON_HAND_QTY PM_PART_HARSH.ON_HAND_QTY%TYPE,
    ENGG_PREFIX PM_PART_HARSH.ENGG_PREFIX%TYPE,
    ENGG_BASE PM_PART_HARSH.ENGG_BASE%TYPE,
    ENGG_SUFFIX PM_PART_HARSH.ENGG_SUFFIX%TYPE);
    TYPE TB_PART IS TABLE OF REC_PART;
    TYPE REC_DATE IS RECORD(
    START_DATE DATE,
    END_DATE DATE);
    TYPE TB_MONTH IS TABLE OF REC_DATE;
    PROCEDURE MAIN IS
    /* To be called in Scheduler Programs Action */
    BEGIN
    /* Initializing package global variables;*/
    IFMAINT.V_PROG_NAME := 'FULL_INVENTORY';
    IFMAINT.V_ERR_LOG_TAB := 'UDB_ERR_FINV';
    IFMAINT.V_HIST_TAB := 'UDB_HT_FINV';
    IFMAINT.V_UTL_DIR_NAME := 'UDB_SEND';
    IFMAINT.V_PROG_TYPE := 'S';
    IFMAINT.V_IF_TYPE := 'U';
    IFMAINT.V_REC_CNT := 0;
    IFMAINT.V_DEL_INS := 'Y';
    IFMAINT.V_KEY_INFO := NULL;
    IFMAINT.V_MSG := NULL;
    IFMAINT.V_ORA_MSG := NULL;
    IFSMAINT.V_FILE_NUM := IFSMAINT.V_FILE_NUM + 1;
    IFMAINT.LOG_ERROR; /*Initialize error log table, delete prev. rows*/
    /*End of initialization section*/
    IFMAINT.SET_INITIAL_PARAM;
    IFMAINT.SET_PROGRAM_PARAM;
    IFMAINT.SET_UTL_DIR_PATH;
    IFMAINT.GET_DEALER_PARAMETERS;
    PV_SEQ_NO := IFSMAINT.GENERATE_FILE_NAME;
    IF NOT CHECK_FILE_EXISTS THEN
    WRITE_FILE;
    END IF;
    IF IFMAINT.V_BACKUP_PATH_SEND IS NOT NULL THEN
    IFMAINT.COPY_FILE(IFMAINT.V_UTL_DIR_PATH,
    IFMAINT.V_FILE_NAME,
    IFMAINT.V_BACKUP_PATH_SEND);
    END IF;
    IFMAINT.MOVE_FILE(IFMAINT.V_UTL_DIR_PATH,
    IFMAINT.V_FILE_NAME,
    IFMAINT.V_FILE_DEST_PATH);
    COMMIT;
    EXCEPTION
    WHEN IFMAINT.E_TERMINATE THEN
    IFMAINT.V_DEL_INS := 'N';
    IFMAINT.LOG_ERROR;
    ROLLBACK;
    UTL_FILE.FCLOSE(F);
    IFMAINT.DELETE_FILE(IFMAINT.V_UTL_DIR_PATH, IFMAINT.V_FILE_NAME);
    RAISE_APPLICATION_ERROR(IFMAINT.V_USER_ERRCODE, IFMAINT.V_ORA_MSG);
    WHEN OTHERS THEN
    IFMAINT.V_DEL_INS := 'N';
    IFMAINT.V_MSG := 'ERROR IN MAIN PROCEDURE ||IFMAINT.V_PROG_NAME';
    IFMAINT.V_ORA_MSG := SUBSTR(SQLERRM, 1, 255);
    IFMAINT.V_USER_ERRCODE := -20101;
    IFMAINT.LOG_ERROR;
    ROLLBACK;
    UTL_FILE.FCLOSE(F);
    IFMAINT.DELETE_FILE(IFMAINT.V_UTL_DIR_PATH, IFMAINT.V_FILE_NAME);
    RAISE_APPLICATION_ERROR(IFMAINT.V_USER_ERRCODE, IFMAINT.V_ORA_MSG);
    END;
    PROCEDURE WRITE_FILE IS
    CURSOR CR_PART IS
    SELECT A.PART_NUM, ON_HAND_QTY, ENGG_PREFIX, ENGG_BASE, ENGG_SUFFIX
    FROM PM_PART_HARSH A;
    lv_cursor TB_PART;
    LV_CURR_MONTH NUMBER;
    LV_MONTH_1 NUMBER := NULL;
    LV_MONTH_2 NUMBER := NULL;
    LV_MONTH_3 NUMBER := NULL;
    LV_MONTH_4 NUMBER := NULL;
    LV_MONTH_5 NUMBER := NULL;
    LV_MONTH_6 NUMBER := NULL;
    LV_MONTH_7 NUMBER := NULL;
    LV_MONTH_8 NUMBER := NULL;
    LV_MONTH_9 NUMBER := NULL;
    LV_MONTH_10 NUMBER := NULL;
    LV_MONTH_11 NUMBER := NULL;
    LV_MONTH_12 NUMBER := NULL;
    lv_month TB_MONTH := TB_MONTH();
    BEGIN
    IF CR_PART%ISOPEN THEN
    CLOSE CR_PART;
    END IF;
    FOR K IN 1 .. 12 LOOP
    lv_month.EXTEND();
    lv_month(k).start_date := ADD_MONTHS(TRUNC(SYSDATE, 'MM'), - (K + 1));
    lv_month(k).end_date := (ADD_MONTHS(TRUNC(SYSDATE, 'MM'), -K) - 1);
    END LOOP;
    F := utl_file.fopen(IFMAINT.V_UTL_DIR_NAME, IFMAINT.V_FILE_NAME, 'W');
    IF UTL_FILE.IS_OPEN(F) THEN
    /*FILE HEADER*/
    utl_file.put_line(F,
    RPAD('$CUD-', 5, ' ') ||
    RPAD(SUBSTR(IFMAINT.V_PANDA_CD, 1, 5), 5, ' ') ||
    RPAD('-136-', 5, ' ') || RPAD('000000', 6, ' ') ||
    RPAD('-REDFLEX-KA-', 13, ' ') ||
    RPAD('00000000-', 9, ' ') ||
    RPAD(IFMAINT.V_CDS_SPEC_REL_NUM, 5, ' ') ||
    RPAD('CD', 2, ' ') ||
    RPAD(TO_CHAR(SYSDATE, 'MMDDYY'), 6, ' ') ||
    LPAD(IFSMAINT.V_FILE_NUM, 2, 0) ||
    RPAD('-', 1, ' ') || RPAD(' ', 9, ' ') ||
    RPAD('-', 1, ' ') || RPAD(' ', 17, ' ') ||
    RPAD('CD230', 5, ' ') ||
    RPAD(TO_CHAR(SYSDATE, 'MMDDYY'), 6, ' ') ||
    LPAD(IFSMAINT.V_FILE_NUM, 2, 0) ||
    LPAD(PV_REC_CNT, 8, 0) || RPAD(' ', 5, ' ') ||
    RPAD('00000000', 8, ' ') || RPAD('CUD', 3, ' ') ||
    RPAD(IFMAINT.V_CDS_SPEC_REL_NUM, 5, ' ') ||
    RPAD(IFMAINT.V_GEO_SALES_AREA_CD, 3, ' ') ||
    RPAD(IFMAINT.V_FRANCHISE_CD, 2, ' ') ||
    RPAD(IFMAINT.V_DSP_REL_NUM, 9, ' ') ||
    RPAD('00136REDFLEX', 12, ' ') || RPAD(' ', 1, ' ') ||
    RPAD('KA', 2, ' ') || RPAD('000000', 6, ' ') ||
    RPAD('00D', 3, ' ') ||
    RPAD(IFMAINT.V_VENDOR_ID, 6, ' ') ||
    RPAD(IFSMAINT.V_FILE_TYPE, 1, ' ') ||
    RPAD('>', 1, ' ') || PV_CRLF);
    /*LINE ITEMS*/
    OPEN CR_PART;
    FETCH CR_PART BULK COLLECT
    INTO lv_cursor limit 1000;
    FOR I IN lv_cursor.FIRST .. lv_cursor.LAST LOOP
    SELECT SUM(A.BILL_QTY)
    INTO LV_CURR_MONTH
    FROM PD_ISSUE A, PH_ISSUE B
    WHERE A.DOC_TYPE IN ('CRI', 'RRI', 'RSI', 'CSI')
    AND A.DOC_NUM = B.DOC_NUM
    AND B.DOC_DATE BETWEEN TRUNC(SYSDATE, 'MM') AND SYSDATE
    AND A.PART_NUM = LV_CURSOR(i).PART_NUM;
    FOR J IN 1 .. 12 LOOP
    SELECT SUM(A.BILL_QTY)
    INTO LV_MONTH_1
    FROM PD_ISSUE A, PH_ISSUE B
    WHERE A.DOC_TYPE IN ('CRI', 'RRI', 'RSI', 'CSI')
    AND A.DOC_NUM = B.DOC_NUM
    AND B.DOC_DATE BETWEEN lv_month(J).start_date and lv_month(J)
    .end_date
    AND A.PART_NUM = LV_CURSOR(i).PART_NUM;
    END LOOP;
    utl_file.put_line(F,
    RPAD('IL', 2, ' ') ||
    RPAD(TO_CHAR(SYSDATE, 'RRRRMMDD'), 8, ' ') ||
    RPAD(LV_CURSOR(I).ENGG_PREFIX, 6, ' ') ||
    RPAD(LV_CURSOR(I).ENGG_BASE, 8, ' ') ||
    RPAD(LV_CURSOR(I).ENGG_SUFFIX, 6, ' ') ||
    LPAD(LV_CURSOR(I).ON_HAND_QTY, 7, 0) ||
    LPAD(NVL(LV_CURR_MONTH, 0), 7, 0) ||
    LPAD(LV_MONTH_1, 7, 0) || LPAD(LV_MONTH_2, 7, 0) ||
    LPAD(LV_MONTH_3, 7, 0) || LPAD(LV_MONTH_4, 7, 0) ||
    LPAD(LV_MONTH_5, 7, 0) || LPAD(LV_MONTH_6, 7, 0) ||
    LPAD(LV_MONTH_7, 7, 0) || LPAD(LV_MONTH_8, 7, 0) ||
    LPAD(LV_MONTH_9, 7, 0) || LPAD(LV_MONTH_10, 7, 0) ||
    LPAD(LV_MONTH_11, 7, 0) ||
    LPAD(LV_MONTH_12, 7, 0));
    IFMAINT.V_REC_CNT := IFMAINT.V_REC_CNT + 1;
    END LOOP;
    CLOSE CR_PART;
    /*TRAILER*/
    utl_file.put_line(F,
    RPAD('$EOF-', 5, ' ') || RPAD('320R', 4, ' ') ||
    RPAD(SUBSTR(IFMAINT.V_PANDA_CD, 1, 5), 5, ' ') ||
    RPAD(' ', 5, ' ') ||
    RPAD(IFMAINT.V_GEO_SALES_AREA_CD, 3, ' ') ||
    RPAD(TO_CHAR(SYSDATE, 'MM-DD-RR'), 6, ' ') ||
    LPAD(IFSMAINT.V_FILE_NUM, 2, 0) ||
    LPAD(IFMAINT.V_REC_CNT, 8, 0) || 'H' || '>' ||
    IFMAINT.V_REC_CNT);
    utl_file.fclose(F);
    IFMAINT.INSERT_HISTORY;
    END IF;
    END;
    FUNCTION CHECK_FILE_EXISTS RETURN BOOLEAN IS
    LB_FILE_EXIST BOOLEAN := FALSE;
    LN_FILE_LENGTH NUMBER;
    LN_BLOCK_SIZE NUMBER;
    BEGIN
    UTL_FILE.FGETATTR(IFMAINT.V_UTL_DIR_NAME,
    IFMAINT.V_FILE_NAME,
    LB_FILE_EXIST,
    LN_FILE_LENGTH,
    LN_BLOCK_SIZE);
    IF LB_FILE_EXIST THEN
    RETURN TRUE;
    END IF;
    RETURN FALSE;
    EXCEPTION
    WHEN OTHERS THEN
    RETURN FALSE;
    END;
    END;

    Try this:
    OPEN CR_PART;
    loop
    FETCH CR_PART BULK COLLECT
    INTO lv_cursor limit 1000;
    exit when CR_PART%notfound;
    FOR I IN lv_cursor.FIRST .. lv_cursor.LAST LOOP
    SELECT SUM(A.BILL_QTY)
    INTO LV_CURR_MONTH
    FROM PD_ISSUE A, PH_ISSUE B
    WHERE A.DOC_TYPE IN ('CRI', 'RRI', 'RSI', 'CSI')
    AND A.DOC_NUM = B.DOC_NUM
    AND B.DOC_DATE BETWEEN TRUNC(SYSDATE, 'MM') AND SYSDATE
    AND A.PART_NUM = LV_CURSOR(i).PART_NUM;
    FOR J IN 1 .. 12 LOOP
    SELECT SUM(A.BILL_QTY)
    INTO LV_MONTH_1
    FROM PD_ISSUE A, PH_ISSUE B
    WHERE A.DOC_TYPE IN ('CRI', 'RRI', 'RSI', 'CSI')
    AND A.DOC_NUM = B.DOC_NUM
    AND B.DOC_DATE BETWEEN lv_month(J).start_date and lv_month(J)
    .end_date
    AND A.PART_NUM = LV_CURSOR(i).PART_NUM;
    END LOOP;
    utl_file.put_line(F,
    RPAD('IL', 2, ' ') ||
    RPAD(TO_CHAR(SYSDATE, 'RRRRMMDD'), 8, ' ') ||
    RPAD(LV_CURSOR(I).ENGG_PREFIX, 6, ' ') ||
    RPAD(LV_CURSOR(I).ENGG_BASE, 8, ' ') ||
    RPAD(LV_CURSOR(I).ENGG_SUFFIX, 6, ' ') ||
    LPAD(LV_CURSOR(I).ON_HAND_QTY, 7, 0) ||
    LPAD(NVL(LV_CURR_MONTH, 0), 7, 0) ||
    LPAD(LV_MONTH_1, 7, 0) || LPAD(LV_MONTH_2, 7, 0) ||
    LPAD(LV_MONTH_3, 7, 0) || LPAD(LV_MONTH_4, 7, 0) ||
    LPAD(LV_MONTH_5, 7, 0) || LPAD(LV_MONTH_6, 7, 0) ||
    LPAD(LV_MONTH_7, 7, 0) || LPAD(LV_MONTH_8, 7, 0) ||
    LPAD(LV_MONTH_9, 7, 0) || LPAD(LV_MONTH_10, 7, 0) ||
    LPAD(LV_MONTH_11, 7, 0) ||
    LPAD(LV_MONTH_12, 7, 0));
    IFMAINT.V_REC_CNT := IFMAINT.V_REC_CNT + 1;
    END LOOP;
    end loop;
    CLOSE CR_PART;

  • Archiving Purchase Info Records

    Hi,
    Can you please let me know what are the prerequsites that need to be completed before archiving a Purchase Info Record?
    do we need to check for any other dependency/flag other than the Mark for deletion (table field LOEKZ   'X')
    Please any one of you sugest me the solution for the above questions;
    Thanks,
    Jeevan.

    Hi Jeevan,
    In order to be able to delete an info record, you have to mark the Info Records for deletion. 
    To do so, use transaction ME15 and flag record for deletion. After you have deleted all info records, use archiving program RM06IW30 or follow the menu path Logistics - Materials Management u2013Purchasing - Master Data - Info Record - Follow on Functions - Archive. 
    Make sure that you only archive Records Marked for deletion as all records may be deleted! 
    Create Archive File: Info Record:
    a) Select Action: Archive and enter a new Variant, for example: Z_EINA_ARCH_ 01, press 'Maintain'
    b) On selection screen enter the data range (Vendor, Material, etc.) you want to archive.
    c) Deselect the 'Test' flag if you don't want to test first.
    d) Press green back-arrow and enter the description of this new variantt on the screen which follows.
    e) Save the variantt which brings you again to the selection screen. Press green back arrow again.
    f) To start archiving process (batch-job), press the 'Start Date' button and select the time when you want to start this process. Select 'Immediate' for instant processing and press the 'Save' button on the bottom of the Start Time' window.
    g) Select the 'Spool Parameter' button and save entries. Eventually enter a valid printer to have the result outputted.
    h) You are ready now to start the process. Press the 'Start' button and monitor the success with the 'Job Overview' button You can also go the 'fast path' by using transaction SE38, program RM06IW30 to archive info records. For large data archiving, use the background jobs and run those during off-peak times. If you run the program online, you will see a confirmation on the status bar telling 'New Archive file created:.... ' Delete Archived Records: Info Record
    a) Follow the menu path: Tools - Administration - Administration - Archiving
    b) Select the Object Name MM_EINA for info records
    c) Select the menu button 'Delete'
    d) Select the menu button: 'Archive Selection'
    e) Click the archive created in previous step
    f) Select Start Date for process and Spool Parameters for output
    g) Submit selection.
    h) Check status by pressing the Job Overview button
    Reg,
    Ashok
    Assign points if useful.

  • Archive Source Files with Errors

    Hi All,
    I need to archive my sending file in repective directory, if it has errors..
    It is mentioned in library that A permanent error occurs either during the conversion of the file content, or in a module in the module processor.
    What is permanent error?
    For example if my input file name is abc.txt which containg 3 fields id,name,address.
    i have configured sending file adapter with FCC .
    now i have given some faulty entries in abc.txt like for example 4 fields separated by comma,
    but this is a error.Asper FCC i have given only 3 fields..
    so this error file has to archive in so and so folder.
    If i check <b>Archive Source Files with Errors</b> parameter in sending file adapter.
    and specified the specified directory.
    but its not working for me. file is not picking but its not coming in the archive folder.
    In library it is mentioned to see message transfer bean link.
    there i found <b>Transform.PermanentErrors: True/False</b>.
    If we have not used message transfer bean then, in that case will this parameter <b>archive source files with errors</b> will work for FCC?
    Please suggest me .
    I would really appreciate if anyone could send a complete blog for this how to archive source files with errors in FCC?
    Thanks in Advance..
    Regards,
    Reehan

    Reehan,
    I am not sure if you are still stuck with your issue of archiving faulty source files but I came across the same issue. Please check your mode of processing. The faulty source files archiving feature will work good in ARCHIVE mode but not in test or delete mode. I haven't check for set for read-only mode.
    Good luck!

  • Use of "Archive Source Files with Errors" for BIC module error

    Hi All,
    I have Edi file to Idoc scenario , where i am using SEEBURGER BICXIADAPTER.MODULE.
    My question is if the input file caught in error with BIC Module e.g
    "MP: exception caught with cause Error in BICMODULE-module:Temporary error: BIC XI Adapter call failed. Reason: SegmentDescription: checkAndResetChildrenCounter() not enough repetitions of the segment UNH found: 0 instead of 1 offset :80 DESCRIPTION: SegmentDescription Warning: Not enough repetitions of the segment UNH ([not specified]/[not specified])."
    can we move these errornous files to another directory using Processing parameter  "Archive Source Files with Errors" available with Sender File Adapter?
    In File Adpater my Module tab looks like :
    1     localejbs/CallBicXIRaBean     Local Enterprise Bean     bic
    2     localejbs/CallSapAdapter     Local Enterprise Bean     0
    In my scenario  , its not working..  do we have any other option to achive this?

    Hi Jyoti,
    I had a raised a similar case with SAP and came to know that archiving only works if the error raised by the module is so called "Permanent". However, bic doesn't raise permanent error due to which archiving of error files won't be possible. Seeburger haven't came up with a solution yet.
    Regards,
    Prateek

  • INSERT works for only last record of the Internal table ...??

    I am trying to insert data from an internal table JTAB to a Databse Table
    in CRM. The name of the databse table in CRM is CRMD_PARTNER
    For this i first declared an internal table JTAB with same structure as that of
    the databse table CRMD_PARTNER
    DATA: BEGIN OF JTAB OCCURS 0,
         INCLUDE STRUCTURE CRMD_PARTNER,
          END OF JTAB.
    Then I filled my JTAB with the required entries. For filling the JTAB i am
    getting data from couple of other tables and filling the internal table JTAB
         loop at ktab.
    here i am filling values in JTAB and then i say APPEND JTAB.
            endloop.
         Till here every thing works well and my internal table JTAB has all values that
    I need to insert to the CRMD_PARTNER table. Also my JTAB has vales for all primary key fields.
    Now I write a condition like below.
    if not jtab[] inital.
    MODIFY CRMD_PARTNER from TABLE JTAB.
    COMMIT WORK.
    endif.
    Here comes the problem...once this code is executed and once program totally executes...
    i always see that only last record in JTAB is being inserted to the databse table CRMD_PARTNER.
    When i check in the debugger ..i see that the loop is being executed only once
    and the SY-TABIX of JTAB is being always set to the length of JTAB and
    that is why only last  record is being inseretd to the table CRMD_PARTNER.
    Now what should i do ..to insert all records of JTAB to CRMD_PARTNER table?
    is there a way i can restet the SY-TABIX and make the modify statement work for all records of JTAB?
    Finally i hardcoded the values in JTAB then i dont have any problem and all records of JTAB are being
    inserted into the database table.
    only if i am dynamically fetching the values into JTAB(by Putting in loop..endloop statements and
    pushing values to JTAB and APPENDING JTAB ) i have a problem .
    in this case also my JTAB is being correctly filled with values ..but when i try to
    insert theses values to CRMD_PARTNER with MODIFY stmt ..only always last record is being inserted.
    i tried with INSERT instead of MODIFY and i am getting runtime error.
    kindly pease help.
    Regards,
    Jessica Sam

    yes a@s i am really struggling from last 3 days.
    I actually want to create a sales order in CRM and want to
    assign ship to party for each line item in a sales order in CRM
    for that i used the standard bapi BAPI_BUSPROCESSND_CREATEMULTI and i am able
    to create an order successfully, but i see that the ship toparty
    that i give at header is being copied to each line item.
    but this is not what is expected.
    So finally i tracked in which table the ship to party gets stored in each line item
    and then tried to insert the records directly into databse. and it worked
    but only if i am hardcoding..but not when i am dynamically fecthing the values into JTAB for insertion,
    Can yoy help ..if you have any idea why insert/modify is processing only last record?
    any help will be highly appreciated.
    Regards,
    Jessica Sam

  • Duplicate Records error when processing transaction file....BPC 7.0

    Hi All,
    I have a situation. I am using BPC NW 7.0 and I have updated my dimension files. When I try to validate my transaction file every single record is validated successfully. But when I try to import the flat file into my application, I am getting a lot of dupplication records error and these are my questions.
    1. Will we get duplicate records in transaction files?
    2. Even if there are duplication, since it is a cube it should summarize not display that as a error and reject records?
    3. Is there something I can do to accept duplicates (I have checked the Replace option in the data package, to overwrite the simillar records, but it is only for account, category and entity only.
    5. In mycase I see identical values in all my dimension and the $value is the only difference. Why is it not summing up.
    Your quickest reply is much appreciated.
    Thanks,
    Alex.

    Hi,
    I have the same problem.
    In my case the file that I want to upload has different row that differ for the nature column. In the conversion file I map different nature to one internal nature.
    ES: cost1 --> cost
          cost2 --> cost
          cost3 --> cost
    In my desire was that in BPC the nature cost assume the result  cost = cost1 + cost2 + cost3.
    The result is that only the first record is uploaded and all other recorda are rejected as duplicate.
    Any suggestion?

  • Archive Purchase Info Record MM_EINA : problem with CDHDR

    Hello,
    I have notice that when i archive an Info Record, Only CDPOS is deleted and not CDHDR.
    In the customizing, Object Class CHANGEDOCU is set.
    Can someone explain why datas in CDHDR are not deleted ?
    Thank you,
    Cipriano

    Yes, I agree with Jurgen.  You should open a message with SAP.  As, the CDHDR records should be archived along with the CDPOS records.  That is how it works when I archive purchase info records with MM_EINA.
    Best Regards,
    Karin Tillotson

  • Duplicate records error?

    hello all
    while extracting master data am getting duplicate records error?
    how do i rectify this?
    in infopackage screen in processing tab,will i get the  option " ignore double data records",?
    when will this option will be enable?
    regards

    Hello
    This option will be available only for Master Data and not for Transactional Data. You could control the Duplicate Records for Transactional Data in ODS, there is an option in the ODS Settings.
    ***F1 Help
    Flag: Handling of duplicate data records
    From BW 3.0 you can determine for DataSources for master data attributes and texts whether the extractor transfers more than one data record in a request for a value belonging to time-independent master data.
    Independently of the extractor settings (the extractor potentially delivers duplicate data records) you can use this indicator to tell the BW whether or not you want it to handle any duplicate records.
    This is useful if the setting telling the extractor how to handle duplicate records is not active, but the system is told from another party that duplicate records are being transferred (for example, when data is loaded from flat files).
    Sankar

  • PPro CS4 "unknown recorder error"

    Premiere Pro CS4 and Premiere Elements 7.0 are both having the same problem on this machine.  During the digitizing process (F5 - Capture window) it will randomly stop digitizing and a window titled Recorder Error pops up with the unhelpful message "Unknown recorder error." and only an "OK" button.
    Here's the machine details:
    XP
    Professional
    Version 2002
    SP3
    HP Compaq dc7600 Convertible Intel(R)
    Pentium(R) D CPU 3.40GHz
    3.39 Ghz, 3.00 GB of RAM
    NVIDIA GeForce 7300 LE Graphics card
    We're digitizing video from a Canon GL2 via Firewire:
    Texas Instruments OHCI Compliant IEEE 1394 Host Controller
    What other info would a trouble-shooter want to know?
    This issue is consistent for mulitple users (this machine has multiple users with unique log-ins and user profiles in Windows).
    It's worked fine until a couple of months ago.  An identical machine is runing Production Suite CS4 just fine.  That machine has only one primary user (me).  The other gets used by a half a dozen people, but the admin. functions are all locked by our IT department.  Can anyone provide links to trusted trouble-shooting instructions or give me some idea of what the problem might be?
    Any assistance or insight would be greatly appreciated.
    Anthony Roberts
    JournalStar.com

    We hauled three of our systems to the Mirage in Las Vegas last week where we had to do a live show record to disk and create a bunch of DVD's for
    the event attendees. We ran component from the staging company's equipment into the back of two DV Cam decks and from there we ran firewire cables into our 2 computers.  We were getting the dreaded "unknown record error" on both systems at first.  On both systems, I "Ran" msconfig and disabled all unnecessary startup items.  One of the systems never had an "unknown record error" after that. The other system continued to have the error approx 3-5 times during a 3 hour capture. OnLocation crashed when we hit the record button so we captured everything through Premiere CS4. I think the problem may have been the drive we were capturing to. We were bad.  Harm can probably guess what we did.
    BTW, on the last day before my flight home, I netted a whopping $6 at a craps table.  One of the young ladies I work with won $1100 placing daring but foolish bets.  I watched a guy, not with our party lose over $6000 in a matter of minutes. I asked one of the guys working at the table what was the most he ever saw someone win at the craps table. he said 1.4 million.

  • Unknown recorder error CS6 premiere pro

    I know work with CS6 for over 7 months. Always smoothly till this week. I wanted to capture a film via tape (sony HVR-V1E), my system diretly gave the following message: unknown recorder error. I klicked "okay" and gladly, I thought,  it started to capture the whole tape. It looked great on my canvas.
    I have edit the film but when I look at the final film on HD TV screen, I see a lot of noise, the footages are in bad shape. Cant sent that to my customer.
    Also Premiere Pro crashed a couple of times while editting. It never did that before/
    Some extra information. Maybe it has nothing to do with the above mentioned problem but still I want to write it down, one never knows. This is the only thing that is changed on my system:
    Working on an iMac (with SSD), Quicktime needed an extra mpeg2 component for footages from a camcorder. Have replace the quicktime with a backup. Does that cost the problem?
    I am on Lion Mountain. Have an SSD. The Suite is on the SSD, the footages on another internal drive.
    What I have done already to solve this:
    - checked google and checked this forum and checked, read and searched for hrs.
    - checked if there were faillures on my SSD or the other HD drive (HD2, where I keep my footages)
    - Updated my suite
    - project sequence is HD 1440x1080
    - Checked the footages if they were HD, and they were. (I am in Europe, so in PAL, 25 frames_
    File Path: /Volumes/Macintosh HD 2/Video/project/xxxx.mpeg
    Type: MPEG Movie
    File Size: 67,3 MB
    Image Size: 1440 x 1080
    Frame Rate: 25,00
    Source Audio Format: 48000 Hz - compressed - Stereo
    Project Audio Format: 48000 Hz - 32 bit floating point - Stereo
    Total Duration: 00:00:20:16
    Pixel Aspect Ratio: 1,3333
    - firewire is 800 (from camera connected to imac with an adapter from FW400 to FW 800, was never a problem).
    - Have checked the above mentioned values with earlier projects and they are the same values. Earlier projects are very clear and really HD, nice to watch.
    - loaded the files into iMovie and also into FCP 7. output was the same problem
    - exported the final film to H.264
    - Exported the final film via Media Encoder
    - HD (both) have >125 gig space not used.
    - Put the footages on the SSD, thats where PPR is installed
    - cleaned the media caches
    - checked it on our HD smart TV and the output of the tape is fine, sharp.
    What can I try more. Please your help, the deadline is friday.. so I have one day to solve this problem/
    Thanks in advance.

    INEF1- Have you got an answer to this-Premiere pro cs 6 have started with me:
    It comes up with Blackmagic design intensity shuttle usb 3.0 and with the direct firewire into the pc.
    I have:
    -Installed latest updates
    - change the firewire
    -Taking out the firewire card and re inserted it
    -reset bios
    -Tried capturing through the live editing desk DATAVIDEO SE-500-that captured for about 39minutes but then same problem.
    2nd-It gives a low level error occurred capture device..I've look through the forums but don't find any related to this  capture device error
    thanks

Maybe you are looking for