Table splitting during upgradation

Hi,
Please read the SAP Note below.
In upgrade SAP had made changes to USR03 and SADR tables. The data that was previously stored in USR03 (SAP user) or in SADRP/SADR (office user) was merged in the tables of central address management (ADRP, ADRC, ADCP, ADR2, ADR3 and so on). The programs which are using those tables will not give syntax errors, but will not generate any output. As per the OSS note 96607:
  1. Replace accesses to table USR03 by calling function module SUSR_USER_ADDRESS_READ or BAPI_USER_GET_DETAIL.
2 Replace accesses to the addresses of an SAP application in table SADR by calling function module ADDR_GET.
Can you please let me know where this splitting or merging of tables information will be stored?
The means the data that was in USR03 or in SADR was merged in tables ADRP, ADRC, ADCP, ADR2, ADR3 .
Where this merging tables information will be stored?
Thanks,
Murthy

Hi Chitra,
Table splitting time depends upon number of tables for which table splitting needs to be done, no of parallel jobs and version of R3ta, R3load and R3szchk files.
How many tables are your splitting ?
How many parallel jobs have u defined ?
Have you updated R3ta, R3load and R3szchk ?
Regards,
Deepak Kori

Similar Messages

  • Table Maintenance, during upgrades

    Experts,
    How do we handle table maintenance generator screens and t-codes during upgrades (for Z tables)? Do they come up in the SPAU list post upgrade or is there anything to do in SPDD during ACT_<version> stage?
    Tx, in adv.

    Hi,
    Shaji is correct. Table maintenances are corrected after the upgrade is applied by the BASIS team. You will just have to re-test those table maintenances.
    If you are also doing Unicode Conversion as well. Then you need to make sure to set Unicode Checks Active for the function pool related to your table maintenance.
    i.e you have in table maintenance stored in function group ZMYFUNC. Then you need to go to the "generated" function pool SAPLZMYFUNC and set Unicode Checks Active.
    Unless your table maintenances have lots of custom logic in them, you generally would not encounter any errors at all.

  • T512w Table during upgrade

    hello,
    we are upgrading from 46c to ecc60,
    In HCM,can some one explain me how to handle the table t512w during the upgrade...
    any inputs are appreciated
    cheers

    Dear Sandeep,
    Release notes of ECC6.0 for Payroll :
    http://help.sap.com/saphelp_erp2005/helpdata/en/43/44c490d36157c0e10000000a155369/Chap%2025%20PY%20Payroll.pdf
    http://solutionbrowser.erp.sap.fmpmedia.com/ for differences between versions (give source and target version)
    Hope this will help.
    Regards,
    Naveen.

  • Error during Upgrade from 4.6c to ECC 6.0

    Hi All,
      We are facing an error when upgrading from 4.6c to ECC 6.0. We are facing this error on the table COEP - runtime object inconsistancy. What we found is there is ERP upgrade has created new extra fields in the table. In log file the error is specified as : Duplicate Field name, But we not able to find the duplicate field name in the table.  Please kindly help as early as possible. The upgrade process is stuck.
    Regards
    Anil Kumar K

    Hi Anil,
    Is this issue fixed? Can i know how you fixed it?
    replied to your message Re: How to adopt the index changes during upgrade.
    Thanks,
    Somar

  • Error in transaction KEA0 during upgrade

    Dear All,
    During upgrading the custmer system from 4.7 to ECC 6.0 we found some error in the transaction KEA0 while trying to activate the cross client for an operating concern.
    The activation logs says that there are the syntax errors in the gereted subroutine pool: RK2O0100.
    I however corrected the error and activated the code but again when I am trying to execute the same it gave me the same error and when I go to the sub routine pool I found that the changes are again undone and the previos version is activated. This happens repeatedly that the code is corrected and activated by me and says no error and ones I log off and login again it says the same error.
    Help will be appriciated.
    Thanks iin advance....
    Abhi...

    Dear Raymond,
    Thankds for your help..........
    I tring to run RKEAGENV it shows the error message with STOP button saying that the field WTGBTR not contained in the nametab for table CE1E_B1 and when Io run RKEAGENF it gives a short dump saying that entry for CE10100 is not allowed for TTABS.....
    Please help.

  • Unicode export:Table-splitting and package splitting

    Hi SAP experts,
    I know there are lot of forums related to this topic, but I have some new questions and hence posting a new thread.
    We are in the process of doing unicode conversion in our landscape(CRM 7.0 system based on NW 7.01) and we are running on AIX 6.1 and DB2 9.5. The database size is around 1.5 TB and hence, we want to go in for optimization for export and import in order to reduce the downtime.As a part of the process, we have tried to do table-splitting and parallel export-import to reduce the downtime.
    However, we are having some doubts whether this table-splitting has actually worked in our scenario,as the export has exeucted for nearly 28 hours.
    The steps followed by us :
    1.) Doing the export preparation using SAPINST
    2.) Doing table splitting preparation, by creating a table input file having entries in the format <tablename>%<No.of splits>.Also, we have used the latest R3ta file and the dbdb6slib.o(belonging to version 7.20 even though our system is on 7.01) using SAPINST.
    3.) Starting with the export using SAPINST.
    some observations and questions:
    1.) After completion of tablesplitting preparation, there were .WHR files that were generated for each of the tables in DATA directory of export location. However, how many .WHR files should be created and on what basis are they created?
    2.) I will take an example of a table PRCD_CLUST(cluster table) in our environment, which we had split. We had 29 *.WHR files that were created for this particular table. The number of splits given for this table was 36 and the table size is around 72 GB.Also, we noticed that the first 28 .WHR files for this table, had lots of records but the last 29th .WHR file, had only 1 record. But we also noticed that, the packages/splits for the 1st 28 splits were created quite fast but the last one,29th one took a long time(serveral hours) to get completed.Also,lots of packages were generated(around 56) of size 1 GB each for this 29th split. Also, there was only one R3load which was running for this 29th split, and was generating packages one by one.
    3.) Also,Our question here is that is there any thumb rule for deciding on the number of splits for a table.Also, during the export, are there any things that need to be specified, while giving the inputs when we use table splitting,in the screen?
    4.) Also, what exactly is the difference between table-splitting and package-splitting? Are they both effective together?
    If you have any questions and or need any clarifications and further inputs, please let me know.
    It would be great, if we could get any insights on this whole procedure, as we know a lot of things are taken care by SAPINST itself in the background, but we just want to be certain that we have done the right thing and this is the way it should work.
    Regards,
    Santosh Bhat

    Hi,
    First of all please ignore my very first response ... i have accidentally posted a response to some other thread...sorry for that 
    Now coming you your questions...
    > 1.) Can package splitting and table-splitting be used together? If yes or no, what exactly is the procedure to be followed. As I observed that, the packages also have entries of the tables that we decided to split. So, does package splitting or table-splitting override the other, and only one of the two can be effective at a time?
    Package splitting and table splitting works together, because both serve a different purpose
    My way of doing is ...
    When i do package split i choose packageLimit 1000 and also split out the tables (which i selected for table split)  into seperate package (one package per table). I do it because that helps me to track those table.
    Once the above is done i follow it up with the R3ta and wheresplitter for those tables.
    Followed by manual migration monitor to do export/import , as mentioned in the previous reply above you need to ensure you sequenced you package properly ... large tables are exported first , use sections in the package list file , etc
    > 2.) If you are well versed with table splitting procedure, could you describe maybe in brief the exact procedure?
    Well i would say run R3ta (it will create multiple select queries) followed by wheresplitter (which will just split each of the select into multiple WHR files)  ...  
    Best would go thought some document on table spliting and let me know if you have specific query. Dont miss the role of hints file.
    > 3.) Also, I have mentioned about the version of R3ta and library file in my original post. Is this likely to be an issue?Also, is there a thumb rule to decide on the no.of splits for a table.
    Rule is use executable of the kernel version supported by your system version. I am not well versed with 7.01 and 7.2 support ... to give you an example i should not use 700 R3ta on 640 system , although it works.
    >1.) After completion of tablesplitting preparation, there were .WHR files that were generated for each of the tables in DATA directory of export location. However, how many .WHR files should be created and on what basis are they created?
    If you ask for 10 split .... you will get 10 splits or in some case 11 also, the reason might be the field it is using to split the table (the where clause). But not 100% sure about it.
    > 2) I will take an example of a table PRCD_CLUST(cluster table) in our environment, which we had split. We had 29 *.WHR files that were created for this particular table. The number of splits given for this table was 36 and the table size is around 72 GB.Also, we noticed that the first 28 .WHR files for this table, had lots of records but the last 29th .WHR file, had only 1 record. But we also noticed that, the packages/splits for the 1st 28 splits were created quite fast but the last one,29th one took a long time(serveral hours) to get completed.Also,lots of packages were generated(around 56) of size 1 GB each for this 29th plit. Also, there was only one R3load which was running for this 29th split, and was generating packages one by one.
    Not sure why you got 29 split when you asked for 36, one reason might be the field (key) used for split didn't have more than 28 unique records. I dont know how is PRCD_CLUST  split , you need to check the hints file for "key". One example can be suppose my table is split using company code, i have 10 company codes so even if i ask for 20 splits i will get only 10 splits (WHR's).
    Yes the 29th file will always have less records, if you open the 29th WHR you will see that it has the "greater than clause". The 1st and the last WHR file has the "less than" and "greater than" clause , kind of a safety which allows you to prepare for the split even before you have downtime has started. This 2 WHR's ensures  that no record gets missed, though you might have prepared your WHR files week before the actual migration.
    > 3) Also,Our question here is that is there any thumb rule for deciding on the number of splits for a table.Also, during the export, are there any things that need to be specified, while giving the inputs when we use table splitting,in the screen?
    Not aware any thumb rule. First iteration you might choose something like 10 for 50 GB , 20 for 100 GB. If any of the tables overshoots the window. They you can give a try by  increase or decrease the number of splits for the table. For me couple of times the total export/import  time have improved by reducing the splits of some tables (i suppose i was oversplitting those tables).
    Regards,
    Neel
    Edited by: Neelabha Banerjee on Nov 30, 2011 11:12 PM

  • MAIN_SHDRUN/ACT_UPG ends during upgrade

    During upgrade from CRM 6.0 to 7.0 EHP1 MAIN_SHDRUN/ACT_UPG phase ends.
    In ACTUPG.ELG file I see:
    *******   LIST OF ERRORS AND RETURN CODES  *******
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    DDIC ACTIVATION ERRORS and RETURN CODE in SAPACCI702.CRT
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    1EEDO519X"Table" "CRMST_QUERY_OPP_BTIL" could not be activated
    1 ETP111 exit code
    I read SAPACCI702.CRT:
    3 DDT012XAktiviere Tabelle "CRMST_QUERY_OPP_BTIL"
    3 DDT402 Append-Struktur "ZCRMST_QUERY_OPP_BTIL" an Tabelle "CRMST_QUERY_OPP_BTIL" angefügt
    2WDDT191 Erweiterungskat. "3" möglich, aber Include oder Subtyp ist noch unklassif.
    1EDDT963 Feld "PROD_HIERARCHY" in Tabelle "CRMST_QUERY_OPP_BTIL" ist doppelt (Bitte überprüfen)
    3 DDT013 Tabelle "CRMST_QUERY_OPP_BTIL" wurde nicht aktiviert
    3 DDT014XAktiviere abhängige Tabelle "CRMST_QUERY_OPP_BTIL"
    3 DDT402 Append-Struktur "ZCRMST_QUERY_OPP_BTIL" an Tabelle "CRMST_QUERY_OPP_BTIL" angefügt
    2WDDT191 Erweiterungskat. "3" möglich, aber Include oder Subtyp ist noch unklassif.
    3 DDT038 Aktivierung abhängige Tabelle "CRMST_QUERY_OPP_BTIL" erfolgreich
    Oh... in german language... it's a problem for me. How can I check this structure? how?
    Have you got any suggest?
    Regards.

    Hi Ganimede,
    i have the similar kind of error can you please explain the steps for deleting the entry you have specified...
    would be greate if you coulkd reply earliest
    Thanks and Regards
    Amar Ghuge

  • Behaviour of ztables during upgrade

    Hi all
    I am trying to rationalise our ztables and would like to change the Delivery Class of some of our tables from 'W' to 'C' so that data can be transported through our landscape rather than manually entered on each system.
    I've read the SAP help on the Delivery Class and note that the behaviour during Upgrade differs between 'W' and 'C'...:
    <b>Client Dependant</b>
    Class A and C: Data is only imported into client 000. Existing data records are overwritten.
    Class E, S and W: Data is imported into all clients. Existing data records are overwritten.
    <b>Client Independant</b>
    Classes A, L and C: No data is imported.
    Classes E, S, and W: Data is imported. Existing data records with the same key are overwritten.
    ...but I don't really understand what it means in practice.  Is anybody able to explain to me what data is being imported and what the relevance of Client 000 is in this particular instance?
    Please don't just copy the SAP help as I have already read it!!!
    Thanks
    Andy

    Hi Andrew,
    I can tell you something from what I have observed.
    When we give class A the changes need to be done by us in all systems individualy.
    When Class C, it asks for a transport and we can have same value transported
    In this context client 000 is nothing but the client which you are using.So this would mean with A and C data is only in that client where you are using it or updating it.
    So for other client you need to either make data manually as in case of A or in case of C you need to import the Transport.
    From this behaviour it seems that for Class E,S and W the data that is being created in one client get's copied to other clients too on that system so it is like a system wide change in the data.
    Hope this helps in getting some idea.
    Regards
    Nishant

  • Mass structure and table Inconsistencies after upgrade

    Hi all,
    I'm on an upgrade project upgrading system from R/3 4.6 to SAP ECC 6.0 EHP4. After the upgrade, apparently there's a mass inconsistencies that happen in the table. There's at least 1000 table which can be activated because of this inconsistencies between runtime and DDIC based on the checking run using program RUTMSJOB.
    I've tried applying note 1248769 but it does not fix the problem. I've also tried the mass checking and activation feature of report RUTMSJOB which also does not fix the problem.
    Has anybody encounter this problem before? What's your solution on this problem?
    Here's the pic of the condition of the table in my SE11
    http://img683.imageshack.us/i/64131370.jpg/
    Here's the inconsistencies which is found:
    http://img375.imageshack.us/i/47550099.jpg/
    Kindly note that the same error happen even after I implement the note correction 1248769.
    Thanks
    -Suwardi-
    Edited by: Suwardi Nursalim on Jul 9, 2010 9:40 AM - Change pic link.

    Hello Suwardi,
    please inform which actions you took in SPDD and SPAU - this is key for a problem like yours.
    Did you have a transport request ready with modifications from a previous upgrade? Please check these notes:
    #68678:   Exporting modification adjustment for second upgrade
    #610311: Importing modification adjustment in second upgrade from 620
    Lastly, did you run the upgrade with latest sapup version AND latest upgrade fix?
    Now in order to fix the problem, you'll likely have to restore from backup and perform the upgrade again. Alternatively, please check note #1406740 (are you using IBM's DB2 as your database?). If you faced the problem described on this note (which points to errors in table DD03L during EU_IMP2 phase), you need to apply FP8 like on this note.
    Then you could proceed like the following to fix the problem:
    1. Restore the backup from phase MODPROF immediately before the downtime including the /usr/sap/P01/upg/abap directory
    2. reset the Upgrade with report RSUPGRES
    3. clear the upgrade directory
    4. Start the upgrade from the very beginning after installing Fix Pack 8 for your database.
    Please let us know if this helped.
    Tomas Black

  • Oracle packages - nested table cursor - 11g upgrade testing

    I have a package call Crisis_Contacts. This package has a function(generate_file) in it that has a create cusor statement which calls on another function(split) in another package called NCF_UTL. In testing our code for a 11g upgrade the cursor statement calling on the split function is throwing a Error: ORA-00600: internal error code, arguments: [4814], [5], [0], [0], [], [], [], [], [], [], [], [] when compiling the package, which it was not doing prior to upgrade. Can you suggest a possible workaround or mod to this code to stop the error from occuring.
    Basically, this code is reading html files and spliting them into section and loading them into the cursor. Then the cursor is read to write them to a table. Then it reads the table and transmits the lines of html to another location.
    CREATE OR REPLACE package body ACADMGR.crisis_contacts as
    function generate_file( file_type in varchar2,
    person_pidm in integer := null,
    person_id in varchar2 := null,
    dry_run in varchar2 := 'N',
    pidm_filter in varchar2 := null,
    include_placeholders in varchar2 := 'N',
    comments in varchar2 := null ) return integer is
    cursor typeDef is
    select * from crisis_contact_file_type ccft
    where ccft.file_type = generate_file.file_type;
    file_type_def typeDef%rowtype;
    cursor fieldDef is
    select * from crisis_contact_file_field ccff
    where ccff.file_type = generate_file.file_type
    order by seq asc;
    type r_typ is ref cursor;
    r r_typ;
    datasel varchar2(32767) := '';
    r_pidm integer;
    pers_pidm integer := person_pidm;
    prev_pidm integer;
    r_tag varchar2(255);
    r_val varchar2(4000);
    type field_list is table of varchar2(256);
    type field_index is table of integer index by varchar2(256);
    type val_list is table of varchar2(4000);
    type req_tbl is table of boolean index by varchar2(256);
    req_fields field_list := field_list();
    all_req_seen boolean;
    fields field_list := field_list();
    vals val_list := val_list();
    field_idx field_index;
    out_id integer;
    out_line integer := 0;
    buf varchar2(4000);
    saved_out_line integer := 0;
    xml_section_lf_delims ncf_utl.split_tbl := ncf_utl.split_tbl( ncf_utl.crlf, chr(13), chr(10) );
    cursor splitcur( txt in varchar2 ) is
    select column_value line from table(ncf_utl.split( txt, xml_section_lf_delims ));
    cursor get_output is
    select text from crisis_contact_file_line
    where output_id = out_id
    order by line_no asc;
    log_time date := sysdate;
    log_seq integer := 1;
    procedure outp( txt in varchar2 ) is
    begin
    --dbms_output.put_line( txt );
    insert into crisis_contact_file_line
    ( output_id, line_no, text )
    values
    ( out_id, out_line, txt );
    out_line := out_line + 1;
    end;
    procedure split_outp( txt in varchar2 ) is
    begin
    if txt is not null then
    for r in splitCur( txt ) loop
    outp( r.line );
    end loop;
    end if;
    end;
    begin
    open typeDef;
    fetch typeDef into file_type_def;
    close typeDef;
    if pers_pidm is null and person_id is not null then
    select distinct(spriden_pidm) into pers_pidm from spriden
    where spriden_id = person_id;
    end if;
    insert into crisis_contact_file_out ( file_type, comments ) values ( file_type, comments )
    returning output_id into out_id;
    datasel := gen_outfile_sql( file_type, pers_pidm, pidm_filter, include_placeholders );
    if dry_run = 'Y' then
    ncf_utl.output_text( datasel );
    else
    split_outp( file_type_def.preamble );
    buf := '';
    for f in fieldDef loop
    fields.extend;
    fields(fields.last) := f.alias;
    field_idx(f.output_field) := fields.last;
    vals.extend();
    if f.mandatory = 'Y' then
    req_fields.extend;
    req_fields(req_fields.last) := f.output_field;
    end if;
    if file_type_def.file_mode = 'DELIM' and file_type_def.print_heading = 'Y' then
    buf := buf || case when fieldDef%rowcount > 1 then nvl(file_type_def.delimiter,',') else '' end ||
    nvl(file_type_def.quote,'') || nvl(f.alias,f.output_field) || nvl(file_type_def.quote,'');
    end if;
    end loop;
    if length(buf) > 0 then
    outp(buf);
    end if;
    --outp( '<?xml version="1.0" encoding="UTF-8"?>' );
    --outp( '<recipients>' );
    open r for datasel;
    loop
    fetch r into r_pidm, r_tag, r_val;
    exit when r%notfound;
    if prev_pidm is null or prev_pidm != r_pidm then
    if prev_pidm is not null then
    split_outp( file_type_def.record_end );
    all_req_seen := true;
    --check and reset 'seen' flags for required fields
    for i in 1..req_fields.count loop
    if vals(field_idx(req_fields(i))) is null then
    all_req_seen := false;
    end if;
    end loop;
    --if not all_req_seen then
    -- rollback to USER_START;
    -- out_line := saved_out_line;
    --end if;
    if all_req_seen then
    buf := '';
    for i in 1..fields.count loop
    if file_type_def.file_mode = 'XML' then
    outp( ' <' || fields(i) || '>' || vals(i) || '</' || fields(i) || '>' );
    elsif file_type_def.file_mode = 'DELIM' then
    buf := buf || case when i > 1 then nvl(file_type_def.delimiter,',') else '' end ||
    nvl(file_type_def.quote,'') || vals(i) || nvl(file_type_def.quote,'');
    end if;
    end loop;
    if file_type_def.file_mode = 'DELIM' then
    outp(buf);
    end if;
    end if;
    end if;
    --savepoint USER_START;
    --saved_out_line := out_line;
    if r_tag != 'endofdata' then
    split_outp( file_type_def.record_start );
    end if;
    for i in 1..fields.count loop
    vals(i) := null;
    end loop;
    end if;
    prev_pidm := r_pidm;
    if field_idx.exists(r_tag) then
    if r_val is not null then
    vals(field_idx(r_tag)) := r_val;
    end if;
    end if;
    end loop;
    if pers_pidm is not null then
    split_outp( file_type_def.record_end );
    end if;
    split_outp( file_type_def.postamble );
    --for o in get_output loop
    -- dbms_output.put_line( o.text );
    --end loop;
    end if;
    return out_id;
    end generate_file;
    procedure gen_and_transmit( file_type in varchar2,
    username in varchar2, pass in varchar2,
    pidm_filter in varchar2 := null,
    include_placeholders in varchar2 := 'N',
    comments in varchar2 := null,
    url_base in varchar2 := null, url_path in varchar2 := null,
    ssl_wallet_path in varchar2 := null, ssl_wallet_pass in varchar2 := null ) is
    out_id integer;
    begin
    out_id := generate_file( file_type, pidm_filter => pidm_filter,
    include_placeholders => include_placeholders,
    comments => comments );
    --transmit_file( out_id, username, pass,
    transmit_file( out_id, 'mhaywood', '------------',
    url_base => url_base, url_path => url_path,
    ssl_wallet_path => ssl_wallet_path, ssl_wallet_pass => ssl_wallet_pass );
    end gen_and_transmit;
    begin
    UTL_HTTP.set_persistent_conn_support( true, 10 );
    end;
    create or replace
    PACKAGE BODY NCF_UTL IS
    PROCEDURE OUTPUT_TEXT( BUFFER IN VARCHAR2, LINE_LEN IN INTEGER := 80 ) IS
    TYPE CHRLIST IS TABLE OF VARCHAR2(1);
    cr VARCHAR2(1) := CHR(13);
         lf VARCHAR2(1) := CHR(10);
         tab VARCHAR2(1) := CHR(9);
         whitespace CHRLIST := CHRLIST( ' ', tab );
         pos INTEGER := 1;
         len INTEGER;
         brk INTEGER;
         curbrk INTEGER;
         lnbrk INTEGER;
         whtbrk INTEGER;
         i INTEGER;
    BEGIN
    len := LENGTH(BUFFER);
         WHILE pos <= len
         LOOP
         WHILE SUBSTR( BUFFER, pos, 1 ) IN ( cr, lf )
         LOOP
         pos := pos + 1;
         END LOOP;
         --find the next newline type char, or pos+250 if one isn't found
    lnbrk := LEAST( NVL(NULLIF(INSTR( BUFFER, cr, pos ),0),pos+250),
                             NVL(NULLIF(INSTR( BUFFER, lf, pos ),0),pos+250) );
         --after the loop:
         -- whtbrk should be the last whitespace char before LINE_LEN + pos
         -- (or == pos, if none found)
         -- curbrk should be == 0 or the 1st whitespace after LINE_LEN + pos
         whtbrk := pos;
         curbrk := pos;
         FOR i IN 1..whitespace.COUNT
         LOOP
         LOOP
         curbrk := INSTR( BUFFER, whitespace(i), whtbrk + 1 );
              IF curbrk <= pos + LINE_LEN AND curbrk >= whtbrk THEN
              whtbrk := curbrk;
              END IF;
              EXIT WHEN curbrk > pos + LINE_LEN OR curbrk = 0;
              END LOOP;
              EXIT WHEN curbrk > pos + LINE_LEN;
         END LOOP;
         brk := LEAST( len + 1, lnbrk, NVL( NULLIF( whtbrk, pos ), pos + 250 ),
                        NVL( NULLIF( curbrk, 0 ), pos + 250 ), pos + 250 );
         dbms_output.put_line( SUBSTR( BUFFER, pos, brk - pos ) );
         pos := brk;
         END LOOP;
    END;
    function split( str in varchar2, d0 in varchar2, d1 in varchar2 := null,
    d2 in varchar2 := null, d3 in varchar2 := null,
    d4 in varchar2 := null, d5 in varchar2 := null,
    d6 in varchar2 := null, d7 in varchar2 := null,
    d8 in varchar2 := null, d9 in varchar2 := null )
    return split_tbl pipelined is
    pos integer := 1;
    curidx integer;
    idx integer;
    delims split_tbl;
    idx_delim varchar2(32767);
    len integer := length( str );
    begin
    if d0 is null then
    --split on whitespace
    delims := split_tbl( ' ', chr(9), chr(10), chr(13) );
    else
    delims := split_tbl( d0, d1, d2, d3, d4, d5, d6, d7, d8, d9 );
    end if;
    loop
    idx := -1;
    --we want the lowest value of IDX > 0
    for i in 1..delims.count loop
    curidx := instr( str, delims(i), pos, 1 );
    if curidx > 0 then
    if idx = -1 or curidx < idx then
    idx := curidx;
    idx_delim := delims(i);
    end if;
    end if;
    end loop;
    if idx > 0 then
    --if splitting on whitespace, treat any amount of it as
    --a single delimiter
    if d0 is not null or idx > pos then
    pipe row( substr( str, pos, idx - pos ) );
    end if;
    else
    -- also make sure to ignore trailing whitespace
    if d0 is not null or pos <= len then
    pipe row( substr( str, pos ) );
    end if;
    exit;
    end if;
    pos := idx + length( idx_delim );
    end loop;
    return;
    end split;
    function split( str in varchar2, delimiters in split_tbl := null )
    return split_tbl pipelined is
    i integer;
    cursor spcur( d0 in varchar2 := null, d1 in varchar2 := null,
    d2 in varchar2 := null, d3 in varchar2 := null,
    d4 in varchar2 := null, d5 in varchar2 := null,
    d6 in varchar2 := null, d7 in varchar2 := null,
    d8 in varchar2 := null, d9 in varchar2 := null ) is
    select column_value v from table (split(str,d0,d1,d2,d3,d4,d5,d6,d7,d8,d9));
    function x( i in integer ) return varchar2 is
    begin
    if delimiters.exists(i) then
    return delimiters(i);
    else
    return null;
    end if;
    end;
    begin
    for r in spcur( x(0),x(1),x(2),x(3),x(4),x(5),x(6),x(7),x(8),x(9) ) loop
    pipe row( r.v );
    end loop;
    end split;
    END;
    Thanks

    Curious behavior, but irrelevant, IMHO, since one never switches back and forth between two releases in the real world. All that STARTUP UPGRADE does is that it enables upgrade scripts to be run - it does not prevent statements like CREATE USER from being executed since the upgrade may create new needed accounts
    http://docs.oracle.com/cd/E11882_01/server.112/e16604/ch_twelve045.htm#SQPUG128
    HTH
    Srini

  • How to adopt the index changes during upgrade

    Hi All,
    Please let me know how to adopt the standard index changes during upgrade. We are not able to change using SPDD.
    Regards
    Anil Kumar K

    Hi Micheal,
           I have posted one query can you answer for that also.
    Actually we are facing a problem  in activating the table COEP. It is saying duplicate field name exist for GEBER. But we have checked all the tables and structures.But no duplicates.
    Please help us..
    Regards
    Anil Kumar K

  • Archivelog mode during upgrade

    HI All,
    During upgrading the database from 10g to 11g... I will put my database from noarchivelog mode to archivelog mode to save the disk space.
    Does this task improves or speedsup the process of upgradation or any impact.
    Suggestions please.
    regards,
    TEJAS

    when DB is archivelog mode and you put DB in startup upgrade mode.. and execute catupgrd.sql script it starts updating dictionary objects and generates archivelogs ..
    Hence i put the DB in noarchive log mode..
    Any performance impact by putting the upgrade process in noarchivelog mode in terms of time duration the backup is performed.No performance impact, But recommended to put in Archivelog mode, even though if you have good backup.
    Also apart from this thread, have a question, The catupgrd.sql script execution duration depends upon the data disctionary contents right not on the DB size ? i am having this question. plz answer.The upgrade script creates and alters certain data dictionary tables. It also upgrades or installs the following database components in the new Oracle Database 11g Release 2 (11.2) database:
    Source:-
    Answer is here in detail.. http://docs.oracle.com/cd/E11882_01/server.112/e10819/upgrade.htm#sthref184

  • Error during upgradation

    Hi,
      During upgradation I got error at SACT_700 with following log
    Activate lock object "E_BAMMAPNG"
    4EEDI751 Base table "BAMMAPFCGR" of the lock object is not active
    4 EDI813 Lock object "E_BAMMAPNG" not activated
    4 EDI806XActivate lock object "E_BAMUIBT"
    4EEDI751 Base table "BAMUIB" of the lock object is not active
    4EEDI751 Base table "BAMUIBT" of the lock object is not active
    4 EDI813 Lock object "E_BAMUIBT" not activated
       When I looked at the base table it is already active.
    And also with the following error log
    Activate search help "Z_KSTAR"
    4EEDH163 Search help parameter "SPRAS" does not occur in the selection method
    4WEDH187 Default value "'EN'" does not suit the type of search help parameter "SPRAS"
    4 EDH109 Search help "Z_KSTAR" was not activated
    4 EDH102XActivate search help "Z_SKA1"
    4EEDH163 Search help parameter "SPRAS" does not occur in the selection method
    4WEDH187 Default value "'EN'" does not suit the type of search help parameter "SPRAS"
    4 EDH109 Search help "Z_SKA1" was not activated
    but when I checked the search help it is consistent  & I want default value EN to be specified.
      EDT402 Append structure "ZAKKAPARA_TA" appended to table "KKAPARA_TA"
    4WEDT191 Enhancement category "3" possible, but include or subty. not yet classified
    3EEDT963 Field "MAKTX" in table "KKAPARA_TA" is specified twice. Please check
    3 EDT015 Dependent table "KKAPARA_TA" was not activated
        but the field MAKTX is specified once in KKAPARA_TA but from appended structure  ZAKKAPARA_TA.
    Regards,
    Karthik.k

    Hi Karthik,
    Kindly check the versions of  E_BAMMAPNG and E_BAMUIBT. if you find Modif and Active version, then lock objects are not in active mode.
    open the object in edit mode, check the object and activate it.
    in search help Z_KSTAR and Z_SKA1, check if SPRAS is used as IMPORTING and/or EXPORTING parameter. remove the check of IMPORTING or EXPORTING parameter, save and activate the search help
    MAKTX is available in KKAPARA_TA, hence remove it from ZAKKAPARA_TA. save and activate the structure and table.
    after each modification, adjust and activate the database.
    regards
    Rahul.

  • Urgent: Unicode conversion - table splitting

    Hi all,
    I am having a problem when trying to perform the export step of the unicode conversion on ERP2005, MSSQL server 2005. Due to previously very long runtime, I am trying to use the table splitting option. I have performed the "Table Splitting Preparation" step with what seems to be a success.
    The problem is when I run the actual database instance export.
    First of all:
    Which filepath should I provide SAPinst when it is asking for "Table input file"?
    Second:
    How can I actually determine the Package Unload Order? I tried selecting this option, but I was not given the opportunity to change this order in the subsequent screen. (The F1 help in SAPinst sais that I would be able to...)
    Have anyone of you experience with this?
    Best Regards,
    Thomas

    Hi Thomas,
    As I can imagine from the date of your last posting, you probably have your answers. But just so other people, who are searching on this topic, can find the answer here, I will fill in the blanks.
    First, when using table splitting, you must export using the migration monitor (MIGMON). That means, when you run SAPINST, on the ABAP System -> Database Export screen, select the Export Method: Export using Migration Monitor.
    At this point, as you said, you have already run the Table Splitting option from SAPINST, so the WHR files are in the export DATA directory. This whr.txt file is also there.
    The file input screen you are mentioning does not appear when you select the Export method mentioned above. But you can create such a file for MIGMON. You have to create this on your own and can only do it after the SAPINST has split the STR files.
    When finished with SAPINST, create the .txt file (I usually call it table_order.txt) and add the names of the STR files which you want to make sure are exported first. After MIGMON completes that list, all other STR files are exported in alphabetical order.
    Create your table_order.txt file inserting the filenames of the packages that were created, without the STR/WHR extensions (using some large tables which STR splitter broke out into their own STR files during my last export). You have to look in DB02 and sort descending based on used space to determine the order in which these tables should be exported and listed in the .txt file:
    <SPLIT Table>-1
    <SPLIT Table>-2
    <SPLIT Table>-n
    BSIS
    RFBLG
    CE1VVOC
    CE1WWOC
    DBTABLOG
    CE3VVOC
    COEP
    ARFCSDATA
    GLPCA
    KONV
    SWW_CONT
    SOC3
    CDCLS
    CE3WWOC
    BSE_CLR
    STXL
    EDIDS
    COSP
    BSAD
    EDI40
    ACCTIT
    BSIM
    VBFS
    BSAS
    ACCTCR
    CDHDR
    CE4VVOC_ACCT
    SGOHIST
    MSEG
    In the export_monitor_cmd.properties file, you specify this file name for the 'orderBy=' parameter. Create and keep the table_order.txt file in the same directory as the export_monitor.sh/.bat files.
    For importing, again you will need to use the Migration Monitor. SAPINST will automatically stop and prompt you to start the import using MIGMON. Here, you can specify the same table_order.txt file. You might want to amend it to control when the rest of the packages are imported, if you found one or more tables holding up the completion of the export.
    I hope this helps someone.
    Best Regards,
    Warren Chirhart

  • ORA-1653 during upgrade db

    Hello!
    I see in alert.log during upgrade 10.2.0.5 -11.2.0.3
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEMShould I restart catupgrd.sql ?
    Thanks and regards,
    Pavel

    Pavel wrote:
    Hello!
    I see in alert.log during upgrade 10.2.0.5 -11.2.0.3
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEM
    ORA-1653: unable to extend table SYS.SETTINGS$ by 8 in tablespace SYSTEMShould I restart catupgrd.sql ?
    Thanks and regards,
    Pavelyou should make SYSTEM tablespace larger

Maybe you are looking for