Inconsistency during Unicode conversion

There are some tables inconsistent durning Unicode conversion, is it possible to delete them?

Hi, there are two ways to solve the problem, one is delete this table, one is make the table conssitent. These tables could be deleted if it is useless on the system.

Similar Messages

  • Error during unicode conversion

    Dear all,
    we are facing one issue during unicode conversion. the code is as follows:
    FIELD-SYMBOLS : <n> TYPE X,
                                 <c> TYPE C.
    DATA : a TYPE i.
    DATA : x TYPE x.
    MOVE <n> TO a.
    a = a + 22.
    MOVE a TO x.
    ASSIGN x TO <c> CASTING TYPE c
    and we are getting the following error:
    ERROR : the length of "X" in bytes must be a multiple of the size of a Unicode character (regardless of the size of the unicode character
    please guide on how this can be resolved in ECC.
    Regards,
    N.Jain

    Hi,
    here is some code to do it:
    FIELD-SYMBOLS : <n> TYPE x,
                    <c> TYPE c.
    DATA : a TYPE i.
    DATA : x TYPE x.
    MOVE <n> TO a.
    a = a + 22.
    MOVE a TO x.
    DATA:
      buffer TYPE xstring,
      conv TYPE REF TO cl_abap_conv_in_ce,
      x_char.
    buffer = x.
    conv = cl_abap_conv_in_ce=>create(
           encoding = 'NON-UNICODE'
           input = buffer
    CALL METHOD conv->read(
      EXPORTING
        n    = 1
      IMPORTING
        data = x_char
    ASSIGN x_char TO <c>.
    Regards
    Walter Habich
    Edited by: Walter Habich on Jun 10, 2008 4:09 PM

  • Nametab issue during unicode conversion

    There are some tables inconsistent durning Unicode conversion, is it possible to delete them?
    Moderator message: not directly related to ABAP development, please have a look in the Netweaver forums, alway search for information before asking.
    Edited by: Thomas Zloch on Dec 23, 2010 1:32 PM

    Hi,
    Reposting the question again, as earlier one was incomplete.
    We are doing Unicode Conversion activity in one of our system. After running through UCCHECK transaction, we got around 80 programs in which most of them were resolved manually. When we are trying to edit one object, the error"Namespace /OSJ8C/' does not exists is displayd. Any help to resolve this issue will be highly appreciated .

  • Error in Import during unicode conversion

    Hello,
    I am doing Combined upgrade and unicode conversion to ECC6 SR3.
    Upgrade is completed sucessfully. During import i get an error stating :
    bSl Trace: Error 959 in exec_immediate() from oci_execute_stmt(), orpc=0
    DbSl Trace: ORA-959 occurred when executing SQL stmt (parse error offset=92)
    (DB) ERROR: DDL statement failed
    (CREATE UNIQUE INDEX "REPOLOAD~0" ON "REPOLOAD" ( "PROGNAME", "R3STATE", "MACH" ) TABLESPACE PSAPSID620 STORAGE (INITIAL 209
    7152 NEXT 0000010240K MINEXTENTS 0000000001 MAXEXTENTS 2147483645 PCTINCREASE 0 ) NOLOGGING COMPUTE STATISTICS )
    DbSlExecute: rc = 99
      (SQL error 959)
      error message returned by DbSl:
    ORA-00959: tablespace 'PSAPSID620' does not exist
    This tablespace doesnt exist as it was upgraded to PSAPSID700.
    I checked DBLORA.TPL and it has two references of 620 :
    index storage parameters
    loc: APPL0 PSAP@SCHEMAID@                 0000
         APPL1 PSAP@SCHEMAID@                   0000
         APPL2 PSAP@SCHEMAID@                  0000
         CLUST PSAP@SCHEMAID@                  0000
         DDIM  PSAPDIMI                                      0000
         DFACT PSAPFACTI                                0000
         DODS  PSAPODSI                                   0000
         POOL  PSAP@SCHEMAID@                  0000
         SDIC  PSAP@SCHEMAID@                     0  
         SDOCU PSAP@SCHEMAID@                  0000
         *SLDEF PSAP@SCHEMAID@620              0000*
         *SLEXC PSAP@SCHEMAID@620            0000
         SLOAD PSAP@SCHEMAID@                  0000
         SPROT PSAP@SCHEMAID@                   0000
         SSDEF PSAP@SCHEMAID@700              0000
         SSEXC PSAP@SCHEMAID@700              0000
    Thanks,
    FBK

    Hi Volker,
    Thank you very much. Stupid error :-).
    I also did step 2 with fixr3owns yesterday with SAPINST and then it worked. I got to step 3 "iSeries Installation Step", but got an error in SAPINST when "Creating infrastructure for SAP system BND". I logged off SAPINST, patched CRTR3SYS and started SAPINST again and then got the error in step 2 wirh fixr3owns.
    I´m now past step 2 and in step 3 "iSeries Installation Step". And I get the same error as I got yesterday...
    From the job log:
    TMKSVR request processor: request was successful                      
    Processing request from client 10.96.220.5:4350.                      
    EXECUTE CRTR3SYS SID(BND) GLOBALHOST('SAP002') FORCEPROF(*NO)     
      ADBHOST('SAP002') 1 N                                               
    Keyword GLOBALHOST not valid for this command.                        
    Keyword FORCEPROF not valid for this command.                         
    Keyword ADBHOST not valid for this command.                           
    Error found on CRTR3SYS command.                                      
    Errors occurred in command.                                           
    error: call_AS400_command: "CPF0006" caught after "CRTR3SYS SID(BND)  
      GLOBALHOST('SAP002') FORCEPROF(*NO) ADBHOST('SAP002')"              
    I can run "CRTR3SYS SID(BND) GLOBALHOST('SAP002') FORCEPROF(*NO)" manually.
    Do you have any ideas?
    Best regards
    Henrik Hviid

  • Performance during Unicode Conversion

    Hello,
    currently we are planning to do a Unicode Conversion on a Linux/MaxDB System.
    The DB has a size of approx. 350 GB. The system runs on Fujitsu-Siemens FlexFrame hardware.
    One problem is that during a test conversion the export of the DB took more than 24 hours.
    This seems too much regarding the DB size.
    We know that filesystem and RAM and CPUs influence the export duration.
    What possibilities do we have to tune the export with "software" tools?
    - table splitting
    For example: Our largest table has ~26GB datasize.
    How do we split this table?
    In 10 parts or in 20 or in 200?
    If you could please post your experiences.
    Regards, Michael

    I don´t wanna discourage you but I´m not sure if splitting that single table will help alot.
    The problem with this combination is the fact, that the MaxDB database reads data block-by-block, means, exeucting single I/Os for each block to read. Given that fact and the fact, that each I/O will be ran over an NFS (with the known latency) it´s unlikely that you will get much more speed out of this.
    The problem is NOT the speed between the NAS and the database server, if you will look on the network traffic you will see, that there´s plenty left. The problem is the latency on the client side (the database server) when you put your files on NFS.
    I can export a 250 GB database in under 3 hours (admittedly - without UC conversion) using a direct fibrechannel connection (so non-NFS) so the database itself is certainly not the problem, it´s the combination of the usage of NFS (no matter if v2 or v3) and the way the MaxDB database reads data.
    I´ve been involved in a likewise project and with a NetApp specialist and some in-depth tuning we were able to gain 15 - 20 % speed but not more.
    Markus

  • SAPInst Export Error during Unicode Conversion

    Hello,
    I started SAPInst, to export the source instance (for unicode conversion).
    During the export preparation phase, I get the following error (in SAPInst)
    ERROR
    CJS-00030: Assertion failed: in component: NW_getDBInfo step: fillContextat this point the dbtype has to be != ind
    ERROR
    MUT-03025 Caught ESAPinstException in Modulecall:
    ESAPinstException: error text undefined
    ERROR
    FCO-00011 The step fillContext with step key
    |NW_Prepare_Export| ind| ind | ind | ind |0 |0 |NW_GetSidromProfiles |ind |ind |ind |ind |2 |0 |NW-getDBInfo |ind |ind |ind |4 |0 | fillContext was executed with status ERROR
    Please help me how to rectify this error.
    Thanks,
    Haleem

    Hello,
    I recently experienced this problem, and it seems that when executing this step the <sid>adm account needs to be defined to use C shell.  If its already set up to use C shell, check the <sid>adm logon scripts to make sure all the environment variables are being set correctly.
    Regards,
    Brad

  • Error during Unicode Conversion at DDNTF table

    Hello
    We are doing a Unicode Conversion for our CRT systems. This system now
    is a SAP CRM 7.0 SR1 ABAP.
    We have performed preparatory steps (SPUMG, checks, updating R3load,
    R3ldctl, R3szchk, dboralib, ...),
    but when we start Import process we get an error in SAPSDIC package.
    The error looks like:
    *************************+
    (RTF) ########## WARNING ###########
    Without ORDER BY PRIMARY KEY the exported data may be unusable
    for some databases
    (EXP) INFO: table DDNTF will be exported with sorting
    (RSCP) ERROR: 'DDNTF' in UMGCCTL: Guess has illegal value.
    (RSCP) WARN: DDNTF: Missing in UMGSTAT,
    ERROR in nametab conversion
    conversion of nametab table 'DDNTF ':
    rscpMCStartTab: rc = 128
    (CNV) ERROR: data conversion failed. rc = 2
    (DB) INFO: disconnected from DB
    /usr/sap/CRT/SYS/exe/run/R3load: job finished with 1 error(s)
    /usr/sap/CRT/SYS/exe/run/R3load: END OF LOG: 20091229193651
    Any idea??
    Thanks in advance
    Regards

    Yes I executed this Report using Menu Tool Bar.
    The first execution gave me some errors regarding some tables existing at database level but  not in DDIC. We ignored this errrors because those tables are not important.
    But the last try, we decided solve this inconsistencies deleteting those tables at database level, repeting report RADUCNT (in this case no errors appeared) and repeting the export with the same error..
    Now, we have seen in Tabscript STATUS of SPUMG there are some errors regarding DDNTF inconsistency found at table control. Reason 4.  But this reason means that faillback codepage missing at table control for this table. But as far I know, I can't fill failbak field for this kind of table...
    So know I'm reseting SPUMG, and reestarting again..
    But I don't understand what is the problem
    In a few minutes I will you inform about results
    Meanwhile is anybody have any suggestion or idea... please let me know
    Thanks!

  • Error during Unicode Conversion - Import

    Hi,
    I am doing a Unicode conversion, done with the system export, while doing the
    import during the Import ABAP phase system is throwing the error with R3load.
    R3load -testconnect fails with following error.
    emalonsbx002:wuuadm 18% R3load -testconnect
    sapparam: sapargv( argc, argv) has not been called.
    sapparam(1c): No Profile used.
    sapparam: SAPSYSTEMNAME neither in Profile nor in Commandline
    R3load: START OF LOG: 20071113095228
    R3load: sccsid @(#) $Id: //bas/700_REL/src/R3ld/R3load/R3ldmain.c#14 $ SAP
    R3load: version R7.00/V1.4 [UNICODE]
    Compiled Oct 20 2007 06:20:06
    R3load -testconnect
    DbSl Trace: OCI-call 'OCISessionBegin' failed with rc=1033
    DbSl Trace: CONNECT failed with sql error '1033'
    DbSl Trace: OCI-call 'OCISessionBegin' failed with rc=1033
    DbSl Trace: CONNECT failed with sql error '1033'
    (DB) ERROR: db_connect rc = 256
    DbSl Trace: OCI-call 'OCISessionBegin' failed with rc=1033
    DbSl Trace: CONNECT failed with sql error '1033'
    DbSl Trace: OCI-call 'OCISessionBegin' failed with rc=1033
    DbSl Trace: CONNECT failed with sql error '1033'
    (DB) ERROR: DbSlErrorMsg rc = 99
    R3load: job finished with 1 error(s)
    R3load: END OF LOG: 20071113095230
    Please let me know if anybody have any idea on this.
    Thanks in advance.

    Your database isn´t fully started yet (in shutdown or startup) - check your alert.log
    Also see
    Note 655240 - Composite SAP note: ORA-01033
    Markus

  • Usage of Conversion Classes During Unicode Conversion

    Hi,
    I have code
    "TRANSLATE  BBKF-SGTXT FROM CODE PAGE '0100' TO '0120'"  Now During Unicode i need to Use the Conversion classes
    Can any one help to proceede further on this
    Actually I am using  CL_ABAP_CONV_OUT_CE and I am getting the Data in Binary Format But I want that to be in Character format can any one suggest me
    Thanks & Regards,
    Prasad

    Do you have an EBCDIC system with code page 0100? (check it using SNLS transaction)
    In that case, CL_ABAP_CONV_OUT_CE is okay.
    After having converted to code page 0120 (into xstring), here is how to cast xstring variable to character.
    DATA x TYPE x LENGTH 800. " so that to compile, x length
                              " must be a multiple of 4 (don't ask why)
    FIELD-SYMBOLS <c> TYPE c.
    x = xstring. "make sure x is long enough
    ASSIGN x TO <c> CASTING.
    BBKF-SGTXT = <c>.

  • LRAW data Conversion during Unicode Conversion

    Hi,
    We have upgraded our 46C MDMP system to Unicode system.
    After conversion we found that some of the tables with field of LRAW type are having conversion issue.
    In 46C system, Russian data which enetered in the MS Cyrillic (Code page 1504), is converted to code page 1500 after unicode conversion.
    We have not found any entries for this table in vocabulary, reprocess log or in SUMG.
    For example table STXL. This table contains Purchase order text for the material.
    Report HRUU_CONVERT_IN_UNICODE does not shows LRAW data field for conversion.
    Can you suggest how LRAW data can be converted from 1500 to 1504 code page.
    Thanks and Regards
    Rahul

    Hi Rahul,
    STXL is a language dependent INDX type table. Hence SPUMG will not scan this one with the default settings.
    Please have a look at the description of INDX type table handling in the Unicode conversion Guide.
    In the Unicode system, you can use UMG_ADJUST_STXL for the conversion.
    However please be aware of the restrictions:
    Once you have decided about the code page and converted the object, this cannot be reverted !
    Therefore it is highly recommended to check the display mode first and in addition test it extensively on a sandbox system !
    Best regards,
    Nils Buerckel
    SAP AG

  • Skipped tables during UNICODE conversion

    Hello everybody.
    We are making a unicode conversion in a 701 Ehp4 over iSeries.
    The export has finished succesfully but we have detected that some tables have been skipped. In the Export_monitor.log there are some comments like this one:
    INFO: 2010-04-15 01:30:56 com.sap.inst.migmon.LoadTask processPackage
    Unloading of 'AGKO' export package from database will be skipped.
    Task file '/usr/sap/SAPinst/NW04S/EXP/AGKO.TSK' is empty and contains no tasks.
    Does anyone know why have these tables been skipped?
    Do we have to repeat all the process?
    Thank you in advanced.
    Best regards.
    Rubén Garcí

    Hello
    The problem was related with Inplace method
    Regards
    Ruben

  • R/3 Export & Import during Unicod Conversion via SOCKET Method

    Hi
    We are in the process of upgrading our R/3 Enterprise to ECC 6.0. The size of the database is around 4 TeraBytes.
    Can somebody help us with necessary documentation/ othe rinputs to handle R/3 Export & Import using SOCKET method?  FYI we referred the SAP given doucments on the same. Any particular DOs & DO NOTs would be highly appreciated.
    Thank You
    Sai

    Hi Rahul
    regarding your 'Unicode doubt"' some ideas:
    1) The Upgrade Master Guide SAP ERP 6.0 and the Master Guide SAP ERP 6.0 include introductory information. Among other, these guides reference the SAP Service Marketplace-location http://service.sap.com/unicode@sap.
    2) In Unicode@SAP can you find several (content-mighty) FAQs
    Conclusion from the FAQ: First of all your strategy needs to follow your busienss model (which we can not see from here):
    Example. The "Upgrade to mySAP ERP 2005"-FAQ includes interesting remarks in section "DO CUSTOMERS NEED TO CONVERT TO A UNICODE-COMPLIANT ENVIRONMENT?"
    "...The Unicode conversion depends on the customer situation....
    ... - If your organization runs a single code page system prior to the upgrade to mySAP ERP 2005, then the use of Unicode is not mandatory. ..... However, using Unicode is recommended if the system is deployed globally to facilitate interfaces and connections.
    - If your organization uses Multiple Display Multiple Processing (MDMP) .... the use of Unicode is mandatory for the mySAP ERP 2005 upgrade....."
    In the Technical Unicode FAQ you read under "What are the advantages of Unicode ...", that "Proper usage of JAVA is only possible with Unicode systems (for example, ESS/MSS or interfaces to Enterprise Portal). ....
    => Depending on the fact if your systems support global processes, or depending on your use of Java Applications, your strategy might need to look different
    3) In particular in view of your 3rd option, I recommend you to take a look into these FAQs, if not already done.
    Remark: mySAP ERP 2005 is the former name of the application, which is named SAP ERP 6.0, now
    regards, and HTH, Andreas R

  • Effort estimate (ABAP) during Unicode Conversion project

    Hi ,
    I am working on a CU&UC (Combined Upgrade and Unicode Conversion) project from 46C to ECC6.
    We are having 8 different codepages (MDMP SYSTEM) and around 10000 custom objects.
    Does anyone have any experince in doing such kind of project ? I am looking at the effort estimate on technical side (ABAP) and functional side.
    Any inputs of Technical effort in such a project with related number of codepages and custom objects will be useful (in man days & project duration).
    Thanks
    Dilip

    Hi Dilip,
    the main effort from ABAP side seems to be testing of the new system and creating the Unicode vocabulary. You need test teams and test plans for all locations and for the vocabulary you need people for all your languages. ABAP coding itself is probably not your biggest problem. You have to switch your system to unicode and then check all your sources. This could be done by a report or transaction (check the documentation regarding the Migration). From my point of view this is the main effort you'll have, testing and building the vocabulary. How big this effort is depends on your system. I think six month are a good time frame for the whole project (I assume you migrate a landscape of 3 systems). Because I'm working in the Basis administration I could not tell you the exact man days, but I would estimate up to two weeks for the vocabulary per language. You should have the numbers for the system tests from previous upgrades.
    The technical migration effort depends on your hardware and time requirements. If your system is big (TBytes) and your downtime is very limited it could be necessary to tune the migration by testing it several times. If your requirements are not that high, then a single test migration could be sufficient. I've done two bigger migrations with 2.5 and 3.5 TB in less than 24 hours but therefore you have to have fast hardware and you have to tune your process well. If you have a single huge table this could destroy all efforts reaching your time target, if for example the index build could not be done in less then your time target. If you have enough hardware resources and time to tune your migration you could reach almost any time (which makes sense - some hours).
    Best regards
    Ralph Ganszky

  • Minimize downtime during UNICODE conversion

    Hello forum !
    I am looking at the UNICODE conversion of a large (?) SAP Oracle database for ECC. The Database it self is at about 10 TB. UNICODE converting this in one single run will take too long to fit inside the company defined tolerable downtime window.
    When looking at the data I find that the largest 50 tables makes up about 75% of the total database size. Most of the 50 largest tables have data for several years back, and I am able to partition the tables with one partition pr year.
    The good thing about this is that I am able to make the partitions with old data (more than one year old data) READ-ONLY.
    My assumption is that I will be able to UNICODE convert 60-70% and create the indexes for these partitions without taking down the database to be converted.
    Now,- doing the downtime-phase I will be able to exclude the old partitions of the largest tables in the export since I have already created a new UNICODE database with these objects and thereby drastically reducing the required downtime.
    Is there anyone who has any experience with an approach like the one I have schetched out here ?
    Kjell Erik Furnes

    Can you please be a bit more specific on your installation? How long is your possible downtime (export time should be about 2/3 of that)? How long is your actual export estimation? Do you have suitable hardware (CPU/disk)?
    What you absolutely need to have:
    [1043380 - Efficient Table Splitting for Oracle Databases|https://service.sap.com/sap/support/notes/1043380] -> export tables with ROWID splitting
    [855772 - Distribution Monitor|https://service.sap.com/sap/support/notes/855772] -> enables distributed and parallel export/import
    So on the hardware side you will need, a storage subsystem as fast as possible with two copies of the disks (one for exporting and one for importing). A bunch of application servers with fast CPUs helps a lot, everything connected with at least gigabit ethernet.
    If you knew all this already, sorry for the spam. But i achived throughputs at about 400gb/hour on a medium sized hardware, i think throughputs of 1tb/hour should be possible (not talking about cluster tables
    Regards, Michael

  • Unicode Conversion approach

    HI All
    During Unicode approach isit possible to do the conversion of custom and standard program to unicode and make the system live and later on do the data conversion with business downtime.
    The reason i am asking so as to reduce the business down time. We are going for CU&UC and donot want a long business downtime.
    Let me know if there are any approach where we can do the data conversion in the later phase.
    Thanks
    Rajat Sarkar

    Hi,
    please have a look at:
    http://service.sap.com/~form/sapnet?_SHORTKEY=01100035870000380759&_OBJECT=011000358700000620342009E
    > Page 21
    Excerpt begin:
    1) Start release >= SAP NetWeaver AS 6.20
    > UCCHECK can be run before or during Unicode conversion
    2) Start release <= 4.6C (CU&UC or TU&UC*)
    > Execute the whole UCCHECK procedure in the SBX Unicode system
    > Upgrade the DEV system to SAP ERP 6.0 (Non-Unicode or Unicode) and execute UCCHECK (Additional Maintenance System necessary)
    Excerpt End.
    Therefore if you have a 4.6c system (or lower) as source system, then the Unicode enabling procedure (UCCHECK) cannot be separated from the actual Unicode conversion (Export / Import). In case of NW 6.20 or higher, you can do most of the enabling before the conversion (you will still need a rework after the conversion).
    Best regards,
    Nils Buerckel

Maybe you are looking for

  • Clamshell mode issue when waking from sleep

    Just got an rMBP and hooked it up to a ViewSonic monitor via DVI-thunderbolt to use in clamshell/desktop mode in my home office. All good – except that whenever I wake the MBP from sleep, the ViewSonic just shows snow for about 30 secs to 1 min befor

  • REX Add Organisation errors.

    I'm adding trying to "Create Cloud Administration Organization", and get the following error message in the Create Organisation Unit task. Has anyone seen this before? com.newscale.is.core.RoutingException: Routing exception found: Reference Field no

  • Audio of .mp4 movie files getting distorted by 1 hour mark.

    I converted some of my fave movies from DVD to .mp4 file format that my iPod supports. After I did so, I copied several flicks in .mp4 format to my iPod. Each of those movies is at least 2 hours in length. But the problem with iPods playing .mp4 vids

  • Configuration of VPN in WRVS4400N?

    Hi! I'm quite new to this remote access and VPN overall, so please give me easy enough instructions...;^) I'm just bought the Cisco (Linksys) WRVS4400N Wireless-N Gigabit Security Router - VPN v2.0. I'm trying to configure it to provide a remote acce

  • Unrecgnized ActionEvent

    I have been programming in ANSI/ISO C++ for a while, and just started to learn Java so that I could write a GUI to a set of tools that currently only has a command line interface. I got the GUI designed OK, and am now adding ActionListeners and such