INI: XOQ-01600: OLAP DML Error "ORA-01438: value larger than specified prec

I have created a Time dimension in AWM 11.1.0.7.0B.
I have added two Hierarchies to it. One Hierarchy has All Years, Year, Week, day. The second hierarchy has All Years, Year, Quarter.
When I maintain the dimension, the following error occurs:
An error has occurred on the server
Error class: Express Failure
Server error descriptions:
INI: Error creating a definition manager, Generic at TxsOqConnection::generic<BuildProcess>
INI: XOQ-01600: OLAP DML Error "ORA-01438: value larger than specified precision allowed for this column
" while executing DML "SYS.AWXML!R11_LOAD_DIM('TIME.ALL_GREGORIAN_YEARS.LEVEL' SYS.AWXML!___R11_LONG_ARG_VALUE(SYS.AWXML!___R11_LONG_ARG_DIM 1) 'TIME.END_DATE.ATTRIBUTE' 'TIME.TIME_SPAN.ATTRIBUTE' 'TIME.LONG_DESCRIPTION.ATTRIBUTE' 'TIME.SHORT_DESCRIPTION.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_END_DATE.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_TIME_SPA.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_LONG_DES.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_SHORT_DE.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_TIME_SPA1.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_TIME_SPA2.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_TIME_SPA3.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_LONG_DES1.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_SHORT_DE1.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_TIME_SPA4.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_LONG_DES2.ATTRIBUTE' 'TIME.ALL_GREGORIAN_YEARS_SHORT_DE2.ATTRIBUTE' 'TIME.GREGORIAN_QUARTER_END_DATE.ATTRIBUTE' 'TIME.GREGORIAN_QUARTER_TIME_SPAN.ATTRIBUTE' 'TIME.GREGORIAN_QUAOLAP DML Error "%(1)s" while executing DML "%(2)s", Generic at TxsOqStdFormCommand::execute
at oracle.olapi.data.source.DataProvider.callGeneric(Unknown Source)
at oracle.olapi.data.source.DataProvider.callGeneric(Unknown Source)
at oracle.olapi.data.source.DataProvider.executeBuild(Unknown Source)
at oracle.olap.awm.wizard.awbuild.UBuildWizardHelper$1.construct(Unknown Source)
at oracle.olap.awm.ui.SwingWorker$2.run(Unknown Source)
at java.lang.Thread.run(Thread.java:595)

The most likely explanation is that you have a numeric attribute whose data type precision is less than that of the column it is mapped to. For example if you had an attribute with data type NUMBER(2) and you mapped it to a column with data type NUMBER(5), then you could get this error on load. Note that some of the attributes may not be visible in AWM. The offending SQL statement should be in the OUTPUT column of the CUBE_BUILD_LOG, but you may be able to fix this by just describing the view associated with the dimension. For example, here is the description of a TIME view created in 11.2..
  Name                            Null?    Type
DIM_KEY                             VARCHAR2(60)
LEVEL_NAME                             VARCHAR2(30)
MEMBER_TYPE                             VARCHAR2(1)
DIM_ORDER                             NUMBER
END_DATE                             DATE
TIME_SPAN                             NUMBER
LONG_DESCRIPTION                        VARCHAR2(60 CHAR)
SHORT_DESCRIPTION                        VARCHAR2(60 CHAR)
MONTH_END_DATE                         DATE
MONTH_TIME_SPAN                        NUMBER(5)
MONTH_LONG_DESCRIPTION                    VARCHAR2(60 CHAR)
MONTH_SHORT_DESCRIPTION                   VARCHAR2(60 CHAR)
FISCAL_QUARTER_END_DATE                   DATE
FISCAL_QUARTER_TIME_SPAN                   NUMBER(5)
FISCAL_QUARTER_LONG_DESC                   VARCHAR2(60 CHAR)
FISCAL_QUARTER_SHORT_DES                   VARCHAR2(60 CHAR)
FISCAL_YEAR_END_DATE                        DATE
FISCAL_YEAR_TIME_SPAN                        NUMBER(5)
FISCAL_YEAR_LONG_DESCRIP                   VARCHAR2(60 CHAR)
FISCAL_YEAR_SHORT_DESCRI                   VARCHAR2(60 CHAR)
CALENDAR_QUARTER_END_DAT                   DATE
CALENDAR_QUARTER_TIME_SP                   NUMBER(5)
CALENDAR_QUARTER_LONG_DE                   VARCHAR2(60 CHAR)
CALENDAR_QUARTER_SHORT_D                   VARCHAR2(60 CHAR)
CALENDAR_YEAR_END_DATE                    DATE
CALENDAR_YEAR_TIME_SPAN                   NUMBER(5)
CALENDAR_YEAR_LONG_DESCR                   VARCHAR2(60 CHAR)
CALENDAR_YEAR_SHORT_DESC                   VARCHAR2(60 CHAR)You can get the same information from user_cube_attributes
SQL> select attribute_name, data_precision from user_cube_attributes where dimension_name = 'TIME' and data_type = 'NUMBER';
ATTRIBUTE_NAME                 DATA_PRECISION
TIME_SPAN
MONTH_TIME_SPAN                    5
FISCAL_QUARTER_TIME_SPAN              5
FISCAL_YEAR_TIME_SPAN                   5
CALENDAR_QUARTER_TIME_SPAN              5
CALENDAR_YEAR_TIME_SPAN               5

Similar Messages

  • INI: XOQ-01600: OLAP DML error "ORA-4030: out of process memory" OLAP PGA S

    Hi All ,
    While executing the cube generation I am getting an error. Anybody knows the reason ? I have amended the olap_page_pool_size to 200MB and it doesn't help at all .
    INI: error creating a definition manager, Generic at TxsOqConnection::generic<BuildProcess>INI: XOQ-01600: OLAP DML error "ORA-4030: out of process memory when trying to allocate 82860 bytes (OLAP PGA Stack,xsVPBlankParm: PPARM)" while executing DML "SYS.AWXML!R11_COMPILE_PARTITIONS('TIME.DIMENSION')", Generic at TxsOqStdFormCommand::execute
    Thanks in advance,
    Debashis

    HI David ,
    Thanks for the reply.
    My Time Dimension having 10 years of data in day level granularity and Fact table is not partitioned and having only one month of data as 299 records .
    Just to let you know that we define two hierarchy level under TIMES one is "ALL levels" and another is Detail where END_DATE has been defaulted with some value and TIME_SPAN is set mapped to the Times table column having distinct value 1 for each records .Also the Member specified as ROW_WID of the Time table.
    Just to let you know we have ran(Maintain from Dimension hierarchy) 'Product' and 'Position' dimension individually and it works fine i.e Load ,Compile and Sync process works fine but while run Times it is throwing issue :
    ORA-4030: out of process memory when trying to allocate 59340 bytes (OLAP PGA Stack,xsVPBlankParm: PPARM)&quot; while executing DML
    we run the Times hierarchy from OLAPTRAIN and it was perfectly fine . Not sure with our time Dim definition .
    Any clue ?
    Many Thanks,
    Debashis

  • CDC Failed with Error ORA-01438: value larger than specified precision

    Hi,
    I have created Asynchronous Distributed Change Data Capture Set-up as per the Guildines from Oracle.
    Now My Change Set has become Invalid with error: ORA-01438: value larger than specified precision allowed for this column.
    I Compared Change Table Structures with Source Table on Source Database and the structure is matching.
    Now I am trying to Run below script:
    Begin
    dbms_cdc_publish.alter_change_set (Change_Set_name =><<Change Set Name>>,
    recover_after_error => 'Y'
    End;
    I am getting the below error:
    ORA-01438: value larger than specified precision allowed for this column
    ORA-06512: at "SYS.DBMS_APPLY_ERROR", line 147
    ORA-06512: at "SYS.DBMS_APPLY_ERROR", line 301
    ORA-06512: at "SYS.DBMS_APPLY_ADM", line 490
    ORA-06512: at line 1
    ORA-06512: at "SYS.DBMS_CDC_PUBLISH", line 580
    ORA-06512: at line 2
    Still the error is existing, I couldn't find any issues with Change Table structure as it is in Line with Source Table Structure.
    Can anyone please help me on this.
    Staging Database details:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    TNS for Linux: Version 10.2.0.4.0 - Production
    Source Database details:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    TNS for HPUX: Version 10.2.0.4.0 - Production
    Also Can anyone let me know if it is possible to run CDC Set-up without DBA Privileges. i,e once the CDC Set-up is done, if we revoke the DBA Privilege will the Capture work properly. I know that for creating Change Table, Set and Source we require DBA Access. But once the set-up is done, do we really require DBA Privileges?
    Thanks,
    Shashi
    Edited by: user8254952 on 09-Mar-2010 05:57
    Edited by: user8254952 on 09-Mar-2010 06:00

    Hi Shashi,
    As CDC is built on top of Oracle streams framework, for better responses you can post this ques under {forum:id=70} category.
    "ORA-01438: value larger than specified precision allowed for this column" is usually encountered for NUMBER datatype columns. At what level did you check the datatypes? Please confirm if the precision(for NUMBER datatype) are also similar.
    Please attach output for the following queries:
    select error_message from dba_capture;
    select error_number, error_message from dba_apply_error;Please re-validate the data types and then run the same API with remove_ddl = 'Y' option, like this:
    exec DBMS_CDC_PUBLISH.ALTER_CHANGE_SET(change_set_name=>'CHANGE_SET_NAME' , recover_after_error=>'Y' , remove_ddl=>'Y');
    Hope it helps!
    Cheers,
    AA

  • Dump with ORA-01438: value larger than specified prec

    Hello All,
    I am getting a dump in production with the above said error and the ST22 gives me further analysis that
    "If the error occurred in Open SQL, there is probably an inconsistency
    between the NAMETAB and the ABAP/4 Dictionary.                      
    Compare the length specifications of the fields in the NAMETAB with 
    those in the ABAP/4 Dictionary and contact someone who is able to   
    perform a consistency check and eliminate the problem.              
    In most cases, this person will be your SAP consultant.             
    Please make a note of the actions and input which caused the error. 
    Please make a note of the actions and input which caused the error. 
    To resolve the problem, contact your                                
    SAP system administrator.                                                                               
    Choose "Print" for a hard coopy of the termination message. You can 
    display and adminster short dump messages using Transaction ST22.                                                                               
    Error analysis                                                                               
    The problem has arisen because, within the database interface,      
    one of the data buffers made available for the INSERT (UPDATE)      
    is longer than the maximum defined in the database.                 
    On the other hand, it may be that the length in the NAMETAB         
    does not match the maximum length defined in the database.          
    (In this case, the length in the NAMETAB is longer.)                 "
    First time it gave me one record so I dropped that record and again executed the program but next time it gave me other record.
    So I am not sure it is related to data error or database error. I can not debug the program in production and dev also as it is difficult to repeat the data in dev.
    Any clues on this ?
    Thanks in advance
    Sameer

    Looks like a precision error:
    ORA-01438: value larger than specified precision allowed for this column
    Cause: When inserting or updating records, a numeric value was entered that exceeded the precision defined for the column.
    Action: Enter a value that complies with the numeric column's precision, or use the MODIFY option with the ALTER TABLE command to expand the precision.
    Rob

  • XOQ-01600: OLAP DML error "ORA-33858: 11g Cube

    Hi All ,
    I was trying to apply the cube changes after changing the sparsity definition against one of the dimension under Storage tab and it is throwing below error .
    Any reason ?
    An error has occurred on the server
    Error class: Express Failure
    Server error descriptions:
    DPR: cannot create server cursor, Generic at TxsOqDefinitionManager::generic<CommitRoot>
    INI: XOQ-01600: OLAP DML error "ORA-33858: The value of the ampersand-substitution expression is NA." while executing DML "SYS.AWXML!R11_MANAGE_CUBE('MARKET_SALES_CUBE_WORKING.CUBE' 'ALTER' 'NUMBER' SYS.AWXML!___R11_LONG_ARG_VALUE(SYS.AWXML!___R11_LONG_ARG_DIM 1) SYS.AWXML!___R11_LONG_ARG_VALUE(SYS.AWXML!___R11_LONG_ARG_DIM 2) 'TIME.DIMENSION' 'TIME.CALENDER.HIERARCHY' 'TIME.MONTHLY.LEVEL' 'COMPRESSED' 'YES' 'YES' 'MARKET_SALES_CUBE_WORKING.SOLVE.AGGREGATIONDEFINITION' 'NO')", Generic at TxsOqStdFormCommand::execute
    Thanks in advance,
    DxP

    I have seen this kind of error show up if you rename an object (e.g. a measure) and then make further modifications. Did that happen in your case? If you export the cube to XML, delete it from AWM, and then recreate it from XML, then it may resolve the problem. If not, and if it is preventing your from making progress, then I would open a service request since we usually need to enable tracing to resolve this class of error.

  • Strange ORA-01438: value larger than specified precision...error

    When i issue a query like
    SELECT * FROM shp_dtls WHERE shp_locn='AL';
    i get the error
    ORA-01438: value larger than specified precision allowed for this columnWhy do i get an error like this for a SELECT query. Isn't this an INSERT related error?

    Normally this error happens on INSERT or UPDATE statement. There are 2 possibilities I think for it to happen on SELECT:
    1) your FGA policy defined on select for this table that inserts data into wrong column (you should also be getting the violating procedure name in your error message);
    2) wrong data has been inserted into table using OCI (in which case Oracle does not verify data correctness) or corrupted export file was impored.

  • ORA-01438: value larger than specified precisio

    Hi All,
    I have a problem when using OCCI.
    Oracle DB table A has one column, data type is NUMBER(10,0).
    c++ code>>
    Query from table
    rs = stmt->executeQuery(...select * from A...);
    long long n = (long long)(long double)rs->getNumber(1);
    Insert to table-
    Number nVal= Number((long double)n);               
    stmt->setNumber(1, nVal);
    --Error: ORA-01438: value larger than specified precisio
    How shall I do this insert(c++ data type is long long, DB column data type is NUMBER(10, 0))?
    Message was edited by:
    user591149

    Try to convert __int64 to std::string, then
    Number num = 0;
    num.fromText(env, strNumber, '999999999999');
    and use this occi Number in your statement.

  • ORA-01438: value larger than specified precision

    I'm using a web application, and when try to insert a value larger than the column precision, I get the error ORA-01438.
    I need to capture this error and show my own error when this happen, is any way to do this?
    .

    You can write your own message using syntax like
    PRAGMA EXCEPTION_INIT(exception_name, -Oracle_error_number);
    Detailed information is here
    http://download-west.oracle.com/docs/cd/B10501_01/appdev.920/a96624/07_errs.htm#707

  • XOQ-01600: OLAP DML error while executing DML "SYS.AWXML!R11_LOAD_DIM"

    hi all,
    I am trying to do the tutorial Building OLAP 11g Cubes (http://st-curriculum.oracle.com/obe/db/11g/r1/olap/cube/buildicubes.htm), but when I try to "Maintain Cube SALES_CUBE" I get the following error:
    An error has occurred on the server
    Error class: Express Failure
    Server error descriptions:
    INI: error creating a definition manager, Generic at TxsOqConnection::generic<BuildProcess>
    INI: ORA-35571: The maximum number of load errors has occurred. No changes from this step were committed to the database.
    XOQ-01600: OLAP DML error while executing DML "SYS.AWXML!R11_LOAD_DIM", Generic at TxsOqStdFormCommand::execute
    INI: XOQ-01601: error while loading data for Cube Dimension "OLAPTRAIN.PRODUCT" into the analytic workspace, Generic at TxsOqStdFormCommand::callR11LoadDim
    at oracle.olapi.data.source.DataProvider.callGeneric(Unknown Source)
    at oracle.olapi.data.source.DataProvider.callGeneric(Unknown Source)
    at oracle.olapi.data.source.DataProvider.executeBuild(Unknown Source)
    at oracle.olap.awm.wizard.awbuild.UBuildWizardHelper$2.construct(Unknown Source)
    at oracle.olap.awm.ui.SwingWorker$2.run(Unknown Source)
    at java.lang.Thread.run(Thread.java:662)
    Can anyone help me?
    Thanks,
    Carlos

    I have seen this (in 11.1.0.7, I believe) when the metadata cache ("kgl") gets out of synch with the data dictionary. Specifically there is a flag that determines if prefixes get added to dimension members ("use surrogates" in AWM terms) that becomes false instead of true. If you look at the generated SQL in the OUTPUT column of the CUBE_BUILD_LOG you may find that sometimes a prefix is added to dimension members (e.g. "LEAF_LEVEL_" || dim_table.leaf_column) and othertimes it is not (e.g. just dim_table.leaf_column). A workaround if this is the case may be to execute the following (as dba) before building the dimension.
    alter system flush shared_pool;

  • Catching ORA-01438 live / value larger than specified precision allowed for

    I've yet another question! Is it possible to catch the ORA-01438 error right away before the user submits the data?
    I have a db column that is used to store numbers of length 3,0. When entering 4 digits, the ORA-01438 is raised.
    The Apexlib doesn't work, since I have some conditional regions on that particular page!
    Thanks for posting your thoughts and hints.
    Regards,
    Seb

    Seb,
    simply create a validation and check the length of user input.
    brgds,
    Peter
    Blog: http://www.oracle-and-apex.com
    ApexLib: http://apexlib.oracleapex.info
    Work: http://www.click-click.at
    Training: http://www.click-click.at/apex-4-0-workshops

  • SQL Error: ORA-12899: value too large for column

    Hi,
    I'm trying to understand the above error. It occurs when we are migrating data from one oracle database to another:
    Error report:
    SQL Error: ORA-12899: value too large for column "USER_XYZ"."TAB_XYZ"."COL_XYZ" (actual: 10, maximum: 8)
    12899. 00000 - "value too large for column %s (actual: %s, maximum: %s)"
    *Cause:    An attempt was made to insert or update a column with a value
    which is too wide for the width of the destination column.
    The name of the column is given, along with the actual width
    of the value, and the maximum allowed width of the column.
    Note that widths are reported in characters if character length
    semantics are in effect for the column, otherwise widths are
    reported in bytes.
    *Action:   Examine the SQL statement for correctness.  Check source
    and destination column data types.
    Either make the destination column wider, or use a subset
    of the source column (i.e. use substring).
    The source database runs - Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    The target database runs - Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production
    The source and target table are identical and the column definitions are exactly the same. The column we get the error on is of CHAR(8). To migrate the data we use either a dblink or oracle datapump, both result in the same error. The data in the column is a fixed length string of 8 characters.
    To resolve the error the column "COL_XYZ" gets widened by:
    alter table TAB_XYZ modify (COL_XYZ varchar2(10));
    -alter table TAB_XYZ succeeded.
    We now move the data from the source into the target table without problem and then run:
    select max(length(COL_XYZ)) from TAB_XYZ;
    -8
    So the maximal string length for this column is 8 characters. To reduce the column width back to its original 8, we then run:
    alter table TAB_XYZ modify (COL_XYZ varchar2(8));
    -Error report:
    SQL Error: ORA-01441: cannot decrease column length because some value is too big
    01441. 00000 - "cannot decrease column length because some value is too big"
    *Cause:   
    *Action:
    So we leave the column width at 10, but the curious thing is - once we have the data in the target table, we can then truncate the same table at source (ie. get rid of all the data) and move the data back in the original table (with COL_XYZ set at CHAR(8)) - without any issue.
    My guess the error has something to do with the storage on the target database, but I would like to understand why. If anybody has an idea or suggestion what to look for - much appreciated.
    Cheers.

    843217 wrote:
    Note that widths are reported in characters if character length
    semantics are in effect for the column, otherwise widths are
    reported in bytes.You are looking at character lengths vs byte lengths.
    The data in the column is a fixed length string of 8 characters.
    select max(length(COL_XYZ)) from TAB_XYZ;
    -8
    So the maximal string length for this column is 8 characters. To reduce the column width back to its original 8, we then run:
    alter table TAB_XYZ modify (COL_XYZ varchar2(8));varchar2(8 byte) or varchar2(8 char)?
    Use SQL Reference for datatype specification, length function, etc.
    For more info, reference {forum:id=50} forum on the topic. And of course, the Globalization support guide.

  • I am getting error "ORA-12899: value too large for column".

    I am getting error "ORA-12899: value too large for column" after upgrading to 10.2.0.4.0
    Field is updating only through trigger with hard coded value.
    This happens randomly not everytime.
    select * from v$version
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    PL/SQL Release 10.2.0.4.0 - Production
    CORE     10.2.0.4.0     Production
    TNS for Linux: Version 10.2.0.4.0 - Production
    NLSRTL Version 10.2.0.4.0 - Production
    Table Structure
    desc customer
    Name Null? Type
    CTRY_CODE NOT NULL CHAR(3 Byte)
    CO_CODE NOT NULL CHAR(3 Byte)
    CUST_NBR NOT NULL NUMBER(10)
    CUST_NAME CHAR(40 Byte)
    RECORD_STATUS CHAR(1 Byte)
    Trigger on the table
    CREATE OR REPLACE TRIGGER CUST_INSUPD
    BEFORE INSERT OR UPDATE
    ON CUSTOMER FOR EACH ROW
    BEGIN
    IF INSERTING THEN
    :NEW.RECORD_STATUS := 'I';
    ELSIF UPDATING THEN
    :NEW.RECORD_STATUS := 'U';
    END IF;
    END;
    ERROR at line 1:
    ORA-01001: invalid cursor
    ORA-06512: at "UPDATE_CUSTOMER", line 1320
    ORA-12899: value too large for column "CUSTOMER"."RECORD_STATUS" (actual: 3,
    maximum: 1)
    ORA-06512: at line 1
    Edited by: user4211491 on Nov 25, 2009 9:30 PM
    Edited by: user4211491 on Nov 25, 2009 9:32 PM

    SQL> create table customer(
      2  CTRY_CODE  CHAR(3 Byte) not null,
      3  CO_CODE  CHAR(3 Byte) not null,
      4  CUST_NBR NUMBER(10) not null,
      5  CUST_NAME CHAR(40 Byte) ,
      6  RECORD_STATUS CHAR(1 Byte)
      7  );
    Table created.
    SQL> CREATE OR REPLACE TRIGGER CUST_INSUPD
      2  BEFORE INSERT OR UPDATE
      3  ON CUSTOMER FOR EACH ROW
      4  BEGIN
      5  IF INSERTING THEN
      6  :NEW.RECORD_STATUS := 'I';
      7  ELSIF UPDATING THEN
      8  :NEW.RECORD_STATUS := 'U';
      9  END IF;
    10  END;
    11  /
    Trigger created.
    SQL> insert into customer(CTRY_CODE,CO_CODE,CUST_NBR,CUST_NAME,RECORD_STATUS)
      2                values('12','13','1','Mahesh Kaila','UPD');
                  values('12','13','1','Mahesh Kaila','UPD')
    ERROR at line 2:
    ORA-12899: value too large for column "HPVPPM"."CUSTOMER"."RECORD_STATUS"
    (actual: 3, maximum: 1)
    SQL> insert into customer(CTRY_CODE,CO_CODE,CUST_NBR,CUST_NAME)
      2                values('12','13','1','Mahesh Kaila');
    1 row created.
    SQL> set linesize 200
    SQL> select * from customer;
    CTR CO_   CUST_NBR CUST_NAME                                R
    12  13           1 Mahesh Kaila                             I
    SQL> update customer set cust_name='tst';
    1 row updated.
    SQL> select * from customer;
    CTR CO_   CUST_NBR CUST_NAME                                R
    12  13           1 tst                                      Urecheck your code once again..somewhere you are using record_status column for insertion or updation.
    Ravi Kumar

  • Replicat error: ORA-12899: value too large for column ...

    Hi,
    In our system Source and Target are on the same physical server and in the same Oracle instance. Just different schemes.
    Tables on the target were created as 'create table ... as select * from ... source_table', so they have a similar structure. Table names are also similar.
    I started replicat, it worked fine for several hours, but when I inserted Chinese symbols into the source table I got an error:
    WARNING OGG-00869 Oracle GoldenGate Delivery for Oracle, OGGEX1.prm: OCI Error ORA-12899: value too large for column "MY_TARGET_SCHEMA"."TABLE1"."*FIRSTNAME*" (actual: 93, maximum: 40) (status = 12899), SQL <INSERT INTO "MY_TARGET_SCHEMA"."TABLE1" ("USERID","USERNAME","FIRSTNAME","LASTNAME",....>.
    FIRSTNAME is Varchar2(40 char) field.
    I suppose the problem probably is our database is running with NLS_LENGTH_SEMANTICS='CHAR'
    I've double checked tables structure on the target - it's identical with the source.
    I also tried to manually insert this record into the target table using 'insert into ... select * from ... ' statement - it works. The problem seems to be in the replicat.
    How to fix this error?
    Thanks in advance!
    Oracle GoldenGate version: 11.1.1.1
    Oracle Database version: 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production
    NLS_LANG: AMERICAN_AMERICA.AL32UTF8
    NLS_LENGTH_SEMANTICS='CHAR'
    Edited by: DeniK on Jun 20, 2012 11:49 PM
    Edited by: DeniK on Jun 23, 2012 12:05 PM
    Edited by: DeniK on Jun 25, 2012 1:55 PM

    I've created the definition files and compared them. They are absolutely identical, apart from source and target schema names:
    Source definition file:
    Definition for table MY_SOURCE_SCHEMA.TABLE1
    Record length: 1632
    Syskey: 0
    Columns: 30
    USERID 134 11 0 0 0 1 0 8 8 8 0 0 0 0 1 0 1 3
    USERNAME 64 80 12 0 0 1 0 80 80 0 0 0 0 0 1 0 0 0
    FIRSTNAME 64 160 98 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    LASTNAME 64 160 264 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    PASSWORD 64 160 430 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    TITLE 64 160 596 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    Target definition file:
    Definition for table MY_TAEGET_SCHEMA.TABLE1
    Record length: 1632
    Syskey: 0
    Columns: 30
    USERID 134 11 0 0 0 1 0 8 8 8 0 0 0 0 1 0 1 3
    USERNAME 64 80 12 0 0 1 0 80 80 0 0 0 0 0 1 0 0 0
    FIRSTNAME 64 160 98 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    LASTNAME 64 160 264 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    PASSWORD 64 160 430 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    TITLE 64 160 596 0 0 1 0 160 160 0 0 0 0 0 1 0 0 0
    Edited by: DeniK on Jun 25, 2012 1:56 PM
    Edited by: DeniK on Jun 25, 2012 1:57 PM

  • File_To_RT data truncation ODI error ORA-12899: value too large for colum

    Hi,
    Could you please provide me some idea so that I can truncate the source data grater than max length before inserting into target table.
    Prtoblem details:-
    For my scenario read data from source .txt file and insert the data into target table.suppose source file data length exceeds max col length of the target table.Then How will I truncate the data so that data migration will be successful and also can avoid the ODI error " ORA-12899: value too large for column".
    Thanks
    Anindya

    Bhabani wrote:
    In which step you are getting this error ? If its loading step then try increasing the length for that column from datastore and use substr in the mapping expression.Hi Bhabani,
    You are right.It is for Loading SrcSet0 Load data.I have increased the column length for target table data store
    and then apply the substring function but it results the same.
    If you wanted to say to increase the length for source file data store then please tell me which length ?Physical length or
    logical length?
    Thanks
    Anindya

  • ERROR ITMS-9000: "Images larger than 2000000 pixels are not allowed in books

    How to reduce image to resolve following issue ?
    ERROR ITMS-9000: "Images larger than 2000000 pixels are not allowed in books
    Thanks.

    Hi there djking!
    I have an article for you here that can help you with that issue. The article is all about adding images to your books, and can be found here:
    iBooks Author: Add and edit photos and other images
    http://support.apple.com/kb/PH2797
    Thanks for using the Apple Support Communities!
    Cheers,
    Braden

Maybe you are looking for

  • How do you find your blocked numbers in Contacts or on your iphone?

    How do you find your blocked numbers in Contacts or on your iphone? Is there a way to get a list of the blocked numbers?

  • Remove 0's after decimal.

    Hi, I have column value which returns values something like  123456.78 . This is sales information. I want the value to be returned something like this. "123,456". No 0's after decimal. The datatype for field stored in DB is Money. Can you any one he

  • Disable "Ledger Group" in FBB1

    Hi, I've performed the t code FAGLF101 for Accounts Payable, using the Sort Method "SAP". Using this sort method, the batch input session that is created gets the Ledger Group as "0L", which is the default ledger. In my case, I won't post anything in

  • 10.4.6 - SLOW INTERNET

    I'm going nuts trying to figure this out...but ever since I installed 10.4.6 my internet has been running slower than ever. It's happening in ALL BROWSERS as well as iTunes and anything else that is powered by the internet. I hear this is happening t

  • Performance OSX "Snow Leopard" vs. Windows 7

    Hello, after a few days of experience with CP5 on OSX I've loaded the Windows Version of CP5 as a part of ELS2. After installation I start CP 5 in a virtual machine. The VM has 2 processors from 4 in core i7 an 2GB RAM of 8GB on a MacBook Pro. Now I