Decimal conversion error (Data Integrator 11.7.000)

Did someone encoutered this error? Any solution?
1004     3488     RUN-050802     06/01/2009 10:43:23     Cannot convert data <-.91> into type <DECIMAL>. Context: Column <Project Column AIColumn7>.
It happens when I try to perform a multiplication of a number with -0.91
Thanks

this seems to be happening for a specific optimization plan
tried with 11.7.3.0 and oracle 9i, since the query is not getting pushed down so most likely not a issue with database
I am trying to reproduce this using 2 source table and multiplying 2 columns in mapping by selecting a int column from one table and decimal (6,2) column from other table and output is also decimal (6,2)
is there anything else in the mapping expression for the other column, which might be creating multiple nodes for optimizer
can you give some more information about the DF, like outer join, where condition
since you mentioned both the source are oracle, why is the expression not getting pushed down ?

Similar Messages

  • Decimal Conversion Error During Export

    Hi All
    We are migrating our production system from HP-UX to SOLARIS and Informix to Oracle.
    During the export phase for the process SAPAPPL1 terminates with error :
    DbSl Trace: Error in decimal conversion. Length: 5, Decimals: 0
    (EXP) ERROR: DbSlExeRead: rc = 2, table "S011"
    (SQL error 0)
    error message returned by DbSl:
    I am not able to get hold of it.
    Kindly help, if someone has experience on the same.
    Regards
    Rahul

    Hi Rahul,
    What is your kernel release? Please confirm that you are in latest kernel and using the latest version of R3load.
    Regards,
    Satyabrat

  • Importing data error: SIGNEDDATA conversion error

    The complete description is:
    The number of failing rows exceeds the maximum specified. (Microsoft Data Transformation Services (DTS) Data Pump (8004202c): TransformCopy 'DTSTransformation__13' conversion error:  General conversion failure on column pair 1 (source column 'SIGNEDDATA' (DBTYPE_STR), destination column 'SIGNEDDATA' (DBTYPE_NUMERIC)).)
    We are trying to convert a value that come with miles separators, use the character "." (point) as separator and "," (comma) as decimal separator, we use a vb function (replace) to change the points for null string, but it doesn't work.  Do you have any suggestion to solve this?
    We are using BPC 7.0 MS SP3.
    Thanks in advance,
    Mariana Rodriguez.
    Edited by: Mariana Rodriguez on Mar 23, 2011 4:15 PM

    Hi Mariana,
    I've had the same kind of problem, which gave me the
    General conversion failure on column pair 1 (source column 'SIGNEDDATA' (DBTYPE_STR), destination column 'SIGNEDDATA' (DBTYPE_NUMERIC)).)
    errror message
    My issue was resolved by changing the CREDITNEGATIVE=NO option to CREDITNEGATIVE=YES, because BPC probably converted negative values as 'string' values in stead of 'numeric' when importing SAP FI data with the minus sign '-' at the end (100-) instead of SQL signeddata (-100)
    But I have to be carefull, my Account Dimension setup is as follows:
    Balance Sheet Asset => AccType=AST
    Balance Sheet Liabilities => AccType=LEQ
    Profit and Loss Revenues => AccType =INC
    Profit and Loss Expenses => AccType=INC
    My flat file seperator is a comma (,) and decimalpoint is a point (.)
    We are using BPC 7.5 MS SP07 Patch 02 and I'm not using any VBA to convert the amount.
    Maybe it can help?
    Regards,
    Peter van Drunen - FPM Solutions

  • BestPractices Error - Set Decimal Notation and Date Format

    Hello everyone,
    first of all, I don't know, if this is the right forum for my question, but I hope so
    We're trying to implement the Best Practices Scenarios to a new SAP ERP System.
    We did the steps of the following url, where everything worked fine: http://help.sap.com/bp_dm603/DM_DE/html/index_DE.htm
    Then we integrated Best Practices 603V8 via Solution Builder (TA: /n/SMB/BBI). But now, when we switch to Implementation Assistant and try to activate Scenario DE_109 --> 100(DE) --> Folder USERSETTINGS, we receive the following error in report /SMB99/SU3_O009_B32: Set Decimal Notation and Date Format.
    So we called transaction SU3 and changed Decimal Notation and Date Format, but the error still exists.
    Anybody knows what to do?
    Thanks for your help!
    Best regards
    Patrick
    PS: We executed Testscript /SMB99/SU3_C009_B32 via TA SECATT and there is the error:
    TCD SU3 0,292 sec N Maintain Users Own Data

    Hi,
    Follow below link to verify whatever you have done is correct:
    http://help.sap.com/bp_bblibrary/600/Documentation/B32_BB_ConfigGuide_EN_MX.doc.
    Thanks
    Sunny

  • Data conversion error. Overflowed the specific type.

    Hi guys, as usual I am struggling with the data conversion. I have got a column BS as Float in SQL, the usual values is 10.5656445, 899.66552366 etc. I'll try to move in Excel the data but I wish to have only one decimal. I try decimal in the data conversion
    with precision 1 but it returns me the same value now I am trying
    numeric, precision 1 but it returns the error "Conversion failed because the data value overflowed the specified type". Any suggestion? Thanks

    Could you tell me what should be the result for 10.5656445, 899.66552366?  Is that 10.5, 899.6 ?
    If yes then, set Precision: 4, Scale: 1
    (This is a total of 4 digits, 3 digits for the whole number and 1 for the fractional)
    Cheers,
    Vaibhav Chaudhari

  • Most simple query on Event Hub stream (json) constantly gives Data Conversion Errors

    Hello all,
    Been playing with ASA in December and didn't have any issues, my queries kept working and outputted the data as needed.  However, since January, I created a new demo, where I now constantly get Data Conversion errors.  The scenario is described
    below, but I have the following questions:
    Where can I get detailed information on the data conversion errors?  I don't get any point now (not in the operation logs and not in the table storage of my diagnostic storage account)
    What could be wrong in my scenario and could be causing these issues
    The scenario I have implemented is the following:
    My local devices send EventData objects, serialized through Json.Net to an Event Hub with 32 partitions.
    I define my query input as Event Hub Stream and define the data as json/utf8.  I give it the name TelemetryReadings
    Then I write my query as SELECT * FROM TelemetryReadings
    In the output, I create an output on blob with CSV/UTF8 encoding
    After that, I start the job
    The result is an empty blob container (no output written) and tons of data conversion errors in the monitoring graph.  What should I do to get this solved?
    Thanks
    Sam Vanhoutte - CTO Codit - VTS-P BizTalk - Windows Azure Integration: www.integrationcloud.eu

    So, apparently the issue was related to the incoming objects, I had.  I was sending unsupported data types (boolean and Dictionary).  I changed my code to remove these from the json and that worked out well.  There was a change that got deployed
    that (instead of marking the unsupported fields as null, they were throwing an exception).  That's why things worked earlier.
    So, it had to do with the limitation that I mentioned in my earlier comment:
    https://github.com/Azure/azure-content/blob/master/articles/stream-analytics-limitations.md
    Unsupported type conversions result in NULL values
    Any event vales with type conversions not supported in the Data Types section of Azure Stream Analytics Query Language
    Reference will result in a NULL value. In this preview release no error logging is in place for these conversion exceptions.
    I am creating a blog post on this one
    Sam Vanhoutte - CTO Codit - VTS-P BizTalk - Windows Azure Integration: www.integrationcloud.eu

  • Finding Error while creating Data Integrator Repositor

    Hi,
    I am working in SAP BO Data Integrator. I have created databeses and its login. Installed SQL Server and while installing Data Integrator I got an error for creating a new repository which is as follows
    Cannot open connection to the repository.  The error message from the underlying DBMS is <ODBC call <SQLDriverConnect> for data source <MSSERVER\SQLEXPRESS> failed: <[Microsoft][ODBC SQL Server Driver][SQL Server]Login failed for user 'AIO_REPO_IDOC'. The user is not associated with a trusted SQL Server connection.>. Notify Customer Support.>. (BODI-20006)
    Can anyone resolve this problem.
    Edited by: sap_beginnner on Aug 9, 2010 4:10 PM

    Hi,
    I used SQL Server Authentication to logon to databases and for DI Version I am using SAP BusinessObjects XI 3.2.
    I tried it again by deleting all databases and then create the same databases but on creating repository for table AIO_REPO_IDOC in repository manager gives error which is as follows
    Cannot open connection to the repository.  The error message from the underlying DBMS is <ODBC call <SQLDriverConnect> for data source <IDHASOFT238933\SQLEXPRESS> failed: <[Microsoft][ODBC SQL Server Driver][SQL Server]Login failed for user 'AIO_REPO_IDOC'. The user is not associated with a trusted SQL Server connection.>. Notify Customer Support.>. (BODI-20006) 
    An error occurred during creation of the local repository. (BODI-300054)
    I also tried through another tool to acces the databases. It is working successfully.
    Edited by: sap_beginnner on Aug 10, 2010 8:58 AM

  • Unit Conversion Error in Direct Input method for data transfer

    Hi Experts,
    I am getting a error "E MG 427: Conversion error: field BMMH6-MEINH; content PAK" When I am uploading Alt. UoM using BMHH6 structure in Direct Input. I checked value for UoM in converted data it is "PAC". I think system is internally converting it to PAK so the error is coming.
    Please let me know what need to be done to avoid this error.
    Thanks in Advance..
    -Harkamal

    Hi
    Before passing this unit to the program
    check the conversion Exit in the Domain of the Field
    and use the fun module
    CONVERSION_EXIT_ALPHA_INPUT and pass that value and see how it takes
    otherwise use the fun module UNIT_CONVERSION_SIMPLE and pass the value to program
    Regards
    Anji

  • Data Conversion Errors for the last week

    We've been running a simple Stream Analytics job for little over a month now with a very light workload. Input is Event hub and output SQL Server. We noticed today that we haven't received anything into SQL Server since 2014-12-08 (we don't receive events
    every day so we only know that everything still worked on the 8th of December), so we checked the job's logs. It seems that job is failing to process all the messages: The value of "Data Conversion Errors" is high.
    I wonder what could have happened? We haven't touched the client since we started the job so it's still sending the messages in same format. And we haven't touched the job's query either.
    Has there been an update to either to Stream Analytics or to Events Hub which could cause the issue we're seeing?

    I've followed word for word the TollApp Instructions (except the thing with NamespaceType "Messaging" that has been added to New-AzureSBNamespace).
    I have 0 line in output, and this is the service log:
    Correlation ID:
    e94f5b9e-d755-4160-b49e-c8225ceced0c
    Error:
    Message:
    After deserialization, 0 rows have been found. Possible reasons could be a missing header or malformed CSV input.
    Message Time:
    2015-01-21 10:35:15Z
    Microsoft.Resources/EventNameV2:
    sharedNode92F920DE-290E-4B4C-861A-F85A4EC01D82.entrystream_0_c76f7247_25b7_4ca6_a3b6_c7bf192ba44a#0.output
    Microsoft.Resources/Operation:
    Information
    Microsoft.Resources/ResourceUri:
    /subscriptions/eb880f80-0028-49db-b956-464f8439270f/resourceGroups/StreamAnalytics-Default-West-Europe/providers/Microsoft.StreamAnalytics/streamingjobs/TollData
    Type:
    CsvParserError
    Then I stopped the job, and connected to the event hub with a console app and received that:
    Message received. Partition: '11', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    85,21/01/2015 10:24:56,QBQ 1188,OR,Toyota,4x4,1,0,4,361203677
    Message received. Partition: '11', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    33,21/01/2015 10:25:42,BSE 3166,PA,Toyota,Rav4,1,0,6,603558073
    Message received. Partition: '11', Data: 'TollId,EntryTime,LiMessage received. Partition: '10', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    59,21/01/2015 10:23:59,AXD 1469,CA,Toyota,Camry,1,0,6,150568526
    Message received. Partition: '10', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    25,21/01/2015 10:24:17,OLW 6671,NJ,Honda,Civic,1,0,5,729503344
    Message received. Partition: '10', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    51,21/01/2015 10:24:23,LTV 6699,CA,Honda,CRV,1,0,5,169341662
    Note the bug on the 3rd message. In my opinion it's unrelated, it could be the writeline that can't keep up with the stream in the console application. And at worst it's in the stream, but then I should see at least some lines in output for the correctly
    formatted messages.

  • A conversion error occurred while the program -- display data on the screen

    HI all,
    Iam getting a dump error described below:
    A conversion error occurred while the program was trying to
    display data on the screen.
    The ABAP output field and the screen field may not have the
    same format.
    Some field types require more characters on the screen than
    in the ABAP program. For example, a date field on a screen needs
    two characters more than it would in the program. When attempting to
    display the date on the screen, an error will occur that triggers the
    error message.
                  Screen name.............. " Ztable_MM_MRQ "
                  Screen number............ 0100
                  Screen field............. "WA_PO_ITEMS-MENGE"
                  Error text............... "FX015: Sign lost."
    Further data:
    Give us step by step procedure to rectify the same with T.codes
    Thanks
    Regards
    Siraj

    Raymond
    please give details in se51 where i have to put a "V"  to allow negative amounts
    whether it is in Text or I/O templates
    name                                     type of screen element         Text or I/O field
    WA_PO_ITEMS-MENGE     Text                           PO_quantity__     
    regards

  • Bulk load data conversion error (type mismatch or invalid character for the specified codepage) for row 1, column 3 (NumberOfMultipleMatches).

    Hi,
    I have a file where fields are wrapped with ".
    =========== file sample
    "asdsa","asdsadasdas","1123"
    "asdsa","asdsadasdas","1123"
    "asdsa","asdsadasdas","1123"
    "asdsa","asdsadasdas","1123"
    ==========
    I am having a .net method to remove the wrap characters and write out a file without wrap characters.
    ======================
    asdsa,asdsadasdas,1123
    asdsa,asdsadasdas,1123
    asdsa,asdsadasdas,1123
    asdsa,asdsadasdas,1123
    ======================
    the .net code is here.
    ========================================
    public static string RemoveCharacter(string sFileName, char cRemoveChar)
                object objLock = new object();
                //VirtualStream objInputStream = null;
                //VirtualStream objOutStream = null;
                FileStream objInputFile = null, objOutFile = null;
                lock(objLock)
                    try
                        objInputFile = new FileStream(sFileName, FileMode.Open);
                        //objInputStream = new VirtualStream(objInputFile);
                        objOutFile = new FileStream(sFileName.Substring(0, sFileName.LastIndexOf('\\')) + "\\" + Guid.NewGuid().ToString(), FileMode.Create);
                        //objOutStream = new VirtualStream(objOutFile);
                        int nByteRead;
                        while ((nByteRead = objInputFile.ReadByte()) != -1)
                            if (nByteRead != (int)cRemoveChar)
                                objOutFile.WriteByte((byte)nByteRead);
                    finally
                        objInputFile.Close();
                        objOutFile.Close();
                    return sFileName.Substring(0, sFileName.LastIndexOf('\\')) + "\\" + Guid.NewGuid().ToString();
    ==================================
    however when I run the bulk load utility I get the error 
    =======================================
    Bulk load data conversion error (type mismatch or invalid character for the specified codepage) for row 1, column 3 (NumberOfMultipleMatches).
    ==========================================
    the bulk insert statement is as follows
    =========================================
     BULK INSERT Temp  
     FROM '<file name>' WITH  
      FIELDTERMINATOR = ','  
      , KEEPNULLS  
    ==========================================
    Does anybody know what is happening and what needs to be done ?
    PLEASE HELP
    Thanks in advance 
    Vikram

    To load that file with BULK INSERT, use this format file:
    9.0
    4
    1 SQLCHAR 0 0 "\""      0 ""    ""
    2 SQLCHAR 0 0 "\",\""   1 col1  Latin1_General_CI_AS
    3 SQLCHAR 0 0 "\",\""   2 col2  Latin1_General_CI_AS
    4 SQLCHAR 0 0 "\"\r\n"  3 col3  Latin1_General_CI_AS
    Note that the format file defines four fields while the fileonly seems to have three. The format file defines an empty field before the first quote.
    Or, since you already have a .NET program, use a stored procedure with table-valued parameter instead. I have an example of how to do this here:
    http://www.sommarskog.se/arrays-in-sql-2008.html
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Date field - no date entered - conversion error

    Hi,
    i've a input text field with a datepattern attached in my jsp, whose requrired attribute is set to false. If i submit the page and no date is entered, JSF returns "Conversion error occured" for the date field.
    For testing purposes i made a simple JSP which contains only this date field and a reload button, which only submits the form. When the button is clicked the conversion error is displayed as described.
    What am i doing wrong here?
    thx
    brunft
    <html>
    <%@ page contentType="text/html" %>
    <%@ taglib uri="http://java.sun.com/jsf/html" prefix="h" %>
    <%@ taglib uri="http://java.sun.com/jsf/core" prefix="f" %>
    <f:view locale="de">
    <html>
        <head><title>Date test</title></head>
        <body>
        <h:form id="datetest">
          <p> </p>
          <h:inputText id="someDate" styleClass="formText" value="#{testDate.someDate}">
            <f:convertDateTime pattern="#{testDate.datePattern}" />
          </h:inputText>
          <p> </p>
          <h:commandButton id="refresh" styleClass="formButton"
                  value="Reload"
                  action="#{testDate.actionReload}" />
          <p> </p>
          <p style="color:red;"><h:messages /></p>
        </h:form>
        </body>
    </html>
    </f:view>The bean:
    public class TestDate {
      private Date someDate;
      private static String datePattern = "dd.MM.yyyy";
      public TestDate() {
      public Date getSomeDate() {
        return this.someDate;
      public void setSomeDate(Date someDate) {
        this.someDate = someDate;
      public String getDatePattern() {
        return this.datePattern;
      public String actionReload() {
        return "success";
    }faces-config.xml:
      <managed-bean>
        <description>Date test.</description>
        <managed-bean-name>testDate</managed-bean-name>
        <managed-bean-class>test.TestDate</managed-bean-class>
        <managed-bean-scope>request</managed-bean-scope>
      </managed-bean>

    You can try by deleting "pattern=="#{testDate.datePattern}"
    <f:convertDateTime pattern="#{testDate.datePattern}" />

  • Data Integrator Error ASSIGN ... CASTING

    I am consistantly getting an ASSIGN ... CASTING error on the second R/3 data flow that runs in any of my SAP jobs.  The first R/3 data flow will run correctly.  I am using Data Integrator version 11.0.2.5.  Has anyone experienced this error when attempting to run R/3 data flows?

    Greetings Post Originator,
    This post is older than 60 days and there are no entries in the past 30 days.  Based on the content discussed, it appears that you question has been answered. This message is being marked as answered and points are being assigned if available where possible. 
    Thank you for being an active participant in the SAP Forums,
    Rob Siegele
    Forum Moderator
    SAP Americas

  • Data Integrator - Bulk Load Append ERROR

    Hi,
    Im working in a SYBASE database and i have the following error:
    My source table for example have 1000 rows and i set my target table with the bulk load append option every 100 rows to commit, but every time when the process ends in my target i have 900 rows, i think the data integrator dont commit the last 100 rows.
    I have a "clean" pass of data, no where, no joins is a simple copy from one table to another table.
    Anyone can give me a solution or something to solve this problem?
    Thanks a lot...

    Sybase ASE or IQ and which version ?
    what is the Data Integrator version that you are using ?
    is the job failing with errors ? may be the last batch is not committed because of constraints on the target table, did you try inserting the same set of rows without bulk load ? try inserting the rows without bulk load and see if it gives any errors

  • File Content Conversion error ? Can anyone help?

    The Error that shows is :
    Conversion of file content to XML failed at position 0: java.lang.Exception: ERROR consistency check in recordset structure validation (line no. 6: missing structure(s) before type 'headerData'
    What could be the possible reason for this error? It is urgent and poits to be rewarded. Thanks in advance.
    My structure is like the one below :
    <HEADER>
         <.Field1..>
         <  Field2 >
    </HEADER>
    <headerData>
       <fields>
    </headerData>
    <itemData>
    <fields>
    </itemData>
    <textData>
      <fields>
    </textData>
    <Footer>
        <fields>
    </Footer>

    I checked the source file and the first field is having value "4". Also there is no other line bofore it.
    &#65279;4GIPS2SAP.DAT   GIPS ME  P.R .TRANSMISSION    20071002200710021901GIPS ME TORINO 0000000#
    10027879961  00662522  PL11619    12PL11031000        PL11031000ZDUM     33610       PL11EUR
    227879961    120071002N.2 BANCHI PER RACCOLTA STILLICIDIO     KONIECZNY 000000000001000EA 2007113000000500000
    327879961    1001FORNITURA N°2 BANCHI PER RACCOLTA STILLICIDIO PER                 #
    327879961    1002LINEA LAVORAZIONE TESTE CILINDRI.                                 #
    10027879958  00662522  PL11619    14PL11031000        PL11031000ZDUM     33610       PL11EUR
    227879958    120071002NUOVI SAG PER LINEA BASAMENTO MOTORE    KONIECZNY 000000000001000EA 2007113000000500000
    327879958    1001Fornitura scaffali a gravita per materiali della                  #
    327879958    1002linea di lavorazione besamento motore.                            #
    10027879959  00662522  PL11619    13PL11031000        PL11031000ZDUM     33210       PL11EUR
    227879959    120071002FORNITURA PEDANE ANTIFORUNISTICHE       KONIECZNY 000000000001000EA 2007113000000500000
    327879959    1001FORNITURA E POSA DELLE PEDANE ANTIFORTUNISTICHE PER               #
    327879959    1002LINEE DI LAVORAZIONE BASAMENTO MOTORE E TESTE                     #
    327879959    1003CILINDRI COME DA SPECIFICA DEL 10/09/2007.                        #
    5GIPS2SAP.DAT   GIPS ME TEXAS 000 0 0 0000015##################################################################

Maybe you are looking for