Mapping conversion error in Dates

Hello Experts,
I am doing date conversion in Message Mapping from ECC date will come in YYYYMMDD for target i need to convert to YYYY_MM_DD for this i am using date transfer function.Issue is when date is coming like 20110908 it is converting properly like 2011-09-08
But when from ECC date is coming like 00000000 then by default it converting to 0002-11-30 how this date is coming i am not understanding actually it should convert like 0000-00-00 .
Please help me

Hi,
I am not quite sure why the conversion works like that. That said, the date 0000-00-00 is not a valid date, as that date has never existed. That may be the reason for the strange behaviour by PI.
I suggest you add fixed-values conversion before your dateConvert step in the mapping. Here you check for date 00000000 and map it to e.g. 00010101.
This should solve the issue, if you can live with a result date of 0001-01-01 in these cases. 
Br,
Kenneth

Similar Messages

  • Beware: conversion error with dates

    When I converted my pdf to xlsx format, some dates in a table were not converted correctly. In the pdf file, the dates were in mm/dd/yyyy format, but if the day was between 1-12, Export PDF treated the date as if was in dd/mm/yyyy format. It couldn't do that for dates with day 13-31 because that would have been an invalid month so these dates were converted correctly. This problem does not occur when converting the pdf file to Word docx format.
    Is this a known problem?

    Hi,
    Would you send your PDF document to [email protected] as an email attachment? 
    I will check it from my end.Please add the link to this forum post for reference.
    Regards,
    Florence

  • Crystal Data Conversion Issue (Error converting data type varchar to datetime)

    Hi,
    I can run stored procedure without error in SQL Server using my personal credentials as well as database credentials.
    I can also run Crystal Report after connecting to Stored procedure without error on my desktop using my personal credentials as well as database credentials.
    But when I upload the crystal report in BOBJDEV and when I run using database credentials report fails saying that "Error in File ~tmp1d1480b8e70fd90.rpt: Unable to connect: incorrect log on parameters. Details: [Database Vendor Code: 18456 ]" but I can run the crystal report successfully on BOBJDEV using my personal credentials.
    I googled (Data Conversion Error Message) about this issue & lot of people asked to do "Verify Database" in Crystal Report. So I did that, but when I do it I am getting a error message like this:
    Error converting data type varchar to datetime.
    Where do you think the error might be occurring? Did anyone faced this kind of issue before? If so, how to resolve it?
    (FYI, I am using Crystal Reports 2008, & for stored procedure I have used SSMS 2012 )
    Please help me with this issue.
    Thanks & Regards.
    Naveen.

    hello Naveen,
    since the report works fine in the cr designer / desktop, we need to figure out where you should post this question.
    by bobjdev do you mean businessobjects enterprise or crystal reports server? if so please post this question to the bi platform space.
    -jamie

  • Unit Conversion Error in Direct Input method for data transfer

    Hi Experts,
    I am getting a error "E MG 427: Conversion error: field BMMH6-MEINH; content PAK" When I am uploading Alt. UoM using BMHH6 structure in Direct Input. I checked value for UoM in converted data it is "PAC". I think system is internally converting it to PAK so the error is coming.
    Please let me know what need to be done to avoid this error.
    Thanks in Advance..
    -Harkamal

    Hi
    Before passing this unit to the program
    check the conversion Exit in the Domain of the Field
    and use the fun module
    CONVERSION_EXIT_ALPHA_INPUT and pass that value and see how it takes
    otherwise use the fun module UNIT_CONVERSION_SIMPLE and pass the value to program
    Regards
    Anji

  • Data Conversion Errors for the last week

    We've been running a simple Stream Analytics job for little over a month now with a very light workload. Input is Event hub and output SQL Server. We noticed today that we haven't received anything into SQL Server since 2014-12-08 (we don't receive events
    every day so we only know that everything still worked on the 8th of December), so we checked the job's logs. It seems that job is failing to process all the messages: The value of "Data Conversion Errors" is high.
    I wonder what could have happened? We haven't touched the client since we started the job so it's still sending the messages in same format. And we haven't touched the job's query either.
    Has there been an update to either to Stream Analytics or to Events Hub which could cause the issue we're seeing?

    I've followed word for word the TollApp Instructions (except the thing with NamespaceType "Messaging" that has been added to New-AzureSBNamespace).
    I have 0 line in output, and this is the service log:
    Correlation ID:
    e94f5b9e-d755-4160-b49e-c8225ceced0c
    Error:
    Message:
    After deserialization, 0 rows have been found. Possible reasons could be a missing header or malformed CSV input.
    Message Time:
    2015-01-21 10:35:15Z
    Microsoft.Resources/EventNameV2:
    sharedNode92F920DE-290E-4B4C-861A-F85A4EC01D82.entrystream_0_c76f7247_25b7_4ca6_a3b6_c7bf192ba44a#0.output
    Microsoft.Resources/Operation:
    Information
    Microsoft.Resources/ResourceUri:
    /subscriptions/eb880f80-0028-49db-b956-464f8439270f/resourceGroups/StreamAnalytics-Default-West-Europe/providers/Microsoft.StreamAnalytics/streamingjobs/TollData
    Type:
    CsvParserError
    Then I stopped the job, and connected to the event hub with a console app and received that:
    Message received. Partition: '11', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    85,21/01/2015 10:24:56,QBQ 1188,OR,Toyota,4x4,1,0,4,361203677
    Message received. Partition: '11', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    33,21/01/2015 10:25:42,BSE 3166,PA,Toyota,Rav4,1,0,6,603558073
    Message received. Partition: '11', Data: 'TollId,EntryTime,LiMessage received. Partition: '10', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    59,21/01/2015 10:23:59,AXD 1469,CA,Toyota,Camry,1,0,6,150568526
    Message received. Partition: '10', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    25,21/01/2015 10:24:17,OLW 6671,NJ,Honda,Civic,1,0,5,729503344
    Message received. Partition: '10', Data: 'TollId,EntryTime,LicensePlate,State,Make,Model,VehicleType,VehicleWeight,Toll,Tag
    51,21/01/2015 10:24:23,LTV 6699,CA,Honda,CRV,1,0,5,169341662
    Note the bug on the 3rd message. In my opinion it's unrelated, it could be the writeline that can't keep up with the stream in the console application. And at worst it's in the stream, but then I should see at least some lines in output for the correctly
    formatted messages.

  • Importing data error: SIGNEDDATA conversion error

    The complete description is:
    The number of failing rows exceeds the maximum specified. (Microsoft Data Transformation Services (DTS) Data Pump (8004202c): TransformCopy 'DTSTransformation__13' conversion error:  General conversion failure on column pair 1 (source column 'SIGNEDDATA' (DBTYPE_STR), destination column 'SIGNEDDATA' (DBTYPE_NUMERIC)).)
    We are trying to convert a value that come with miles separators, use the character "." (point) as separator and "," (comma) as decimal separator, we use a vb function (replace) to change the points for null string, but it doesn't work.  Do you have any suggestion to solve this?
    We are using BPC 7.0 MS SP3.
    Thanks in advance,
    Mariana Rodriguez.
    Edited by: Mariana Rodriguez on Mar 23, 2011 4:15 PM

    Hi Mariana,
    I've had the same kind of problem, which gave me the
    General conversion failure on column pair 1 (source column 'SIGNEDDATA' (DBTYPE_STR), destination column 'SIGNEDDATA' (DBTYPE_NUMERIC)).)
    errror message
    My issue was resolved by changing the CREDITNEGATIVE=NO option to CREDITNEGATIVE=YES, because BPC probably converted negative values as 'string' values in stead of 'numeric' when importing SAP FI data with the minus sign '-' at the end (100-) instead of SQL signeddata (-100)
    But I have to be carefull, my Account Dimension setup is as follows:
    Balance Sheet Asset => AccType=AST
    Balance Sheet Liabilities => AccType=LEQ
    Profit and Loss Revenues => AccType =INC
    Profit and Loss Expenses => AccType=INC
    My flat file seperator is a comma (,) and decimalpoint is a point (.)
    We are using BPC 7.5 MS SP07 Patch 02 and I'm not using any VBA to convert the amount.
    Maybe it can help?
    Regards,
    Peter van Drunen - FPM Solutions

  • A conversion error occurred while the program -- display data on the screen

    HI all,
    Iam getting a dump error described below:
    A conversion error occurred while the program was trying to
    display data on the screen.
    The ABAP output field and the screen field may not have the
    same format.
    Some field types require more characters on the screen than
    in the ABAP program. For example, a date field on a screen needs
    two characters more than it would in the program. When attempting to
    display the date on the screen, an error will occur that triggers the
    error message.
                  Screen name.............. " Ztable_MM_MRQ "
                  Screen number............ 0100
                  Screen field............. "WA_PO_ITEMS-MENGE"
                  Error text............... "FX015: Sign lost."
    Further data:
    Give us step by step procedure to rectify the same with T.codes
    Thanks
    Regards
    Siraj

    Raymond
    please give details in se51 where i have to put a "V"  to allow negative amounts
    whether it is in Text or I/O templates
    name                                     type of screen element         Text or I/O field
    WA_PO_ITEMS-MENGE     Text                           PO_quantity__     
    regards

  • Bulk load data conversion error (type mismatch or invalid character for the specified codepage) for row 1, column 3 (NumberOfMultipleMatches).

    Hi,
    I have a file where fields are wrapped with ".
    =========== file sample
    "asdsa","asdsadasdas","1123"
    "asdsa","asdsadasdas","1123"
    "asdsa","asdsadasdas","1123"
    "asdsa","asdsadasdas","1123"
    ==========
    I am having a .net method to remove the wrap characters and write out a file without wrap characters.
    ======================
    asdsa,asdsadasdas,1123
    asdsa,asdsadasdas,1123
    asdsa,asdsadasdas,1123
    asdsa,asdsadasdas,1123
    ======================
    the .net code is here.
    ========================================
    public static string RemoveCharacter(string sFileName, char cRemoveChar)
                object objLock = new object();
                //VirtualStream objInputStream = null;
                //VirtualStream objOutStream = null;
                FileStream objInputFile = null, objOutFile = null;
                lock(objLock)
                    try
                        objInputFile = new FileStream(sFileName, FileMode.Open);
                        //objInputStream = new VirtualStream(objInputFile);
                        objOutFile = new FileStream(sFileName.Substring(0, sFileName.LastIndexOf('\\')) + "\\" + Guid.NewGuid().ToString(), FileMode.Create);
                        //objOutStream = new VirtualStream(objOutFile);
                        int nByteRead;
                        while ((nByteRead = objInputFile.ReadByte()) != -1)
                            if (nByteRead != (int)cRemoveChar)
                                objOutFile.WriteByte((byte)nByteRead);
                    finally
                        objInputFile.Close();
                        objOutFile.Close();
                    return sFileName.Substring(0, sFileName.LastIndexOf('\\')) + "\\" + Guid.NewGuid().ToString();
    ==================================
    however when I run the bulk load utility I get the error 
    =======================================
    Bulk load data conversion error (type mismatch or invalid character for the specified codepage) for row 1, column 3 (NumberOfMultipleMatches).
    ==========================================
    the bulk insert statement is as follows
    =========================================
     BULK INSERT Temp  
     FROM '<file name>' WITH  
      FIELDTERMINATOR = ','  
      , KEEPNULLS  
    ==========================================
    Does anybody know what is happening and what needs to be done ?
    PLEASE HELP
    Thanks in advance 
    Vikram

    To load that file with BULK INSERT, use this format file:
    9.0
    4
    1 SQLCHAR 0 0 "\""      0 ""    ""
    2 SQLCHAR 0 0 "\",\""   1 col1  Latin1_General_CI_AS
    3 SQLCHAR 0 0 "\",\""   2 col2  Latin1_General_CI_AS
    4 SQLCHAR 0 0 "\"\r\n"  3 col3  Latin1_General_CI_AS
    Note that the format file defines four fields while the fileonly seems to have three. The format file defines an empty field before the first quote.
    Or, since you already have a .NET program, use a stored procedure with table-valued parameter instead. I have an example of how to do this here:
    http://www.sommarskog.se/arrays-in-sql-2008.html
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Date field - no date entered - conversion error

    Hi,
    i've a input text field with a datepattern attached in my jsp, whose requrired attribute is set to false. If i submit the page and no date is entered, JSF returns "Conversion error occured" for the date field.
    For testing purposes i made a simple JSP which contains only this date field and a reload button, which only submits the form. When the button is clicked the conversion error is displayed as described.
    What am i doing wrong here?
    thx
    brunft
    <html>
    <%@ page contentType="text/html" %>
    <%@ taglib uri="http://java.sun.com/jsf/html" prefix="h" %>
    <%@ taglib uri="http://java.sun.com/jsf/core" prefix="f" %>
    <f:view locale="de">
    <html>
        <head><title>Date test</title></head>
        <body>
        <h:form id="datetest">
          <p> </p>
          <h:inputText id="someDate" styleClass="formText" value="#{testDate.someDate}">
            <f:convertDateTime pattern="#{testDate.datePattern}" />
          </h:inputText>
          <p> </p>
          <h:commandButton id="refresh" styleClass="formButton"
                  value="Reload"
                  action="#{testDate.actionReload}" />
          <p> </p>
          <p style="color:red;"><h:messages /></p>
        </h:form>
        </body>
    </html>
    </f:view>The bean:
    public class TestDate {
      private Date someDate;
      private static String datePattern = "dd.MM.yyyy";
      public TestDate() {
      public Date getSomeDate() {
        return this.someDate;
      public void setSomeDate(Date someDate) {
        this.someDate = someDate;
      public String getDatePattern() {
        return this.datePattern;
      public String actionReload() {
        return "success";
    }faces-config.xml:
      <managed-bean>
        <description>Date test.</description>
        <managed-bean-name>testDate</managed-bean-name>
        <managed-bean-class>test.TestDate</managed-bean-class>
        <managed-bean-scope>request</managed-bean-scope>
      </managed-bean>

    You can try by deleting "pattern=="#{testDate.datePattern}"
    <f:convertDateTime pattern="#{testDate.datePattern}" />

  • Mapping debbuging error:DBG1006: Error while initializing test data

    Hi,
    I have created a mapping in which I am trying to delete data from a table based on exisitence of rows in another table with a filter condtion in between them.When i try to run the debbugger, it throws following error:
    ============
    Analyzing map for debug...
    Retrieving Control Center connection info...
    Connecting to Control Center schema...
    Checking character set of Control Center schema...
    Configuring sources and targets...
    Mapping Debugger Error:
    oracle.wh.service.sdk.mapping.debugger.WBMappingDebuggerException: DBG1006: Error while initializing test data for sources and targets:
    oracle.wh.repos.sdk.CMPException: Trying to access invalid Object.
    Element ID: 110138
    Status: 4
    Owning FCO: 98409
    ==================
    Any idea what could be the possible cause?
    Query which I am trying to simulate is of following type:
    Delete from TableA t1
    where exits(select 1 from TableB t2 where t1.col1 = t2.col1 and t2.col2 = 'delete')
    Thanks
    AJ.

    Oops..thats weird.
    Tell me one thing...then how gud is the code generated in terms of working...if someone has to really test the code pre-hand for expected business logicWhy i am raising this concern is becoz there are many features for which OWB provides support in indirect way(workaround) like exists/not exists.
    Do we need to deploy mapping each time and then test through a run from Control centre?
    Other thing, As far i understand, the code generated is not in such a format that one can directly copy paste it in any sql/plsql editor and test it...
    Regards
    AJ.

  • Data conversion error

    Hi All,
    I am getting below error in my SQL server agent job.
     Source: Data Flow Task Data Conversion [31]     Description: Data conversion failed while converting column "Partner" (935) to column "Copy of Partner" (940).  The conversion returned status value 4 and status text "Text
    was truncated or one or more characters had no match in the target code page.".  End Error  Error: 2013-09-25 15:16:31.32     Code: 0xC020902A     Source: Data Flow Task Data Conversion [31]     Description: The
    "output column "Copy of Partner" (940)" failed because truncation occurred, and the truncation row disposition on "output column "Copy of Partner" (940)" specifies failure on truncation. A truncation error occurred on
    the specified object of the specified component.  End Error  Error: 2013-09-25 15:16:31.33     Code: 0xC0047022     Source: Data Flow Task SSIS.Pipeline     Description: SSIS Error Code DTS_E_PROCESSINPUTFAILED.  The
    ProcessInput method on component "Data Conversion" (31) failed with error code 0xC020902A while processing input "Data Conversion Input" (32). The identified component returned an error from the ProcessInput method. The error is specific
    to the component, but the error is fatal and will cause the Data Flow task to stop running.  There may be error messages posted before this with more information about the failure.  End Error  Error: 2013-09-25 15:16:31.35     Code:
    0xC02020C4     Source: Data Flow Task Excel Source [1]     Description: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.  End Error  Error: 2013-09-25 15:16:31.36     Code: 0xC0047038
        Source: Data Flow Task SSIS.Pipeline     Description: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "Excel Source" (1) returned error code 0xC02020C4.  The component returned a failure
    code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the
    failure.  End Error  DTExec: The package execution returned DTSER_FAILURE (1).  Started:  3:10:23 PM  Finished: 3:16:38 PM  Elapsed:  374.184 seconds.  The package execution failed.  The step failed.
    Kindly help me to debug this

    Possible cause of error: Source has more characters than what is allowed by the conversion.
    i.e. Conversion allows DT_STR(50) (50 characters) but a source record has more than 50 characters for that column.
    Possible solution: increase the conversion character limit if possible. Or, under Error Outputs > Truncation > switch from Fail Component to Redirect Row or Ignore failure (you will then have to decide what you wish to do for such issue). Or, make
    sure the source can not violate this character limit.

  • Most simple query on Event Hub stream (json) constantly gives Data Conversion Errors

    Hello all,
    Been playing with ASA in December and didn't have any issues, my queries kept working and outputted the data as needed.  However, since January, I created a new demo, where I now constantly get Data Conversion errors.  The scenario is described
    below, but I have the following questions:
    Where can I get detailed information on the data conversion errors?  I don't get any point now (not in the operation logs and not in the table storage of my diagnostic storage account)
    What could be wrong in my scenario and could be causing these issues
    The scenario I have implemented is the following:
    My local devices send EventData objects, serialized through Json.Net to an Event Hub with 32 partitions.
    I define my query input as Event Hub Stream and define the data as json/utf8.  I give it the name TelemetryReadings
    Then I write my query as SELECT * FROM TelemetryReadings
    In the output, I create an output on blob with CSV/UTF8 encoding
    After that, I start the job
    The result is an empty blob container (no output written) and tons of data conversion errors in the monitoring graph.  What should I do to get this solved?
    Thanks
    Sam Vanhoutte - CTO Codit - VTS-P BizTalk - Windows Azure Integration: www.integrationcloud.eu

    So, apparently the issue was related to the incoming objects, I had.  I was sending unsupported data types (boolean and Dictionary).  I changed my code to remove these from the json and that worked out well.  There was a change that got deployed
    that (instead of marking the unsupported fields as null, they were throwing an exception).  That's why things worked earlier.
    So, it had to do with the limitation that I mentioned in my earlier comment:
    https://github.com/Azure/azure-content/blob/master/articles/stream-analytics-limitations.md
    Unsupported type conversions result in NULL values
    Any event vales with type conversions not supported in the Data Types section of Azure Stream Analytics Query Language
    Reference will result in a NULL value. In this preview release no error logging is in place for these conversion exceptions.
    I am creating a blog post on this one
    Sam Vanhoutte - CTO Codit - VTS-P BizTalk - Windows Azure Integration: www.integrationcloud.eu

  • Data Conversion Error Message

    I am getting the following error when running a report in Crystal Reports 2008:
    'Database Connection Error: '42000:[Microsoft][ODBC SQL Server Driver][SQL Server]Error converting data type varchar to datetime. [Database Vendor Code: 8114 ]'
    I can run my stored procedure without error in SQL Server.  When I add the stored procedure to a new, simple Crystal Report, it runs without error and data is returned.  The Crystal Report that is getting the error has been converted from Crystal Reports 8 to Crystal Reports 2008.  There has to be something in the way the report is passing data to the stored procedure that is causing the issue.  Any ideas?

    Hi, 
    After you open the report in CR2008, go to Database | Verify Database.  You should get prompted again for the parameters.  Enter them and try running the report. 
    There is a big difference in how CR8 and 2008 work with SQL.  Verifying the database usually clears these issues up.  If it doesn't fix, try browsing each field datetime to see which field Crystal is seeing as character. 
    Good luck,
    Brian

  • Arithmatic conversion error in query

    I have a stored procedure for a report . The underlying table has a decimal(18,2) column.
    In the Stored Procedure code, the conversion to decimal (12,2) gives an error as one of the records has a value that exceeds (12,2).
    I am running this report for a userID that does not have permission to this project (so bad row will not be included in the result).
    When running the Stored procedure from query analyser, this row with bad data is not selected based on #TblPermission and SP returns other rows without errors.
    The issue is when running the SP from the application, i am getting an arithmatic conversion error.
    It appears that the SELECT conversions are happening even before join on the temp table is satisfied.
    Shouldnt it filter data based on the join first and then do the conversions (like when run from query analyser) ?
    We will be fixing the conversion issue at a later point.
    But any ideas/suggestions on how to address this to stop it from erroring when called from the UI ?
    CREATE PROCEDURE [dbo].[Test_Report]
    @_user varchar(10)
    AS
    BEGIN
    SELECT PID INTO #TblPermission FROM dbo.tbl_permissions(nolock) WHERE userid= @_user
    SELECT A.NAME,
     CAST(A.ProjectAmount as decimal(12,2))  Amt -- giving an arithmatic conversion error
    FROM TableA A
    Join #TblPermission UP on UP.PID= A.PID
    WHERE A.ProjectType ='mapped'
    end

    It appears that the SELECT conversions are happening even before join on the temp table is satisfied.
    Shouldnt it filter data based on the join first and then do the conversions (like when run from query analyser) ?
    The answer to that question is no.  SQL is allowed to process a query in whatever order it believes will be most efficient.  So it can do the conversion, and later do the WHERE clause.  So you have to write the code so that it is safe if SQL
    does that.  One way is to use a case statement to only do the conversion when the value will fit in decimal(12,2), so you could do
    SELECT A.NAME,
    Case When A.ProjectAmount Between -9999999999.99 And 9999999999.99 Then CAST(A.ProjectAmount as decimal(12,2)) End Amt
    FROM TableA A
    Join #TblPermission UP on UP.PID= A.PID
    WHERE A.ProjectType ='mapped'
    That will work on all releases of SQL.
    If you are on SQL 2012 or later, you can use the new TRY_CONVERT function which will do conversions without causing errors.  Instead of an error when the valued being converted is invalid, you get NULL.  So that would be
    SELECT A.NAME,
    TRY_CONVERT(decimal(12,2), A.ProjectAmount) Amt
    FROM TableA A
    Join #TblPermission UP on UP.PID= A.PID
    WHERE A.ProjectType ='mapped'
    Tom

  • Conversion error in process flow

    I've build a process flow with the following construction
    1. Function call (DB) which returns a date via a OUT parameter (Checked)
    2. Bound the OUT parameter to a Process-flow parameter of type date
    3. Use the proces-flow parameter to bind a input parameter in a following mapping.
    All is well in validation, generation and deployment, but upon running I stumble onto a conversion error. Somehow the process-flow parameter is represented in a format (2006-09-21 00:00:00.000000000 +02:00) which is not accepted in the mapping (tries to convert it using 'yyyy-mm-dd'). I cannot find the conversion in the coding.
    I've tried to see if either in the mapping or in the process-flow I'm able to input the format of the date, but was not successful.
    Anyone any idea

    Chris,
    Looks like a request has not got deleted properly from the table :
    Options that you could look at :
    1. Delete all the data from the ODS and load again ( if that is an option )
    2. There are some entries you need to delete from the RSODSACTREQ table- but for this the suggestion would be to raise an OSS note for SAP to look into the same - we had a similar issue and raised an OSS note for the same and then got some notes from sap.
    Also please mention your BI version and patch level.
    Arun

Maybe you are looking for

  • Disabling toolbar in browser

    i have to disable the toolbar in the browser when the application is running.ie back and other buttons. can this be done. i don't know java script or jsp much as i'm a beginner. does it require java script or jsp. i have a java script code but i dont

  • Can't install Nokia Podcasting application on the ...

    Is Nokia Podcasting compatible with the N86? Always get an error when installing it. If not, what other Podcasting applications are available? Thanks  Solved! Go to Solution.

  • Cannot install itunes because old version will not delete

    When installing itunes I get an error message tells me the the older version of itunes cannot be deleted. I can not find any older versions of itunes on my PC to delete. Please advise Thanks

  • Safe Boot restart problem

    In SnowLeopard I never had this problem... but in Lion 10.7.2 on my 27" iMac after doing a Safe Boot a simple restart doesn't work. The system hangs with a blank gray screen (not the gray linen login screen)(and a moveable cursor... i.e. system does

  • Crystal report 2008 installation on top of BO XI R2

    Hi,       I would like to install the Crystal Reports 2008 with SP3 and Xcelsius2008 on top of BO XI R2 Version.       Is there any disadvantage by doing this activity on XI R2?       Could you please tell me the installation steps for crystal report