Data is being truncated or clipped

I have a report that has a Stored Procedure as the data source. One of the columns that is returned, by the stored procedure has been changed from a single field to the result of a case statement. The original field, a string, was 26 characters in length. When the result of the case statement returns a value greater than 26 characters the string is truncated to 26 characters. I have double checked the stored procedure and it does return strings greater than 26 characters. What change do I need to make in my report so that it will display more than 26 characters? I have checked the grow property and the number of lines is set to "0". The result was the data wrapped to the next line, but the data was still truncated to 26 characters.

Crystal is very smart, it reads the Stored Procedure and/or database where you must define the length of the field, a must for any database structure or variable. If you change the source then you have to tell Crystal you did so. Verifying updates not just the reference to the field length but if that field is used in any formula or other parts of the report objects they all get updated also so CR can allocate more space, and in your case, to hold more data.
So yes it would be convenient in your case only to just right click and define but there are consequences when changes are done that must be propagated through the report.
It actually easier to just click on Database and then Verify, 2 clicks and you are done.
Another option is in the Report Options where you can tell CR to Verify the database every time the report is opened. It causes delays but until you lock your database and SP's it is a convenient option so you don't have to do it manually. Comes in handy if more than one DBA is making changes.
Thanks again
Don

Similar Messages

  • Data being truncated in output - even though field large enough

    I have a rdf report. I see that one of the columns is not displayed completely - even though it is big enough to accomodate the data that is being returned from the query. I increased the size of the field in the designer - still the data is being truncated. What could be going on here?

    beside the enlargment of the visible area of the column in the layout editor, You should also increase the width property of that column from the property inspector.
    Regards
    Mostafa

  • ALV to EXCEL, data being truncated

    I am trying to download my ALV output into excel spreadsheet, but for one of the field, the column is being truncated. EG the no 123456 is coming as 12345.. please help
    Thanks
    Nirmal

    Hi Nirmal,
    this is well known problem, try doing this..
    while building the fieldcat for that field, give this also
    <b>fieldcat-no_zero = 'X'.</b>
    this solves your Problem
    Regards
    vijay

  • File is being truncated by Sender File Adapter

    Hello,
    My input file is very wide. It has got 175 fields and is around 4000+ char wide. When file adapter loads it, the inbound payload does not contain all the fields (i.e. it is being truncated). Any idea that how can I load this too wide file!
    Thanks,

    Hey Dehra,
    <i>Firstly, the number of records in your file is not a problem at all. Your limitation is the length of one record. See question 3 in note 545923. It says that your transfer structure cannot be more that 1962 bytes including the system fields.
    Now to calculate how many fields will fit into 1962 bytes. For this, take the first field. If the infoobject is defined as CHAR 9 then it is 9 bytes, a key-figure is 10 bytes most of the time. So add up. Remember even if in your flat file there are 2 chars but it is mapped into a char 9 field, it still counts as char 9. I don't think number of columns is a factor.
    You can split your file into multiple files of 1962 bytes size but keeping the unique key of the records common across all the files. So that you can merge then in an ODS maybe. Again remember that fields like request number, date packet number, partition ID and record number has to be there in every PSA and a part of the 1962 limit.</i>
    Information shared by  Somnath Kumar in Truncated flat file
    This might help you to solve your problem
    thanks,
    pooja

  • Stored Proc output being truncated with latest ojdbc14 and classes12.jar

    Hello,
    I downloaded the latest ojdbc14.jar from Oracle website and used it with my Java program. I noticed a peculiar behavior. If my stored proc is returning a VARCHAR with large number of characters, the output is being truncated. This worked fine with the older ojdbc14.jar (and the older classes12.jar). It does not work properly with the new ojdbc14 and classes12.jar. Is there something that I am missing here?
    My Oracle Server version is - Oracle Database 10g Enterprise Edition Release 10.2.0.1.0.
    Below is the sample stored proc and the Java file to execute the same.
    Am I missing something here?
    CREATE OR REPLACE PROCEDURE TestOracleDriver (returnVal OUT VARCHAR2) IS
    BEGIN
    returnVal := '';
    FOR tmpVar in 1..2000
    Loop
              returnVal := returnVal || to_char(tmpVar) || ',';          
    END LOOP;
    END;
    import java.sql.*;
    import java.util.*;
    import java.util.Hashtable;
    import oracle.jdbc.driver.OracleTypes;
    public class OracleDriverTest
         public static void main(String argv[]) throws Exception
              Connection c = null;
              try
                   Driver dr = new oracle.jdbc.OracleDriver();
                   Properties props = new Properties();
                   props.put("user", "scott");
                   props.put("password", "tiger");
                   c = dr.connect("jdbc:oracle:thin:@SERVER_NAME:PORT:SID", props);
                   System.out.println("GOT A PLAIN CON" );
                   System.out.println("The driver is " + c.getMetaData().getDriverVersion() );
                   System.out.println("The DBMS is " + c.getMetaData().getDatabaseProductVersion() );
                   execute(c);
              catch (Exception e)
              e.printStackTrace();
              finally { try {c.close();} catch (Exception ign){} }
         private static String execute(Connection connection )
                   String procName = "TestOracleDriver";
                   CallableStatement cstmt=null;
                   ResultSet rs = null;
                   String returnedData = null;
                   try
                        cstmt = connection.prepareCall("{call " + procName+" (?)}");
                        cstmt.registerOutParameter(1,OracleTypes.VARCHAR);
                        cstmt.execute();
                        returnedData = cstmt.getString(1);
                        System.out.println("ProcTest ::execute --> Returned data is \n"+ returnedData);
                        return returnedData;
              catch (Exception e) {
                   System.out.println("ProcTest ::execute -->Error while executing "+procName);
                   e.printStackTrace();
                   return null;
              finally
                   try
                        connection.close();
                   catch (Exception e1)
                        System.out.println("ProcTest ::execute -->Error while closing connection ");
                             e1.printStackTrace();
                   }//end of catch
         }//end of finally
    }//End of execute method
    Regards,
    Hari
    Message was edited by:
    user553601
    Message was edited by:
    user553601

    Hari,
    If you think it is a bug, you can report it to Oracle Support via the MetaLink Web site.
    Good Luck,
    Avi.

  • Sporadically getting error "string or binary data would be truncated" in SQL server 2008 while inserting in a Table Type object

    I am facing a strange SQL exception:-
    The code flow is like this:
    .Net 4.0 --> Entity Framework --> SQL 2008 ( StoredProc --> Function {Exception})
    In the SQL Table-Valued Function, I am selecting a column (nvarchar(50)) from an existing table and (after some filtration using inner joins and where clauses) inserting the values in a Table Type Object having a column (nvarchar(50))
    This flow was working fine in SQL 2008 but now all of sudden the Insert into @TableType is throwing  "string or binary data would be truncated"  exception. 
    Insert Into @ObjTableType
    Select * From dbo.Table
    The max length of data in the source column is 24 but even then the insert statement into nvarchar temp column is failing.
    Moreover, the same issue started coming up few weeks back and I was unable to find the root cause, but back then it started working properly after few hours
    (issue reported at 10 AM EST and was automatically resolved post 8 PM EST). No refresh activity was performed on the database.
    This time however the issue is still coming up (even after 2 days) but is not coming up in every scenario. The data set, for which the error is thrown, is valid and every value in the function is fetched from existing tables. 
    Due to its sporadic nature, I am unable to recreate it now :( , but still unable to determine why it started coming up or how can i prevent such things to happen again.
    It is difficult to even explain the weirdness of this bug but any help or guidance in finding the root cause will be very helpful.
    I also Tried by using nvarchar(max) in the table type object but it didn't work.
    Here is a code similar to the function which I am using:
    BEGIN
    TRAN
    DECLARE @PID
    int = 483
    DECLARE @retExcludables
    TABLE
        PID
    int NOT
    NULL,
        ENumber
    nvarchar(50)
    NOT NULL,
        CNumber
    nvarchar(50)
    NOT NULL,
        AId
    uniqueidentifier NOT
    NULL
    declare @PSCount int;
    select @PSCount =
    count('x')
    from tblProjSur ps
    where ps.PID
    = @PID;
    if (@PSCount = 0)
    begin
    return;
    end;
    declare @ExcludableTempValue table (
            PID
    int,
            ENumber
    nvarchar(max),
            CNumber
    nvarchar(max),
            AId
    uniqueidentifier,
            SIds
    int,
            SCSymb
    nvarchar(10),
            SurCSymb
    nvarchar(10)
    with SurCSymbs as (
    select ps.PID,
                   ps.SIds,              
                   csl.CSymb
    from tblProjSur ps
                right
    outer join tblProjSurCSymb pscs
    on pscs.tblProjSurId
    = ps.tblProjSurId
    inner join CSymbLookup csl
    on csl.CSymbId
    = pscs.CSymbId 
    where ps.PID
    = @PID
        AssignedValues
    as (
    select psr.PID,
                   psr.ENumber,
                   psr.CNumber,
                   psmd.MetaDataValue
    as ClaimSymbol,
                   psau.UserId
    as AId,
                   psus.SIds
    from PSRow psr
    inner join PSMetadata psmd
    on psmd.PSRowId
    = psr.SampleRowId
    inner join MetaDataLookup mdl
    on mdl.MetaDataId
    = psmd.MetaDataId
    inner join PSAUser psau
    on psau.PSRowId
    = psr.SampleRowId
                inner
    join PSUserSur psus
    on psus.SampleAssignedUserId
    = psau.ProjectSampleUserId
    where psr.PID
    = @PID
    and mdl.MetaDataCommonName
    = 'CorrectValue'
    and psus.SIds
    in (select
    distinct SIds from SurCSymbs)         
        FullDetails
    as (
    select asurv.PID,
    Convert(NVarchar(50),asurv.ENumber)
    as ENumber,
    Convert(NVarchar(50),asurv.CNumber)
    as CNumber,
                   asurv.AId,
                   asurv.SIds,
                   asurv.CSymb
    as SCSymb,
                   scs.CSymb
    as SurCSymb
    from AssignedValues asurv
    left outer
    join SurCSymbs scs
    on    scs.PID
    = asurv.PID
    and scs.SIds
    = asurv.SIds
    and scs.CSymb
    = asurv.CSymb
    --Error is thrown at this statement
    insert into @ExcludableTempValue
    select *
    from FullDetails;
    with SurHavingSym as (   
    select distinct est.PID,
                            est.ENumber,
                            est.CNumber,
                            est.AId
    from @ExcludableTempValue est
    where est.SurCSymb
    is not
    null
    delete @ExcludableTempValue
    from @ExcludableTempValue est
    inner join SurHavingSym shs
    on    shs.PID
    = est.PID
    and shs.ENumber
    = est.ENumber
    and shs.CNumber
    = est.CNumber
    and shs.AId
    = est.AId;
    insert @retExcludables(PID, ENumber, CNumber, AId)
    select distinct est.PID,
    Convert(nvarchar(50),est.ENumber)
    ENumber,
    Convert(nvarchar(50),est.CNumber)
    CNumber,
                            est.AId      
    from @ExcludableTempValue est 
    RETURN
    ROLLBACK
    TRAN
    I have tried by converting the columns and also validated the input data set for any white spaces or special characters.
    For the same input data, it was working fine till yesterday but suddenly it started throwing the exception.

    Remember, the CTE isn't executing the SQL exactly in the order you read it as a human (don't get too picky about that statement, it's at least partly true enough to say it's partly true), nor are the line numbers or error messages easy to read: a mismatch
    in any of the joins along the way leading up to your insert could be the cause too.  I would suggest posting the table definition/DDL for:
    - PSMetadata, in particular PSRowID, but just post it all
    - tblProjectSur, in particularcolumns CSymbID and TblProjSurSurID
    - cSymbLookup, in particular column CSymbID
    - PSRow, in particular columns SampleRowID, PID,
    - PSAuser and PSUserSur, in particualr all the USERID and RowID columns
    - SurCSymbs, in particular colum SIDs
    Also, a diagnostic query along these lines, repeat for each of your tables, each of the columns used in joins leading up to your insert:
    Select count(asurv.sid) as count all
    , count(case when asurv.sid between 0 and 9999999999 then 1 else null end) as ctIsaNumber
    from SurvCsymb
    The sporadic nature would imply that the optimizer usually chooses one path to the data, but sometimes others, and the fact that it occurs during the insert could be irrelevant, any of the preceding joins could be the cause, not the data targeted to be inserted.

  • Date not being displayed in the custom format

    Hi ,
      I have a problem wherein i am unable to change the format of the date.
    I need the date to be in the format 'January 10,2010.' however the date is being displayed as '1/10/10'. I have tried all options for the date formatting. I specified the custom format as 'MMMM D,YYYY' as mentioned in the SAP Library. I am trying to use the std custom options like YYYY-MM-DD, the date is not being dispalyed in the custom format provided by SAP either.
    I have binded date from the context, used the Current Date field provided by SAP.  Nothing works.
    Its not working on the new form. However the old forms have the same kind of formatting and are being displayed as desired.
    Please suggest as to what the issue could be.
    Thanks,
    Soumya.

    Hi All,
    Even I'm encountering the same issue. I'm trying to control the outputted date format programmatically.Please have a look.
    data: v_int_date like sy-datum,
            v_ext_date(10).
    v_int_date = '20110201'
    write v_int_date to v_ext_date MM/DD/YYYY.
    Here I've tried to convert the date to external format using a fixed format(MM/DD/YYYY). But still it's getting outputted in the format DD/MM/YYYY only as that's the way its defined in the User Master!
    Is there any solution for this? How can I 'override' defaults in the SAP User Master?
    Thanks,
    Mahesh

  • My performance is very slow when I run graphs. How do I increase the speed at which I can do other things while the data is being updated and displayed on the graphs?

    I am doing an an aquisition and displaying the data on graphs. When I run the program it is slow. I think because I have the number of scans to read associated with my scan rate. It takes the number of seconds I want to display on the chart times the scan rate and feeds that into the number of samples to read at a time from the AI read. The problem is that it stalls until the data points are aquired and displayed so I cannot click or change values on the front panel until the updates occur on the graph. What can I do to be able to help this?

    On Fri, 15 Aug 2003 11:55:03 -0500 (CDT), HAL wrote:
    >My performance is very slow when I run graphs. How do I increase the
    >speed at which I can do other things while the data is being updated
    >and displayed on the graphs?
    >
    >I am doing an an aquisition and displaying the data on graphs. When I
    >run the program it is slow. I think because I have the number of
    >scans to read associated with my scan rate. It takes the number of
    >seconds I want to display on the chart times the scan rate and feeds
    >that into the number of samples to read at a time from the AI read.
    >The problem is that it stalls until the data points are aquired and
    >displayed so I cannot click or change values on the front panel until
    >the updates occur on the graph. What can I do to be a
    ble to help
    >this?
    It may also be your graphics card. LabVIEW can max the CPU and you
    screen may not be refreshing very fast.
    --Ray
    "There are very few problems that cannot be solved by
    orders ending with 'or die.' " -Alistair J.R Young

  • Recovering Data from a truncated table in OraXE?

    A funny thing happened, i truncated a table containing BLOBs and other stuff, with over 700 entries!!, now i wanna know if it´s feasible to recover that truncated data, i'm just wondering about it, i know it sounds silly i know, but the storage space didn´t get shorter after the data was truncated from the table, so after this´s been said, can any body explain me what happened, can i retrieved this truncated data that i truncated from the table (without any previous backup)?
    Message was edited by:
    efebo_abel2002
    Message was edited by:
    efebo_abel2002
    Mensaje editado por:
    efebo_abel2002

    Truncate simply resets the pointer that marks the 'end of data' in a table to the beginning of the table. The storage is still allocated, on the assumption that you want to reuse the table and want to avoid the overhead associated with extending the table. The system knows there are never any rows past the 'end of data' and it won;t bother looking past that point.
    Recovering that data from the actual current database - about as likely as recovering data from the Windows recycle bin when you've set the max allowable storage in the recycle bin to '0'. Technically possible, but needing someone with a lot of experience and patience ... and absolutely no changes to the database after truncate. An expensive proposition.
    You have, of course, switched to archive log mode and been taking regular backups. (Have you not?) In which case, a 'point in time' recover, to just before the truncate, would be the solution. This brings back the database (or at least a tablespace) back to a former 'life', and is quite different from restoring a table.

  • Entire Scenario how the data is being process.

    Hi,
    I need the full scenario in detail, when the sender adapter pick the file from the source directory, how the data is passed to IS and how the data is passed to adapter engine and how the adapter engine process the data and how the data is send to adapter framework and wat all the steps adapter framework perform and on wat step the audit logs is maintain, how messaging, logging and queing will done in AFW and after processing the data in adpter engine how the data is being passed to Integration Engine and how the pipeline steps will get execute and how the data is been transfered to receiver.
    All others steps being process while sending the data from sender system to receving steps and how the data is process internally and where audit log is maintain etc.

    Hi,
    Please see the below links
    see the below links to helps you lot
    http://help.sap.com/saphelp_nw2004s/helpdata/en/fd/16e140a786702ae10000000a155106/content.htm
    /people/siva.maranani/blog/2005/05/25/understanding-message-flow-in-xi
    http://help.sap.com/saphelp_nw2004s/helpdata/en/6a/a12241c20af16fe10000000a1550b0/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/e4/6019419efeef6fe10000000a1550b0/content.htm
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/327dc490-0201-0010-d49e-e10f3e6cd3d8
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/34a1e590-0201-0010-2c82-9b6229cf4a41
    Regards
    Chilla

  • Need to understand how the PGI date is being calculated

    Hi,
    Need to understand how the PGI date is being calculated.
    Does the PGI date get recalculated if the cusotme delivery date is changed.
    Thanks,
    Raki

    Hi,
    Thanks for the information and please let us know is there any calculation logic exist for that?
    I have changed delivery date in the sales order and based on that-
    - Transportation planning/material availability/loading/schedule line and confirmed quantity date was chaged.
    Is this the - GI date - Transist time= PGI?
    Thanks,

  • How much data is being used when i mirror my ipad to my tv through the apple tv?

    how much data is being used when i mirror my ipad to my tv through the apple tv?

    wwcswapmeet wrote:
    I will try that. Thank you.  I am hoping to buy a wifi hotspot, but wondered what type of data plan I should get if I were to use it consistantly for 8 hours a day (only on saturdays and sundays).  A great tool for work & presentations!!
    It will use minimal internet bandwidth unless your content is streaming from web>iPad>AppleTV. 
    Local network comms for content already on the iPad (eg presentations) will not generally use internet bandwidth apart from protected iTunes material that requires brief internet authorisation.
    AC

  • Error occurred while data was being converted.

    hi all,
    i am gettting a run-time error (dump) in this line of code,
                netval = wa_tb_str_ekpo-ktmng  *  wa_tb_str_ekpo-netpr.
    ktmng is data type QUAN and has reference field MEINS of EKPO table
    while netpr is data type CURR and has reference field WAERS of EKKO table.
    ABAP ERROR ANALYSIS SAYS :
    Error occurred while data was being converted.
    Please help me.
    Thanks.

    hi prasad,
                 you  can pass this value to the char field. then pass it to the field. and let me know the result.
    regards,
    venkat.

  • How can I find out what type of data is being sent over one of the phones on my plan?

    I have a few phones under a shared family plan.
    I wanted to know what type of data is being sent over them, wether this is VOIP, Streaming Audio or Video.

    I think the only way you can find out is if you look at the device itself.

  • CfLayout tabs are being truncated in IE 9

    After installing IE 9 I've noticed that my cflayout tabs are now being truncated.  They look file in IE 8, FF and Chrome, but in 9 some of the tabs are truncated (exampl...).  Has anyone else had this problem or come across a solution for this?
    Please no IE bashing, my app has to work for all major browsers =/

    Fixed.  Hold ctrl + MouseWheelDown.  IE was zoomed in...  >.<

Maybe you are looking for

  • Oracle 10g on slackware 10.1 or suse?

    dear gurus, I would like to asked, can i installed Oracle 10g on slackware 10.1 or suse new relase 10... what it makes differnece from red hat linux... as we could not find support of red hat in ksa, and red hat become more commerical flavor and aski

  • Texting problem with Apple 6

    My SMS text messages will not deliver from Apple 6

  • How to use NOORDER in PL/SQL

    Hi, I am using SELECT clause to fetch some hard coded values and want to display them in the same order. For eg. SELECT DISTINCT DECODE (grnt.PLAN_TYPE, 0, 'A', 1, 'D', 2, 'B', 3, 'C', 4, 'E','') AS CD FROM X; I want o/p as : CD A D B C E But its com

  • CTS+ is integrated in Charm, but can't seem to use it....

    We have CTS+ working for our Portal system.  We were able to define an Implementation Project in ChaRM and define the Portal DEV, QAS and Prod systems with no problems.  We created the task list and the Portal systems show up.  But now when we create

  • About state in ejbs

    where does a stateful session bean holds its state between method calls? I have read it in book as application variables. stateless session beans also have application variables then why dont they maintain state?