Identify data type

jow can i identify the data type of a variable that is returned by the getParameter method?
suppose there is an int myint=5;
that is submited from an html form, or sent by another page.
i get the variable with request.getParameter("myint")
and ofcourse put it in a String.
How can i check what it was before?

Answered in crosspost
http://forum.java.sun.com/thread.jspa?threadID=769930

Similar Messages

  • Identify bind variable data type 121?

    Looks like the OCI forum is not popular. Reposting here with hopes of better luck... Or perhaps I wasn't specific enough for the forum? OK, this is obviously a binary data type. The problem is that I want to know exactly which of them it is. Or I'd like to know if someone has improperly coded an OCI application. Here's the original post.
    I received a level 12 SQL trace file that referenced data type 121 for a bind variable value. I have searched unsuccessfully for documentation in many places. Can anyone tell me what data type this is? Following is the relevant excerpt from the trace file.
    bind 18: dty=121 mxl=4000(4000) mal=00 scl=00 pre=00 oacflg=00 oacfl2=1 size=4000 offset=0
    bfp=022c1e80 bln=4000 avl=2196 flg=15
    value=
    Dump of memory from 0x22C1E80 to 0x22C2714
    22C1E80 01A214BC 022C1E88 022F1F88 01E0A954 [.....,.../.....T]

    Hi
    The following are the internal datatypes used by oracle related to 121.
    KOTAD, KOTMD, KOTMI, KOTTB, KOTTD
    Not sure about their exact use but these datatypes might be used internally for LOB's.
    Hope it helps.
    Diwakar

  • Does new 1402 Activity object in Cloud for Customer have wrong data type namespaces?

    I see that the individual "activity" objects (e.g., ActivityTask, AppointmentActivity, etc) are being deprecated as of 1405 and moving to a new Activity object with a TypeCode. That makes a lot of sense.
    In reviewing this new Activity object, I've noticed that many of the node elements utilize Data Types in the http://sap.com/xi/Common/DataTypes namespace instead of the http://sap.com/xi/AP/Common/GDT. For example, the root node element UUID has a data type UUID in the http://sap.com/xi/Common/DataTypes namespace, unlike all the other objects in the Public Solution Model.
    There are many more elements in this new Activity object that are mismatched like this to other common usage in other objects. Is this a new trend or a mistake?

    Hello Gregory, Alessandro,
    just for clarification:
    SAP wanted to eliminate the superflous supplementary components from the data types used in the Business Objects.
    Take the GDT Identifier as an example. It has the following supplementary components:
    SCHEME_ID
    SCHEME_VERSION_ID
    SCHEME_AGENCY_ID
    SCHEME_AGENCY_SCHEME_ID
    SCHEME_AGENCY_SCHEME_AGENCY_ID
    CONTENT
    You very seldom need the SCHEME_ID or the SCHEME_AGENCY_ID (if the ID is a GTIN or DUNS number then you need them). I never came over a situation to use the SCHEME_AGENCY_SCHEME_AGENCY_ID.
    So the data types from the "new" namespace omit those components. But SAP noticed the effort to adapt all the coding.
    Therefore it was decided that only new objects will use the data types from the new namepace.
    In the case of the ActivityTask, AppointmentActivity, etc. this wil requires - as Alessandro mentioned - some adoption effort.
    HTH,
          Horst

  • How to insert data in a column with uniqueidefier data type

    Guys,
    I need insert data in a column with uniqueidefier data type, when i am trying to that getting error.
    error message says: "Conversion failed when converting from a character string to uniqueidentifier."
    I have data in table a col1,col2,col3,col4 - col3,col4 has datatype as varchar and i am updating table b columns col1,col2 with table a col3 and col4.
    Please guide how to do it.

    Hi,
    Not any String can be convert to uniqueidentifier.
    1. you have to make sure u use a value which is fir to be uniqueidentifier
    2. Use convert or cast in the insert query in order to convert the string into uniqueidentifier
    insert X ... convert(uniqueidentifier, 'string which fit to be convert to uniqueidentifier')
    Please post DDL+DML for more specific help
    DDL = Data Definition Language. In our case that is, CREATE TABLE statements for your tables and other definitions that are needed to understand your tables structure and there for let us to test and reproduce the problem in our server. Without DDL no one
    can execute any query.
    How to get DDL: Right click on the table in Object Explorer and select script table as CREATE. Post these create table scripts here.
    DML = data manipulation language is a family of queries used for manipulating the data it self like: inserting, deleting and updating data. In our case we need some sample data in order to check the query and get result, so we need some indert query for
    sample data.
    If you post a "create query" for the tables and "insert query" with some sample, then we could help you without Assuming/Guessing. There is a reason that DDL is generally asked for and expected when discussing query problems - it helps
    to identify issues, clarify terminology and prevent incorrect assumptions.  Sample data also provides a common point of reference for the discussion. A script that can be used to illustrate or reproduce the issue you have, will encourage others to help.
    [Personal Site] [Blog] [Facebook]

  • Issue with Info Object Transport after changing Data Type

    Hi Experts,
    We have a DSO which is running past 5 years. And recently(3 days back) we added new fields into that DSO. Delta loaded for last three days and new info object (Say XXX) data populated into DSO.
    Now the problem is, We identified that, info object data type is not correct. We used data type NUMC  instead of CHAR.So Character information is missing for that field.
    Example:
    Data from ECC: ABC123
    Data Loaded to DSO: 123 (Missed character ABC)
    So we deleted data from DSO and changed the info object data type in Development system.
    And also I have deleted only last three days delta records from DSO and transported my info object to Production. But its giving error as: Info object contains data in DSO".
    But that info object field is empty in DSO. I have already deleted last three days delta.
    Do I need to Delete all 5 years data from DSO to change the data type of recently added new info object ?
    Please give me your solutions and ideas to solve this issue.
    Thank you,
    Best Regards,
    Santhosh

    Hi Raman,
    Thank you for your answer.
    When I changed the data type of info object in Dev, I deleted the content of DSO. And same thing transported to QA as well. Before importing changed info object, I just deleted the content of DSO from QA. So transport done successfully QA.
    But in Production we have a history of past 5 years. So I cant delete all contents of DSO.
    So just deleted only the delta request's which contains data for that info object(3 days back, we moved that info object(NUMC data type) to production,So last 3 days delta only loaded for that info object). And tried to transport it. But it was failed. I am sure we need to drop all data from DSO, if I am interested to go with same info object.
    I have some comments on your first approach,
    1. If we delete info objects from Dev DSO and transport to Production will give transport failure, Why because, my previous transports errors clearly saying that info object contains data in DSO . So it wont allow to remove that info object from DSO.
    2.We are using same info object technical name in BO Data federator also. So if we change add new info objects again we need to make changes with BO as well. I am thinking this point as my last option if i cant find any other solution.
    Thank you.
    Best Regards.

  • What is the data type for the Data Dashboard XY Graph?

    Hey all,
    I am playing with the new Data Dashboard (!!) on my iPad 2 and am having trouble setting up the shared variables that host data to the XY Graph.
    On my PC, in my Project (inside of a variable library) I have set up shared variables for all of the other controls/indicators available in the app; They all work great. However, I have not been able to identify the data type that is for the XY Graph in Data Dashboard. Of course, inside LabVIEW, I use a cluster of 2 arrays; this is not an available data type that I can select while creating a new variable on the hosting side. I have even gone so far as to create a custom control from my XYGraph in LabVIEW and then creating the data type "From Custom Control..." with no success.
    What gives? Am I missing something obvious? Thanks for the help!!
    Tom
    Solved!
    Go to Solution.

    Great question, Aerogoob.
    The XY graph can be bound to a 1d array of "points", where each point is a cluster of two numerics (X and Y). To create a shared variable of this type, you can set the data type to "From Custom Control..." in the shared variable properties dialog. Of course, first you'll have to build the custom control of the correct type: array of cluster of two numerics.
    If any of that doesn't make sense, please post back and we can walk you through it in more detail.
    Also, just for completeness, the chart indicator can be bound to a scalar numeric or to an array of numerics. The graph indicator can only be bound to an array of numerics.

  • Date type in xml schema

    The date type in schema processor doesn't work. I can replace the date in the xml with any character and still get the xml validated. Same problem with year and month type.

    Hi,
    The behaviour I posted above has been identified as a bug by Oracle and a patch(#6053186) now exists for this problem. This patch is available for database version 10.2.0.3 and the bug would eventually be fixed only in database version 11g according to the details published for the bug.
    Rgds,
    Rakesh.

  • SUS - Added Data Type Enhancement and Higher level Proxies are not active

    Hello,
    I've added a field to our current data type enhancement Z_Purchase_Order_Item.  Once I regenerate the proxy on the enhancement and activate it the field appears as it should in the high level items that use the enhancement (PurchaseOrderRequest_In).  But those proxies have become inactive and when I try to activate them I get this message:
    Interface II_BBPX1_SUS_PO was repaired before the Modification Assistant was enabled. 
    All Modification Assistant functions only apply to future modifications, not to those already
    undertaken.  This means:
    -The modification overview only displays future modifications.
    -When resetting to the standard, the system will reset all objects to their current version, since
    the actual standard can no longer be identified by the Modification Assistant.
    -Support for adjustment after an upgrade will only be available for future modifications. 
    Modifications that already exist must be re-made manually using version management.
    The next message says:
    Object can only be created in SAP package.
    Then the status bar shows "Proxy Activated".  But when I close and reopen the proxy I see that it is once again inactive. 
    Does any know what I need to do to activate this proxy? 
    Thanks,
    Matt

    In SPROXY you can open your proxy and then view the Activation Log under the GoTo menu.  The log will explain better what the problems might be.  In my case I needed to activate another data type enhancement first.
    Thanks,
    Matt

  • DATE TYPE

    I am using this code:
    DROP TABLE "SYSTEM"."DATE_TBL";
    CREATE TABLE "SYSTEM"."DATE_TBL"
    "JEDAN" DATE,
    "DVA" VARCHAR2(50 BYTE),
    "TRI" VARCHAR2(50 BYTE)
    TRUNCATE TABLE date_tbl;
    ALTER SESSION SET NLS_DATE_FORMAT="DD-MM-YYYY HH24:MI:SS";
    DECLARE
    CURSOR c1 IS
    SELECT * FROM LDAP_DATA;
    event_rec c1%ROWTYPE;
    BEGIN
    OPEN c1;
    LOOP
    FETCH c1 INTO EVENT_REC;
    EXIT WHEN (c1%NOTFOUND);
    INSERT INTO DATE_TBL VALUES (
    to_date( event_rec.eventtime, 'mm/dd/yyyy HH12:mi:ss PM' ),
    event_rec.sourceIP,
    event_rec.sourceport);
    END LOOP;
    CLOSE c1;
    END;
    select * from date_tbl
    when I do the last SELECT order I get this in SQL Developer:
    JEDAN DVA TRI
    26-04-0010 16:02:38 127.0.0.1 46132
    And that is the result I want do get.
    BUT, when I export table I get this result:
    JEDAN      DVA      TRI
    26.04.10 16:02:38,000000000      127.0.0.1      46132
    How can I remove these nine zeros in column JEDAN (type DATE)?

    Hi,
    fibonacci wrote:
    The thing is that I want to have column with DATE type that holds dates.
    I want to do export in CSV format, but I have nine zeros. I do this - right click on the table -> export -> csv
    Is there any other way to export data from table in CSV format where I could use your suggestion?Is this some SQL Developer utility you're using?
    It sounds like it automatically applies a standard format to all DATEs, but the same format is designed to handle TIMESTAMPs (which can include fractions of a second) as well.
    There's probably a way to tell SQL Developer to use the format you want, 'DD.MM.YY HH24:MI:SS', as its default format, but I don't know what that is. Try posting a question in the SQL Developer forum:
    SQL Developer
    A work-around is to convert the DATE to a string, using TO_CHAR. I'm sure the tool won't try to re-format strings. You may need to create a view that has the formatted VARCHAR2 in place of the DATE. (Do NOT actually change the table to make the column a VARCHAR2; I mean use a query or a view to produce a VARCHAR2 column where your real table has a DATE.)
    How can I create my own scheme? I have Oracle XE 10g and I am using SQL DeveloperLog in as SYSTEM. Use the CREATE USER command to make a new user, and the GRANT command to give it some basic privileges.
    For example:
    CREATE USER          fibonacci
    IDENTIFIED BY          leonardo
    DEFAULT TABLESPACE     users
    GRANT     CONNECT, RESOURCE     TO fibonacci;After that, log in as FIBONACCI (password LEONARDO) to create tables, and do most anything else.
    The only time you should connect as SYSTEM is to to things like the commands above.

  • Function returning string.  Data type question

    Hello all,
    Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production
    Our database has a parent record master_member_record_id & and children of those records member_record_id. I wrote a function which returns the master record for the child.
    eg. get_master(member_record_id). simple enough.
    I have wrote the opposite of this also, get_member_records(master_member_record_id). Obviously this function has multiple records to return so I have it set to return a listagg string with ',' separation. They want to be able to use this function as follows:
    select * from member_table where member_record_id in (get_member_records(master_member_record_id));
    or something similar, I realize this is a data type issue here, wondering if this is even possible. What do you think?
    Thanks in advance for your criticism/help,
    Chris

    My disagreement is with how pipeline table functionality is used.
    Instead of this (what it sounds like the OP is doing but returning CSV format)
    create or replace type TClientIdArray is table of number;
    create or replace function GetClientIDs( parentID number ) return TClientIdArray is
      array TClientIdArray;
    begin
      select
        client_id bulk collect into array
      from master_table
      where parent_id = parentID;
      return( array );
    end;A pipeline would look as follows:
    create or replace type TClientIdArray is table of number;
    create or replace function GetClientIDs( parentID number ) return TClientIdArray pipelined is
    begin
      for c in (select
                   client_id
                 from master_table
                 where parent_id = parentID ) loop
        pipe row( c.client_id );
      end loop;
      return;
    end;The first method is in fact more sensible in this case, especially when used from PL/SQL. A single SQL call/context switch to get a list of client identifiers. The issues with the first method are
    - cannot effectively deal with a large data set of client identifiers
    - would be suboptimal to use this function in subsequent SQL
    - if called by a client, a ref cursor (not collection) should be returned
    But assuming that the list of client identifiers are needed to be stepped through via complex PL/SQL app processing, using a (small) array is fine (assuming that concurrency/locking is not needed via a FOR UPDATE on the select that identifies the client identifiers to process).
    A pipeline in this case does not do anything better. It introduces more moving parts - as a native PL/SQL can no longer be used to get the collection and step through it. A TABLE() SQL function is needed, a SQL statement, and more context switching.
    The pipeline is simply an unnecessary layer between the code wanting to use the client identifiers and the SQL statement supplying the client identifiers. This approach of SQL -calls-> PIPELINE -calls-> MORE SQL is invariable wrong - unless the pipeline (PL/SQL code) does something very funky with the data from its SQL call, prior to piping it, that cannot be done using SQL.
    If the "user" of that client identifiers is SQL code, then both the above methods are flawed. As this code is already SQL, why use PL/SQL to call more SQL code? Simply use the SQL code/logic that supplies the client identifier list directly in the SQL code that needs the client identifiers. This means something like SQL JOIN, EXISTS, or IN clauses.

  • Conversion failed when converting the varchar value 'undefined' to data typ

    Conversion failed when converting the varchar value 'undefined' to data type int.
    hi, i installed oracle insbridge following the instruction in the manual. in rate manager, when i tried to create a new "Normal rating" or "Underwriting", im getting the following exception
    Server Error in '/RM' Application.
    Conversion failed when converting the varchar value 'undefined' to data type int.
    Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code.
    Exception Details: System.Data.SqlClient.SqlException: Conversion failed when converting the varchar value 'undefined' to data type int.
    Source Error:
    An unhandled exception was generated during the execution of the current web request. Information regarding the origin and location of the exception can be identified using the exception stack trace below.
    Stack Trace:
    [SqlException (0x80131904): Conversion failed when converting the varchar value 'undefined' to data type int.]
    System.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection) +1948826
    System.Data.SqlClient.SqlInternalConnection.OnError(SqlException exception, Boolean breakConnection) +4844747
    System.Data.SqlClient.TdsParser.ThrowExceptionAndWarning(TdsParserStateObject stateObj) +194
    System.Data.SqlClient.TdsParser.Run(RunBehavior runBehavior, SqlCommand cmdHandler, SqlDataReader dataStream, BulkCopySimpleResultSet bulkCopyHandler, TdsParserStateObject stateObj) +2392
    System.Data.SqlClient.SqlDataReader.HasMoreRows() +157
    System.Data.SqlClient.SqlDataReader.ReadInternal(Boolean setTimeout) +197
    System.Data.SqlClient.SqlDataReader.Read() +9
    System.Data.SqlClient.SqlCommand.CompleteExecuteScalar(SqlDataReader ds, Boolean returnSqlValue) +50
    System.Data.SqlClient.SqlCommand.ExecuteScalar() +150
    Insbridge.Net.Fwk.DAO.DataAccess.ScalarQuery(String connectionString, String command, Transaction transType, Object[] procParams) +110
    [Exception: Cannot Execute SQL Command: Conversion failed when converting the varchar value 'undefined' to data type int.]
    Insbridge.Net.Fwk.DAO.DataAccess.ScalarQuery(String connectionString, String command, Transaction transType, Object[] procParams) +265
    Insbridge.Net.Fwk.DAO.SqlProcessor.ExecuteScalarQueryProc(String subscriber, String datastore, String identifier, String command, Transaction transType, Object[] procParams) +101
    Insbridge.Net.Fwk.DAO.SqlProcessor.ExecuteScalarQuery(String subscriber, String identifier, String command) +22
    Insbridge.Net.RM.IBRM.ExeScalar(String cmd, Object[] paramsList) +99
    Insbridge.Net.RM.Components.Algorithms.AlgEdit.Page_Load(Object sender, EventArgs e) +663
    System.Web.Util.CalliHelper.EventArgFunctionCaller(IntPtr fp, Object o, Object t, EventArgs e) +14
    System.Web.Util.CalliEventHandlerDelegateProxy.Callback(Object sender, EventArgs e) +35
    System.Web.UI.Control.OnLoad(EventArgs e) +99
    System.Web.UI.Control.LoadRecursive() +50
    System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) +627
    my insbridge versions are as follows
    IBRU 4.0.0 Copyright ©2010, Oracle. All rights reserved. - Version Listing
    RMBUILD.DLL 4.0.0.0 (x86)
    SRLOAD.DLL 3.13.0 (x86)
    IBRM v04.00.0.17
    IB_CLIENT v04.00.0.00
    RM.DLL 4.00.0 (x86)
    IBFA 3.0.2
    OS: Windows Server 2003
    DB: Sql Server 2008
    Browser: IE8
    how do i solve this, please help

    This is an error due to conversion failed from character string to int datatype. Infact, the table column contains "NO" value which you are trying to convert to Int/SUM which is illegal. Without your code and table structure, its difficult to pinpoint your
    actual issue. But check, all columns for value "NO". 

  • Problems dealing with Boolean data types(SQL).

    Greetings.
    Below is an sql code i am trying to execute.
    select Band, Memb_contact =
    CASE 
    when contact_person >= 0 then 'Yes'
    else 'No' 
    end, Club_Name, Club_contact
    from Members, Club
    where contact_person = 'Yes'
    My column contact_person has a data type of boolean. It stores 1 and -1 only.
    I'm am trying to execute the select statement to return The Band Name and contact number only if the person
    is identified as a contact person in the members table.
    It return the following error:
    Msg 245, Level 16, State 1, Line 1
    Conversion failed when converting the varchar value 'Yes' to data type bit.
    Please help.

    Try the below: You can not have 'Yes' instead use 1.
    select Band, Memb_contact =
    CASE 
    when contact_person >= 0 then 'Yes'
    else 'No' 
    end, Club_Name, Club_contact
    from Members, Club
    where contact_person = 1

  • Changing a VARCHAR(2) to BLOB Data type

    I am trying to change a VARCHAR(2) data type into a BLOB. I have researched the web and found what I thought was the solution. That was to create a new column that was a BLOB, copy the information from the column I what to modify into it, then delete the original column and rename the BLOB column to that original name.
    However I am getting error messages such as…
    ORA-00904: "CUSTOMER2"."Cust_Firstname": invalid identifier
    When I issue this statement:
    UPDATE CUSTOMER2 SET "CUSTOMER2"."Cust_Firstname"=Temp_Name;
    or
    ORA-00904: "CUSTOMER2"."Temp_Name": invalid identifier
    When I issue this statement:
    UPDATE CUSTOMER2 SET "CUSTOMER2"."Cust_Firstname"="CUSTOMER2"."Temp_Name";
    Or
    And I have gotten an error message saying update failed was expecting a numeric got BLOB.
    I have UPDATED the table…
    SQL>Alter table CUSTOMER2 ADD Temp_Name BLOB;
    Table altered.
    0.74 seconds
    Does anyone have any ideas? All help would be very much appreciated!
    -Robert

    Hello Robert,
    first, I don't see which part of your question concerns APEX. There is {forum:id=75} where you probably get best input for your problem.
    Anyhow, just a few things you might think about before posting there.
    I am trying to change a VARCHAR(2) data type into a BLOBI'm not sure if you really want this to be a BLOB. Ususally, if text fields get too small, you'd use CLOB. Again, I'd guess the 4000 bytes you can easily store in a VARCHAR2 (not VARCHAR(2), right? this would be a two byte text field) would be sufficient for what I would expect to find in a column named "Cust_Firstname".
    UPDATE CUSTOMER2 SET "CUSTOMER2"."Cust_Firstname"=Temp_Name;Is this the actual code? I guess you didn't create the column "Cust_Firstname" with double quotes. Double quotes enable case sensitive handling, so "Cust_Firstname" is different from "cust_firstname" or "CUST_FIRSTNAME". The latter version would be the default if no quotes are used, i.e. everything is handled as it were in upper case.
    Concerning your assignment order: If "Temp_Name" is the new BLOB column, you wouldn't want "Cust_Firstname" to be overwritten with that empty value.
    If I understood you right, you want to transfer the value from "Cust_Firstname" (which is VARCHAR2) to "Temp_Name", which you created as BLOB. You should change the order in the update statement then.
    -Udo

  • Alter mount database failing: Intel SVR4 UNIX Error: 79: Value too large for defined data type

    Hi there,
    I am having a kind of weird issues with my oracle enterprise db which was perfectly working since 2009. After having had some trouble with my network switch (replaced the switch) the all network came back and all subnet devices are functioning perfect.
    This is an NFS for oracle db backup and the oracle is not starting in mount/alter etc.
    Here the details of my server:
    - SunOS 5.10 Generic_141445-09 i86pc i386 i86pc
    - Oracle Database 10g Enterprise Edition Release 10.2.0.2.0
    - 38TB disk space (plenty free)
    - 4GB RAM
    And when I attempt to start the db, here the logs:
    Starting up ORACLE RDBMS Version: 10.2.0.2.0.
    System parameters with non-default values:
      processes                = 150
      shared_pool_size         = 209715200
      control_files            = /opt/oracle/oradata/CATL/control01.ctl, /opt/oracle/oradata/CATL/control02.ctl, /opt/oracle/oradata/CATL/control03.ctl
      db_cache_size            = 104857600
      compatible               = 10.2.0
      log_archive_dest         = /opt/oracle/oradata/CATL/archive
      log_buffer               = 2867200
      db_files                 = 80
      db_file_multiblock_read_count= 32
      undo_management          = AUTO
      global_names             = TRUE
      instance_name            = CATL
      parallel_max_servers     = 5
      background_dump_dest     = /opt/oracle/admin/CATL/bdump
      user_dump_dest           = /opt/oracle/admin/CATL/udump
      max_dump_file_size       = 10240
      core_dump_dest           = /opt/oracle/admin/CATL/cdump
      db_name                  = CATL
      open_cursors             = 300
    PMON started with pid=2, OS id=10751
    PSP0 started with pid=3, OS id=10753
    MMAN started with pid=4, OS id=10755
    DBW0 started with pid=5, OS id=10757
    LGWR started with pid=6, OS id=10759
    CKPT started with pid=7, OS id=10761
    SMON started with pid=8, OS id=10763
    RECO started with pid=9, OS id=10765
    MMON started with pid=10, OS id=10767
    MMNL started with pid=11, OS id=10769
    Thu Nov 28 05:49:02 2013
    ALTER DATABASE   MOUNT
    Thu Nov 28 05:49:02 2013
    ORA-00202: control file: '/opt/oracle/oradata/CATL/control01.ctl'
    ORA-27037: unable to obtain file status
    Intel SVR4 UNIX Error: 79: Value too large for defined data type
    Additional information: 45
    Trying to start db without mount it starts without issues:
    SQL> startup nomount
    ORACLE instance started.
    Total System Global Area  343932928 bytes
    Fixed Size                  1280132 bytes
    Variable Size             234882940 bytes
    Database Buffers          104857600 bytes
    Redo Buffers                2912256 bytes
    SQL>
    But when I try to mount or alter db:
    SQL> alter database mount;
    alter database mount
    ERROR at line 1:
    ORA-00205: error in identifying control file, check alert log for more info
    SQL>
    From the logs again:
    alter database mount
    Thu Nov 28 06:00:20 2013
    ORA-00202: control file: '/opt/oracle/oradata/CATL/control01.ctl'
    ORA-27037: unable to obtain file status
    Intel SVR4 UNIX Error: 79: Value too large for defined data type
    Additional information: 45
    Thu Nov 28 06:00:20 2013
    ORA-205 signalled during: alter database mount
    We have already checked in everywhere in the system, got oracle support as well without success. The control files are in the place and checked with strings, they are correct.
    Can somebody give a clue please?
    Maybe somebody had similar issue here....
    Thanks in advance.

    Did the touch to update the date, but no joy either....
    These are further logs, so maybe can give a clue:
    Wed Nov 20 05:58:27 2013
    Errors in file /opt/oracle/admin/CATL/bdump/catl_j000_7304.trc:
    ORA-12012: error on auto execute of job 5324
    ORA-27468: "SYS.PURGE_LOG" is locked by another process
    Sun Nov 24 20:13:40 2013
    Starting ORACLE instance (normal)
    control_files = /opt/oracle/oradata/CATL/control01.ctl, /opt/oracle/oradata/CATL/control02.ctl, /opt/oracle/oradata/CATL/control03.ctl
    Sun Nov 24 20:15:42 2013
    alter database mount
    Sun Nov 24 20:15:42 2013
    ORA-00202: control file: '/opt/oracle/oradata/CATL/control01.ctl'
    ORA-27037: unable to obtain file status
    Intel SVR4 UNIX Error: 79: Value too large for defined data type
    Additional information: 45
    Sun Nov 24 20:15:42 2013
    ORA-205 signalled during: alter database mount

  • Data types in pi 7.1

    Hi,
    when do you use core, free and aggregated data types?
    Thanks in advance
    regards,
    Ramya Shenoy

    This may help you,
    Introduction
    SAP delivers descriptions of Global Data Types (GDTs) as Enterprise Services Repository (ESR) content to customers. Customers can create their own data types in the ES Repository and use SAP GDTs.
    u2022
    Two ways of modeling data types are supported: Classical XSD style and new CCTS (Core Component Technical Specification) style.
    u2022
    Data types are classified as free-style (classical), core (atoms of CCTS modeling), or aggregate (complex CCTS)
    u2022
    Standard fault data types are no longer created automatically for every namespace (only when fault message type is created).
    u2022
    Additional functions in data type editor provide better support for structural changes (e.g., moving subtrees).
    More Information:
    SAP Help Portal: Global Data Types
    According to this, we distinguish between SAP Core Data Type and SAP Global Data Type. The SAP Core Data Type as well as the W3C Type (XSD type) does not have any business semantics. The SAP GDT has business semantics, it actually can be based on either a core or an aggregated DT.
    Free-style data types are directly based on the primitive data types, e.g., xsd:decimal. They do not need any further parameters to define themselves.
    More Information:
    u2022
    SAP Help Portal: Core Data Types and Aggregated Data Types Recommendation
    When to use free, core and aggregated data types
    Core data types are based on CCTS specification. Aggregated data types have business semantics, are based on CDTs, and are the basis for application-specific data types SAP-wide.
    Customers who adhere to the CCTS specification or have business scenarios dealing with SAP backend systems, should try and reuse the existing core and aggregated data types delivered by SAP. However, customers who wish to use their own schema for defining the data types, free-style data type is the preferred approach.
    While transferring the business scenarios from XI 3.0/PI 7.0 to PI 7.1, all the data types in the design objects will be transferred as free-style data types by default.
    While creating new objects in PI 7.1x, it is recommended to reuse the core and aggregated global data types provided by SAP.
    The detailed list of SAP delivered data types is available on SAP Developer Network (SDN) at ES Workplace.
    https://www.sdn.sap.com/irj/sdn/esworkplace
    Core data types
    Core data types are based on representation terms which are ISO 15000 5 (ebCCTS) compliant.
    SAP delivers core data types with the same name as representation terms. Applying the standard methodology simplifies the implementation of cross-company processes.
    CDTs are intermediate data types, and are a level above primitive data types like decimal, string, etc. They are defined by representation terms such as Amount, Identifier, etc. They are themselves not yet application-specific and are therefore referred to as context-free.
    While creating core data type based on any given representation term, we get an option to either use a default XSD type or reuse an existing core data types.
    For using the standard XSD types we select Type as u2018XSD Typeu2019
    For reusing already existing core data type we select Type as u2018Data Typeu2019 Prerequisite
    Define a dependency on the SWC SAP Global to reuse any of the standard core data types shipped by SAP.
    xample

Maybe you are looking for

  • The persistent store record 6,810 could not be found

    Under heavy load, we get an exception stack trace in the log, that says that an expired message could not be deleted. This message is repeated every 15-30 minutes for other records. 5 hours later, it says that a thread is stuck and the server is bloc

  • Pages to ePub... why so difficult? where are my page breaks?

    So my book is at the editors and I wanted to start to look into publishing to iBooks. So from pages I exported to ePub and the final book looked terrible. So I went through all the forums and I have got to say that there is no step by step guide to a

  • Getting error message 1402 & 1500 could not open key

    I cannot download update error message 1500, cannot remove or repair program adobe 9,3,1 error message 1402 could not open key HKEY_LOCAL_MACHINE/SOFTWARE error 1500 canot install update

  • Enabling Hot Keys during Order Processing

    Is there a way that Hot Keys can be created and assigned so that during order processing a user can "jump" from screen to screen? For example, if the user was on the Item Text Tab and wanted to go to the Additional Data B Tab, is there a way to enabl

  • Help!The synchronization of master and detail tables

    Is there a simple method to synchronize zhe primary key of detail table with master table?