Split Index from data

Hello together,
I have a strange question to you where I'm not sure, if that work and if how that will work.
Let's say I have a Oracle DB 11.2.0.3, I think OS will be not important in that case.
I have a tablespace with some tables in there. On that tables I've created Indexes for the performance. But now, for the overview or whatever, i want to create a second tablespace and split the Index from the data so that the Index itself with it's runtime data will save in the new tablespace, but the data for which the tablespace is lies in the old tablespace.
I think, that is not possible, because an index is just a logical unit and the tablespace is a level to high. But a friend of me means it is possible.
What can you tell me to this?
Thanks and regards,
David

Thanks a lot for quick answering!
Don't misunderstand me for such a question: that was a discussion, how we could increase the performance of a tool, that is using a 11.2.0.3 DB. It is a kind of monitoring tool, that stores data in a DB and how greater the tablespace / datafile gets so longer the tool needs to analyze. But I think also the way to split the idexes won't bring some performance because if you have some documents printed out in one big box and you search just one side in the box normally you would take the same time to find that document as you split the documents into two or three or four… boxes and you have to search in each on them. Because you can’t separate, that the document is in box 1 or 2 or 3…..
Thanks a lot!

Similar Messages

  • Drop Indexes from Data Object in DOE

    Hi,
              I have created the index for data object by using Mobile Appli Attrib tab in DOE workbench. But now I need to remove it. Everytime I remove checkbox and index name and save it. Now I open the data object again index will be there,it will not be removed. 
    Thanks & Regards,
    Manisha Dabour.

    Hi Monisha,
    I tried the same in 710 SP11 system and I could delete the  index name and uncheck check boxes.
    Below are the steps I did.
    -- Selected the root node of a DO.
    -- Clicked on "Mobile Appln Attribute.
    -- Chose client type in the popup as "JAVACLNT" .
    -- Checked few fields and entered a index name.
    -- Saved the changes.
    -- repeated the above and uncheckd fields/removed index and did some changes.
    Observed that everytime my latest changes were captured.
    Can you please open a OSS message for the same.
    Regards
    Rohith

  • Split number from data string

    I have user inputs like J1.A1 to J1.A20 I want to get the range from the given text 1 to 20,
    Kindly share me the logic
    Regards
    Vikram Kumar P
    Solved!
    Go to Solution.

    I would use the Scan From String with the format set to "J1.A%d".  You will get the numeric value as an output.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Split ''Upload from PSA" to separate Data Targets

    Hi,
    I have the following situation:
    1 ODS as datasource
    2 ODS and 1 CUBE as data targets.
    I'm updating my data targets with deltas with one infopackage using process chain.
    As the CUBE is also a data source for another CUBE I want to speed up the upload process by separating the load to data targets.
    So:
    data from 1st ODS to CUBE 1 as first step then I can process data to CUBE 2 with different process chain;
    2nd step would be upload data again form 1st ODS to two remaining ODSs.
    The point is that I'm using one info package and cannot split loads per data targets. I tried to create two infopackages: one with CUBE 1 as data target and second with my ODSs as data targets. But the second infopackage doesn't return delta any more.
    Is there any solution for such a case?

    Thanks Arun for your input, the problem is that I don't want to create new transfer rules or update rules or any other objects. Only question is, if data from PSA can be loaded to the specified target with use of process chain. I know I can use the specified request to upload data targets separately but it involves manual work and I need to use info package variants.
    Message was edited by: A. Pazucha

  • Error message: User data is not indexed, yet. Index user data first

    Greeting All,
    Each time I try to index a 'Prepared' searchObjectConnector in the cockpit.. I get a message <
    User data is not indexed, yet. Index user data first>. even that I ran the z_ESH_AU_UPLOAD_AUTH_RFC_720 report in R/3 system.. and checked the auth /user data in the admin cockpit.. and it was indexed..
    1.any idea why I keep getting this error
    2.any idea how can I clean the buffer or the message Queu.. my guess this might  be not error but a q message..
    3. any suggestion please
    Thanks

    Hi Areege,
      I noticed that you created an OSS message on this issue. That was a good idea.
    There is an answer for you in OSS from yesterday.
    Kind regards, Klaus

  • Can you customize report "page" tabs so they are named automatically from data in the file, and can you add "pages" dynamically?

    I am looking to develop a dynamic report template for our internal reporting use.  Namely, I am looking for the ability to do two things:
    1)  Update the names of the tabs at the bottom of the report template (lets call them the "pages" of the report) from data stored within the file (for instance, label a tab with the "description" property from a given group or channel)
    2)  Add additional tabs with associated graphs based on the number of groups a particular file has (for instance, a particular component could have a test life of 2000 hrs, but have been put through several different tests along the way.  It would be nice to show an overall life summary on the first tab, and then break out the data from the individual tests on seperate tabs with labels specific to that test, such as test #, etc.)
    Any ideas on how this may (or may not) be done?

    Hi SethRow187,
    Let's take this one step at a time, starting with the one REPORT sheet per Data Portal Group request.  This is something that I do all the time, and I have developed lots of standard functions that I can re-use quickly and easily.  DIAdem has the basic functions which enable this, but it takes quite a few commands to get this to work the way you and I want with absolute control.
    In the "Group Sheets General" example I'm attaching below, the same REPORT layout TDR file is loaded once for each Data Portal Group, and then subsequently altered programmatically so that all the Group references in the REPORT objects (graphs, tables, text boxes) are replaced with references to the current Group.  This results in N REPORT sheets where each sheet has explicit references to a particular Group.
    In the "Group Sheets Simple" example I'm attaching below, again the same REPORT layout TDR file is loaded once for each Data Portal Group, but in this case no further change is made to each REPORT sheet after it is loaded.  This results in N REPORT sheets which are absolutely identical in content to each other and which all reference the default Group.  They appear different when refreshed because in each sheet, under its "Settings>>Layout Setup>>Worksheet Parameters" dialog there is a command configured to run each time the sheet is refreshed-- this command uses the current REPORT sheet index to set the corresponding Group to be the default Group.
    In both cases the name of the Group is used for the REPORT sheet, which is convenient in that both Group names and REPORT sheet names are required to be unique, but you could use the Group description or any other property instead as long as you ensure that it has not already been used to name a previous REPORT sheet.
    I also have scripts to concatenate the contents of N Groups into a new Group, located at the top of the Data Portal, and this combined with the below attached applications would give you the lead off overview REPORT sheet you mentioned.
    Brad Turpin
    DIAdem Product Support Engineer
    National Instruments
    Attachments:
    Group Sheets Examples.zip ‏58 KB

  • How to process data in the past based from data in the present

    hello guys,
    i have a problem in my labview programs. how to process data in the past based from data in the present ?
    i have a formula self-organizing maps
    this formula is looking for D1, D1 is neuron index that will be searched for the smallest value.and the result are D1=2 ,D2=5, D3=17 from calculating with formula  .it means the smallest value is 2, "2" from weight [2 2] in file attached.
    and then it will be in other formula
    it mean [2 2] + 0.5 ( [1 1]-[2 2] ) = [1.5 1.5]
    and the weight will be  [1.5 2 2 ] in matrix
                                              1.5 3 5
    I would appreciate any input/help on solving this
    thanks
    Attachments:
    dika.vi ‏16 KB
    weight.txt ‏1 KB
    data .txt ‏1 KB

    Hi Ronny Hanks,
    Moving your records from internal table into the database table depends upon various scenarios :-
    1. If you use INSERT statement.
    INSERT <database_table> FROM TABLE <internal_table>.
    But in this case, you need to make sure that you don't have any duplicate entries in your internal table that violates data entry into database table, else you will get a dump.
    INSERT <database_table> FROM TABLE <internal_table> ACCEPTING DUPLICATE KEYS.
    In this case, you are forcefully inserting duplicate records into your database table which may lead to data redundancy in your database table.
    2. If you use UPDATE statement.
    UPDATE <database_table> FROM TABLE <internal_table>.
    This will update the existing records in your database table from the internal table.
    3. If you use MODIFY statement.
    MODIFY <database_table> FROM TABLE <internal_table>.
    This statement works both in combination of INSERT & UPDATE statements.
    Existing records (in database table) will be eventually updated/modified and new records (not in database table currently) will be successfully inserted into the database table.
    Hope this solves your problem.
    Thanks & Regards.
    Tarun Gambhir.

  • Indexes in data dictionary

    Hi all,
                can anyone tell me what is the process of creating a secondary index.?
                can we create a secondary index if a field is assigned with primary key?
                 can we create secondary index for a mara table field?
                when we click on INDEXES in data dictionary ---> database table by selecting a field then we get unique index and non-unquie index, what does that mean?
    regards,
    satish

    What is Secondary Index:
    first i will tell u little about primary index.
    The primary index contains the key fields of the table and a pointer to the non-key fields of the table. The primary index is created automatically when the table is created in the database.
    You can create further indexes on a table in the ABAP Dictionary. These are called secondary indexes.
    PURPOSE:
    This is necessary if the table is frequently accessed in a way that does not take advantage of the sorting of the primary index for the access. Different indexes on the same table are distinguished with a three-place index identifier.
    Generally it is used when field which u want is not a primary field .
    Advatange:
    1. faster access to data when field is not a primary key field.
    Disadvatnage:
    1. Slower down database table inset, update delete commands.
    You go for secondary index when you have any SELECT statement which has a WHERE clause which is not effectively using the existing index(s). But this decision of additional index has to be made in sync with the Basis and Functional team. Every additional index will be a burden on the database and the system resources.
    The secondary index is a type of index in database table. Indexes are used to improve the performances of data retrieval from the table. This happens by passing the values to these indices during SELECT query in the WHERE clause.
    There are different types of indices. Primary key is default present for all the tables in database. Only ONE Primary key can exist in a database table. It results in unique records in the database table.
    The other type of index is secondary index. Multiple fields can be grouped to make a secondary index of table. If all the field values is passed during SELECT query, the performance increases drastically.
    Not many fields should be included in secondary index. Large number of fields reduces the performance of INSERT query of database.
    Indexes are used to speed up data selection from the database. System chosse and use the index based on the fields given in WHERE and HAVING clause.
    The primary index is always created automatically in the R/3 System. It consists of the primary key fields of the database table. This means that for each combination of fields in the index, there is a maximum of one line in the table.
    If you are searching on non-key fields, you can create your own secondary index, which can restrict the number of table entries searched to form the result set.
    Improper use of Secondary index can slow down the performance. Discuss with Basis team before creating a secondary index.
    Check the threds -
    secondary index
    Secondary Index
    Primary index: the primary index contains key fiels of a table and a pointer to non-key fields of the table. The primary index is created automatically when a table is created in database and moreover you can further define reference to the primary index which are known as Secondary index.
    How many indexes can be created for a table? -> 9
    When you are selecting the data on non primary key,obviously system will give output with bad performance ,if you create secondary index then it gives better performance
    GOTO SE11 -> give ur table name -> click on display button -> click on index button ->click on create button -> give id ( z98 like ) -> give field names MANDT ,then ur field names
    save and activate
    Regards
    Vasu

  • Is there a way to remove the index from the XML output?

    Hi,
    It is easy to remove the index from the WebHelp output.
    We also want to remove it from the XML output.
    How do I do that?
    Thanks,
    Rakefet

    @Jeff: In the Single Source Layouts recipe box is a layout for XML. Although, I'm really not sure why folks would use it, other than to possibly export the project data to another system of some sort.
    @Rakefet: One thought is to try creating a new blank Index. Then choose the empty Index when you use XML Output.
    Cheers... Rick

  • Export Index from PDF Portfolio

    I need to compare two large PDF Portfolio documents. From what I can tell, there is currently no third party software around that will do this so I'm trying to figure out a way around it. Each Portfolio that I'm working with has over a thousand email messages exported from Outlook and the corresponding attachments. I know that an "Index" was created when I exported the messages because I can see the pane that lists all the typical details in an index with sortable column headers (i.e., From, Subject, Date, Size, etc...). I'm trying to figure out a way to export only this indexed information into a .csv or other delimited file so I can create unique IDs and compare the documents using Excel. From what I've read, the "Edit" button is one way to try this but for some reason, when Outlook converts messages to a PDF Portfolio, it also removes the ability to use the "Edit" tool. Does anyone know of a way to export just the index information or if there's a hidden index file somewhere? Please let me know.
    Thanks,

    You are correct that the Portfolios created by the mail client plugins are not editable. They use a special layout which does not exist as an editable master layout in Acrobat.
    It's possible to export the rows of data from the 'details' view using JavaScript, but those fields don't contain the body of the message so you will only get the headers (to, from, date, etc.). If this is all you need, follow these instructions:
    Switch to the Cover Sheet view from the View > Portfolio menu
    Open the javaScript console (Ctrl-J) and clear it with the trashcan icon
    Paste in the code which follows, select it all, then press Ctrl+Enter to execute it.
    // code follows
    var d = this.dataObjects;
    var csv = '';
        if (d) {
        cfl = this.collection.fields.length;
        tx_cols = new Array();
        tx_names = new Array();
        for (i = 0; i < cfl; i++) {
          tx_cols.push(this.collection.fields[i].text);
          tx_names.push(this.collection.fields[i].name);
            csv = tx_cols.join() + "\r\n";
            for (var i = 0; i < d.length; i++) {
          tx_data = new Array();
          for (var r = 0; r < cfl; r++) {
            tx_data.push(d[i].getFieldValue(tx_names[r]));
          csv += tx_data.join() + "\r\n";
        this.createDataObject({cName:"EmailData.csv", cValue:csv,cMIMEType:'text/csv'});
        this.exportDataObject("EmailData.csv");
        this.removeDataObject("EmailData.csv");

  • SAX parser splits up character data; I expected Ign. whitesp

    Im am working on a XML parser for loading data from some back
    end systems into an Oracle 8i database. I am using the SAX
    parser for this purpose. After doing some tests with larger
    amounts of XML data (> 1M), I found some unexpected behaviour.
    The parser sometimes splits up character data into two chunks of
    data. The XML looks as follows:
    <TAGNAME> this is the character data </TAGNAME>
    The parser raises the following events:
    1 startElement name = "TAGNAME"
    2 characters chbuf = " "
    3 characters chbuf = "this is the character data "
    4 endElement name = "TAGNAME"
    I expected an ignorableWhitespace event at step 2. The XML
    document contains repetitive tagnames. The strange thing about
    the parse process is that the parser splits up the character
    data only sometimes, and I can't determine any kind of logica
    for this behaviour. Most occurrences of exactly the same tagname
    and character data are parsed correctly (that is, as I
    expected).
    Am I dealing with correct behaviour here, or is it a bug??
    Rolf.
    null

    Oracle XML Team wrote:
    : Rolf van Deursen (guest) wrote:
    : : Im am working on a XML parser for loading data from some
    back
    : : end systems into an Oracle 8i database. I am using the SAX
    : : parser for this purpose. After doing some tests with larger
    : : amounts of XML data (> 1M), I found some unexpected
    behaviour.
    : : The parser sometimes splits up character data into two
    chunks
    : of
    : : data. The XML looks as follows:
    : : <TAGNAME> this is the character data </TAGNAME>
    : : The parser raises the following events:
    : : 1 startElement name = "TAGNAME"
    : : 2 characters chbuf = " "
    : : 3 characters chbuf = "this is the character data "
    : : 4 endElement name = "TAGNAME"
    : : I expected an ignorableWhitespace event at step 2. The XML
    : : document contains repetitive tagnames. The strange thing
    about
    : : the parse process is that the parser splits up the character
    : : data only sometimes, and I can't determine any kind of
    logica
    : : for this behaviour. Most occurrences of exactly the same
    : tagname
    : : and character data are parsed correctly (that is, as I
    : : expected).
    : : Am I dealing with correct behaviour here, or is it a bug??
    : : Rolf.
    : The behavior is expected and correct. Could you elaborate on
    why
    : you would expect the parser to treat the whitespace signalled
    in
    : step 2 as ignorable?
    : Oracle XML Team
    : http://technet.oracle.com
    : Oracle Technology Network
    Thank you for your quick response.
    In my test XML, there are about 27500 tags containing character
    data. All character data starts with a whitespace character.
    After parsing the xml, the whitespace of only 5 (!) tags is
    treated as a seperate character event (so two character events
    are raised in succession). The remaining tags all raise only ONE
    character event for the entire character data. I can't explain
    the difference in treatment.
    null

  • Split Output from Select into 2 different Cols

    Hi,
    I am new to oracle dev. Env Oracle 10g R2
    I need a help with the following.
    I have a select query that returns comma separated values.
    Column LIST1 in table svp_1.
    This needs to hold that data returned by the select query.
    Datatype of Column LIST1 varchar2(4000)
    I get the follwing error.
    ORA-19011: Character string buffer too small
    CREATE OR REPLACE PROCEDURE
    sale_temp( pType in Char,lDate in date) as
    cnt number;
    v_error_code NUMBER;
    v_error_msg VARCHAR2(2000);
    cursor c1 is select svp
    from sale
    where type = pType
    var1 c1%ROWTYPE;
    BEGIN
    UPDATE svp_1
    SET LIST1 = ( SELECT LTRIM( xmlagg (xmlelement (c, RTRIM(RELATED_SVP) || ',')).extract ('//text()'), ',' ) AS RELATED_SVP from (
    SELECT rtrim(BASE_SVP) BASE_SVP,
    rtrim(RELATED_SVP) RELATED_SVP
    from
    svp_enc se
    where
    se.effective date<= pDate
    and (se.expiration_date is null or se.expiration_date > plDate )
    and se.base_svp in
    ( var1.svp
    and se.RELATION_TYPE = '4'
    )group by base_svp)
    WHERE type = pType
    commit;
    dbms_output.put_line('UPDATE done' );
    end;
    end loop;
    close c1;
    END ;
    My requirement :
    In need to capture the data from the select query and update the column.
    The data needs to be returned as comma separated values.
    This data can also be split into 2 columns.....ie data split and updated into 2 cols ie LIST1 AND LIST2 ( can be added to existing table).
    What is the best way to do this?
    Is the below approach possilble?
    LOOP through cursor ...check the length of data....for the select query ....check length .if > 3990 byte.....then write into 2 plsql variables.....and update 2 cols....
    Is this apporach right, if so how to do that in the procedure ? Please give an example...
    Are there any other approaches...?
    All I need is to Split Output from Select into 2 different columns and avoid the ORA-19011: Character string buffer too small error.
    Thanks in advance.

    Re: ORA-19011: Character string buffer too small

  • Function based indexes in Data Modeler 3.0

    Trying to create a function based index in DATA MODELER 3.0 but cannot find a way to do it. When I reverse engineer (import from DATA DICTIONNARY) an existing table that contains such an object, the index appears in the physical model as a regular index. I guess this functionnality is not yet available...
    Thanks.

    Hi,
    You can create a function-based Index by going to the Properties dialog for the Index in the Relational model (not the Physical model). If you select the Index Expression tick box, you can then enter the expression into the Expression text area.
    David

  • Missing IN or OUT parameter at index on Data Modeler

    Hi,
    I'm using Data Modeler 2.0.0 build 570 (with Java 1.5.0 (build 1.5.0_15-b04)).
    Usually I import tables from Data Dictionary (Oracle Database 10g Enterprise Edition Release 10.2.0.3.0) into a relational model, I modify them and then I export relational model informations into a Reporting Schema (different oracle schema inside same db).
    I cannot understand why sometimes, during these exports, I get the error "Missing IN or OUT parameter at index"
    Here an extract of the Data Modeler log file:
    2009-09-15 16:00:36,601 [Thread-6] ERROR ReportsHandler - Error Exporting to Reporting Schema:
    java.sql.SQLException: Missing IN or OUT parameter at index:: 10
         at oracle.jdbc.driver.SQLStateMapping.newSQLException(SQLStateMapping.java:70)
         at oracle.jdbc.driver.DatabaseError.newSQLException(DatabaseError.java:110)
         at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:171)
         at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:227)
         at oracle.jdbc.driver.OraclePreparedStatement.processCompletedBindRow(OraclePreparedStatement.java:1737)
         at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3376)
         at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3462)
         at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeUpdate(OraclePreparedStatementWrapper.java:1061)
         at oracle.dbtools.crest.exports.reports.RSCheckConstraint.export(Unknown Source)
         at oracle.dbtools.crest.exports.reports.RSColumns.export(Unknown Source)
         at oracle.dbtools.crest.exports.reports.RSRelationalModel.export(Unknown Source)
         at oracle.dbtools.crest.exports.reports.ReportsHandler.export(Unknown Source)
         at oracle.dbtools.crest.swingui.ControllerApplication$ExportToReportsSchema$1.run(Unknown Source)
    Does anyone know how to solve this problem?

    Hi,
    the problem is solved in latest published version (build 584). What is your version?
    If you still have issue with build 584 please provide information from log file.
    The constraint is generic if its syntax is not bound to specific database (vendor or version). So if you use syntax introduced in Oracle 10g then you have to define constraint as "Oracle 10g" related and it'll be generated for Oracle 10g and next versions of database.
    Philip

  • I cannot Split Deliv. Date in Sale Order

    Hi expert
    Can anyone give some suggestions.
    - Sale order is MTO and have production order that generated from sale order but production order was closed by user.
    - User create new production order by tcode CO08.
    Users want to change Deliv. Date in Sale Order by refer to new production order but cannot change.
    I cannot edit and split Deliv. date because date of old production order and ATP check not found new split item.
    regards,

    Hi, Allen Qiu2010 
    Thanks for your suggestion, after check at shipping tab than Partial delivery per item is blank.
    After change type, i can split  new delivery date but it refer to the old prod.ord, which was closed.
    Clould you tell me about new item which split delivery?
    How to refer between new Deliv.date and prod.ord. that create by CO08?
       -  first item of Deliv.date, and confirmed dty. have 40 PC (refer to old prod.ord which was closed)
         ->i checked material on MB52 don't have stock of this sale order . (sale order is make to order)
         -> prod. ord which create by CO08 unfinished process.
    Sorry, if i wrong understand.
    regrads,

Maybe you are looking for

  • 2 ALVs in a single view

    How to use 2 ALVs in a Single view. I need to take refernce of them individually, so that i can confgure them, independant of each other Thanks, Selvakumar M.

  • Can't Print

    I recently upgraded to Tiger and when I try to print a calendar (using iCal 2.0.1) on my HP OfficeJet 6110, only two lines print then the paper ejects. I've even tried printing from Preview and saving as a PDF and get the same results. No other appli

  • Already bought iWork. How do I get the App Store to recognize App purchases that I've made over the past few years?

    I know I'm not the only person experiencing this issue, but I can't find who else has a solution. Essentially, I've been buying Mac Apps over the past few years and using them frequently. Now the App Store exists. Awesome. But how do I get the App St

  • How to implement custom methods on form submit

    I'm a newbie to ADF environment, coming from pl/sql world. I have already tried default drag and drop functionality for forms. But, with new requirements, I need to write some custom code to create and update form. I read about managed beans, but did

  • Can i add data for current month only

    can I add data for current month only?