Partitioning of Data

Hi,
I need help regarding Partitioning of data.
1. What is Partitioning?
2. How it is going to help in handling the data?
3. What is Physical and Logical Partitioning?
4. If we do Partitioning, we will view the data in different Stages? If so how can we view the data and where?
5. Is Partitioning is going to help the Performance Issue?
6. If we do Partitioning of data is it going to help in Reporting and how?
7. I came to know that Partitioning will be done base on Calendar Year and Fiscal Year?
8.  What is Calendar and Fiscal Year? How there are useful to us?
9. Which is the better Option for Partitioning (Calendar of Fiscal Year)? Can you provide me detail information of best option?
10. How to do the Partitioning and what are the steps involved in doing the Partitioning? Can any one give me step - by- step?
I need to answer all this question to my Client and i need to prepare the document with the best options on the Partitioning.
Thanks & Regards
Jagannadha Raju

Hi Raju,
search forums with these topics you will find many posts answered by Experts
check these links,
http://help.sap.com/saphelp_nw04s/helpdata/en/33/dc2038aa3bcd23e10000009b38f8cf/content.htm
http://help.sap.com/saphelp_nw04/helpdata/en/e0/b60b404b2b1e07e10000000a1550b0/content.htm
https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/cccad390-0201-0010-5093-fd9ec8157802
partitioning
Hope this helps,
Sudhakar.

Similar Messages

  • Is there a way of partitioning the data in the cubes

    Hello BPC Experts,
    we are currently running an Appset with 4 Applciations. Anyway two of these are getting really big.
    In BPC for MS there is a way to partitioning the data as I saw in the How tos.
    In NW Versions the BPC queries the Multiprovider. Is there a way to split the underlying Basis Cube to several (split by time or Legal Entity).
    I think this would help to increase the speed a lot as data could be read in parallel.
    Help is very much appreciated.
    Daniel
    Edited by: Daniel Schäfer on Feb 12, 2010 2:16 PM

    Hi Daniel,
    The short answer to your question is that, no, there is not a way to manually partition the infocubes at the BW level. The longer answer comes in several parts:
    1. BW automatically partitions the underlying database tables for BPC cubes based on request ID, depending on the BW setting for the cube and the underlying database.
    2. BW InfoCubes are very different from MS SQL server cubes (ROLAP approach in BW vs. MOLAP approach usually used in Analysis Services cubes). This results in BW cubes being a lot smaller, reads and writes being highly parallel, and no need for a large rollup operation if the underlying data changes. In other words, you probably wouldn't gain much from semantic partitioning of the BW cubes underlying BPC, except possibly in query performance, and only then if you have very high data volumes (>100 million records).
    3. BWA is an option for very large cubes. It is expensive, but if you are talking 100s of millions of records you should probably consider it. It uses a completely different data model than ROLAP or MOLAP and it is highly partition-able, though this is transparent to the BW administrator.
    4. In some circumstances it is useful to partition BW cubes. In the BW world, this is usually called "semantic partitioning". For example, you might want to partition cubes by company, time, or category. In BW this is currently supported through manually creating several basic cubes under a multiprovider. In BPC, this approach is not supported. It is highly recommended to not change the BPC-generated Infocubes or Queries in any way.
    5. If you have determined that you really need to semantically partition to manage data volumes in BPC, the current best way is probably to have multiple BPC applications with identical dimensions. In other words, partition in the application layer instead of in the data layer.
    Hopefully that's helpful to you.
    Ethan

  • Importing partitioned table data into non-partitioned table

    Hi Friends,
    SOURCE SERVER
    OS:Linux
    Database Version:10.2.0.2.0
    i have exported one partition of my partitioned table like below..
    expdp system/manager DIRECTORY=DIR4 DUMPFILE=mapping.dmp LOGFILE=mapping_exp.log TABLES=MAPPING.MAPPING:DATASET_NAPTARGET SERVER
    OS:Linux
    Database Version:10.2.0.4.0
    Now when i am importing into another server i am getting below error
    Import: Release 10.2.0.4.0 - 64bit Production on Tuesday, 17 January, 2012 11:22:32
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "MAPPING"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
    Starting "MAPPING"."SYS_IMPORT_FULL_01":  MAPPING/******** DIRECTORY=DIR3 DUMPFILE=mapping.dmp LOGFILE=mapping_imp.log TABLE_EXISTS_ACTION=APPEND
    Processing object type TABLE_EXPORT/TABLE/TABLE
    ORA-39083: Object type TABLE failed to create with error:
    ORA-00959: tablespace 'MAPPING_ABC' does not exist
    Failing sql is:
    CREATE TABLE "MAPPING"."MAPPING" ("SAP_ID" NUMBER(38,0) NOT NULL ENABLE, "TG_ID" NUMBER(38,0) NOT NULL ENABLE, "TT_ID" NUMBER(38,0) NOT NULL ENABLE, "PARENT_CT_ID" NUMBER(38,0), "MAPPINGTIME" TIMESTAMP (6) WITH TIME ZONE NOT NULL ENABLE, "CLASS" NUMBER(38,0) NOT NULL ENABLE, "TYPE" NUMBER(38,0) NOT NULL ENABLE, "ID" NUMBER(38,0) NOT NULL ENABLE, "UREID"
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_TG_ID" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."PK_MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_UREID" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_V2" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_PARENT_CT" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    ORA-39112: Dependent object type CONSTRAINT:"MAPPING"."CKC_SMAPPING_MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"MAPPING"."PK_MAPPING_ITM" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_TG_ID" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."PK_MAPPING" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_UREID" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_V2" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_PARENT_CT" creation failed
    Processing object type TABLE_EXPORT/TABLE/COMMENT
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/REF_CONSTRAINT
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_MAPPING_MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_MAPPING_CT" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_TG" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_TT" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_PART" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_TIME_T" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_DAY" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_BTMP" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_TG_ID" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_V2_T" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."PK_MAPPING" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_PARENT_CT" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_UREID" creation failed
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    ORA-39112: Dependent object type TABLE_STATISTICS skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Job "MAPPING"."SYS_IMPORT_FULL_01" completed with 52 error(s) at 11:22:39Please help..!!
    Regards
    Umesh Gupta

    yes, i have tried that option as well.
    but when i write one tablespace name in REMAP_TABLESPACE clause, it gives error for second one.. n if i include 1st and 2nd tablespace it will give error for 3rd one..
    one option, what i know write all tablespace name in REMAP_TABLESPACE, but that too lengthy process..is there any other way possible????
    Regards
    UmeshAFAIK the option you have is what i recommend you ... through it is lengthy :-(
    Wait for some EXPERT and GURU's review on this issue .........
    Good luck ....
    --neeraj                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • How to use a 2nd partition for data storage?

    The 40 GB hard disk on my satellite L10 was delivered with two 20 GB partitions (drive C and D). I now understand (by reading this forum) that it is preferred to use one partition for data storage.
    Until now there was no need to use the D-drive. But by now I need the extra space.
    How can I use the folders structure in windows XP home (My Documents etc) and put the data on the D-drive.
    We work on this computer with several accounts.
    I tried to move one folder out of My Documents to the D-drive, but then it can be reached by the other users. Should I use shortcuts towards the D-drive? I hope there is a simple solution.
    Thanks in advance.
    (I tried my best to write this in English....)

    Hi
    I didnt tried it before but you can try to make your personal folder on D partition. The notebook administrator should set the security settings that can be found under folder properties. Go to Security tab and check all permissions for other users. Choose Deny option for all of them.
    I believe that on this way you can use your data folder alone and be sure that other users can not read or execute any of your private data.
    Please let me know if it works properly.
    Bye

  • Partition by Date and PK

    I am designing a new laboratory database.
    My primary data tables will have at least <tt>id (PK NUMBER)</tt> and <tt>created_on (DATE)</tt>. Also, for any two entries, the entry with a higher <tt>id</tt> will have a later <tt>created_on</tt> date.
    I plan to partition by <tt>created_on</tt> to increase performance on recently entered data. Since the columns increase together, the table would also be partitioned by <tt>id</tt>, implicitly. However, Oracle would not know about the implied partitioning by <tt>id</tt> to take advantage of the partitioning of table joins on <tt>id</tt>.
    Two questions:
    1. How do I enforce both columns increasing together?
    2. How can I take advantage of this implicit partitioning for table joins?

    1. How do I enforce both columns increasing together?With a trigger on insert?
    2. How can I take advantage of this implicit partitioning for table joins?Not to worry, PK will allways be GLOBAL index, and fall (when seqential read) into correct partition of date(created_on) performance-wise.
    All other indexes you create LOCAL.
    :p

  • Does hash partition distribute data evenly across partitions?

    As per Oracle documentation, it is mentioned that hash partitioning uses oracle hashing algorithms to assign a hash value to each rows partitioning key and place it in the appropriate partition. And the data will be evenly distributed across the partitions. Ofcourse following following conditions :
    1. Partition count should follow 2^n logic
    2. Data in partition key column should have high cardinality.
    I have used hash partitioning in some of our application tables, but data isn't distributed evenly across partitions. To verify it, i performed a small test :
    Table script :
    Create table ch_acct_mast_hash(
    Cod_acct_no number)
    Partition by hash(cod_acct_no)
    PARTITIONS 128;
    Data population script :
    declare
    i number;
    l number;
    begin
    i := 1000000000000000;
    for l in 1 .. 100000 loop
    insert into ch_acct_mast_hash values (i);
    i := i + 1;
    end loop;
    commit;
    end;
    Row-count check :
    select count(1) from Ch_Acct_Mast_hash ; --rowcount is 100000
    Gather stats script :
    begin
    dbms_stats.gather_table_stats('C43HDEV', 'CH_ACCT_MAST_HASH');
    end;
    Data distribution check :
    Select min(num_rows), max(num_rows) from dba_tab_partitions
    where table_name = 'CH_ACCT_MAST_HASH';
    Result is :
    min(num_rows) = 700
    max(num_rows) = 853
    As per the result, it seems there is lot of skewness in data distribution across partitions. Maybe I am missing something, or something is not right.
    Can anybody help me to understand this behavior?
    Edited by: Kshitij Kasliwal on Nov 2, 2012 4:49 AM

    >
    I have used hash partitioning in some of our application tables, but data isn't distributed evenly across partitions.
    >
    All keys with the same data value will also have the same hash value and so will be in the same partition.
    So the actual hash distribution in any particular case will depend on the actual data distribution. And, as Iordan showed, the data distribution depends not only on cardinality but on the standard deviation of the key values.
    To use a shorter version of that examle consider these data samples which each have 10 values. There is a calculator here
    http://easycalculation.com/statistics/standard-deviation.php
    0,1,0,2,0,3,0,4,0,5 - total 10, distinct 6, %distinct 60, mean 1.5, stan deviation 1.9, variance 3.6 - similar to Iordan's example
    0,5,0,5,0,5,0,5,0,5 - total 10, distinct 2, %distinct 20, mean 2.5, stan dev. 2.64, variance 6.9
    5,5,5,5,5,5,5,5,5,5 - total 10, distinct 1, %distinct 10, mean 5, stan dev. 0, variance 0
    0,1,2,3,4,5,6,7,8,9 - total 10, distinct 10, %distinct 100, mean 4.5, stan dev. 3.03, variance 9.2
    The first and last examples have the highest cardinality but only the last has unique values (i.e. 100% distinct).
    Note that the first example is lower for all other attributes but that doesn't mean it would hash more evenly.
    Also note that the last example, the unique values, has the highest variance.
    So this is no single attribute that is controlling. As Iordan showed the first example has a high %distinct but all of those '0' values will hash to the same partition so even using a perfect hash the data would use 6 partitions.

  • Block level backup engine Samsung SSD 840 EVO partition labeled "data"

    Replaced the hard drive in my Dell 1545 laptop with a Samsung SSD 840 EVO.
    Everything beautiful.
    Am running Windows 8.1 - again everything beautiful.
    Thought I would run "System Image Backup" (off the File History page).
    failed - Microsoft error code 0x80780119
     because Samsung Migration tool created a first partition labeled "data"
    (don't know whose data this might be suspect it has nothing to do with the OS but is required for the SSD, perhaps for MBR boot purposed.)
    This "data" partition is very small and 99% full: no room for the VSS backup requires.
    Bypassed that problem with:
    PS C:\> wbAdmin start backup -backupTarget:H: -include:C:\
    which runs perfectly, except for the warning,
    " Note: the list of volumes included for backup does not include all the volumes that contain operating system components. This backup cannot be used to perform a system recovery. However. you can recover other items if the destination media type Supports
    it."

    Hi,
    Glad to hear that you've found a workground to backup the data and thanks for sharing the experience here. Regarding to the Samsung "data" partition, yes, as you mentioned, it is related with Samsung Migration software, so I think members working
    on Samsung will be familiar with this partition.
    http://www.samsung.com/in/support/category/storage/solidstatedrive/
    NOTE
    This response contains a reference to a third party World Wide Web site.
    Microsoft is providing this information as a convenience to you. Microsoft does not control
    these sites and has not tested any software or information found on these sites.
    Yolanda Zhu
    TechNet Community Support

  • Format my partition without data loss

    Can you create new partitions once it has been formatted and the operating system loaded?
    Lastly, I looked for the solutions from internet, I found MiniAide Magic Partition can help me. It can create new partitions and format partitions without data loss. It totally helped me.
    MIniaide Magic Partition Tool
    Share it here and hope it can also help you manage your partitions/disks.

    Windows has a capability to paritions without the loss of data. Refer the following weblinks:
    http://www.sevenforums.com/tutorials/2672-partitio​n-volume-shrink.html
    http://www.howtogeek.com/howto/windows-vista/resiz​e-a-partition-for-free-in-windows-vista/
    I am an HP employee.
    Regards,
    Vidya
    Make it easier for other people to find solutions, by marking my answer “Accept as Solution” if it solves your problem.
    ***Click on "Thumbs up" button to the bottom right side of my post to say thanks!***

  • Easy to install Lion with BootCamp + Win 8 + NTFS partition for data?

    Hey, I have a MacBook Pro (2011) currently on OS X 10.6.8, but I bought Lion 10.7 and want to use that.
    Here's what I want to do:
    * install Lion (fresh, not upgrade)
    * install BootCamp
    * install Win 8 (DevEd) (for games and some other apps)
    And here are my questions:
    1) What's the best (free) way to save/"ghost" the current OS so I can quickly get back to that in case something happens when I install Lion?
    2) What order do I install BootCamp and Win 8 in? Any fool-proof tutorial for how to do this?
    3) If I do a backup with Time Machine as the last thing I do before installing Lion, I just copy the data back from the TM drive after install, right?
    4) I would like to use Win 8 and some programs there to read and write to the HFS+ OS X partition - is there a free and/or good way to do that?
    5) On PC I can use WinAmp, the great plugin "DeleteCurrentSong" and a shortcut key to delete the currently playing track. Is there a way to do something similar in any program in OS X?
    6) If not, would it be possible to - early in the install process - make a big NTFS partition and put most of my data there so it can be accessed from both OS X (with Paragon) and Windows? And Time Machine wouldn't back it up as long as it's NTFS, right?
    Thanks a lot for any answers!

    Answer to question 1:
    a. Get a USB disc the same size or better larger than your internal HD
    b. Download Carbon Copy Cloner (free)
    c. Clone the internal HD to the external disc
    d. Reboot with the 'option' key held down. Choose the external disc to check that the clone was successful and is bootable.
    e. If yes, reboot again, hodling down the 'option' key. Choose the internal HDD.
    f. Dismount/eject the external HD and disconnect it from your computer.
    g. Go to app store and download Lion.
    h. When the download is complete, DO NOT run it. First, copy the file (Applications > Lion Install.app) to another location on your disc (the downloads folder will do). This is because after you run the installer it will delete itself from the Applications folder, and you may very well either want/need to do a re-install later without wanting to wait for hours for the download.
    i. Run Applications > Lion Install.app

  • Merge o/s partition with data partition on macbook pro, osx 10.6.8

    hi there
    i have a macbook pro running snow leopard, 10.6.8. the internal hard drive is divided into three partitions, #1 is osx, #2 is also osx, and was meant to be another osx to boot into, #3 is a data drive.
    i never use the 2nd partition to boot into, and would like to trash it and merge it with #1. that is, make the main osx partition larger.
    can i do this without having to reinstall the o/s?
    cheers

    Depends on how the partitons are arranged. If they're exactly as quoted, then yes.
    Open Disk Utility and select the partition tab.
    In the diagram, the top partition needs to be the one you're booted from, and partition 2 the middle one.
    Select Partiton 2 so that it's highlighted, and click the '-' button below the diagram.
    Drag the bottom right corner of partition 1 down to meet partition 3.

  • How to Implement 30 days Range Partitioning with Date column. Not Interval

    Hi,
    I am using the db:
    Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit
    Current table structure is:
    CREATE TABLE A
    a NUMBER,
    CreationDate DATE
    PARTITION BY RANGE (CreationDate)
    INTERVAL ( NUMTODSINTERVAL (30, 'DAY') )
    (PARTITION P_FIRST
    VALUES LESS THAN (TIMESTAMP ' 2001-01-01 00:00:00'))
    How can I define virtual column based partitioning with RANGE partitioning without using INTERVAL partitioning.
    And that is with Intervals of 30 days.
    For monthly I am trying as
    CREATE TABLE A
    a NUMBER,
    CreationDate DATE,
    monthly_interval date as (to_char(CreationDate,'MM-YYYY')) VIRTUAL
    PARTITION BY RANGE (monthly_interval)
    partition p_AUG12 values less than (to_date('08-2012','mm-yyyy')),
    partition p_SEP12 values less than (to_date('09-2012','mm-yyyy')),
    partition p_OCT12 values less than (to_date('10-2012','mm-yyyy'))
    Enable ROw Movement
    BUT CAN'T INSERT the data even for that:
    Insert into a (a, CreationDate)
    Values (1, '12-10-2012')
    Insert into a (a, CreationDate)
    Values (1, '12-10-2012')
    Please suggest..

    Hi rp,
    Interval Partitioned to Range. Created Daily Partitions from Monthly Part. got complicated so I am posting here.
    Basically,
    I know Interval Partitioning is a kind of Range partitioning. But explicitly for Interval Partitioned tables XML Indexes are not allowed as discussed here:
    XMLIndexes on an Interval Partitioned Table??
    I can do monthly partitions as :
    CREATE TABLE A
    a NUMBER,
    CreationDate DATE,
    monthly_interval varchar2(8) as (to_char(CreationDate,'MM-YYYY')) VIRTUAL
    PARTITION BY RANGE (monthly_interval)
    partition p_AUG12 values less than ('09-2012'),
    partition p_SEP12 values less than ('10-2012'),
    partition p_OCT12 values less than ('11-2012')
    ) Enable ROw Movement
    Insert into a (a, CreationDate)
    Values (1, '12-SEP-2012')
    Insert into a (a, CreationDate)
    Values (1, '14-SEP-2012')
    Select * from A partition (p_SEP12)
    Select * from A partition (p_AUG12)
    Select * from A partition (p_OCT12)
    Can we do it for 30 days partitions, instead of the monthly partitions. ANY suggestions..
    Thanks..

  • How to partition the data dynamically

    Hi,
    I've the data which goes on increasingly day by day. So, for the sake of performance i wanted to go for partitioning. i want to divide the data into partitions by months. How can i acheive this one. I do not want to hard code it to the year like 2007 or 2008 like this.The problem with the hard coding is again i've to make it to 2008 in the next year.
    Regards,
    Alok Dubey

    Long ago have tried something like that:
    Create Procedure:
    BEGIN
    1. looking for you date
    2. If date … then
    3. EXECUTE EMEDIATE ‘create partition ….’
    EXECUTE EMEDIATE ‘create partition on idex….’
    4. Drop old partition if needed
    END
    Setup scheduler for you procedure

  • Materialized view on a Partitioned Table (Data through Exchange Partition)

    Hello,
    We have a scenario to create a MV on a Partitioned Table which get data through Exchange Partition strategy. Obviously, with exchange partition snap shot logs are not updated and FAST refreshes are not possible. Also, this partitioned table being a 450 million rows, COMPLETE refresh is not an option for us.
    I would like to know the alternatives for this approach,
    Any suggestions would be appreciated,
    thank you

    From your post it seems that you are trying to create a fast refresh view (as you are creating MV log). There are limitations on fast refresh which is documented in the oracle documentation.
    http://docs.oracle.com/cd/B28359_01/server.111/b28313/basicmv.htm#i1007028
    If you are not planning to do a fast refresh then as already mentioned by Solomon it is a valid approach used in multiple scenarios.
    Thanks,
    Jayadeep

  • Large data sets and table partitioning : removing data

    Hi,
    I have to delete lines into a big table using partition.
    Someone says me it is more efficient for the performances to drop a whole partition (or subpartition if I use a composite partition) than to delete a line at the same time.
    He says me that data access (in my partition) should be very bad if I delete lines progressively (in this partition) instead of keeping the lines and deleting the whole partition when all its lines are no more used.
    What do you think about it?
    Thanks
    Sandrine

    Hi Sandrine,
    I agree with what you're being told. I'll be much more efficient to "clone" the data you want to keep from your patition somewhere (clone table, ...) and then drop the whole partition.
    Main thing is if you drop an object there's no BEFORE IMAGE stored in your UNDO structures (UNDO TS / RBS). SO you'll have way less disk I/Os.
    Hope this helps ^^

  • Partition Key Date Format

    Hi All
    We have table which is range partitioned on a Date field with the following date format
    PARTITION "P200905" VALUES LESS THAN (TO_DATE(' 2008-08-03 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN')
    But whenever the user is inserting the data he is using this to_date format
    TO_DATE(?,'YYYY-MM-DD HH24:MI:SS')
    Here while inserting the date he is not using the format which we used to create the partition I mean 'SYYYY-MM-DD HH24:MI:SS' this format instead the user is using 'YYYY-MM-DD HH24:MI:SS' is this will cause any error to insert data into the particular partitions ?
    Particular error we are getting is the
    Original thrown object message: java.sql.SQLException: ORA-14400: inserted partition key does not map to any partition

    Hi Neel
    You can use conversion routines for the display purpose
    For that particular keyfigure, you have to give a suitable conversion routine and activate the infoobject.....
    http://help.sap.com/saphelp_nw70/helpdata/en/3a/b9c03aa187de41a781d9afad665ce2/frameset.htm
    Refer other routines also
    Further you can hide first two digits using macros
    Regards
    N Ganesh

Maybe you are looking for

  • FRM-47023: NO SUCH PARAMETER NAMED SESSIONID EXISTS

    Hi all, I'm getting FRM-47023: NO SUCH PARAMETER NAMED SESSIONID EXISTS while doing OM Superuser > Enter Orders. This issue is specific when we login to EBS in Dutch and does not appear in English! Language specifi error - FRM-47023: parameter met de

  • How to maintain format in html email using the send email....

    Does anyone know how to maintain the format in a HTML email using the Send E-mail To Recipients From Recordset behavior? When text only is selected instead of HTML text on the Options tab the format of the input is maintained, but when HTML text is s

  • Positioning

    i am adding some animations to keynote. i would like the animations to come from the left, or the top. here is my problem : i can't seem to get the slide to "detach" from the upper-left side of the screen, when i am zoomed out. i have looked everywhe

  • "Syntax error, insert "AssignmentOperator Expression" to complete ForInit"?

    I am getting this unknown error and do not know how to solve it. here is my code: it gives me the error for the first part of both for loops. int THISHOLDSTHEVALUEOFTHENUMBEROFTIMESTOLOOPINTHEFIRSTLOOP = 3;           int THISHOLDSTHEVALUEOFTHENUMBERO

  • Delimited objects displayed in RH_STRUC_GET

    Hi,   I am calling RH_STRUC_GET for a org unit ad trying to get all org units under it. The issue is its getting even delimited org units even though I have specified today's date as begda and endda. I assumed the begda and endda were to display org