Logical Partition of infocube

We have a cube logical partitioned by sales org.  We now would like to add additional sales org's to the list. Is this possible??  Currently with data in the cube it is blocked.  Does anyone know steps to re-logical partition a cube??  With/or without data re-loading?  Any options at the database level (oracle)?
Thanks!!

Hi,
Yes that is correct you cant do cube partitioning with existing data.
Another important thing is based upon selection conditions only you can do the partitioning.  There are two selection conditions they are as follows.
               1. 0CALMONTH
               2. 0FISCAL YEAR PERIOD
if these two time characteristics were not used in the cube means you can't partition the cube. 
STEPS TO PARTITION THE INFOCUBE:
for example if you have Sales and Distribution cube you want to do logical partitioning on that cube. 
we have to create infocube with same structure ( this we can do by giving the technical name of the cube in copy from option) at the time of creation of cube. Activate the cube and come back.
then you should create update rules for that cube.
select the first cube right click on it choose the option generate export datasource.
the system will finish generation process.
select the infosource option on the left side of the AWB.
in that we have DATAMART select it choose refresh icon. 
you get the infosource with datasource assignment.
select the data source right click on it choose create infopackage.
it will take you to the maintain infopackage screen. 
here you need to select the datatarget tab and select the target into which the data need to be updated.
schedule it and start.
now manage the data target and check the data has been updated or not.
after doing all this process go to infoprovider select the first cube delete the data.
now depending upon your requirement do the partitioning.
double click on the cube it will take you to the edit infocube screen from the menu bar select extras it will display the partitioning option select it.
then it will popup one small wizard showing the time characteristics based upon your requirement choose which ever you want check that box and continue the process.
Hope this will help you
Thanks and Regards
Vara Prasad

Similar Messages

  • DSO Logical Partitioning

    I've seen lots of discussions on Logical Partitioning of InfoCubes, but not a lot on Logical Partitioning of DSO Objects. 
    Can anyone talk to Logical Partitioning of DSO's or point me to some examples?
    Thanks

    Kenneth,
    Logical partitioning refers to splitting the data into smaller logical chunks by way of design ... for example having two cubes with one cube for data for 2008 and one for 2009 would be an example of logical partitioning.
    Similar approaches can be taken for DSO's as well - have separate DSOs one for each country... things like that.
    Also please note that this is not the same as physical partitioning - which refers to the usual cube compression etc....

  • 0IC_C03 related Inventory Process - Logical Partitioning (Vs) Physical Part

    Hello Everyone,
    After going through multiple postings throughout the form and documentation from SAP, it states that the 0IC_C03 InfoCube when used with Non Cumulative keyfigures is not recommended to be partitioned logically by physical year/calendar year as the query will read all the data sets due to the stock marker logic.
    In our specific scenario,
    1. After the InfoCube (0IC_C03) was enhanced with additional characterisitcs such as Doc Number, Movement type and so on due to business requirements I was not able to actually use the Non Cumulative Keyfigures as they were not populated within the report.
    2. So, we decided not to use the Non Cumulative keyfigures but rather create two cumulative keyfigures (Issue Stock Quantity - Receipt Stock Quantity) and (Issue Valuated Stock Value - Receipt Valuated Stock Value) and both of these are available in the InfoCube and are calculated during the update process.
    These two keyfigures are cumulative with exception aggregation of LAST based on 0CALDAY.
    The question is,
    Since we are not using the actual Non Cumulative Keyfigures (even though we are not using these, we still have the stock marker updated and data compressed based on this along with Validity table defined) can we do logical partitioning on the InfoCube based on Calendar year.
    Thanks....

    Hello Elango,
    Appreciate your response.
    First of..I do understand the difference between logical and physical partitioning and the question is not about joining them together.
    I am sorry, if others cannot understand the detailed issue posted. My apologies was a part of polite gesture, and please do respond back with proper precise answer if you think you did actually understand the question....
    The question here is about how I can leverage the performance and administrative performance by logically breakingdown the data.
    The issues due to which I am trying to look into different aspects of logical partitioning are:
    1. If I do logical partitioning by Plant due to the stock marker logic then I cannot do archiving as a Plant and its related data cannot be archived by time characteristic as the partitioning is not done by time characteristic.
    2. The reason I would have to have document number and movement type in the InfoCube is due to the kind of reporting users perform.
    We have a third party system whose data needs to be reconciled to the data in the plants and storage locations.
    And in order to do so, the first step users would be running the report is plant, storage location and sku. From here on for the storage locations which have balance they would like to drill down on to the document number and movement type to see what the actual activity is.
    So, to support this requirement I would have to have the above characterisitcs in the InfoCube.
    The question again is,.....what is the exact list of issues I would be having doing the logical partitioning by time characteristic.
    Once again, even though the non cumulative keyfigures are available in the InfoCube we are not using them for any reporting purpose....so please keep that in consideration while replying back.
    Thanks
    Dharma.

  • Physical Vs Logical Partitioning

    We have 2 million records in the sales infocube for 3 years. We are currently discussing the pros and cons of using Logical partitioning Vs Physical Partitioning. Please give your inputs.

    hi
    there are two types of partitioning generally talked about with SAP BW, logical and physical partitioning.
    Logical partitioning - instead of having all your data in a single cube, you might break into separate cubes, with each cube holding aspecific year's data, e.g. you could have 5 sales cubes, one for each year 2001 thru 2005.
    You would then create a Multi-Provider that allowed you to query all of them together.
    A query that needs data from all 5 years would then automatically (you can control this) be split into 5 separate queries, one against each cube, running at the same time. The system automatically merges the results from the 5 queries into a single result set.
    So it's easy to see when this could be a benefit. If your queries however are primarily run just for a single year, then you don't receive the benefit of the parallel processing. In non-Oracle DBs, splitting the data like this may still be a benefit by reducing the amount of rows in the fact table that must be read, but does not provide as much value to an Oracle DB since Infocube queries are using a Star_Transformation.
    Physical Partitioning - I believe only Oracle and Informix currently support Range partitioning. This is a separately licensed option in Oracle.
    Physical partitioning allows you to split an Infocube into smaller pieces. The pieces, or partitions, can only be created by 0FISCPER or 0CALMONTH for an InfoCube (ODSs can be partitioned, but require a DBAs involvement). The DB can then take advantage of this partitioning by "pruning" partitions during a query, e.g. a query only needs data form June 2005
    The DB is smart enough to restrict the indices and data it will read to the June 2005 partition. This assumes your query restricts/filters on the partitioning characteristic. It can apply this pruning to a range of partitions as well, e.g. 0FISCPER 001/2005 thru 003/2005 would only look at the 3 partitions.
    It is NOTsmart enough, however, to figure out that if your restrict to 0FISCYEAR = 2005, that it should only read 000/2005 thru 016/2005 since 0FISCYEAR is NOT the partitioning characteristic.
    An InfoCube MUST be empty in order to physically partition it. At this time, there is no way to add additional partitions thru AWB, so you want to make sure that you create partitions out into the future for at least a of couple of years.
    If the base cube is partitioned, any aggregates that contain the partitioning characteristic (0CALMONTH or 0FISCPER) will automatically be partitioned.
    In summary, you need to figure out if you want to use physical or logical partitioning on the cube(s), or both, as they are not mutually exclusive.
    So you would need to know how the data will be queried, and the volume of data. It would make little sense to partition cubes that will not be very large.
    physical partitioning is done at database level and logical partitioning done at data target level.
    Cube partitioning with time characteristics 0calmonth Fiscper is physical partitioning.
    Logical partitioning is u partition ur cube by year or month that is u divide the cube into different cubes and create a multiprovider on top of it.
    logical Vs physical partitions ?

  • Logical Partition.

    Hello All,
    I have a lil knowledge about logical partition. Could you please send me some document related to logical partition?
    Where and how it can be used in BW?
    Thanks.

    Hi ,
    1.Partitioning the cube in both physical and logical way helps you to increase its performance on Queriying.
    Partitioning InfoCubes Using the Characteristic 0FISCPER (At Infocube maintainance)
    Prerequisites
    When partitioning using 0FISCPER values, values are calculated within the partitioning interval that you specified in the InfoCube maintenance. To do this, the value for 0FISCVARNT must be known at the time of partitioning; it must be set to constant.
    Procedure
    1. The InfoCube maintenance is displayed. Set the value for the 0FISCVARNT characteristic to constant. Carry out the following steps:
    a. Choose the Time Characteristics tab page.
    b. In the context menu of the dimension folder, choose Object specific InfoObject properties.
    c. Specify a constant for the characteristic 0FISCVARNT. Choose Continue.
    2. Choose Extras -->DB Performance --> Partitioning. The Determine Partitioning Conditions dialog box appears. You can now select the 0FISCPER characteristic under Slctn. Choose Continue.
    3. The Value Range (Partitioning Condition) dialog box appears. Enter the required data.
    Pls chk the below link for partioning..
    http://help.sap.com/saphelp_bw33/helpdata/en/33/dc2038aa3bcd23e10000009b38f8cf/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/0a/cd6e3a30aac013e10000000a114084/frameset.htm
    Re: logical Vs physical partitions ?
    Regarding Partitioning
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    /message/4204952#4204952 [original link is broken]
    How i can partioning the  BW?
    Partioning and ETable
    What is the use of cube partition?
    Re: logical Vs physical partitions ?
    *Hope this helps*
    Regards
    CSM reddy

  • Logical Partition of Cube

    Hi BW experts,
    Can anyone explain the steps to create logical partitioning of Cube data?
    I want to do it per fiscal year. If we create 05 cubes(Same structure) each for say fiscal year 04' 05'...08'
    now how will the data flow into respective cubes.
    Kindly explain the steps.
    thanks,
    Nipun Sharma
    P.S. points will be rewarded.

    Hi Nipun,
    1.Partitioning the cube in both physical and logical way helps you to increase its performance on Queriying.
    Partitioning InfoCubes Using the Characteristic 0FISCPER (At Infocube maintainance)
    Prerequisites
    When partitioning using 0FISCPER values, values are calculated within the partitioning interval that you specified in the InfoCube maintenance. To do this, the value for 0FISCVARNT must be known at the time of partitioning; it must be set to constant.
    Procedure
    1. The InfoCube maintenance is displayed. Set the value for the 0FISCVARNT characteristic to constant. Carry out the following steps:
    a. Choose the Time Characteristics tab page.
    b. In the context menu of the dimension folder, choose Object specific InfoObject properties.
    c. Specify a constant for the characteristic 0FISCVARNT. Choose Continue.
    2. Choose Extras -->DB Performance --> Partitioning. The Determine Partitioning Conditions dialog box appears. You can now select the 0FISCPER characteristic under Slctn. Choose Continue.
    3. The Value Range (Partitioning Condition) dialog box appears. Enter the required data.
    Pls chk the below link for partioning..
    http://help.sap.com/saphelp_bw33/helpdata/en/33/dc2038aa3bcd23e10000009b38f8cf/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/0a/cd6e3a30aac013e10000000a114084/frameset.htm
    Re: logical Vs physical partitions ?
    Regarding Partitioning
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    /message/4204952#4204952 [original link is broken]
    How i can partioning the  BW?
    Partioning and ETable
    What is the use of cube partition?
    *pls assign points,if info is useful*
    Regards
    CSM reddy

  • Any real reason for logical partitioning over physical?

    Hi!
    I have seen a number of scenarios where SAP BI (assuming BI 7.0 for the rest of the discussion), running in high volume scenarios, have been cluttered by a lot of logically partitioned cubes joined by multi providers....
    Obviously the disadvantage of using logical partitions is that it increases maintenance efforts: you need a new update rule for each logical partition (cube) , then you need to manually add/delte cubes from the multiprovider, filtering data in the update rules to reach the correct cube based on time characteristic etc etc...
    I have seen one clear advantage which is the parallelization of queries run against a multiprovider - assuming you wan't to all underlying cubes ... but are there any other advantages which overcome the mainenance overhead?
    For me it feels like using physical database partitions in the same cube would be the correct decision in 90% of the cases. It seems to me that the underlying RDBMS should be able to handle itself to:
    1) Parallellize a query over several physical partitions if needed.
    2) Be smart enough to only query the needed partition if the query is restricted based on the partitioning characteristic.
    Please correct me anyone? - When is logical partitions really motivated?
    Best regards,
    Christian
    Edited by: Christian on May 15, 2008 3:55 PM

    Hi,
    This is a great question. Generally it is very difficult to understand the real motivation for the physical partioning - multiple cubes. You are right, it definitely increases the maintenance overhead. And you have already pointed out both the advantages and disadvantages.
    Physical Partitioning is more useful where we have huge amounts of data. Imagine a cube with 3 or 4 GB of data - which are not usual - but possible.  The Table Partioning is useful with small infocubes, less than 1 GB. With bigger Infocubes, Table level partitioning may not provide the required level of performance. If we have too many small partitions, that would also reduce the perfomance. If we have too few  partitions, the query performance will not be as much as we want. In this scenario, we can use Physical partitioning (Multiple Cubes) combined with Table Level Partitioning to achieve the required performance levels. On top, we can even think of using Aggregates for further betterment of the performance.
    While all the above seems to be relevant for older versions of BW (upto 3.5), BI 7.0 has the BIA (BI Accelerator), which works on the Blade Server with all the data cached directly on the main memory. I am not sure how much this would impact the data modeling - I have not started working on the BIA as yet.
    rgds
    naga

  • Impact of logical partitioning on BIA

    We are on BI release 701, SP 5. We are planning to create logical partition for some of the infocubes in our system. These cubes are already BIA enabled, so will creation of logical indexes have any impact on the BIA or improve the BIA rollup runtime

    Hi Leonel,
    Logical partitioning will have impact on BIA in terms of performance .
    Current cube is already indexed on BIA. Now if you divide current cube data into different cubes and create multiprovider on it then each cube will have its own F table index on BIA.
    you have to put new cubes to BIA and Execute Initial filling step on BIA and place rollups for respective cubes.
    Point to be noted :
    As Data will be deleted from Current cube and move to other cubes , Indexes will not get deleted
    from corresponding F table indexes in BIA . There will be indexes for records which are not present in cube. Thus it is always a good practice to flush BIA which will remove all the indexes from BIA for current cube and create new Indexes on BIA.
    In this case , we will have consistent indexes on BIA which will not hamper performance.
    This will also improve rollup time as data will be less in each cube after logical partitioning. For rollup
    improvement time , we can implement Delta indexing on BIA as well.
    Question : Why do we want to create logical partitioning for cubes which are present on BIA as queries will never come to Cubes in BI system ?
    Regards,
    Kishanlal Kumawat.

  • Logical Partitioning Infoprovider

    Hi Experts
    Could anyone tell me how to create multiple Infocubes on the top of one particular Infosource such that say one infocube contains data for Europe ,the other for Asia,the other for Americas and So on.That is they should be partitioned by region.
    Points will be Assigned
    Regards
    Ank

    Hi......
    U can do it by following Method
    1. To partition a Info cube go to info cube manage screen
    2. In EXTRAS --> Partitioning
    3.Activate
    Condition : Cube should not have data while partitioning
    E.g if u have data for month specific months jan , Feb, Mar....
    And u r adding monthwise data ,Suppose ur in Feb n u have added missing data for Jan month.this record get added after Feb records in internal table.
    Also check this :
    Only one InfoProvider?  Using Multiprovider you could have lower TCO due to Logical Partitioning
    Hope this helps you.....
    Regards,
    Debjani.......
    Edited by: Debjani  Mukherjee on Oct 10, 2008 11:01 AM

  • Data flow of logical partitions via MultiProvider

    Hi experts,
    I need to report on an InfoCube which will hold 2 years of data totals about 120 million records..So I would like to use logical partition to distribute the data to other InfoCubes based on time characteristic e.g. month..But I am not sure how can I achieve this? What I would like to know is the procedure, e.g. how to set up DTPs, how to use Process Chain to automate the data load, recommended strategy/approach etc...
    Your advice is highly appreciated!
    Thanks in advance.
    Regards,
    Meng

    Hi Joon,
    In case of logical partitioning the very first import thing is reporting will always be done on Multiprovider so as to keep the reporting unchanged with respect to underlying infoprovider changes.
    Now you can have different approaches, let us say in your example you will have to create two different infocubes one for say current year data (I assume already existing) and one for previous year data.
    Create transformation and DTP between current cube and previous year data cube and move all the previous year data from current cube to newly created cube using delta method by giving month or year selection in DTP selections. Once you move the data validate the data in target cube and do the selective deletion on the current cube, so as to avoide the data duplication. Compression on current data cube will be good as most of the requests will be empty after selective deletion only holding request ID's.
    Till now you have separated the current year and previous year data in two separate cubes, create a multiprovider on top of these cubes and you are free to create reports on top of multiprovider.
    Now when you are loading data from source system you have two options,
    1) Keep the current data flow as it is for current cube from source system and load all the history data in current cube and let it be as it is in current cube, being in the same multiprovider your overall reporting will not have any impact. But if you want you can move this data from current cube to previous cube using earlier used delta DTP and do the selective deletion again. You can even automate this process.
    2) In second scenario you can create individual flows to current and history cube and do the selective loading from source system. Your data will be directly loaded to both these cubes but the extraction from source system will be twice from the same source.
    Regards,
    Durgesh.

  • Finding whole mapping from database file - filesystems - logical volume manager - logical partitions

    Hello,
    Trying to make reverse engeneering of database files and their physical carriers on logical partitions ( fdisk ).
    And not able to make whole path from filesystem down to partitions with intermediate logical volumes.
    1. select from dba_data_files ...
    2. df -k
    to get the listing of filesystems
    3. vgdisplay
    4. lvdisplay
    5. cat /proc/partitions
    6. fdisk /dev/sda -l
       fdisk /dev/sdb -l
    Problem I have is that not able to determine which partitions are consisten in logical volumes. And then which logical volumens are consisted in filesystem.
    Thank you for hint or direction.

    Hello Wadhah,
    Before start the discussion let me explain I am newcommer to Oracle Linux. My genetic with dba experience of Oracle is from IBM UNIX ( AIX 6.1 ) and Oracle 11gr2.
    First task is to get the complete picture of one database on Oracle Linux for future maintenance tasks and make database more flexible and
    preparing for more intense work:
    -adding datafiles,
    -optimize/replace archive redolog files on separated filesystem from ORACLE_BASE
    - separating auditing log files from $ORACLE_BASE to own filesystem
    - separating diag directory on separated file system ( logging, tracing )
    - adding/enlarging TEMP ts
    - adding/enlarging undo
    - enlarging redo for higher transaction rate ( to reduce number of switched per time perceived in alert_SID.log )
    - adding online redo and control files mirrors
    So in this context try to inspect content of the disk space from the highest logical level in V$, DBA views down to fdisk partitions.
    The idea was to go in these steps:
    1. select paths of present online redo groups, datafiles, controlfiles, temp, undo
       from V$, dba views
    2. For the paths got from the step 1
       locate filesystems and for those filesystems inspect which are on logical volumens and which are directly on partitions.
    3. For all used logical volumes locate the logical partitions and their disks /dev/sda, /dev/sdb, ...

  • Sql to create logical partitions

    Oracle: 10.2.0.5
    I am working with another group and they are pulling data from one of the databases I work on. They are using what they call 'logical partitions'. Basically it is a sql statement with a MOD function in the where clause.
    select *
    from table
    where mod(field,10) = 0This allows them to divide the table up into 10 chunks. So they run 10 sessions to pull data across the network. They are using array processing(1000 records at a time) in a 3rd party tool to pull the data and write it to teradata. I have no ability to change this process to something else. They are not using a cursor. its just a fetch of 1000 recorsd at a time. I checked that first.
    The MOD function forces a full table scan. Before I go and and add a bunch of function based indexes to support this, does anyone know of another way to write these sqls without having to have a function on the left side of the where clause and get it to use an index? I want an index in part because 10 sessions is too slow to pull the data in an acceptable time so I want to increase the number of sessions i can handle. We are pulling from a number of tables so if its all full table scans I am far more constrained on my side.
    I am hoping there is a way to on the fly chunk a table with buckets or something and use an index. So I can ramp this up to say 20-30 sessions per table so each session gets 1/20 or 1/30 of the table.

    Guess2 wrote:
    Oracle: 10.2.0.5
    I am working with another group and they are pulling data from one of the databases I work on. They are using what they call 'logical partitions'. Basically it is a sql statement with a MOD function in the where clause.
    select *
    from table
    where mod(field,10) = 0This allows them to divide the table up into 10 chunks. So they run 10 sessions to pull data across the network. They are using array processing(1000 records at a time) in a 3rd party tool to pull the data and write it to teradata. I have no ability to change this process to something else. They are not using a cursor. its just a fetch of 1000 recorsd at a time. I checked that first.
    The MOD function forces a full table scan. Before I go and and add a bunch of function based indexes to support this, does anyone know of another way to write these sqls without having to have a function on the left side of the where clause and get it to use an index? I want an index in part because 10 sessions is too slow to pull the data in an acceptable time so I want to increase the number of sessions i can handle. We are pulling from a number of tables so if its all full table scans I am far more constrained on my side.
    I am hoping there is a way to on the fly chunk a table with buckets or something and use an index. So I can ramp this up to say 20-30 sessions per table so each session gets 1/20 or 1/30 of the table.From the school of thought that if some is good, then more is better.
    I suspect that the spindle upon which this table resides will be saturated
    with I/O requests long before 20 is reached.
    Session (CPU) is 100 - 1000 times faster than mechanical disk.
    Sessions as few as a half dozen can overwhelm single disk drive.

  • Use of Logical Partition in a Oracle Table...

    What is the use of Logical Partition in a Oracle Table as Target.  Techincal Manual does not say any significance.
    My question is:
    If the Table has no partitions and if we add Logical Partitions using Data Service, what purpose will it serve?
    We are planning to load 30 Million records a day into a Oracle Table. As of now the Target table has no partition and we are planning to have that soon. Is there a better way to load the data into Target Table, using Partition, Bulk Loading(API), Degree of Parallelism, etc., We have not dealt data of that volume, inputs are highly appreciated.
    Regards.
    Santosh.

    Initial Value:
    Indicator that NOT NULL is forced for this field
    Use
    Select this flag if a field to be inserted in the database is to be filled with initial values. The initial value used depends on the data type of the field.
    Please note that fields in the database for which the this flag is not set can also be filled with initial values.
    When you create a table, all fields of the table can be defined as NOT NULL and filled with an initial value. The same applies when converting the table. Only when new fields are added or inserted, are these filled with initial values. An exception is key fields. These are always filled automatically with initial values.
    Restrictions and notes:
    The initial value cannot be set for fields of data types LCHR, LRAW, and RAW. If the field length is greater than 32, the initial flag cannot be set for fields of data type NUMC.
    If a new field is inserted in the table and the initial flag is set, the complete table is scanned on activation and an UPDATE is made to the new field. This can be very time-consuming.
    If the initial flag is set for an included structure, this means that the attributes from the structure are transferred. That is, exactly those fields which are marked as initial in the definition have this attribute in the table as well.
    hope it helps,
    Saipriya

  • Logical partitioning, pass-through layer, query pruning

    Hi,
    I am dealing with performance guidelines for BW and encountered few interesting topics, which however I do not fully undestand.
    1. Mainetance of logical partitioning.
    Let's assume logical partitioning is performed on year. Does it mean that every year or so it is necessary to create additional cube/transformation and modify multiprovider? Is there any automatic procedure by SAP that supports creation of new objects, or it is fully manual?
    2.Pass- though layer.
    There are very few information about this basic concept.  Anyway:
    - is pass through DSO write optimized one? Does it store only one loading - last one? Is it deleted after lading is sucessfully finished (or before new load starts)? And - does this deletion do not destroy delta mechanism? Is the DSO replacing PSAfunctionally (i.e. PSA can be deleted every load as well)?
    3. Query pruning
    Does this happen automatically on DB level, or additional developments with exits variables, steering tables and FMs is required?
    4. DSOs for master data loads
    What is the benefit of using full MD extraction and DSO delta insetad of MD delta extraction?
    Thanks,
    Marcin

    1. Mainetance of logical partitioning.
    Let's assume logical partitioning is performed on year. Does it mean that every year or so it is necessary to create additional cube/transformation and modify multiprovider? Is there any automatic procedure by SAP that supports creation of new objects, or it is fully manual?
    Logical partitioning is when you have separate ODS / Cubes for separate Years etc ....
    There is no automated way - however if you want to you can physically partition the cubes using time periods and extend them regularly using the repartitioning options provided.
    2.Pass- though layer.
    There are very few information about this basic concept. Anyway:
    - is pass through DSO write optimized one? Does it store only one loading - last one? Is it deleted after lading is sucessfully finished (or before new load starts)? And - does this deletion do not destroy delta mechanism? Is the DSO replacing PSAfunctionally (i.e. PSA can be deleted every load as well)?
    Usually a pass through layer is used to
    1. Ensure data consistency
    2. Possibly use Deltas
    3. Additional transformations
    In a write optimized DSo - the request ID is key and hence delta is based on request ID. If you do not have any additional transformations - then a Write optimized DSO is essentially like your PSA.
    3. Query pruning
    Does this happen automatically on DB level, or additional developments with exits variables, steering tables and FMs is required?
    The query pruning - depends on the run based and cost based optimizers within the DB and not much control over how well you can influence the execution of a query other than havin up to date statistics , building aggregates etc etc.
    4. DSOs for master data loads
    What is the benefit of using full MD extraction and DSO delta insetad of MD delta extraction?
    It depends more on the data volumes and also the number of transformation required...
    If you have multiple levels of transformations - use a DSO or if you have very high data volumes and want to identify changed records - then use a DSO.

  • Standalone report server not found on the network between logical partitions on AIX

    Hello,
    Here s our architecture:
    forms/reports11gr2(patchset 1)
    weblogic 10.3.6
    on IBM AIX 7.1
    Server JRE
    Java(TM) SE Runtime Environment (build pap6460sr13ifix-20130303_02(SR13+IV37419)                                                                             )
    IBM J9 VM (build 2.4, JRE 1.6.0 IBM J9 2.4 AIX ppc64-64 jvmap6460sr13-20130114_1
    JRE -client - 1.6.0_27
    We have 2 logical partitions separated in 2 different physical m/c where cluster of forms /reports is installed.
    if i have a report service repsrv11g on one logical partition say, box 100 on Physical box 6000, the other logical partition box101's forms server on physical box 7000 is not able to look up the report service when calling from Forms using the Run_report_object.
    Gives, FRM -41213 error.
    If i just run the URL(use 2nd box) with http://101:8888/reports/rwservlet/showjobs?server=repsrv11g, it gives REP-51002: Bind to Reports Server repsrv11g failed
    We thought/read that as long as they re on the same network / domain, report service is available.
    Also did rwdiag.sh on one partition, its not able to find the other one.
    Ran the test form which Oracle provides and it s also not able to find the report server on the network when run on the other lpar.
    Temporarily, we created another report service on the other lpar but still using loadbalancing dns while doing web.show_document, so, it could potentially fail to bring up a report if load balancer redirects from one form's server to report server on the other partition.
    Any thoughts would be greatly appreciated.
    Thanks.

    Hello,
    Any inputs on this pls?

Maybe you are looking for

  • Update Long Text of Production Order

    Hi ABAP Gurus, We want to update some information in long text automatically while releasing production order. We try PPCO0007 We found that we must update AUFK-LTEXT as 'E' then we can see long text via CO02, CO03. Without updating AUFK-LTEXT, even

  • ISight problem, time for repair?

    iSight suddenly stopped working. No green light. Won't work in iChat, Gmail, or Photo Booth. Message "No camera connected to this computer". I followed the advice given here. Restarted. I checked Utilities >> System Profiler >> USB and saw iSight lis

  • Query on Foreign Currency Valuation(FCV)

    Hello Friends, Foreign currency valuation(FCV) program unable to post the entries when we maintain an Open item managed GL account as Balance sheet adjustment account in OB09. Why this program unable to post the entries to a open item managed GL acco

  • Re Initialisation for 0CRM_SRV_IbASE_ATTR

    Hi Gurus, I have tried doing re init for the 0CRM_SRV_IbASE_ATTR but I have got error says delete the init request. But I deleted the init from the shedule in the infopackage and no data in the Datatarget as well. Please help me out from this. Thanks

  • PDFs made through 'reader' button have a narrow margin in Safari 7.0.4

    Hi One thing I like about Safari is its handy feature of converting most webpages to uncluttered, neat PDFs which include only the main text passage with its relevant illustrations. But after the recent upgrade, I noticed two things the second of whi