Impact of logical partitioning on BIA

We are on BI release 701, SP 5. We are planning to create logical partition for some of the infocubes in our system. These cubes are already BIA enabled, so will creation of logical indexes have any impact on the BIA or improve the BIA rollup runtime

Hi Leonel,
Logical partitioning will have impact on BIA in terms of performance .
Current cube is already indexed on BIA. Now if you divide current cube data into different cubes and create multiprovider on it then each cube will have its own F table index on BIA.
you have to put new cubes to BIA and Execute Initial filling step on BIA and place rollups for respective cubes.
Point to be noted :
As Data will be deleted from Current cube and move to other cubes , Indexes will not get deleted
from corresponding F table indexes in BIA . There will be indexes for records which are not present in cube. Thus it is always a good practice to flush BIA which will remove all the indexes from BIA for current cube and create new Indexes on BIA.
In this case , we will have consistent indexes on BIA which will not hamper performance.
This will also improve rollup time as data will be less in each cube after logical partitioning. For rollup
improvement time , we can implement Delta indexing on BIA as well.
Question : Why do we want to create logical partitioning for cubes which are present on BIA as queries will never come to Cubes in BI system ?
Regards,
Kishanlal Kumawat.

Similar Messages

  • Any real reason for logical partitioning over physical?

    Hi!
    I have seen a number of scenarios where SAP BI (assuming BI 7.0 for the rest of the discussion), running in high volume scenarios, have been cluttered by a lot of logically partitioned cubes joined by multi providers....
    Obviously the disadvantage of using logical partitions is that it increases maintenance efforts: you need a new update rule for each logical partition (cube) , then you need to manually add/delte cubes from the multiprovider, filtering data in the update rules to reach the correct cube based on time characteristic etc etc...
    I have seen one clear advantage which is the parallelization of queries run against a multiprovider - assuming you wan't to all underlying cubes ... but are there any other advantages which overcome the mainenance overhead?
    For me it feels like using physical database partitions in the same cube would be the correct decision in 90% of the cases. It seems to me that the underlying RDBMS should be able to handle itself to:
    1) Parallellize a query over several physical partitions if needed.
    2) Be smart enough to only query the needed partition if the query is restricted based on the partitioning characteristic.
    Please correct me anyone? - When is logical partitions really motivated?
    Best regards,
    Christian
    Edited by: Christian on May 15, 2008 3:55 PM

    Hi,
    This is a great question. Generally it is very difficult to understand the real motivation for the physical partioning - multiple cubes. You are right, it definitely increases the maintenance overhead. And you have already pointed out both the advantages and disadvantages.
    Physical Partitioning is more useful where we have huge amounts of data. Imagine a cube with 3 or 4 GB of data - which are not usual - but possible.  The Table Partioning is useful with small infocubes, less than 1 GB. With bigger Infocubes, Table level partitioning may not provide the required level of performance. If we have too many small partitions, that would also reduce the perfomance. If we have too few  partitions, the query performance will not be as much as we want. In this scenario, we can use Physical partitioning (Multiple Cubes) combined with Table Level Partitioning to achieve the required performance levels. On top, we can even think of using Aggregates for further betterment of the performance.
    While all the above seems to be relevant for older versions of BW (upto 3.5), BI 7.0 has the BIA (BI Accelerator), which works on the Blade Server with all the data cached directly on the main memory. I am not sure how much this would impact the data modeling - I have not started working on the BIA as yet.
    rgds
    naga

  • Data flow of logical partitions via MultiProvider

    Hi experts,
    I need to report on an InfoCube which will hold 2 years of data totals about 120 million records..So I would like to use logical partition to distribute the data to other InfoCubes based on time characteristic e.g. month..But I am not sure how can I achieve this? What I would like to know is the procedure, e.g. how to set up DTPs, how to use Process Chain to automate the data load, recommended strategy/approach etc...
    Your advice is highly appreciated!
    Thanks in advance.
    Regards,
    Meng

    Hi Joon,
    In case of logical partitioning the very first import thing is reporting will always be done on Multiprovider so as to keep the reporting unchanged with respect to underlying infoprovider changes.
    Now you can have different approaches, let us say in your example you will have to create two different infocubes one for say current year data (I assume already existing) and one for previous year data.
    Create transformation and DTP between current cube and previous year data cube and move all the previous year data from current cube to newly created cube using delta method by giving month or year selection in DTP selections. Once you move the data validate the data in target cube and do the selective deletion on the current cube, so as to avoide the data duplication. Compression on current data cube will be good as most of the requests will be empty after selective deletion only holding request ID's.
    Till now you have separated the current year and previous year data in two separate cubes, create a multiprovider on top of these cubes and you are free to create reports on top of multiprovider.
    Now when you are loading data from source system you have two options,
    1) Keep the current data flow as it is for current cube from source system and load all the history data in current cube and let it be as it is in current cube, being in the same multiprovider your overall reporting will not have any impact. But if you want you can move this data from current cube to previous cube using earlier used delta DTP and do the selective deletion again. You can even automate this process.
    2) In second scenario you can create individual flows to current and history cube and do the selective loading from source system. Your data will be directly loaded to both these cubes but the extraction from source system will be twice from the same source.
    Regards,
    Durgesh.

  • Finding whole mapping from database file - filesystems - logical volume manager - logical partitions

    Hello,
    Trying to make reverse engeneering of database files and their physical carriers on logical partitions ( fdisk ).
    And not able to make whole path from filesystem down to partitions with intermediate logical volumes.
    1. select from dba_data_files ...
    2. df -k
    to get the listing of filesystems
    3. vgdisplay
    4. lvdisplay
    5. cat /proc/partitions
    6. fdisk /dev/sda -l
       fdisk /dev/sdb -l
    Problem I have is that not able to determine which partitions are consisten in logical volumes. And then which logical volumens are consisted in filesystem.
    Thank you for hint or direction.

    Hello Wadhah,
    Before start the discussion let me explain I am newcommer to Oracle Linux. My genetic with dba experience of Oracle is from IBM UNIX ( AIX 6.1 ) and Oracle 11gr2.
    First task is to get the complete picture of one database on Oracle Linux for future maintenance tasks and make database more flexible and
    preparing for more intense work:
    -adding datafiles,
    -optimize/replace archive redolog files on separated filesystem from ORACLE_BASE
    - separating auditing log files from $ORACLE_BASE to own filesystem
    - separating diag directory on separated file system ( logging, tracing )
    - adding/enlarging TEMP ts
    - adding/enlarging undo
    - enlarging redo for higher transaction rate ( to reduce number of switched per time perceived in alert_SID.log )
    - adding online redo and control files mirrors
    So in this context try to inspect content of the disk space from the highest logical level in V$, DBA views down to fdisk partitions.
    The idea was to go in these steps:
    1. select paths of present online redo groups, datafiles, controlfiles, temp, undo
       from V$, dba views
    2. For the paths got from the step 1
       locate filesystems and for those filesystems inspect which are on logical volumens and which are directly on partitions.
    3. For all used logical volumes locate the logical partitions and their disks /dev/sda, /dev/sdb, ...

  • Sql to create logical partitions

    Oracle: 10.2.0.5
    I am working with another group and they are pulling data from one of the databases I work on. They are using what they call 'logical partitions'. Basically it is a sql statement with a MOD function in the where clause.
    select *
    from table
    where mod(field,10) = 0This allows them to divide the table up into 10 chunks. So they run 10 sessions to pull data across the network. They are using array processing(1000 records at a time) in a 3rd party tool to pull the data and write it to teradata. I have no ability to change this process to something else. They are not using a cursor. its just a fetch of 1000 recorsd at a time. I checked that first.
    The MOD function forces a full table scan. Before I go and and add a bunch of function based indexes to support this, does anyone know of another way to write these sqls without having to have a function on the left side of the where clause and get it to use an index? I want an index in part because 10 sessions is too slow to pull the data in an acceptable time so I want to increase the number of sessions i can handle. We are pulling from a number of tables so if its all full table scans I am far more constrained on my side.
    I am hoping there is a way to on the fly chunk a table with buckets or something and use an index. So I can ramp this up to say 20-30 sessions per table so each session gets 1/20 or 1/30 of the table.

    Guess2 wrote:
    Oracle: 10.2.0.5
    I am working with another group and they are pulling data from one of the databases I work on. They are using what they call 'logical partitions'. Basically it is a sql statement with a MOD function in the where clause.
    select *
    from table
    where mod(field,10) = 0This allows them to divide the table up into 10 chunks. So they run 10 sessions to pull data across the network. They are using array processing(1000 records at a time) in a 3rd party tool to pull the data and write it to teradata. I have no ability to change this process to something else. They are not using a cursor. its just a fetch of 1000 recorsd at a time. I checked that first.
    The MOD function forces a full table scan. Before I go and and add a bunch of function based indexes to support this, does anyone know of another way to write these sqls without having to have a function on the left side of the where clause and get it to use an index? I want an index in part because 10 sessions is too slow to pull the data in an acceptable time so I want to increase the number of sessions i can handle. We are pulling from a number of tables so if its all full table scans I am far more constrained on my side.
    I am hoping there is a way to on the fly chunk a table with buckets or something and use an index. So I can ramp this up to say 20-30 sessions per table so each session gets 1/20 or 1/30 of the table.From the school of thought that if some is good, then more is better.
    I suspect that the spindle upon which this table resides will be saturated
    with I/O requests long before 20 is reached.
    Session (CPU) is 100 - 1000 times faster than mechanical disk.
    Sessions as few as a half dozen can overwhelm single disk drive.

  • Use of Logical Partition in a Oracle Table...

    What is the use of Logical Partition in a Oracle Table as Target.  Techincal Manual does not say any significance.
    My question is:
    If the Table has no partitions and if we add Logical Partitions using Data Service, what purpose will it serve?
    We are planning to load 30 Million records a day into a Oracle Table. As of now the Target table has no partition and we are planning to have that soon. Is there a better way to load the data into Target Table, using Partition, Bulk Loading(API), Degree of Parallelism, etc., We have not dealt data of that volume, inputs are highly appreciated.
    Regards.
    Santosh.

    Initial Value:
    Indicator that NOT NULL is forced for this field
    Use
    Select this flag if a field to be inserted in the database is to be filled with initial values. The initial value used depends on the data type of the field.
    Please note that fields in the database for which the this flag is not set can also be filled with initial values.
    When you create a table, all fields of the table can be defined as NOT NULL and filled with an initial value. The same applies when converting the table. Only when new fields are added or inserted, are these filled with initial values. An exception is key fields. These are always filled automatically with initial values.
    Restrictions and notes:
    The initial value cannot be set for fields of data types LCHR, LRAW, and RAW. If the field length is greater than 32, the initial flag cannot be set for fields of data type NUMC.
    If a new field is inserted in the table and the initial flag is set, the complete table is scanned on activation and an UPDATE is made to the new field. This can be very time-consuming.
    If the initial flag is set for an included structure, this means that the attributes from the structure are transferred. That is, exactly those fields which are marked as initial in the definition have this attribute in the table as well.
    hope it helps,
    Saipriya

  • Logical Partition of infocube

    We have a cube logical partitioned by sales org.  We now would like to add additional sales org's to the list. Is this possible??  Currently with data in the cube it is blocked.  Does anyone know steps to re-logical partition a cube??  With/or without data re-loading?  Any options at the database level (oracle)?
    Thanks!!

    Hi,
    Yes that is correct you cant do cube partitioning with existing data.
    Another important thing is based upon selection conditions only you can do the partitioning.  There are two selection conditions they are as follows.
                   1. 0CALMONTH
                   2. 0FISCAL YEAR PERIOD
    if these two time characteristics were not used in the cube means you can't partition the cube. 
    STEPS TO PARTITION THE INFOCUBE:
    for example if you have Sales and Distribution cube you want to do logical partitioning on that cube. 
    we have to create infocube with same structure ( this we can do by giving the technical name of the cube in copy from option) at the time of creation of cube. Activate the cube and come back.
    then you should create update rules for that cube.
    select the first cube right click on it choose the option generate export datasource.
    the system will finish generation process.
    select the infosource option on the left side of the AWB.
    in that we have DATAMART select it choose refresh icon. 
    you get the infosource with datasource assignment.
    select the data source right click on it choose create infopackage.
    it will take you to the maintain infopackage screen. 
    here you need to select the datatarget tab and select the target into which the data need to be updated.
    schedule it and start.
    now manage the data target and check the data has been updated or not.
    after doing all this process go to infoprovider select the first cube delete the data.
    now depending upon your requirement do the partitioning.
    double click on the cube it will take you to the edit infocube screen from the menu bar select extras it will display the partitioning option select it.
    then it will popup one small wizard showing the time characteristics based upon your requirement choose which ever you want check that box and continue the process.
    Hope this will help you
    Thanks and Regards
    Vara Prasad

  • Logical partitioning, pass-through layer, query pruning

    Hi,
    I am dealing with performance guidelines for BW and encountered few interesting topics, which however I do not fully undestand.
    1. Mainetance of logical partitioning.
    Let's assume logical partitioning is performed on year. Does it mean that every year or so it is necessary to create additional cube/transformation and modify multiprovider? Is there any automatic procedure by SAP that supports creation of new objects, or it is fully manual?
    2.Pass- though layer.
    There are very few information about this basic concept.  Anyway:
    - is pass through DSO write optimized one? Does it store only one loading - last one? Is it deleted after lading is sucessfully finished (or before new load starts)? And - does this deletion do not destroy delta mechanism? Is the DSO replacing PSAfunctionally (i.e. PSA can be deleted every load as well)?
    3. Query pruning
    Does this happen automatically on DB level, or additional developments with exits variables, steering tables and FMs is required?
    4. DSOs for master data loads
    What is the benefit of using full MD extraction and DSO delta insetad of MD delta extraction?
    Thanks,
    Marcin

    1. Mainetance of logical partitioning.
    Let's assume logical partitioning is performed on year. Does it mean that every year or so it is necessary to create additional cube/transformation and modify multiprovider? Is there any automatic procedure by SAP that supports creation of new objects, or it is fully manual?
    Logical partitioning is when you have separate ODS / Cubes for separate Years etc ....
    There is no automated way - however if you want to you can physically partition the cubes using time periods and extend them regularly using the repartitioning options provided.
    2.Pass- though layer.
    There are very few information about this basic concept. Anyway:
    - is pass through DSO write optimized one? Does it store only one loading - last one? Is it deleted after lading is sucessfully finished (or before new load starts)? And - does this deletion do not destroy delta mechanism? Is the DSO replacing PSAfunctionally (i.e. PSA can be deleted every load as well)?
    Usually a pass through layer is used to
    1. Ensure data consistency
    2. Possibly use Deltas
    3. Additional transformations
    In a write optimized DSo - the request ID is key and hence delta is based on request ID. If you do not have any additional transformations - then a Write optimized DSO is essentially like your PSA.
    3. Query pruning
    Does this happen automatically on DB level, or additional developments with exits variables, steering tables and FMs is required?
    The query pruning - depends on the run based and cost based optimizers within the DB and not much control over how well you can influence the execution of a query other than havin up to date statistics , building aggregates etc etc.
    4. DSOs for master data loads
    What is the benefit of using full MD extraction and DSO delta insetad of MD delta extraction?
    It depends more on the data volumes and also the number of transformation required...
    If you have multiple levels of transformations - use a DSO or if you have very high data volumes and want to identify changed records - then use a DSO.

  • 0IC_C03 related Inventory Process - Logical Partitioning (Vs) Physical Part

    Hello Everyone,
    After going through multiple postings throughout the form and documentation from SAP, it states that the 0IC_C03 InfoCube when used with Non Cumulative keyfigures is not recommended to be partitioned logically by physical year/calendar year as the query will read all the data sets due to the stock marker logic.
    In our specific scenario,
    1. After the InfoCube (0IC_C03) was enhanced with additional characterisitcs such as Doc Number, Movement type and so on due to business requirements I was not able to actually use the Non Cumulative Keyfigures as they were not populated within the report.
    2. So, we decided not to use the Non Cumulative keyfigures but rather create two cumulative keyfigures (Issue Stock Quantity - Receipt Stock Quantity) and (Issue Valuated Stock Value - Receipt Valuated Stock Value) and both of these are available in the InfoCube and are calculated during the update process.
    These two keyfigures are cumulative with exception aggregation of LAST based on 0CALDAY.
    The question is,
    Since we are not using the actual Non Cumulative Keyfigures (even though we are not using these, we still have the stock marker updated and data compressed based on this along with Validity table defined) can we do logical partitioning on the InfoCube based on Calendar year.
    Thanks....

    Hello Elango,
    Appreciate your response.
    First of..I do understand the difference between logical and physical partitioning and the question is not about joining them together.
    I am sorry, if others cannot understand the detailed issue posted. My apologies was a part of polite gesture, and please do respond back with proper precise answer if you think you did actually understand the question....
    The question here is about how I can leverage the performance and administrative performance by logically breakingdown the data.
    The issues due to which I am trying to look into different aspects of logical partitioning are:
    1. If I do logical partitioning by Plant due to the stock marker logic then I cannot do archiving as a Plant and its related data cannot be archived by time characteristic as the partitioning is not done by time characteristic.
    2. The reason I would have to have document number and movement type in the InfoCube is due to the kind of reporting users perform.
    We have a third party system whose data needs to be reconciled to the data in the plants and storage locations.
    And in order to do so, the first step users would be running the report is plant, storage location and sku. From here on for the storage locations which have balance they would like to drill down on to the document number and movement type to see what the actual activity is.
    So, to support this requirement I would have to have the above characterisitcs in the InfoCube.
    The question again is,.....what is the exact list of issues I would be having doing the logical partitioning by time characteristic.
    Once again, even though the non cumulative keyfigures are available in the InfoCube we are not using them for any reporting purpose....so please keep that in consideration while replying back.
    Thanks
    Dharma.

  • Standalone report server not found on the network between logical partitions on AIX

    Hello,
    Here s our architecture:
    forms/reports11gr2(patchset 1)
    weblogic 10.3.6
    on IBM AIX 7.1
    Server JRE
    Java(TM) SE Runtime Environment (build pap6460sr13ifix-20130303_02(SR13+IV37419)                                                                             )
    IBM J9 VM (build 2.4, JRE 1.6.0 IBM J9 2.4 AIX ppc64-64 jvmap6460sr13-20130114_1
    JRE -client - 1.6.0_27
    We have 2 logical partitions separated in 2 different physical m/c where cluster of forms /reports is installed.
    if i have a report service repsrv11g on one logical partition say, box 100 on Physical box 6000, the other logical partition box101's forms server on physical box 7000 is not able to look up the report service when calling from Forms using the Run_report_object.
    Gives, FRM -41213 error.
    If i just run the URL(use 2nd box) with http://101:8888/reports/rwservlet/showjobs?server=repsrv11g, it gives REP-51002: Bind to Reports Server repsrv11g failed
    We thought/read that as long as they re on the same network / domain, report service is available.
    Also did rwdiag.sh on one partition, its not able to find the other one.
    Ran the test form which Oracle provides and it s also not able to find the report server on the network when run on the other lpar.
    Temporarily, we created another report service on the other lpar but still using loadbalancing dns while doing web.show_document, so, it could potentially fail to bring up a report if load balancer redirects from one form's server to report server on the other partition.
    Any thoughts would be greatly appreciated.
    Thanks.

    Hello,
    Any inputs on this pls?

  • Geolocation (Logical Partition) on Mobility with Smartphones

    Hello
    We have a new project coming up and the primary requirement is Smartphone Integration with CUCM. (Nokia, Iphone and Android Phones)..
    This project will be deployed in India and now my concern is, since here in India there are DOT regulations for TEHO.
    When I configure a Smartphone to integrate with CUCM, if this smartphone user tries to make a call to another IP phone extension at a different location over the WAN and while this call is on, if he conference this with a local PSTN number from his VOIP line on the smartphone then this is going to be a violation.
    To avoid this problem in India, I know Cisco introduced the Geolocation (Logical partition feature).
    But do these smart phones support the Geolocation configuration.
    I read in the release notes of CUCM that
    " Logical partitioning handling does not take place in the following circumstances:
    • Geolocation or geolocation filter does not associate with the involved devices.
    • No logical partitioning support exists when a dual-mode phone is used. "
    Can you help me with a confirmation on this and if possible a better solution to ensure no violation of any regulation while using Smartphone as IP Phone extensions.
    Thanks in Advance
    Abraham Sebastian

    Windows 8.1 can run on either partition. The biggest issue with MBR over GPT is that MBR is limited to the size od disks supported. 4 partitions of up to 2TB max. GPT has a theoretical limit of 9.4 Zettabytes (10 Billion TB). You may also run into a problem if you are dual booting (or more). If neither of those two things concern you, then just leave well enough alone. About the Virtual hard drive, do you really need them? They will just eat up more space on the real harddrive. Are you going to use a Virtual machine?
    Hoov
    Microsoft MVP - Consumer Security
    SpywareHammer.com

  • Physical Vs Logical Partitioning

    We have 2 million records in the sales infocube for 3 years. We are currently discussing the pros and cons of using Logical partitioning Vs Physical Partitioning. Please give your inputs.

    hi
    there are two types of partitioning generally talked about with SAP BW, logical and physical partitioning.
    Logical partitioning - instead of having all your data in a single cube, you might break into separate cubes, with each cube holding aspecific year's data, e.g. you could have 5 sales cubes, one for each year 2001 thru 2005.
    You would then create a Multi-Provider that allowed you to query all of them together.
    A query that needs data from all 5 years would then automatically (you can control this) be split into 5 separate queries, one against each cube, running at the same time. The system automatically merges the results from the 5 queries into a single result set.
    So it's easy to see when this could be a benefit. If your queries however are primarily run just for a single year, then you don't receive the benefit of the parallel processing. In non-Oracle DBs, splitting the data like this may still be a benefit by reducing the amount of rows in the fact table that must be read, but does not provide as much value to an Oracle DB since Infocube queries are using a Star_Transformation.
    Physical Partitioning - I believe only Oracle and Informix currently support Range partitioning. This is a separately licensed option in Oracle.
    Physical partitioning allows you to split an Infocube into smaller pieces. The pieces, or partitions, can only be created by 0FISCPER or 0CALMONTH for an InfoCube (ODSs can be partitioned, but require a DBAs involvement). The DB can then take advantage of this partitioning by "pruning" partitions during a query, e.g. a query only needs data form June 2005
    The DB is smart enough to restrict the indices and data it will read to the June 2005 partition. This assumes your query restricts/filters on the partitioning characteristic. It can apply this pruning to a range of partitions as well, e.g. 0FISCPER 001/2005 thru 003/2005 would only look at the 3 partitions.
    It is NOTsmart enough, however, to figure out that if your restrict to 0FISCYEAR = 2005, that it should only read 000/2005 thru 016/2005 since 0FISCYEAR is NOT the partitioning characteristic.
    An InfoCube MUST be empty in order to physically partition it. At this time, there is no way to add additional partitions thru AWB, so you want to make sure that you create partitions out into the future for at least a of couple of years.
    If the base cube is partitioned, any aggregates that contain the partitioning characteristic (0CALMONTH or 0FISCPER) will automatically be partitioned.
    In summary, you need to figure out if you want to use physical or logical partitioning on the cube(s), or both, as they are not mutually exclusive.
    So you would need to know how the data will be queried, and the volume of data. It would make little sense to partition cubes that will not be very large.
    physical partitioning is done at database level and logical partitioning done at data target level.
    Cube partitioning with time characteristics 0calmonth Fiscper is physical partitioning.
    Logical partitioning is u partition ur cube by year or month that is u divide the cube into different cubes and create a multiprovider on top of it.
    logical Vs physical partitions ?

  • SoftRaid0: 1 of 3 md's no speed enhancment (using logical partitions)?

    Hi!
    Before I forget to mention it: I don't really know much about raid & stuff (I'm surprised myself I got it running), so don't think a possible explanations is too obvious/simple - I might have missed it anyway
    I've got 3 (+ backup) Raid0's on the same 2 sata hd's. The first two seem to work with the right speed, the third one only normal / half of it. The only difference I could find is that md2 consists of logical partitions - md0 and md1 have their own primary partition each.
    All in infos in order md0/1/2
    /dev/md0 partition 3999992 76416 -1 (<= cat /proc/swaps)
    /dev/md1 on /home type ext4 (rw)
    /dev/md2 on / type ext4 (rw)
    /dev/md0: 3.81GiB raid0 2 devices, 0 spares.
    /dev/md1: 889.97GiB raid0 2 devices, 0 spares.
    /dev/md2: 18.63GiB raid0 2 devices, 0 spares.
    Timing buffered disk reads: 446 MB in 3.01 seconds = 148.08 MB/sec
    Timing buffered disk reads: 470 MB in 3.01 seconds = 156.20 MB/sec
    Timing buffered disk reads: 234 MB in 3.02 seconds = 77.57 MB/sec
    Number Major Minor RaidDevice State
    0 8 2 0 active sync /dev/sda2
    1 8 18 1 active sync /dev/sdb2
    Number Major Minor RaidDevice State
    0 8 3 0 active sync /dev/sda3
    1 8 19 1 active sync /dev/sdb3
    Number Major Minor RaidDevice State
    0 8 5 0 active sync /dev/sda5
    1 8 21 1 active sync /dev/sdb5
    sda & sdb look the same:
    /dev/sda1 * 1 31 248976 83 Linux
    /dev/sda2 32 280 2000092+ fd Linux raid autodetect
    /dev/sda3 281 58369 466599892+ fd Linux raid autodetect
    /dev/sda4 58370 60801 19535040 5 Extended
    /dev/sda5 58370 59585 9767488+ fd Linux raid autodetect
    /dev/sda6 59586 60801 9767488+ fd Linux raid autodetect
    sd*6 (sometimes md3) is a spare raid that shares the same extended partition - and low speed. Most of the time it is not assembled though, which doesn't change anything for md2.
    Any Ideas?

    Still the same, didn't manage to figure anything more out about that.
    Could someone with Extended soft-raid maybe just confirm, if
    a) It's the same as with my "slow partition"
    b) It's also possible to be "fast"
    Or something... anything?
    thx!

  • Cannot create another 2 logical partitions on another physical server

    when i installed the BI 7.0 on AIX/DB2 9 platform. i can create 2 logical partition on the main server , yet i couldn't create another 2 logical parttions on the second server.  the following is the error message
    INFO       2008-02-21 03:49:03.490
    "sapinst_dev.log" [Read only] 20411 lines, 744293 characters
    TRACE      2008-02-21 03:51:28.513 [iaxxejsexp.cpp:199]
               EJS_Installer::writeTraceToLogBook()
    Found Error, error_codes[1] = <db2start dbpartitionnum 5 add dbpartitionnum hostname sapaix08 port 3 without tablespaces
    SQL6073N  Add Node operation failed.  SQLCODE = "-1051".>
    TRACE      2008-02-21 03:51:28.513 [iaxxejsexp.cpp:199]
               EJS_Installer::writeTraceToLogBook()
    During execution of <AddPart.sql>, <2> errors occured.
    ERROR      2008-02-21 03:51:28.513 [iaxxinscbk.cpp:282]
               abortInstallation
    MDB-01999  Error occured, first error is: <SQL6073N  Add Node operation failed.  SQLCODE = "-1051".>
    TRACE      2008-02-21 03:51:28.514 [iaxxejsbas.hpp:388]
               handleException<ESAPinstException>()
    Converting exception into JS Exception Exception.
    ERROR      2008-02-21 03:51:28.515
               CJSlibModule::writeError_impl()
    MUT-03025  Caught ESAPinstException in Modulecall: ESAPinstException: error text undefined.
    TRACE      2008-02-21 03:51:28.515 [iaxxejsbas.hpp:460]
               EJS_Base::dispatchFunctionCall()
    JS Callback has thrown unknown exception. Rethrowing.
    ERROR      2008-02-21 03:51:28.516 [iaxxgenimp.cpp:731]
               showDialog()
    FCO-00011  The step AddDB6Partitions with step key |NW_DB6_DB_ADDPART|ind|ind|ind|ind|0|0|NW_DB6_AddPartitions|ind|ind|ind|ind|12|0|
    AddDB6Partitions was executed with status ERROR .
    TRACE      2008-02-21 03:51:28.539 [iaxxgenimp.cpp:719]
               showDialog()
    the following is my prerequisite for the installation
    1. the user and group id and property is the same as the primary (server1)
    2. the ssh trust relationship has built , i can ssh  server1 from server2 or server2 from server1 with db2sid, sidadm users
    3. i mount the /db2/db2sid , /db2/SID/db2dumps, /sapmnt/SID/exe on server2 as NFS
    4. install the db2 software on /opt/IBM/db2/V9.1   (the location is the same as the primary's)
    HI , DB2 experts. Could you give me some suggestion?   thanks!

    Hi,Thomas,
    Thanks for your help.  the db2 database  desn't use the autostoarge method and the relevant permission is the same as server1.  i checked the db2dialog.log.  the following is the detail information
    "Storage path does not exist or is inaccessible" is the error message . i was wondering which storage path does not exit or inaccessible .
    at the same time i have login on all /db2 with db2sid and run touch to test the permision . it sounds good. I don't know what happens , could you give me some suggestion ? thanks!
    2008-02-21-08.10.56.442000-300 I14165596A287      LEVEL: Event
    PID     : 843832               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:240
    DATA #1 : String, 26 bytes
    Stop phase is in progress.
    2008-02-21-08.10.56.444783-300 I14165884A302      LEVEL: Event
    PID     : 843832               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:250
    DATA #1 : String, 41 bytes
    Requesting system controller termination.
    2008-02-21-08.10.56.450366-300 I14166187A403      LEVEL: Warning
    PID     : 712906               TID  : 1           PROC : db2sysc 5
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, routine_infrastructure, sqlerKillAllFmps, probe:5
    MESSAGE : Bringing down all db2fmp processes as part of db2stop
    DATA #1 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFE400 : 0000 0000                                  ....
    2008-02-21-08.10.56.456345-300 I14166591A304      LEVEL: Event
    PID     : 843832               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:260
    DATA #1 : String, 43 bytes
    System controller termination is completed.
    2008-02-21-08.10.56.461462-300 I14166896A381      LEVEL: Event
    PID     : 843832               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:280
    DATA #1 : String, 24 bytes
    There is no active EDUs.
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFCEE0 : 0000 0000                                  ....
    2008-02-21-08.10.56.504322-300 I14167278A342      LEVEL: Severe
    PID     : 823374               TID  : 1           PROC : db2acd 5
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, routine_infrastructure, sqlerFmpOneTimeInit, probe:100
    DATA #1 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFF5A4 : FFFF FBEE                                  ....
    2008-02-21-08.10.56.654959-300 E14167621A301      LEVEL: Event
    PID     : 843832               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:911
    MESSAGE : ADM7514W  Database manager has stopped.
    STOP    : DB2 DBM
    2008-02-21-08.11.09.664000-300 I14167923A417      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 53 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile FORCE1 0 0
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0022                                  ..."
    2008-02-21-08.11.10.176098-300 I14168341A417      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 53 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile FORCE1 1 1
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0022                                  ..."
    2008-02-21-08.11.10.595702-300 I14168759A417      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 53 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile FORCE1 2 0
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0022                                  ..."
    2008-02-21-08.11.11.124888-300 I14169177A417      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 53 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile FORCE1 3 1
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0022                                  ..."
    2008-02-21-08.11.12.070605-300 I14169595A410      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 46 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile 0 0
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0020                                  ...
    2008-02-21-08.11.12.694723-300 I14170006A410      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 46 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile 1 1
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0020                                  ...
    2008-02-21-08.11.13.115940-300 I14170417A410      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 46 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile 2 0
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0020                                  ...
    2008-02-21-08.11.13.632046-300 I14170828A410      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 46 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile 3 1
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0020                                  ...
    2008-02-21-08.11.14.577056-300 I14171239A418      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 54 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile NODEACT 0 0
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0024                                  ...$
    2008-02-21-08.11.15.004794-300 I14171658A418      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 54 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile NODEACT 1 1
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0024                                  ...$
    2008-02-21-08.11.15.425920-300 I14172077A418      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 54 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile NODEACT 2 0
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0024                                  ...$
    2008-02-21-08.11.15.941622-300 I14172496A418      LEVEL: Event
    PID     : 639410               TID  : 1           PROC : db2stop
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 54 bytes
    /db2/db2ab7/sqllib/adm/db2rstop db2profile NODEACT 3 1
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9CBC : 0000 0024                                  ...$
    2008-02-21-08.11.17.002107-300 I14172915A422      LEVEL: Event
    PID     : 639412               TID  : 1           PROC : db2start
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 57 bytes
    /db2/db2ab7/sqllib/adm/db2rstar db2profile SN ADDNODE 4 2
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9C2C : 0000 0011                                  ....
    2008-02-21-08.11.18.055723-300 E14173338A856      LEVEL: Warning
    PID     : 806940               TID  : 1           PROC : db2star2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, license manager, sqllcRequestAccess, probe:1
    MESSAGE : ADM12007E  There are "80" day(s) left in the evaluation period for
              the product "DB2 Enterprise Server Edition". For evaluation license
              terms and conditions, refer to the IBM License Acceptance and License
              Information document located in the license directory in the
              installation path of this product. If you have licensed this product,
              ensure the license key is properly registered. You can register the
              license via the License Center or db2licm command line utility. The
              license file can be obtained from your licensed product CD.
    2008-02-21-08.11.18.296453-300 E14174195A1040     LEVEL: Event
    PID     : 806940               TID  : 1           PROC : db2star2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, DB2StartMain, probe:911
    MESSAGE : ADM7513W  Database manager has started.
    START   : DB2 DBM
    DATA #1 : Build Level, 152 bytes
    Instance "db2ab7" uses "64" bits and DB2 code release "SQL09012"
    with level identifier "01030107".
    Informational tokens are "DB2 v9.1.0.2", "special_17253", "U810940_17253", Fix Pack "2".
    DATA #2 : System Info, 224 bytes
    System: AIX sapaix08 3 5 00CCD7FE4C00
    CPU: total:8 online:8 Threading degree per core:2
    Physical Memory(MB): total:7744 free:5866
    Virtual  Memory(MB): total:32832 free:30943
    Swap     Memory(MB): total:25088 free:25077
    Kernel   Params: msgMaxMessageSize:4194304 msgMaxQueueSize:4194304
                     shmMax:68719476736 shmMin:1 shmIDs:131072
                     shmSegments:68719476736 semIDs:131072 semNumPerID:65535
                     semOps:1024 semMaxVal:32767 semAdjustOnExit:16384
    2008-02-21-08.11.19.312894-300 I14175236A428      LEVEL: Error
    PID     : 835728               TID  : 1           PROC : db2agent (instance) 4
    INSTANCE: db2ab7               NODE : 004
    APPHDL  : 4-7                  APPID: *LOCAL.db2ab7.080221131118
    FUNCTION: DB2 UDB, base sys utilities, sqleGetAutomaticStorageDetails, probe:111111
    DATA #1 : <preformatted>
    dataSize 752 pMemAlloc 1110cdac0 sizeof(struct sqleAutoStorageCfg) 16
    2008-02-21-08.11.19.346560-300 I14175665A497      LEVEL: Error
    PID     : 835728               TID  : 1           PROC : db2agent (instance) 4
    INSTANCE: db2ab7               NODE : 004
    APPHDL  : 4-7                  APPID: *LOCAL.db2ab7.080221131118
    FUNCTION: DB2 UDB, buffer pool services, sqlbInitStorageGroupFiles, probe:50
    MESSAGE : ZRC=0x800201A5=-2147352155=SQLB_AS_INVALID_STORAGE_PATH
              "Storage path does not exist or is inaccessible."
    DATA #1 : String, 17 bytes
    /db2/AB7/sapdata1
    2008-02-21-08.11.19.349637-300 I14176163A619      LEVEL: Severe
    PID     : 835728               TID  : 1           PROC : db2agent (instance) 4
    INSTANCE: db2ab7               NODE : 004
    APPHDL  : 4-7                  APPID: *LOCAL.db2ab7.080221131118
    FUNCTION: DB2 UDB, buffer pool services, sqlbInitStorageGroupFiles, probe:50
    MESSAGE : ZRC=0x800201A5=-2147352155=SQLB_AS_INVALID_STORAGE_PATH
              "Storage path does not exist or is inaccessible."
    DATA #1 : String, 46 bytes
    Error during storage group file initialization
    DATA #2 : Pointer, 8 bytes
    0x0ffffffffffed006
    DATA #3 : Pointer, 8 bytes
    0x00000001110b3080
    2008-02-21-08.11.19.355029-300 I14176783A435      LEVEL: Error
    PID     : 835728               TID  : 1           PROC : db2agent (instance) 4
    INSTANCE: db2ab7               NODE : 004
    APPHDL  : 4-7                  APPID: *LOCAL.db2ab7.080221131118
    FUNCTION: DB2 UDB, base sys utilities, sqleStartDb, probe:5
    RETCODE : ZRC=0x800201A5=-2147352155=SQLB_AS_INVALID_STORAGE_PATH
              "Storage path does not exist or is inaccessible."
    2008-02-21-08.11.19.357831-300 I14177219A370      LEVEL: Warning
    PID     : 835728               TID  : 1           PROC : db2agent (instance) 4
    INSTANCE: db2ab7               NODE : 004
    APPHDL  : 4-7                  APPID: *LOCAL.db2ab7.080221131118
    FUNCTION: DB2 UDB, base sys utilities, sqle_remap_errors, probe:100
    MESSAGE : ZRC 0x800201a5 remapped to SQLCODE -1051
    2008-02-21-08.11.19.374857-300 I14177590A336      LEVEL: Severe
    PID     : 803022               TID  : 1           PROC : db2sysc 4
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, sqleSysCtrlAddNode, probe:6
    MESSAGE : ADD NODE failed with SQLCODE -1051 MESSAGE TOKEN /db2/AB7/sapdata1 in module SQLECRED
    2008-02-21-08.11.19.381604-300 I14177927A440      LEVEL: Event
    PID     : 639412               TID  : 1           PROC : db2start
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 75 bytes
    DB2NODE=4 DB2LPORT=2 /db2/db2ab7/sqllib/adm/db2rstop db2profile NODEACT 4 2
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9C2C : 0000 0024                                  ...$
    2008-02-21-08.11.20.255191-300 I14178368A287      LEVEL: Event
    PID     : 700804               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:240
    DATA #1 : String, 26 bytes
    Stop phase is in progress.
    2008-02-21-08.11.20.258575-300 I14178656A302      LEVEL: Event
    PID     : 700804               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:250
    DATA #1 : String, 41 bytes
    Requesting system controller termination.
    2008-02-21-08.11.20.265164-300 I14178959A403      LEVEL: Warning
    PID     : 803022               TID  : 1           PROC : db2sysc 4
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, routine_infrastructure, sqlerKillAllFmps, probe:5
    MESSAGE : Bringing down all db2fmp processes as part of db2stop
    DATA #1 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFE400 : 0000 0000                                  ....
    2008-02-21-08.11.20.271570-300 I14179363A304      LEVEL: Event
    PID     : 700804               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:260
    DATA #1 : String, 43 bytes
    System controller termination is completed.
    2008-02-21-08.11.20.276550-300 I14179668A381      LEVEL: Event
    PID     : 700804               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:280
    DATA #1 : String, 24 bytes
    There is no active EDUs.
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFCEE0 : 0000 0000                                  ....
    2008-02-21-08.11.20.312260-300 I14180050A342      LEVEL: Severe
    PID     : 774176               TID  : 1           PROC : db2acd 4
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, routine_infrastructure, sqlerFmpOneTimeInit, probe:100
    DATA #1 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFF5A4 : FFFF FBEE                                  ....
    2008-02-21-08.11.20.474332-300 E14180393A301      LEVEL: Event
    PID     : 700804               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 004
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:911
    MESSAGE : ADM7514W  Database manager has stopped.
    STOP    : DB2 DBM
    2008-02-21-08.11.20.600512-300 I14180695A422      LEVEL: Event
    PID     : 671870               TID  : 1           PROC : db2start
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 57 bytes
    /db2/db2ab7/sqllib/adm/db2rstar db2profile SN ADDNODE 5 3
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9C2C : 0000 0011                                  ....
    2008-02-21-08.11.21.620771-300 E14181118A856      LEVEL: Warning
    PID     : 819454               TID  : 1           PROC : db2star2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, license manager, sqllcRequestAccess, probe:1
    MESSAGE : ADM12007E  There are "80" day(s) left in the evaluation period for
              the product "DB2 Enterprise Server Edition". For evaluation license
              terms and conditions, refer to the IBM License Acceptance and License
              Information document located in the license directory in the
              installation path of this product. If you have licensed this product,
              ensure the license key is properly registered. You can register the
              license via the License Center or db2licm command line utility. The
              license file can be obtained from your licensed product CD.
    2008-02-21-08.11.21.839933-300 E14181975A1040     LEVEL: Event
    PID     : 819454               TID  : 1           PROC : db2star2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StartMain, probe:911
    MESSAGE : ADM7513W  Database manager has started.
    START   : DB2 DBM
    DATA #1 : Build Level, 152 bytes
    Instance "db2ab7" uses "64" bits and DB2 code release "SQL09012"
    with level identifier "01030107".
    Informational tokens are "DB2 v9.1.0.2", "special_17253", "U810940_17253", Fix Pack "2".
    DATA #2 : System Info, 224 bytes
    System: AIX sapaix08 3 5 00CCD7FE4C00
    CPU: total:8 online:8 Threading degree per core:2
    Physical Memory(MB): total:7744 free:5859
    Virtual  Memory(MB): total:32832 free:30936
    Swap     Memory(MB): total:25088 free:25077
    Kernel   Params: msgMaxMessageSize:4194304 msgMaxQueueSize:4194304
                     shmMax:68719476736 shmMin:1 shmIDs:131072
                     shmSegments:68719476736 semIDs:131072 semNumPerID:65535
                     semOps:1024 semMaxVal:32767 semAdjustOnExit:16384
    2008-02-21-08.11.22.860106-300 I14183016A428      LEVEL: Error
    PID     : 37336                TID  : 1           PROC : db2agent (instance) 5
    INSTANCE: db2ab7               NODE : 005
    APPHDL  : 5-7                  APPID: *LOCAL.db2ab7.080221131121
    FUNCTION: DB2 UDB, base sys utilities, sqleGetAutomaticStorageDetails, probe:111111
    DATA #1 : <preformatted>
    dataSize 752 pMemAlloc 11099bac0 sizeof(struct sqleAutoStorageCfg) 16
    2008-02-21-08.11.22.886670-300 I14183445A497      LEVEL: Error
    PID     : 37336                TID  : 1           PROC : db2agent (instance) 5
    INSTANCE: db2ab7               NODE : 005
    APPHDL  : 5-7                  APPID: *LOCAL.db2ab7.080221131121
    FUNCTION: DB2 UDB, buffer pool services, sqlbInitStorageGroupFiles, probe:50
    MESSAGE : ZRC=0x800201A5=-2147352155=SQLB_AS_INVALID_STORAGE_PATH
              "Storage path does not exist or is inaccessible."
    DATA #1 : String, 17 bytes
    /db2/AB7/sapdata1
    2008-02-21-08.11.22.889226-300 I14183943A619      LEVEL: Severe
    PID     : 37336                TID  : 1           PROC : db2agent (instance) 5
    INSTANCE: db2ab7               NODE : 005
    APPHDL  : 5-7                  APPID: *LOCAL.db2ab7.080221131121
    FUNCTION: DB2 UDB, buffer pool services, sqlbInitStorageGroupFiles, probe:50
    MESSAGE : ZRC=0x800201A5=-2147352155=SQLB_AS_INVALID_STORAGE_PATH
              "Storage path does not exist or is inaccessible."
    DATA #1 : String, 46 bytes
    Error during storage group file initialization
    DATA #2 : Pointer, 8 bytes
    0x0ffffffffffed006
    DATA #3 : Pointer, 8 bytes
    0x0000000110981080
    2008-02-21-08.11.22.894826-300 I14184563A435      LEVEL: Error
    PID     : 37336                TID  : 1           PROC : db2agent (instance) 5
    INSTANCE: db2ab7               NODE : 005
    APPHDL  : 5-7                  APPID: *LOCAL.db2ab7.080221131121
    FUNCTION: DB2 UDB, base sys utilities, sqleStartDb, probe:5
    RETCODE : ZRC=0x800201A5=-2147352155=SQLB_AS_INVALID_STORAGE_PATH
              "Storage path does not exist or is inaccessible."
    2008-02-21-08.11.22.897320-300 I14184999A370      LEVEL: Warning
    PID     : 37336                TID  : 1           PROC : db2agent (instance) 5
    INSTANCE: db2ab7               NODE : 005
    APPHDL  : 5-7                  APPID: *LOCAL.db2ab7.080221131121
    FUNCTION: DB2 UDB, base sys utilities, sqle_remap_errors, probe:100
    MESSAGE : ZRC 0x800201a5 remapped to SQLCODE -1051
    2008-02-21-08.11.22.913142-300 I14185370A336      LEVEL: Severe
    PID     : 758092               TID  : 1           PROC : db2sysc 5
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, sqleSysCtrlAddNode, probe:6
    MESSAGE : ADD NODE failed with SQLCODE -1051 MESSAGE TOKEN /db2/AB7/sapdata1 in module SQLECRED
    2008-02-21-08.11.22.918953-300 I14185707A440      LEVEL: Event
    PID     : 671870               TID  : 1           PROC : db2start
    INSTANCE: db2ab7               NODE : 000
    FUNCTION: DB2 UDB, base sys utilities, sqleIssueStartStop, probe:1100
    DATA #1 : String, 75 bytes
    DB2NODE=5 DB2LPORT=3 /db2/db2ab7/sqllib/adm/db2rstop db2profile NODEACT 5 3
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFF9C2C : 0000 0024                                  ...$
    2008-02-21-08.11.23.793386-300 I14186148A287      LEVEL: Event
    PID     : 823654               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:240
    DATA #1 : String, 26 bytes
    Stop phase is in progress.
    2008-02-21-08.11.23.796267-300 I14186436A302      LEVEL: Event
    PID     : 823654               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:250
    DATA #1 : String, 41 bytes
    Requesting system controller termination.
    2008-02-21-08.11.23.802154-300 I14186739A403      LEVEL: Warning
    PID     : 758092               TID  : 1           PROC : db2sysc 5
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, routine_infrastructure, sqlerKillAllFmps, probe:5
    MESSAGE : Bringing down all db2fmp processes as part of db2stop
    DATA #1 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFE400 : 0000 0000                                  ....
    2008-02-21-08.11.23.808100-300 I14187143A304      LEVEL: Event
    PID     : 823654               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:260
    DATA #1 : String, 43 bytes
    System controller termination is completed.
    2008-02-21-08.11.23.812951-300 I14187448A381      LEVEL: Event
    PID     : 823654               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:280
    DATA #1 : String, 24 bytes
    There is no active EDUs.
    DATA #2 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFCEE0 : 0000 0000                                  ....
    2008-02-21-08.11.23.882148-300 I14187830A342      LEVEL: Severe
    PID     : 684418               TID  : 1           PROC : db2acd 5
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, routine_infrastructure, sqlerFmpOneTimeInit, probe:100
    DATA #1 : Hexdump, 4 bytes
    0x0FFFFFFFFFFFF5A4 : FFFF FBEE                                  ....
    2008-02-21-08.11.24.008936-300 E14188173A301      LEVEL: Event
    PID     : 823654               TID  : 1           PROC : db2stop2
    INSTANCE: db2ab7               NODE : 005
    FUNCTION: DB2 UDB, base sys utilities, DB2StopMain, probe:911
    MESSAGE : ADM7514W  Database manager has stopped.
    STOP    : DB2 DBM
    2008-02-21-08.41.01.094426-300 I14188475A371      LEVEL: Warning
    PID     : 741576               TID  : 1           PROC : db2bp
    INSTANCE: db2ab7               NODE : 002
    FUNCTION: DB2 UDB, Connection Manager, sqleUCappImpConnect, probe:150
    RETCODE : ZRC=0x8005006D=-2147155859=SQLE_CA_BUILT
              "SQLCA has been built and saved in component specific control block."
    2008-02-21-08.41.01.109657-300 I14188847A371      LEVEL: Warning
    PID     : 741576               TID  : 1           PROC : db2bp
    INSTANCE: db2ab7               NODE : 002
    FUNCTION: DB2 UDB, Connection Manager, sqleUCappImpConnect, probe:150
    RETCODE : ZRC=0x8005006D=-2147155859=SQLE_CA_BUILT
              "SQLCA has been built and saved in component specific control block."
    2008-02-21-08.41.01.115152-300 I14189219A371      LEVEL: Warning
    PID     : 741576               TID  : 1           PROC : db2bp
    INSTANCE: db2ab7               NODE : 002
    FUNCTION: DB2 UDB, Connection Manager, sqleUCappImpConnect, probe:150
    RETCODE : ZRC=0x8005006D=-2147155859=SQLE_CA_BUILT
              "SQLCA has been built and saved in component specific control block."

  • NW 7.3 specific - Database partitioning on top of logical partitioning

    Hello folks,
    In NW 7.3, I would like to know if it is possible to add a specific database partition rule on top of a logical partitioned cube. For example, if I have a LP cube by fiscal year - I would also like to specifically partition all generated cubes at DB level. I could not find any option in the GUI. In addition, each generated cube can be viewed only (cannot be changed in the GUI). Would anybody know if it is possible?
    Thank you
    Ioan

    Fair point! Let me explain more in details what I am looking for - in 7.0x, a cube can be partitioned at the DB level by fiscal period. Let's suppose my cube has only fiscal year 2011 data. If I partition the cube at the DB level by fiscal period in 12 buckets, I will get 12 distinct partitions (E table only) in the database. If the user runs a query on 06/2012, then the DB will search for the data only in the 06/2012 bucket - this is obviously faster than  browsing entire cube (even with indexes).
    In 7.3, cubes can be logical partitioned (LP). I created a LP by fiscal year - so far so good. Now, I would like to partition at the DB level each individual cube created by the LP. Right now I could not - this means that my fiscal year 2012 cube will have entire data residing in only 1 large partition, so a 06/2012 query will take longer (in theory).
    So my question is --> "Is it possible to partition a cube generated by a LP in fiscal period buckets"? I believe the answers is no right now (Dec 2011).
    By the way, all the above is true in a RDBMS environment - this is not a concern for BWA / HANA since data is column based and stored in RAM (not same technology as RDBMS).
    I hope this clarifies by question
    Thank you
    Ioan

Maybe you are looking for

  • Work center reporting

    Hi, In our system in transaction OP4A lot of 'Shift definations', 'shift sequences' are maintained unnecessary. We want to clean this data (Remove unwanted). Is there any report available for work center like 'where used' function to identify shift p

  • Binding value nodes to component in a view

    Hi,       What is the use of these value nodes and value attributes? How can I bind them to, say a inputfield in my view? Can any change to the value node dynamically affect the inutfield?

  • How to change the product category on the item level of a Sales order

    1) For each product there are Basic Product Hierarchy and Sales Product Hierarchy. 2) when creating sales order, system gets the Sales Product Hierarchy on the Item level with the default logic. So our question is: how to change the logic to let syst

  • SDO_3GL error

    Hi, I'm using Oracle spatial (Oracle Database 10g Release 10.2.0.4.0 - 64bit Production) and GeoServer for a mapping application. GeoServer makes the following request as part of its geom boundary calculations: select sdo_aggr_mbr(geom) from table; b

  • N95 Cell Broadcast Question ?

    Hi. I've been used to when "Cell Broadcast" is enabled for Topic 050 that on the display is shown from where I receive the signal. In my N95 I'm only getting mesages telling me from where the signal is received but nothing is shown on the display. Ca