How to create the fact table

pleae let me know how to create the fatc table by using pl/sql packages
Edited by: 792988 on Sep 6, 2010 3:34 AM

Please let us know something about your fact table.
There's no create_fact_table() procedure that could satisfy everybody.

Similar Messages

  • How to create the custom table?

    Hi, how to create the custom table and how to integrate the table with defferent R/3?
    my requiremnt is i have to create the two tables and those i have to integrate with the existed R/3 and using those R/3 i have to update my custom tables .....can give me some idea?
    Tks
    DPk

    how to create the custom table
    There are two approach in creating a table.
    1. Bottom-up approach
    2. Top-down approach.
    Both are valid and you can choose which approach is suitable for you. I always use the bottom-up approach. Here are the steps to create the tables with this approach.
    1. SE11 will take you to the DDIC and enter the name of the new table to be created. Let us say Zname. Click create.
    2. Enter the short discription of the table and enter the field of the table. If it is primary key and you have to check the box.
    3. Enter the data element and double click it, you will be asked to save and will take you to data element discription page. Enter the short discription of the data element and enter the information of domain like the length of field and type of field.
    4. If you wanted to use the existing domain then its fine, or else, you have to create one. Enter the domain name in the data element page and double click it. Page will ask to save and jump to domain creation page.
    5. In the domain page, you have to save the information which you have already given in the data elements page and check it. Before going to data element page, you have to activate the domain.
    6. Go to data element page and save, check and activate.
    7. Go to main table page and save, check, and activate.
    8. Also, you have to save the technical settings of the table.
    The table is now ready for operation. You can use it in your program or you can use it to enter information.
    Check table: It is the table which will have all the information about the Foreign keys which are the primary keys in the check table.
    It can be created by creating the foreign key from the main table. Click foreign key in the main table and it will take you to a page which will ask for table name and field to which foreign key relation has to be associated. Enter the information and you can create the check table automatically.
    SM30 is used for maintenance of the table, that is to realease the errors occured during the creation of the table.
    how to integrate the table with defferent R/3
    Transport the Table to the another server/client/qas/prd
    Kanagaraja L

  • How to create logical fact table in BMM layer ?

    Hello,
    I have 3 Dimension table - 2 are in one schema and last is another schema. Using this 3 dimension tables, I need to create a logical fact table.
    So, my question is whether we can create this fact table by joining these 3 dimension table which are in 2 different schema s ?
    Thanks

    Fiaz,
    you are correct. We can use tables from different subject area to create a report. However, my question was related to rpd design. Sorry, I was not very clear about the queries earlier.
    Here is the whole scenario in the physical layer of the rpd
    Table name      Databse name      Connection pool name      Schema name
    AV          AV_PXRPAM     AVAILABILITY          CRMODDEV
    OUTAGE          AV_PXRPAM     AVAILABILITY          CRMODDEV
    COMPANY          PXRPAM          PXRPAM_POOL          CRMODDEV
    AV and OUTAGE have the joins already. I want to make a join between COMPANY with OUTAGE. And then I want to include a column from each of above tables to the logical fact table in the BMM layer. then I want to do a star schema with the logical fact table to the above 3 tables in the BMM layer.
    Thanks

  • How to change the fact table in backend query

    Hi
    I have a criteria where f1 is the fact table comming in backend query, how can I change/modify so that if i select same criterian it should come different fact table f2.
    Please suggest.

    Hi Hussain,
    I have a measure 'po amount' which is comming from two fact tables cost_f and line_f from physical layer. I have a implict fact column in presentation layer on a column(internal-row count) from cost_f table.
    Now, when I take only 'po amount ' in criteria, in the backend query it should have cost_f table, but i am seeing line_f table.
    The reason I am checking in this direction is.
    I have criteria with 4 columns and measure column 'po amount' and run the result it is fecting from line_f table.
    The same 4 columns and measure column 'po amount' and one new column 'cost center' is added to the critera the fact table changes to cost_f table.
    In both the cases the result should result from cost_f table, not sure why line_f is comming in backend query.
    Please suggest.

  • How to skip the fact table  /BI0/9AEDFC01 error  while import phase in Heterogeneous migration

    Hi.
    Please  find the below  issue of the fact table while import phase in OS/DB migration and enclosed the below log for  reference.
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140924185259
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
    Compiled Nov 23 2013 13:06:03
    -------------------- Start of patch information ------------------------
    patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
    DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
    --------------------- End of patch information -------------------------
    process id 14248
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job completed
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140924185259
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140924185259
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
    Compiled Nov 23 2013 13:06:03
    -------------------- Start of patch information ------------------------
    patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
    DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
    --------------------- End of patch information -------------------------
    process id 14265
    (DB) INFO: connected to DB
    (DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
    (GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
    (GSI) INFO: vname    = "ORACLE                          "
    (GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
    (GSI) INFO: sysname  = "Linux"
    (GSI) INFO: nodename = "VA1WIPRSCM03"
    (GSI) INFO: release  = "2.6.32-358.el6.x86_64"
    (GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
    (GSI) INFO: machine  = "x86_64"
    (SQL) INFO: Searching for SQL file SQLFiles.LST
    (SQL) INFO: SQLFiles.LST not found
    (SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST opened
    (SQL) INFO: Searching for SQL file DFACT.SQL
    (SQL) INFO: DFACT.SQL not found
    (SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL opened
    (SQL) ERROR: Invalid entry at line 5 in file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) ERROR: SQL list was not built successfully
    (DDL) ERROR: check_sql_list() failed for /BI0/9AEDFC01
    (DB) INFO: disconnected from DB
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140924185259
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925104442
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
    Compiled Nov 23 2013 13:06:03
    -------------------- Start of patch information ------------------------
    patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
    DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
    --------------------- End of patch information -------------------------
    process id 23939
    (DB) INFO: connected to DB
    (DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
    (GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
    (GSI) INFO: vname    = "ORACLE                          "
    (GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
    (GSI) INFO: sysname  = "Linux"
    (GSI) INFO: nodename = "VA1WIPRSCM03"
    (GSI) INFO: release  = "2.6.32-358.el6.x86_64"
    (GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
    (GSI) INFO: machine  = "x86_64"
    (SQL) INFO: Searching for SQL file SQLFiles.LST
    (SQL) INFO: SQLFiles.LST not found
    (SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST opened
    (SQL) INFO: Searching for SQL file DFACT.SQL
    (SQL) INFO: DFACT.SQL not found
    (SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL opened
    (SQL) ERROR: Invalid entry at line 5 in file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) ERROR: SQL list was not built successfully
    (DDL) ERROR: check_sql_list() failed for /BI0/9AEDFC01
    (IMP) INFO: a failed DROP attempt is not necessarily a problem
    (SQL) ERROR: SQL list was not built successfully
    (DDL) ERROR: check_sql_list() failed for /BI0/9AEDFC01
    (DB) INFO: disconnected from DB
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140925104442
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925124401
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
    Compiled Nov 23 2013 13:06:03
    -------------------- Start of patch information ------------------------
    patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
    DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
    --------------------- End of patch information -------------------------
    process id 25323
    (DB) INFO: connected to DB
    (DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
    (GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
    (GSI) INFO: vname    = "ORACLE                          "
    (GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
    (GSI) INFO: sysname  = "Linux"
    (GSI) INFO: nodename = "VA1WIPRSCM03"
    (GSI) INFO: release  = "2.6.32-358.el6.x86_64"
    (GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
    (GSI) INFO: machine  = "x86_64"
    (SQL) INFO: Searching for SQL file SQLFiles.LST
    (SQL) INFO: SQLFiles.LST not found
    (SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
    (SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST opened
    (SQL) INFO: Searching for SQL file DFACT.SQL
    (SQL) INFO: DFACT.SQL not found
    (SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
    (SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL opened
    ------------------ C-STACK ----------------------
    R3load[S](LinStackBacktrace+0x8c)[0x48d167]
    R3load[S](LinStack+0x35)[0x6ca6c5]
    R3load[S](CTrcStack2+0x48)[0x48fba1]
    R3load[S](SigIGenAction+0x212)[0x58b4fb]
    libpthread.so.0[T][0x397680f710]
    R3load[S](check_sql_list+0xab0)[0x61bbb0]
    R3load[S](DBDrop+0xbb)[0x60a8fb]
    R3load[S](import+0xde6)[0x620986]
    R3load[S](main_r3ldmain+0x1cfc)[0x6073bc]
    libc.so.6[T](__libc_start_main+0xfd)[0x397601ed5d]
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925125540
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
    Compiled Nov 23 2013 13:06:03
    -------------------- Start of patch information ------------------------
    patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
    DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
    --------------------- End of patch information -------------------------
    process id 25932
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job completed
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140925125540
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925125540
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
    Compiled Nov 23 2013 13:06:03
    -------------------- Start of patch information ------------------------
    patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
    DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
    --------------------- End of patch information -------------------------
    process id 25955
    (DB) INFO: connected to DB
    (DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
    (GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
    (GSI) INFO: vname    = "ORACLE                          "
    (GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
    (GSI) INFO: sysname  = "Linux"
    (GSI) INFO: nodename = "VA1WIPRSCM03"
    (GSI) INFO: release  = "2.6.32-358.el6.x86_64"
    (GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
    (GSI) INFO: machine  = "x86_64"
    (TSK) ERROR: file /tmp/sapinst_instdir/BS2013SR1/SCM703SR1/ORA/COPY/SYSTEM/STD/AS-ABAP/SAPDFACT_1.TSK.bck already seems to exist
                a previous run may not have been finished cleanly
                file /tmp/sapinst_instdir/BS2013SR1/SCM703SR1/ORA/COPY/SYSTEM/STD/AS-ABAP/SAPDFACT_1.TSK possibly corrupted
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
    /usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140925125540

    Dear Ram Nath,
    It may be late but it will be use full to others.
    I faced similar issue and solved it by commenting the 5th line of DFACT.SQL file.
    The issue was with the indexes not being generated properly to avoid such issues we must ensure to implement SAP note 1991576 in prior, if not we can comment the 5th line.
    Below is the corrected file in my case
    # ORACLE : NATIVE SQL EXPORT GENERATED AT 20150426083248
    #ind:  - commented line (earlier it was just 'ind: ')
    ind: /BI0/E0PPM_VC1~0
    tab: /BI0/F0PPM_VC1
    Please let me know in case of queries if any.
    Regards
    Baranedharan S.

  • How to create solved fact table and corresponding cube

    Hello,
    I want to create cube with solved fact table. It means, i need to feed data for higher levels of dimension also from fact table instead of aggreting from base level.
    I am using star schema and AWM 10g R2 for creating cube.
    If anyone knows how to do this, i would be very benificial.
    Thanks
    Subash

    I have generated parent child script using cwm2_olap_pc_transform.create_script. After running this generated script, it has created 3 table/view.
    My Base Parent Child table is like this:-
    drop table PARENT_CHILD;
    create table PARENT_CHILD (PARENT varchar2(30), CHILD varchar2(30));
    insert into PARENT_CHILD values ('Eligible', 'Compliant');
    insert into PARENT_CHILD values ('Eligible', 'Non-Compliant');
    insert into PARENT_CHILD values ('All', 'Eligible');
    insert into PARENT_CHILD values ('All', 'Ineligible');
    insert into PARENT_CHILD values (null, 'All');
    After running generated script thru cwm2_olap_pc_transform.create_script, it has created :-
    Table - PARENT_CHILD_SOLVED
    View - PARENT_CHILD_SOLVED_view , PARENT_CHILD_view
    This script also inserted data in above table/view. (5 rec in each). Table/View structure is like this:-
    SQL> desc PARENT_CHILD_view
    Name Null? Type
    GID NUMBER
    CHILD1 VARCHAR2(30)
    CHILD2 VARCHAR2(30)
    CHILD3 VARCHAR2(30)
    Data:-
    0 All Eligible Compliant
    0 All Eligible Non-Compliant
    1 All Eligible
    1 All Ineligible
    3 All
    SQL> desc PARENT_CHILD_SOLVED
    Name Null? Type
    GID NUMBER
    CHILD1 VARCHAR2(30)
    CHILD2 VARCHAR2(30)
    CHILD3 VARCHAR2(30)
    Data:-
    0 All Eligible Compliant
    0 All Eligible Non-Compliant
    1 All Eligible
    1 All Ineligible
    3 All
    SQL> desc PARENT_CHILD_SOLVED_view
    Name Null? Type
    GID NUMBER
    SHORT_DESCRIPTION VARCHAR2(30)
    LONG_DESCRIPTION VARCHAR2(30)
    CHILD1 VARCHAR2(30)
    CHILD2 VARCHAR2(30)
    CHILD3 VARCHAR2(30)
    Data:-
    0 Compliant Compliant All Eligible Compliant
    0 Non-Compliant Non-Compliant All Eligible Non-Compliant
    1 Eligible Eligible All Eligible
    1 Ineligible Ineligible All Ineligible
    3 All All All
    I tried to create dim and cube based on this. I am not sure whether its correct or not! Though validate_dimension API call shows that it is valid.
    Script for Dim:-
    DECLARE
    -- variable to hold error message
    errtxt varchar(60);
    BEGIN
    -- To be on safer side just drop dimension before creating new one and catch exceptions
    BEGIN
    cwm2_olap_dimension.drop_dimension('APPS', 'HCP_DIM_PC');
    EXCEPTION
    WHEN OTHERS THEN
    dbms_output.put_line('Dimension HCP_DIM_PC not dropped');
    END;
    cwm2_olap_dimension.create_dimension(
    'APPS',
    'HCP_DIM_PC',
    'Parent Child',
    NULL,
    'Parent Child',
    'Parent Child',
    NULL);
    cwm2_olap_dimension_attribute.create_dimension_attribute(
    'APPS',
    'HCP_DIM_PC',
    'Short Description',
    'Short Description',
    'Short Description',
    'Short Description',
    TRUE);
    cwm2_olap_dimension_attribute.create_dimension_attribute(
    'APPS',
    'HCP_DIM_PC',
    'Long Description',
    'Long Description',
    'Long Description',
    'Long Description',
    TRUE);
    cwm2_olap_dimension_attribute.create_dimension_attribute(
    'APPS',
    'HCP_DIM_PC',
    'Grouping ID',
    'Grouping ID',
    'Grouping ID',
    'Grouping ID',
    TRUE);
    cwm2_olap_hierarchy.create_hierarchy(
    'APPS',
    'HCP_DIM_PC',
    'HCP_DIM_PC_HIER',
    'Standard',
    'Standard',
    'Standard Parent Child Hierarchy',
    'SOLVED LEVEL-BASED');
    cwm2_olap_dimension.SET_DEFAULT_DISPLAY_HIERARCHY ('APPS', 'HCP_DIM_PC', 'HCP_DIM_PC_HIER');
    cwm2_olap_level.create_level(
    'APPS',
    'HCP_DIM_PC',
    'ALL_PARENT_LVL',
    'All Parent Child',
    'All Parent Child',
    'All Parent Child',
    'All Parent Child Level');
    cwm2_olap_level_attribute.create_level_attribute(
    'APPS',
    'HCP_DIM_PC',
    'Short Description',
    'ALL_PARENT_LVL',
    'Short Description',
    'Short Description',
    'Short Description',
    'Short Description',
    TRUE);
    cwm2_olap_level_attribute.create_level_attribute(
    'APPS',
    'HCP_DIM_PC',
    'Long Description',
    'ALL_PARENT_LVL',
    'Long Description',
    'Long Description',
    'Long Description',
    'Long Description',
    TRUE);
    cwm2_olap_level_attribute.create_level_attribute(
    'APPS',
    'HCP_DIM_PC',
    'Grouping ID',
    'ALL_PARENT_LVL',
    'Grouping ID',
    'Grouping ID',
    'Grouping ID',
    'HTB Grouping ID',
    TRUE);
    -- Add all levels one by one to dimension hierarchy. For top most level last parameter is null.
    cwm2_olap_level.add_level_to_hierarchy(
    'APPS',
    'HCP_DIM_PC',
    'HCP_DIM_PC_HIER',
    'ALL_PARENT_LVL',
    NULL);
    BEGIN
    cwm2_olap_table_map.removemap_dimtbl_hierlevel(
    'APPS',
    'HCP_DIM_PC',
    'HCP_DIM_PC_HIER',
    'ALL_PARENT_LVL');
    cwm2_olap_table_map.removemap_DimTbl_HierLevelAttr(
    'APPS', 'HCP_DIM_PC', 'Short Description', 'HCP_DIM_PC_HIER', 'ALL_PARENT_LVL', 'Short Description');
    cwm2_olap_table_map.removemap_DimTbl_HierLevelAttr(
    'APPS', 'HCP_DIM_PC', 'Long Description', 'HCP_DIM_PC_HIER', 'ALL_PARENT_LVL', 'Long Description');
    cwm2_olap_table_map.removemap_DimTbl_HierLevelAttr(
    'APPS', 'HCP_DIM_PC', 'Grouping ID', 'HCP_DIM_PC_HIER', 'ALL_PARENT_LVL', 'Grouping ID');
    EXCEPTION
    WHEN OTHERS THEN
    dbms_output.put_line('Level map for ALL_PARENT_LVL not removed');
    END;
    -- Map ALL_PARENT_LVL level to dimension table. Last parameter is null since it is top most level
    cwm2_olap_table_map.map_dimtbl_hierlevel(
    'APPS',
    'HCP_DIM_PC',
    'HCP_DIM_PC_HIER',
    'ALL_PARENT_LVL',
    'APPS',
    'PARENT_CHILD_SOLVED_VIEW',
    'GID',
    NULL);
    -- one by one map all the level attributes to respective columns in the dimension table.
    cwm2_olap_table_map.Map_DimTbl_HierLevelAttr(
    'APPS',
    'HCP_DIM_PC',
    'Short Description',
    'HCP_DIM_PC_HIER',
    'ALL_PARENT_LVL',
    'Short Description',
    'APPS',
    'PARENT_CHILD_SOLVED_VIEW',
    'SHORT_DESCRIPTION');
    cwm2_olap_table_map.Map_DimTbl_HierLevelAttr(
    'APPS',
    'HCP_DIM_PC',
    'Long Description',
    'HCP_DIM_PC_HIER',
    'ALL_PARENT_LVL',
    'Long Description',
    'APPS',
    'PARENT_CHILD_SOLVED_VIEW',
    'LONG_DESCRIPTION');
    cwm2_olap_table_map.Map_DimTbl_HierLevelAttr(
    'APPS',
    'HCP_DIM_PC',
    'Grouping ID',
    'HCP_DIM_PC_HIER',
    'ALL_PARENT_LVL',
    'Grouping ID',
    'APPS',
    'PARENT_CHILD_SOLVED_VIEW',
    'GID');
    -- Use cwm2_olap_validate.validate_dimension to validate the dimension
    cwm2_olap_validate.validate_dimension('APPS', 'HCP_DIM_PC');
    cwm2_olap_metadata_refresh.mr_refresh;
    -- Rollback if any exception occurs during processing this script
    EXCEPTION
    WHEN OTHERS THEN
    cwm_utility.dump_error;
    errtxt := cwm_utility.get_last_error_description;
    dbms_output.put_line('ERROR: ' || errtxt);
    ROLLBACK;
    RAISE;
    END;
    Script for Cube:-
    declare
    HCP_time_dim number;
    errtxt varchar(60);
    begin
    cwm_utility.collect_garbage;
    begin
    cwm2_olap_cube.drop_cube('APPS', 'HCP_PC_CUBE');
    exception
    when others then
    dbms_output.put_line('No HCP_PC_CUBE to drop');
    end;
    begin
    cwm2_olap_catalog.drop_catalog('HCP_PC_CAT');
    exception
    when others then
    dbms_output.put_line('No HCP_PC_CAT to drop');
    end;
    CWM2_OLAP_CUBE.Create_Cube('APPS', 'HCP_PC_CUBE', 'Parent Child Cube', 'Parent Child Cube','Parent Child Cube');
    cwm2_olap_cube.add_dimension_to_cube('APPS', 'HCP_PC_CUBE','APPS', 'HCP_DIM_PC');
    cwm2_olap_measure.create_measure('APPS', 'HCP_PC_CUBE', 'HCP_PC_MEASURE', 'PC Measure','PC Measure', 'PC Measure Fact');
    cwm2_olap_table_map.map_facttbl_levelkey('APPS','HCP_PC_CUBE','APPS','HCP_PC_FACT','ET', 'DIM:APPS.HCP_DIM_PC/HIER:HCP_DIM_PC_HIER/GID:GID/LVL:ALL_PARENT_LVL/COL:CHILD3;');
    cwm2_olap_table_map.Map_FactTbl_Measure('APPS', 'HCP_PC_CUBE','HCP_PC_MEASURE', 'APPS', 'HCP_PC_FACT', 'MEASURE_COL', 'DIM:APPS.HCP_DIM_PC/HIER:HCP_DIM_PC_HIER/GID:GID/LVL:CHILD3/COL:SHORT_DESCRIPTION;');
    cwm2_olap_catalog.create_catalog('HCP_PC_CAT', 'Parent Child Catalog');
    cwm2_olap_catalog.add_catalog_entity('HCP_PC_CAT', 'APPS', 'HCP_PC_CUBE', 'HCP_PC_MEASURE');
    cwm2_olap_validate.validate_cube('APPS', 'HCP_PC_CUBE','DEFAULT','YES');
    cwm2_olap_metadata_refresh.mr_refresh;
    exception
    when others then
    cwm_utility.dump_error;
    errtxt := cwm_utility.get_last_error_description;
    dbms_output.put_line('ERROR: ' || errtxt);
    rollback;
    raise;
    END;
    My Fact Table is :-
    DROP TABLE HCP_PC_FACT CASCADE CONSTRAINT;
    CREATE TABLE HCP_PC_FACT (
    SHORT_DESCRIPTION VARCHAR2(30) NOT NULL,
    GID NUMBER NOT NULL,
    CHILD1 VARCHAR2(30) NOT NULL,
    CHILD2 VARCHAR2(30) ,
    CHILD3 VARCHAR2(30) ,
    MEASURE_COL NUMBER NOT NULL);
    Data in Fact Table:-
    insert into HCP_PC_FACT values('Compliant',0,'All','Eligible','Compliant', 100);
    insert into HCP_PC_FACT values('Non-Compliant',0,'All','Eligible','Non-Compliant', 200);
    insert into HCP_PC_FACT values('Eligible',1,'All','Eligible',null, 300);
    insert into HCP_PC_FACT values('Ineligible',1,'All','Ineligible',null, 400);
    insert into HCP_PC_FACT values('All',3,'All',null,null, 500);
    I am not sure how to create level, level attr, dim attr for such dim.
    All CWM2 validation API shows that my all dims, cubes are valid but when I try to create presentation thru JDev it hangs after selecting Parent-Child Measure.
    Any complete working exapmle will be helpful.
    P.S. One more query I have can we have one solved and one un-solved dim in the same cube/measure?
    regds
    Prakash

  • How to create the Hierarchial table using ALV.in WD4A

    Hi All,
    I am struggling hard to make the hierarchial ALV,  But could not do it.
    Could you help me.
    Actually I want the column to be hierarchial. For this i am using
    if_salv_wd_column_hierarchy~set_hierarchy_column.
    Also i tried to display the whole table as Hierarchial
        lr_table_settings->set_display_type( if_salv_wd_c_table_settings=>display_type_hierarchy ).
    For both the possibilities the result is not as desired.
    Regards,
    Arti.

    Hi Aleem,
    The link u have provided gives the solution for the Table.
    I want the implementation for ALV.
    Also my requirement is that the first column of the ALV should be hierarchial.
    On expanding the table cell of first column one more table should open of same type.
    This way user can navigate to three diffierent database tables.
    The tree is populated at runtime with the data. So the depth of branches is unknown at design time.
    I hope the requirements are clear to u now.
    Thanks and Regards,
    Arti.

  • How to create the temp table in Sybase with values from Excel ?

    Dear Sir/Madam,
    I know this is not related with oracle DB, however i am in the need of help from you people who worked on Sybase. plz help me..
    I have the excel sheet which contains EmpId and EmpName values. I just wanted to read these two values one by one from an excel sheet and loaded them into one temp table in Sybase. is it possible thru sqlldr ?
    here is the sample code for your reference:
    begin tran
    create table tempdb..empIdName (emp_id int identity,emp_name varchar(50))
    go
    Awaiting for your valuable reply
    thanks
    pannar

    there is some hint provided by sybase user
    If it's Sybase IQ, you can export the excel file to a CSV file, then use a LOAD TABLE statement to get the data into your temporary table.
    For ASE, BCP would provide similar functionality.
    I am using ISQLW tool to work with sybase DB. what query to load the excel data to tem table in sybase. anyone who knows this? help me

  • How to update a fact table when a dimension table is reloaded

    We have implemented BI Apps 796. Insertion into W_EMPLOYEE_D table which stores all the employee information had stopped one year back as some company security policy restricted the informatica worklfows to pick up the data. (PER_ALL_PEOPLE_F was a HRMS table and it contained sensitive information line SSN and salary, was inaccessible to the user which informatica uses and the SDE mapping used to return 0 rows).
    Now we have the approval to see those rows and the dimension table is loaded with some 100 new employees who joined in last one year.
    The ROW_WID of W_EMPLOYEE_D is referenced in lot of fact tables and for all those missing employees the WID in the fact table is 0.
    Now that we have all employees, how to make the FACT table point to the correct WID and not store 0. Has anyone faced this problem before?? Writing an update statement will be a tedious task as there are so many fact tables that join to w_employee_d. Also our company uses Sales, Procurement, Finance modules of OB Apps (which constitutes atleast 20 fact tables)
    Any guidance is appreciated. Thanks in advance

    Hello Kostis,
    thank you for your answer. I don't fully understand you. Can you show me short example, please? I create alias table for time dimension on Physical Layer - original table is TimeDayDim and I create aliases TimeDayDim1, TimeDayDim2, TimeDayDim3, TimeDayDim4. Then I create foreign key Fact.Time1 -> TimeDayDim1, Fact.Time2 -> TimeDayDim2, Fact.Time3 -> TimeDayDim3, Fact.Time4 -> TimeDayDim4. And what now? Must I create these table api Bussines Model and create new time dimensions at bussiness model????
    I need in Answers ONE Time dimension. I think I must split my fact table to four tables ... (time1, place1 ...) (time2, place2 ...) (time3 place3...) (time4 place4...) then link those tables to Time dimension (but I dont know where I can split those tables - on Physical Layer or on Bussines Layer).
    I suppose that I will have in Answers one time dimension and four facts tables and I will be able to query them. (for example: Time.Days, Fact1.Place1, Fact3.Speed, Fact4.Count Criteria: Time.Year = 2008)
    Best Regards Vlada

  • Trying to create a fact table,error:unable to extend temp segement

    Using Oracle 10.2g
    SQL>
    create materialized view facts_table
    (     s_id
    ,     g_id
    ,     sb_id
    ,     sc_id
    ,     y_id)
    refresh with rowid
    as      select s.s_id
    ,     g.g_id
    ,     sb.sb_id
    ,     sc.sc_id
    ,     y.academicyear
    from      student s
    ,     grade g
    ,     subject sb
    ,     school sc
    ,     comqdhb.teachinggroup y;
    ERROR at line 3:
    ORA-01652: unable to extend temp segment by 1024 in tablespace POSTGRADSI am creating the fact table as materialized view because its specified for us.
    I am trying to create a fact table from the dimension tables and its giving this error
    what is the mistake kindly help.
    Also when creating the fact table do all the columns need to be foreign keys compulsorily.
    Edited by: Trooper on Jan 10, 2009 5:25 AM
    Edited by: Trooper on Jan 10, 2009 6:37 AM

    Well basically what your saying is absolutely right
    I realized what i am doing right now is stupendous blunder.
    Basically my aim is there are 5 dimensional tables that are created
    Student->s_id primary key,upn(unique pupil no),name
    Grade->g_id primary key,grade,exam_level,values
    Subject->sb_id primary key,subjectid,subname
    School->sc_id primary key,schoolno,school_name
    year->y_id primary key,year(like 2008)
    s_id,g_id,sb_id,sc_id,y_id are sequences
    select * from student;
    S_ID UPN FNAME COMMONNAME GENDER DOB
    ==============================
    9062 1027 MELISSA ANNE       f  13-OCT-81
    9000 rows selected
    select * from grade;
          G_ID GRADE      E_LEVEL         VALUE
            73 A          a                 120
            74 B          a                 100
            75 C          a                  80
            76 D          a                  60
            77 E          a                  40
            78 F          a                  20
            79 U          a                   0
            80 X          a                   0
    18 rows selectedThese are basically the dimensional views
    Now according to the specification given, need to create a fact table as facts_table which contains all the dim tables primary keys as foreign keys in it.
    The problem is when i say,I am going to consider a smaller example than the actual no of dimension tables 5 lets say there are 2 dim tables student,grade with s_id,g_id as p key.
    create materialized view facts_table(s_id,g_id)
    as
    select  s.s_id,g.g_id
    from   (select distinct s_id from student)s
    ,         (select distinct g_id from grade)gThis results in massive duplication as there is no join between the two tables.But basically there are no common things between the two tables to join,how to solve it?
    Consider it when i do it for 5 tables the amount of duplication being involved, thats why there is not enough tablespace.
    I was hoping if there is no other way then create a fact table with just one column initially
    create materialized view facts_table(s_id)
    as
    select s_id
    from student;then
    alter materialized view facts_table add column g_id number;Then populate this g_id column by fetching all the g_id values from the grade table using some sort of loop even though we should not use pl/sql i dont know if this works?
    Any suggestions.
    Edited by: Trooper on Jan 10, 2009 6:38 AM

  • How and when to create a fact table

    hi every1,
    I am new to oracle and specially to Data Warehousing. I am using OWB 11g on windows XP.
    I have created the dimensional for a database, now i am in the implementation phase.
    i have the following questions.
    1. I don't know how and when to create a fact table. I have already created the dimensions (ROLAP).
    2. Do i need to create a mapping for a Cube. if yes, then how to load them from 4 different dimensions. how will be the join condition, they don't match?
    3. I will also try for aggregation, do i need to use MOLAP for the dimensions or ROLAP works?
    Please, give a quick reply.
    regards
    RF

    You should ask your question here:
    Warehouse Builder
    where it is on topic.

  • How map to my particular table to the fact table in obiee 11g...

    Hi friends,
    I did this simple report in obiee 11g(i.e)
    "NATIONALITY COUNT IN DEPARTMENT WISE"
    For that i used the following tables:
    per_all_assignments_f----->fact table
    hr_all_organization_units----->dim table(containing departments)
    per_all_people_f---------------->dim table(containing nationality)
    I made all the mappings in the physical diagram, as also viewed my report in BI answers
    It shows the following results like
    NATIONALITY---------------------------------------------------------------------COUNT(NATIONALITY)
    AUS------------------------------------------------------------------------------------------------24
    AFR------------------------------------------------------------------------------------------------25
    PHQ_VB-------------------------------------------------------------------------------------------40
    SH_VT----------------------------------------------------------------------------------------------4
    The problem is for me it is showing the above results, but the nationality column is of various codes of the country.
    Since i doesnt want the code of the nationalitian to display in the results..i need the meaning of each and every nationality..
    like,
    AUS------------------------Australian
    AFR-------------------------African
    PHQ_VB----------------------Germanian(assigned)
    Since i know that the meaning for the nationalitian is available in "FND_LOOKUP_VALUES"...okay..
    I can import "FND_LOOKUP_VALUES" table to the physical layer...but how i can able to give the mapping to the fact table in my physical diagram...
    In my report the fact table is "per_all_assignments_f"
    As my fact table doesnt contains any matching column corresponding to the dimension table "FND_LOOKUP_VALUES".....
    Then how i can give mappings to the fact column???? for viewing the full meaning of the nationalitian in my report.....Help me friends...
    Regards,
    Harry...

    Hi bifact,
    I followed the step that U asked me to go but im stuck with later.....
    *) I saved the query that i executed in toad with the columns showing country codes and country meaning to the excel sheet...
    This is the query that i executed in toad and copied data s in excel sheet..
    select z.lookup_code, z.meaning
    from per_all_people_f e, per_all_assignments_f f, hr_all_organization_units h, fnd_lookup_values z where
    e.person_id = f.person_id and f.organization_id = h.organization_id and e.business_group_id = f.business_group_id
    and f.business_group_id = h.business_group_id and f.location_id = h.location_id
    AND z.lookup_type(+) = 'NATIONALITY'AND z.lookup_code(+) = e.nationality and sysdate between
    e.effective_start_date and e.effective_end_date and e.nationality is not null
    *) After that i created a system DSN for the excel drive..
    *) After that when i tried to import metadata of the excel data that is saved, it showed me connection failed...
    For importing this excel data
    what connection details i need to give:--------------------
    as well as user name and also password...
    Soon after importing this excel you said that to give key connection only to fnd_lookup_tables(dim) and that excel data..
    if so, again im not giving key connection to my fact table..
    I think again the same error will occur that no logical mapping is not made to fact so again repository is inconsistent...
    Hi bifact sorry to ask you, but can you see the steps that i followed is correct and after that wat more steps i need to proceed..Can you tell me in breif manner...Thanks for your help...
    Regards,
    Harry...

  • How to create the Table of Content (TOC) in WEB

    Dear Sir
    Please help me
    How to create the Table of content (TOC) in Web Intelligence Report,
    Thanks
    Gnanasekarn.K
    Edited by: Gnanasekaran Kandasamy on Nov 20, 2008 11:12 AM

    Hi Gnanasekarn,
    You can create TOC in webi using Open Document, You can create link on the TOC and you can connect that link to open new report .
    With the help of open document you can open the specific portion of report.
    So you can achive your requirement using URL reporting or Open Document.
    Regards,
    Chaitanya Deshpande

  • How to create a temp table in the memory, not in disk?

    in sql server, you can create a temp table in the memory instead of disk,
    then you can do the insert, delete,update and select on it.
    after finishing, just release it.
    in Oracle,
    I am wonderfing how to create a temp table in the memory, not in disk?
    thanks,

    Thanks for rectifying me Howard.
    I just read your full article on this too and its very well explained here:
    http://www.dizwell.com/prod/node/357
    Few lines from your article
    It is true, of course, that since Version 8.0 Oracle has provided the ability to create a Keep Pool in the Buffer Cache, which certainly sounds like it can do the job... especially since that word 'keep' is used again. But a keep pool is merely a segregated part of the buffer cache, into which you direct blocks from particular tables (by creating them, or altering them, with the BUFFER POOL KEEP clause). So you can tuck the blocks from such tables out of the way, into their own part of the buffer cache... but that is not the same thing as guaranteeing they'll stay there. If you over-populate the Keep Pool, then its LRU mechanism will kick in and age its contents out just as efficiently as an unsegregated buffer cache would.
    Functionally, therefore, there can be no guarantees. The best you can do is create a sufficiently large Keep Pool, and then choose the tables that will use it with care such that they don’t swamp themselves, and start causing each other to age out back to disk.
    Thanks and Regards

  • How to create a dynamic table were the JTable columns keep varying

    How to create a dynamic table were the JTable columns keep varying based on the input to the jtable

    Oooh, I lied. DefaultTableModel has an API for adding and
    removing columns. I didn't know that. You should have read
    the API.
    As for preferring to extend AbstractTableModel rather than
    DefaultTableModel, I think it's more correct. DefaultTableModel
    is a simple implementation of Abstract for basic cases. It isn't
    intended to be extended. I figure most people extending
    DefaultTableModel are also extending JFrame, JPanel, and Thread
    instead of encapsulating the first two and implementing
    Runnable for the third.

Maybe you are looking for

  • HP LaserJet Pro PW1102W Networking Issue

    Hi, We had some issue with our network provider recently and now back online however the wireless function on the printer does not work. I have tried following numerous steps outlined online with no luck, details below. Actions Taken: 1) Uninstall /

  • Cannot restore from Time Capsule after hard drive replacement (recall)!

    Hard drive was replaced on a recall from Apple. I have spent weeks on the phone with Applecare whose conclusion now is that a WD drive I had connected about a year ago (and which was disconnected and software uninstalled) has corrupted the files on t

  • Photo Transfer PC to IMAC

    Just purchased an imac running Mountain Lion OSX 8.3 and am trying to dowload photos from Elements 7 that are on my previous PC running Win XP.  Have just dowloaded Elements 11 on the imac and would appreciate any advice on getting my7000+ photos ont

  • WLC 5508 802.1x with AES

    Hi, We have a staff WLAN on Cisco WLC 5508. We use 802.1x with TKIP with authentication from RADIUS server. We deployed new 802.11n APs but on staff WLAN we cannot enable 802.11n because of the TKIP encryption. Can we just simply change the encryptio

  • Currency and exhange rates data distribution to flat file

    Hi All I have a situation like i am sending the all  BI data to flat files but my problem is how can i send the currency exhange rates(TABLE:TCURR) data to flat files, is it possible using open hub desination or any other methods available please sug