How to skip the fact table  /BI0/9AEDFC01 error  while import phase in Heterogeneous migration

Hi.
Please  find the below  issue of the fact table while import phase in OS/DB migration and enclosed the below log for  reference.
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140924185259
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
Compiled Nov 23 2013 13:06:03
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
--------------------- End of patch information -------------------------
process id 14248
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job completed
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140924185259
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140924185259
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
Compiled Nov 23 2013 13:06:03
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
--------------------- End of patch information -------------------------
process id 14265
(DB) INFO: connected to DB
(DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
(GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
(GSI) INFO: vname    = "ORACLE                          "
(GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
(GSI) INFO: sysname  = "Linux"
(GSI) INFO: nodename = "VA1WIPRSCM03"
(GSI) INFO: release  = "2.6.32-358.el6.x86_64"
(GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
(GSI) INFO: machine  = "x86_64"
(SQL) INFO: Searching for SQL file SQLFiles.LST
(SQL) INFO: SQLFiles.LST not found
(SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST opened
(SQL) INFO: Searching for SQL file DFACT.SQL
(SQL) INFO: DFACT.SQL not found
(SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL opened
(SQL) ERROR: Invalid entry at line 5 in file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) ERROR: SQL list was not built successfully
(DDL) ERROR: check_sql_list() failed for /BI0/9AEDFC01
(DB) INFO: disconnected from DB
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140924185259
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925104442
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
Compiled Nov 23 2013 13:06:03
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
--------------------- End of patch information -------------------------
process id 23939
(DB) INFO: connected to DB
(DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
(GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
(GSI) INFO: vname    = "ORACLE                          "
(GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
(GSI) INFO: sysname  = "Linux"
(GSI) INFO: nodename = "VA1WIPRSCM03"
(GSI) INFO: release  = "2.6.32-358.el6.x86_64"
(GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
(GSI) INFO: machine  = "x86_64"
(SQL) INFO: Searching for SQL file SQLFiles.LST
(SQL) INFO: SQLFiles.LST not found
(SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST opened
(SQL) INFO: Searching for SQL file DFACT.SQL
(SQL) INFO: DFACT.SQL not found
(SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL opened
(SQL) ERROR: Invalid entry at line 5 in file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) ERROR: SQL list was not built successfully
(DDL) ERROR: check_sql_list() failed for /BI0/9AEDFC01
(IMP) INFO: a failed DROP attempt is not necessarily a problem
(SQL) ERROR: SQL list was not built successfully
(DDL) ERROR: check_sql_list() failed for /BI0/9AEDFC01
(DB) INFO: disconnected from DB
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140925104442
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925124401
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
Compiled Nov 23 2013 13:06:03
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
--------------------- End of patch information -------------------------
process id 25323
(DB) INFO: connected to DB
(DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
(GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
(GSI) INFO: vname    = "ORACLE                          "
(GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
(GSI) INFO: sysname  = "Linux"
(GSI) INFO: nodename = "VA1WIPRSCM03"
(GSI) INFO: release  = "2.6.32-358.el6.x86_64"
(GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
(GSI) INFO: machine  = "x86_64"
(SQL) INFO: Searching for SQL file SQLFiles.LST
(SQL) INFO: SQLFiles.LST not found
(SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST
(SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/SQLFiles.LST opened
(SQL) INFO: Searching for SQL file DFACT.SQL
(SQL) INFO: DFACT.SQL not found
(SQL) INFO: Searching for SQL file /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: found /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: Trying to open /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL
(SQL) INFO: /data1/SCMEXPORT/EXPAQ2/ABAP/DB/ORA/DFACT.SQL opened
------------------ C-STACK ----------------------
R3load[S](LinStackBacktrace+0x8c)[0x48d167]
R3load[S](LinStack+0x35)[0x6ca6c5]
R3load[S](CTrcStack2+0x48)[0x48fba1]
R3load[S](SigIGenAction+0x212)[0x58b4fb]
libpthread.so.0[T][0x397680f710]
R3load[S](check_sql_list+0xab0)[0x61bbb0]
R3load[S](DBDrop+0xbb)[0x60a8fb]
R3load[S](import+0xde6)[0x620986]
R3load[S](main_r3ldmain+0x1cfc)[0x6073bc]
libc.so.6[T](__libc_start_main+0xfd)[0x397601ed5d]
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925125540
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
Compiled Nov 23 2013 13:06:03
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
--------------------- End of patch information -------------------------
process id 25932
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job completed
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140925125540
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: START OF LOG: 20140925125540
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#6 $ SAP
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: version R7.40/V1.8 [UNICODE]
Compiled Nov 23 2013 13:06:03
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.009) Support for SUM/ZDM and DMO (note 1778564)
DBSL patchinfo (patches.h): (0.011) DBSL error corrections in 7.41: (4) LOBs (note 1928526)
--------------------- End of patch information -------------------------
process id 25955
(DB) INFO: connected to DB
(DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): UTF16
(GSI) INFO: dbname  = "AQ220140924040917                                                                                                              "
(GSI) INFO: vname    = "ORACLE                          "
(GSI) INFO: hostname = "VA1WIPRSCM03                                                    "
(GSI) INFO: sysname  = "Linux"
(GSI) INFO: nodename = "VA1WIPRSCM03"
(GSI) INFO: release  = "2.6.32-358.el6.x86_64"
(GSI) INFO: version  = "#1 SMP Tue Jan 29 11:47:41 EST 2013"
(GSI) INFO: machine  = "x86_64"
(TSK) ERROR: file /tmp/sapinst_instdir/BS2013SR1/SCM703SR1/ORA/COPY/SYSTEM/STD/AS-ABAP/SAPDFACT_1.TSK.bck already seems to exist
            a previous run may not have been finished cleanly
            file /tmp/sapinst_instdir/BS2013SR1/SCM703SR1/ORA/COPY/SYSTEM/STD/AS-ABAP/SAPDFACT_1.TSK possibly corrupted
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
/usr/sap/AQ2/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140925125540

Dear Ram Nath,
It may be late but it will be use full to others.
I faced similar issue and solved it by commenting the 5th line of DFACT.SQL file.
The issue was with the indexes not being generated properly to avoid such issues we must ensure to implement SAP note 1991576 in prior, if not we can comment the 5th line.
Below is the corrected file in my case
# ORACLE : NATIVE SQL EXPORT GENERATED AT 20150426083248
#ind:  - commented line (earlier it was just 'ind: ')
ind: /BI0/E0PPM_VC1~0
tab: /BI0/F0PPM_VC1
Please let me know in case of queries if any.
Regards
Baranedharan S.

Similar Messages

  • None of the fact tables are compatible error

    Hi All,
    I do see this error (none of the compatible fact table) after setting the content level aggregation on the dimension tables and the fact table. This error i get only when i try to pull the calculated item which is based on a attribute in the fact table. I have an attribute like year in the fact table i need to display like 'CY'||'2013' in a calculated logical column and when i pull this into answers i get this error -
    1) joins are ok ; only one fact table and 3 dimension tables
    2). content level on the fact table are specified at the detail level and also for the dimensions
    any suggestions - thanks for your time

    can anyone please provide some suggestions -
    > i looked at the fact table LTS and specified the logical level for each dimension as the detail
    > specified the LTS for each dimension table
    > I have a column in my fact table which is calendar year and i want to have a derived column like rep_cal_year with 'CY'||cal_year - so when i pull this derived column in my answers i get the error - none of the fact tables are compatibile with the query;
    what could be missign?

  • How to change the fact table in backend query

    Hi
    I have a criteria where f1 is the fact table comming in backend query, how can I change/modify so that if i select same criterian it should come different fact table f2.
    Please suggest.

    Hi Hussain,
    I have a measure 'po amount' which is comming from two fact tables cost_f and line_f from physical layer. I have a implict fact column in presentation layer on a column(internal-row count) from cost_f table.
    Now, when I take only 'po amount ' in criteria, in the backend query it should have cost_f table, but i am seeing line_f table.
    The reason I am checking in this direction is.
    I have criteria with 4 columns and measure column 'po amount' and run the result it is fecting from line_f table.
    The same 4 columns and measure column 'po amount' and one new column 'cost center' is added to the critera the fact table changes to cost_f table.
    In both the cases the result should result from cost_f table, not sure why line_f is comming in backend query.
    Please suggest.

  • How to create the fact table

    pleae let me know how to create the fatc table by using pl/sql packages
    Edited by: 792988 on Sep 6, 2010 3:34 AM

    Please let us know something about your fact table.
    There's no create_fact_table() procedure that could satisfy everybody.

  • Table not found error while importing table defintion in PI 7.1?

    Hi Guys,
    I am trying to import the table structure from DB2 as an external defintion but i am getting the error table not found.
    There are no connection issues with the DB2 and CC is good. Are there any additional settings i need to perform to import the table?
    any help or suggestions would be really appreciated
    Thanks,
    Srini

    Hi Srinivas,
    I think you want to do JDBC lookup.. you must have created the JDBC receiver communication channel. As your communication channel is correct and still you are getting Table not found error.. Check following steps.
    - Check that your Communication channel is activated and working fine.
    - Check that the ID which you are using in JDBC receiver is having proper authorization to import the table definition.
      Just check with all authorization.
    I think the problem is with insufficient authorization to User ID used on JDBC communication channel.
    Thanks,
    Bhupesh.

  • How to pass the Dynamically table name and RELID in IMPORT  x TO y FROM DATABASE dbtab(ar) ID Key

    Hello Expert,
    I wanted to know is there any way through which we can dynamically pass the table name in IMPORT condition.
    For e.g.
    DATA : PCLT TYPE DD03L-TABNAME.
    DATA: SET_RELID TYPE T52RELID-RELID.
    DATA: key TYPE PCLKEY.
    DATA: text_version TYPE PVRSN.
    DATA: BEGIN OF ptext OCCURS 132.
    DATA: line(132).
    DATA: END OF ptext.
    SET_RELID = 'TX'.
    PCLT = 'PCL1'.
    KEY = '00001000004005     9999123119960101000'.
    TEXT_VERSION = '01'.
    IMPORT TEXT_VERSION PTEXT FROM DATABASE <PCLT>(<SET_RELID>) ID KEY.
    If i use this then it gives me error message "The Dictionary structure or table "<PCLT>" is either not active or does not exist."
    Kindly help me in this matter.
    Thanks
    Srikanth

    Hi Srikanth
    I dont know if there is another solution, but this is what i did using a dynamic subroutine, please check it
    just add this lines to your code
    DATA: prog  TYPE c LENGTH 8,
           mess TYPE string,
           sid  TYPE string,
           wrd  TYPE string,
           dir  TYPE trdir,
           OFF(3).
    DATA: lt_code TYPE TABLE OF rssource-line.
    DATA: ls_code TYPE rssource-line.
    perform call_import using pclt set_relid key.
    form call_import USING p_pclt p_set_relid p_key.
       ls_code = 'PROGRAM SUBPOOL.'.
       append ls_code to lt_code.
       ls_code = 'form get_text tables ptext using p_key changing text_version.'.
       append ls_code to lt_code.
       ls_code = 'IMPORT text_version ptext from database'.
       append ls_code to lt_code.
       concatenate p_pclt '(' p_set_relid ')' into ls_code.
       append ls_code to lt_code.
       ls_code = 'id p_key.'.
       append ls_code to lt_code.
       ls_code = 'endform.'.
       append ls_code to lt_code.
       generate subroutine pool lt_code
                           name prog
                           message mess
                           SHORTDUMP-ID sid
                           WORD WRD
                           OFFSET OFF.
       IF SY-SUBRC EQ 0.
         perform GET_TEXT IN PROGRAM (prog) tables ptext
                                              USING p_key
                                              CHANGING text_version .
       ENDIF.
    endform.
    Hope this help
    Regards
    MC

  • How to update a fact table when a dimension table is reloaded

    We have implemented BI Apps 796. Insertion into W_EMPLOYEE_D table which stores all the employee information had stopped one year back as some company security policy restricted the informatica worklfows to pick up the data. (PER_ALL_PEOPLE_F was a HRMS table and it contained sensitive information line SSN and salary, was inaccessible to the user which informatica uses and the SDE mapping used to return 0 rows).
    Now we have the approval to see those rows and the dimension table is loaded with some 100 new employees who joined in last one year.
    The ROW_WID of W_EMPLOYEE_D is referenced in lot of fact tables and for all those missing employees the WID in the fact table is 0.
    Now that we have all employees, how to make the FACT table point to the correct WID and not store 0. Has anyone faced this problem before?? Writing an update statement will be a tedious task as there are so many fact tables that join to w_employee_d. Also our company uses Sales, Procurement, Finance modules of OB Apps (which constitutes atleast 20 fact tables)
    Any guidance is appreciated. Thanks in advance

    Hello Kostis,
    thank you for your answer. I don't fully understand you. Can you show me short example, please? I create alias table for time dimension on Physical Layer - original table is TimeDayDim and I create aliases TimeDayDim1, TimeDayDim2, TimeDayDim3, TimeDayDim4. Then I create foreign key Fact.Time1 -> TimeDayDim1, Fact.Time2 -> TimeDayDim2, Fact.Time3 -> TimeDayDim3, Fact.Time4 -> TimeDayDim4. And what now? Must I create these table api Bussines Model and create new time dimensions at bussiness model????
    I need in Answers ONE Time dimension. I think I must split my fact table to four tables ... (time1, place1 ...) (time2, place2 ...) (time3 place3...) (time4 place4...) then link those tables to Time dimension (but I dont know where I can split those tables - on Physical Layer or on Bussines Layer).
    I suppose that I will have in Answers one time dimension and four facts tables and I will be able to query them. (for example: Time.Days, Fact1.Place1, Fact3.Speed, Fact4.Count Criteria: Time.Year = 2008)
    Best Regards Vlada

  • How map to my particular table to the fact table in obiee 11g...

    Hi friends,
    I did this simple report in obiee 11g(i.e)
    "NATIONALITY COUNT IN DEPARTMENT WISE"
    For that i used the following tables:
    per_all_assignments_f----->fact table
    hr_all_organization_units----->dim table(containing departments)
    per_all_people_f---------------->dim table(containing nationality)
    I made all the mappings in the physical diagram, as also viewed my report in BI answers
    It shows the following results like
    NATIONALITY---------------------------------------------------------------------COUNT(NATIONALITY)
    AUS------------------------------------------------------------------------------------------------24
    AFR------------------------------------------------------------------------------------------------25
    PHQ_VB-------------------------------------------------------------------------------------------40
    SH_VT----------------------------------------------------------------------------------------------4
    The problem is for me it is showing the above results, but the nationality column is of various codes of the country.
    Since i doesnt want the code of the nationalitian to display in the results..i need the meaning of each and every nationality..
    like,
    AUS------------------------Australian
    AFR-------------------------African
    PHQ_VB----------------------Germanian(assigned)
    Since i know that the meaning for the nationalitian is available in "FND_LOOKUP_VALUES"...okay..
    I can import "FND_LOOKUP_VALUES" table to the physical layer...but how i can able to give the mapping to the fact table in my physical diagram...
    In my report the fact table is "per_all_assignments_f"
    As my fact table doesnt contains any matching column corresponding to the dimension table "FND_LOOKUP_VALUES".....
    Then how i can give mappings to the fact column???? for viewing the full meaning of the nationalitian in my report.....Help me friends...
    Regards,
    Harry...

    Hi bifact,
    I followed the step that U asked me to go but im stuck with later.....
    *) I saved the query that i executed in toad with the columns showing country codes and country meaning to the excel sheet...
    This is the query that i executed in toad and copied data s in excel sheet..
    select z.lookup_code, z.meaning
    from per_all_people_f e, per_all_assignments_f f, hr_all_organization_units h, fnd_lookup_values z where
    e.person_id = f.person_id and f.organization_id = h.organization_id and e.business_group_id = f.business_group_id
    and f.business_group_id = h.business_group_id and f.location_id = h.location_id
    AND z.lookup_type(+) = 'NATIONALITY'AND z.lookup_code(+) = e.nationality and sysdate between
    e.effective_start_date and e.effective_end_date and e.nationality is not null
    *) After that i created a system DSN for the excel drive..
    *) After that when i tried to import metadata of the excel data that is saved, it showed me connection failed...
    For importing this excel data
    what connection details i need to give:--------------------
    as well as user name and also password...
    Soon after importing this excel you said that to give key connection only to fnd_lookup_tables(dim) and that excel data..
    if so, again im not giving key connection to my fact table..
    I think again the same error will occur that no logical mapping is not made to fact so again repository is inconsistent...
    Hi bifact sorry to ask you, but can you see the steps that i followed is correct and after that wat more steps i need to proceed..Can you tell me in breif manner...Thanks for your help...
    Regards,
    Harry...

  • How to find Size of the fact table?

    hello all
    how do i find out the size of the fact table?
    Is the size of the fact table is same as size of the cube.
    pls let meknow
    regards
    balji

    Hello Balaji,
    To find out size of fact table
    DB02 ->Detailed anaysis -> Fact table name
    It will show you the details.
    The size of cube is not exactly equal to the size of fact table, but usually as fact table is much more larger than dimension tables, fact table size will give you a fair idea about cube size.
    Hope it helps.
    Regards,
    Praveen

  • How we will know that dimension size is more than the fact table size?

    how we will know that dimension size is more than the fact table size?

    Hi,
    Let us assume that we are going to take Division and distribution channel in a dimension and assume we have 20 distinct values for Division in R/3 and 30 Distinct values for Distribution channel .So Maximum, we can get 20 * 30 records in dimension table and we can take rough estimation of records in the cube by observing the raw data in source system.
    With rgds,
    Anil Kumar Sharma .P

  • Need a document about how to move the fact and dimension table's to different server's

    Hello Experts,
    I need a detailed doc on how to move the fact and dimension tables to different server's.Please help me out from this
           Thanks in advance....

    You still haven't told anyone what products besides Essbase you are using, without which this is an impossible question to answer.
    https://forums.oracle.com/thread/2585515
    https://forums.oracle.com/thread/2585171
    Are you connecting to these tables from Essbase with a load rule / ODBC?  Using Studio?  Using Integration Services?  Any Drill-Through reporting set up?
    This may sound harsh, but if you truly don't know how to answer any of these questions you should probably not be anywhere near this task...

  • How to refer to fact table to update the same fact table

    Here is my scenario. Fact table is having measures basic_sal, Tax, net_sal with dimension key to employee dimension.
    I neet to update the net_sal = basic_sal - tax in fact table.
    I tried to do a map with fact_table as source and also target. but after deploying the map, it's not updating the net_sal to the existing rows. It's inserting the new rows with dimension key and net_sal columns. so now fact table is having the double the existing rows.
    Thanks,
    Srini.

    Ola Srini,
    We have some experience that updating a table while using that same table as source often takes a lot of time... Specially when the table contains a lot of data (which is mostly the case with fact tables).
    You can use two solutions. Expand the mapping u use to fill the fact table... The calculations looks not that complex to me... But I could be wrong.
    Other solution is to create a temp table where you store all sal types. You could load the fact table from this temp table. Disadvantage of this solution is maintenance of the temp table...
    Regards,
    Moscowic

  • Content Tab: None of the fact tables are compatible with the query request

    Hi All,
    **One thing I am not clear yet of all my years with OBIEE is working with the content tab in BMM.**
    I have made a rpd the joins in physical layer as shown below:
    https://picasaweb.google.com/114804305606242416264/OBIEEError#5663056545119428530
    And the BMM layer as:
    https://picasaweb.google.com/114804305606242416264/OBIEEError#5663056519553812930
    Error I am getting when i run a request from the 3 columns from the selected 3 tables is:
    Dim - Comment Code Details
    Fact - Complaint
    Dim - Service Details
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 14020] None of the fact tables are compatible with the query request Sr Num:[DAggr(Fact - Complaint.Sr Num by [ Dim - Service Details.Sr Cat Type Cd, Dim - Comment Code Details.Cmtcode name] )]. (HY000).
    I get no error for consistency.. I read everywhere and I know i need to set the appropriate aggregation levels in the various dims and facts LTS properties to help OBIEE understanding our model, but how to do that.. how do i decide... how should I approach, what should be the aggregation level, what details.
    When i click More button i see different options: Copy, Copy From, Get Levels, Check Level, what do these mean.
    Aggregation Content, group by - Logical Level or Column which one should i choose and how should I decide.
    Can anyone explain the Content Tab in details and from scratch with some example and why we get these errors.... I know many people who are well versed with many other things related to RPD but this. A little efforts of explaining from you guys will really be appreciated.
    Thanks in advance,
    Dev

    Hi Deepak,
    Option 1:
    My tables in physical layer are joined as below:
    D1--> F1 <--D2--> F2 <--D3
    Same way i model it in BMM
    D1--> F1 <-- D2--> F2 <--D3
    Here D1 is non Conformed Dimension for F2 and D3 is non Conformed dim for F1. Later create Dimensional hierarchies, I tried setting up the content levels
    I go Sources>content tab of Fact F1 I set
    Dimensions----------- Logical level
    D1---------------------- D1 Detail
    D2---------------------- D2 Detail
    D3---------------------- D3 Total
    then, I go Sources>content tab of Fact F2 I set
    Dimensions----------- Logical level
    D1---------------------- D1 Total
    D2---------------------- D2 Detail
    D3---------------------- D3 Detail
    Then, I also go in all the dimensions and set their content levels to Details, but it still gives me errors not sure where I am going wrong in setting the content levels.
    I need to know whether the way I have modeled it in BMM is right,
    Option 2:
    I can combine the two facts in a single Logical Fact or the above design should also work.
    (F1&F2)<--D1, D2 , D3 joined separately using complex logical joins.
    what will be the content tab details?
    Thanks,
    Dev

  • Mapping the Fact table to different levels of a dimension

    Hi,
    I have a fact table which stores the data for 4 levels of the dimensions. The aggregation method was taken care by PL/SQL and the fact table will have the data for all the 4 levels. When im trying to map all the levels to a column in the fact table using the OEM, it is generating the F KEY constraints referncing the columns mapped for the various levels of the dimension.
    The problem is that im using a denormalised table for maintaing the values of the dimension. So the columns mapped for the levels(Except for the lowest) can't have the unique key defined on it. The cube is not getting created because of the error in creating the F KEY.
    Can u please suggest how to map this fact table.
    Thnks,
    Manohar Vanama

    I am not exactly clear on your schema but I believe you are trying to map tables which are not strict star or snowflake. This means that you cannot use CWM1 (and OEM), unless you change the structure of the tables. You might be able to map the tables with CWM2. The document below will assist you:
    Oracle9i OLAP User's Guide
    Chapter 4. Designing Your Database for OLAP
    Chapter 5. Creating OLAP Catalog Metadata

  • Error for the fact table while processing the cube - attribute key cannot be found when processing

    Please help as I am new to SSAS and this is urgent requirement. This is a MOLAP cube and below is the error that I am receiving when processing the cube. The cube is set to Prrocess Full. Several similar errors are popped up for various dimensions.
    "Errors in the OLAP storage engine: The attribute key cannot be found when processing: Table: 'Fact_Table', Column: 'ID', Value: '1'. The attribute is 'Id'. Errors in the OLAP storage engine: The attribute key was converted to an unknown member because
    the attribute key was not found. Attribute Id of Dimension: 17 - Ves - PoC Cont from Database: DB, Cube: IPNCube, Measure Group: iSrvy, Partition: Partition1, Record: 1."
    Thanks in advance.

    Thanks for the recommendations David.
    It will be really great if you can clear some of my doubts:
    To my information, all the dimensions need to be processed first and then the fact table will be processed.
    So if the ID's are not present in the dimension tables, then it should not be present in the Fact table either.
    Here we found null values in the dimension table and the ID's were present in the Fact table. What might be the reasons causing such situation?
    Also how frequently the cube needs to be processed? Currently the ETL which processes the cube, is scheduled in a SQL Job Agent on hourly basis everyday. 
    Is there any possibilty that the cube might be under processing state and the SQL job for the next run getting executed trying to access and process the cube while it was still processing?

Maybe you are looking for

  • ACS 5.3 Backup idea

    IS this below alerts  1 .  Incremental Backup Not Configured and Purge Configuration Repository not configured is must ? or it leads to any issues in future ? <TABLE style="BORDER-BOTTOM: #fc8932 1.5pt outset; BORDER-LEFT: #fc8932 1.5pt outset; WIDTH

  • FILE to multiple BAPI/RFC calls

    Hi, I have one input source (CSV with about 100 rows) and I wish to call a custom rfc for each row in my spreadsheet. (The custom RFC has a single flat import structure as opposed to a table or anything complex). I am struggling to do this in XI. I a

  • Sales document types

    Do all sales document types need to have a "Billing type" and "Devlivery type" defined? In our system we have 2 sales order types (Credit and Debit memo) with no information in the Billing and Delivery type columns. Thanks!

  • I want to update to 3.6.4 but don't want to risk the problems container.exe is causing that am reading in this forum.

    There are quite a few having problems related to container.exe due to updating to 3.6.4 as I see from this forum. I don't have video problems, never did, and sounds like if I update, it's going to start having problems.. It is possible for ff to make

  • Sound is coming from the ear-piece and its really really low

    Title pretty much says it. My iPhone is acting really weird ever since this morning, sound isnt coming from the bottom speakers anymore but instead from the ear-piece and its really low. I tried resetting the phone with holding down the home and slee