Stop SAP BW to create primary index after transport

Hi there,
does anyone know how to stop SAP BW to create an index for transported cubes automatically?
Thanks
Oliver
Edited by: Oliver Schropp on Aug 19, 2011 4:04 PM

Hi,
Primary Index is SAP provided.
The primary index contains the key fields of the table and a pointer to the non-key fields of the table. The primary index is created automatically when the table is created in the database.
There is no way to automated deletion of it.
Thanks,

Similar Messages

  • Error in creating BIA index after SPS upgrade

    Hi BIA experts,
    we got issues in BIA after our BI SPS upgrade (current level SPS18):
    when wir try to create new BIA index for a cube (normal basic cube), we got error RSD_TREX121 (index type ICB). This issue occurs since we completed SPS18 upgrade.
    Any help and feedback are highly appreciated!!!
    Regards,
    Sally

    Hi Sally,
    "BIA index has type ICB" is not an error message. It just indicates that the index type has been initialized or changed. Can you describe the symptoms more clearly or provide more details of the job and/or application log?
    Regards,
    Marc
    SAP NetWeaver RIG

  • SAP Homepage customisation: Sercice not showing after transporting

    Hello experts,
    we have a very simple portal role in Dev system (role was copied from SAP ESS role) where added transaction iView. In the backend this is configured using PTARQ tcode and configuring a custom subarea and a custom service.
    This all works fine in Dev system and it also works fine in Qual system.
    However, when we do the same transport (portal role export import + PTARQ normal transport request) into Prod system, then the subarea is NOT shown, instead a record working time iview from the sap standard is shown.
    PCD location etc should all be OK, because, IT IS working fine in Dev and Qual system.
    Would really appreciate if someone could point us into the right direction of how to resolve this please.
    Thanks in advance

    Hi,
    Can you check the back end service configuration . And also check the service which maintained in back end are pointing a  PCD location is standard one or custom one.
    Because this type of issue happened only wron PCD location has maintained only.
    Thanks
    patrick

  • Primary index

    Hi,
    I have created one table(internal table) in se11 and also have assigned some fields are as primary keys.I know that when we define primary key for one field the system will create primary index for that as ' 0 '.now my question is how to see and use that key.
    Regards,
    Krishna.

    Hi..
    Primary Index is Created on the Primary key fields of the Table automatically when the Table is Activated.
    Whenever we use the Select.. where clause on Primary key the Primary index is used by the system. We can't explicitly use indexes.
    To Check whether a query is using index or not we can use SQL Trace ST05.
    <b>reward if Helpful.</b>

  • Is it possible to created secondary indexes on ODS in Production

    Hi,
    Is it possible to created secondary indexes on ODS in Production System. I need to create secondary indexes on ODS but it is already in production. Hence Can I directly create secondary indexes without transportation from dev to production?

    Hi,
    Secondary Indexes for DSO can be transported. For the transport the DSO objects needs to be transported (R3TR ODSO <technical name>.
    Additionally, secondary Indexes are necessary quite often for DSO tables, which can be transported. In few cases you need indexes on other BW tables, but they cannot be transported. Never create additional Indexes on InfoCube tables (like E- and F-Fact tables and Dimension tables)
    Thanks & B.R.
    Vince

  • After Transporting the request RFC Destination is empty in SWCV and BEA

    Hi All,
    we are using SAP NW Mobile 7.1 SP08
    After transporting the SWCV & Backend adapter, in the transported system the RFC Destination is empty
    Please help me to fix this issue.
    Thanks and Regards,
    Joseph Beevin.

    Hi,
    When an SWCV is trnasported to a another system,the backend destination will be empty in the target system.,We need to configure the bakcned destination in the target sytem through the DOE Adimin.
    DOE_ADmin Portal -->Configuration --->Backend Configuration.
    In the backend configuration,select the SWCV for which the destination needs to be configured.Select the parameter "RFC  DESTINATION FOR BACKEND"
    When the SWCV is selected all the active Data Objects will be listed in the table below
    Insert the above parameter to the parameter list(the table below)
    Give the name of the availlable backemnd destination name in the "text value" column of the table for each data object and save it.
    Now the backend destination is configured correctly in the target system.Please try this..
    Regards,
    Muralikrishna T.

  • Test System (BQ1) Objects showing up wrong Source System after Transport.

    Hi Experts,
    We are working on IS-U BI7.0 Project. We have finished with our Development and have transported the DataSources, Transformations and DTPs from Development (BD1) into Test (BQ1) Systems.
    We have entered values in the Conversion of Logical Source Systems correctly.
    When we select the DataSource and Click Object Directory Entry from the Extras Menu, we found that the Original Source System is defined as BQ1 instead of BD1 (Development System), which is where the Object is getting Transported from.
    We have implemented an SAP Note:1008633 (Incorrect Source System after Transport (RSTRAN 514) but couldn't resolve the issue.
    We basically want all the Objects in the BQ1 (Test System) showing the Original Source System as BD1 (Development System) but only the DataSources and it's connected Transformations and DTPs are originating from the BQ1 (Test) System itself (The Transformations when Transported from BD1 to BQ1 are Transported with a different Technical Name which is very strange).
    Had anyone faced similar problem and if so, what was the effective solution implemented.
    Any help in this regard is very much appreciated.
    Thanks in Advance.
    Best Regards,
    Chandu.

    It is resolved by our Basis.
    Thanks,
    Chandu

  • Logical standby stopped when trying to create partitions on primary(Urgent

    RDBMS Version: 10.2.0.3
    Operating System and Version: Solaris 5.9
    Error Number (if applicable): ORA-1119
    Product (i.e. SQL*Loader, Import, etc.): Data Guard on RAC
    Product Version: 10.2.0.3
    logical standby stopped when trying to create partitions on primary(Urgent)
    Primary is a 2node RAC ON ASM, we implemented partitions on primar.
    Logical standby stopped appling logs.
    Below is the alert.log for logical stdby:
    Current log# 4 seq# 860 mem# 0: +RT06_DATA/rt06/onlinelog/group_4.477.635601281
    Current log# 4 seq# 860 mem# 1: +RECO/rt06/onlinelog/group_4.280.635601287
    Fri Oct 19 10:41:34 2007
    create tablespace INVACC200740 logging datafile '+OT06_DATA' size 10M AUTOEXTEND ON NEXT 5M MAXSIZE 1000M EXTENT MANAGEMENT LOCAL
    Fri Oct 19 10:41:34 2007
    ORA-1119 signalled during: create tablespace INVACC200740 logging datafile '+OT06_DATA' size 10M AUTOEXTEND ON NEXT 5M MAXSIZE 1000M EXTENT MANAGEMENT LOCAL...
    LOGSTDBY status: ORA-01119: error in creating database file '+OT06_DATA'
    ORA-17502: ksfdcre:4 Failed to create file +OT06_DATA
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    LOGSTDBY Apply process P004 pid=49 OS id=16403 stopped
    Fri Oct 19 10:41:34 2007
    Errors in file /u01/app/oracle/admin/RT06/bdump/rt06_lsp0_16387.trc:
    ORA-12801: error signaled in parallel query server P004
    ORA-01119: error in creating database file '+OT06_DATA'
    ORA-17502: ksfdcre:4 Failed to create file +OT06_DATA
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    ORA-15001: diskgroup "OT06_DATA" does not exist or
    Here is the trace file info:
    /u01/app/oracle/admin/RT06/bdump/rt06_lsp0_16387.trc
    Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64bit Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u01/app/oracle/product/10.2.0
    System name: SunOS
    Node name: iscsv341.newbreed.com
    Release: 5.9
    Version: Generic_118558-28
    Machine: sun4u
    Instance name: RT06
    Redo thread mounted by this instance: 1
    Oracle process number: 16
    Unix process pid: 16387, image: [email protected] (LSP0)
    *** 2007-10-19 10:41:34.804
    *** SERVICE NAME:(SYS$BACKGROUND) 2007-10-19 10:41:34.802
    *** SESSION ID:(1614.205) 2007-10-19 10:41:34.802
    knahcapplymain: encountered error=12801
    *** 2007-10-19 10:41:34.804
    ksedmp: internal or fatal error
    ORA-12801: error signaled in parallel query server P004
    ORA-01119: error in creating database file '+OT06_DATA'
    ORA-17502: ksfdcre:4 Failed to create file +OT06_DATA
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    ORA-15001: diskgroup "OT06_DATA" does not exist or
    KNACDMP: *******************************************************
    KNACDMP: Dumping apply coordinator's context at 7fffd9e8
    KNACDMP: Apply Engine # 0
    KNACDMP: Apply Engine name
    KNACDMP: Coordinator's Watermarks ------------------------------
    KNACDMP: Apply High Watermark = 0x0000.0132b0bc
    Sorry our primary database file structure is different from stdby, we used db_file_name_convert in the init.ora, it look like this:
    *.db_file_multiblock_read_count=16
    *.db_file_name_convert='+OT06_DATA/OT06TSG001/','+RT06_DATA/RT06/','+RECO/OT06TSG001','+RECO/RT06'
    *.db_files=2000
    *.db_name='OT06'
    *.db_recovery_file_dest='+RECO'
    Is there any thing wrong in this parameter.
    I tried this parameter before for cloning using rman backup. This din't work.
    What exactly must be done? for db_file_name_convert to work.
    Even in this case i think this is the problem its not converting the location and the logical halts.
    help me out.....
    let me know if you have any questions.
    Thanks Regards
    Raghavendra rao Yella.

    Hi reega,
    Thanks for your reply, our logical stdby has '+RT06_DATA/RT06'
    and primary has '+OT06_DATA/OT06TSG001'
    so we are using db_file_name_convert init parameter but it doesn't work.
    Is there any thing particular steps hiding to use this parameter? as i tried this parameter for rman cloning it din't work, as a workaround i used rman set new name command for clonning.
    Let me know if you have any questions.
    Thanks in advance.

  • IMessage stopped working on iOS 5.1.1 after I set up a new Apple ID.  After logging in, it keeps cycling back to the "create ID" screen.  Apple's verified my new ID. I've reset the iPod (3) times--no luck.

    iMessage stopped working on iOS 5.1.1 after I set up a new Apple ID.  After logging in, it keeps cycling back to the "create ID" screen.  Apple's verified my new ID. I've reset the iPod (3) times--no luck.

    I solved my problem by Resetting my iPod and then Restoring my files.  Only speed bump was losing two apps in the Restore, as the current versions now require iOS 7.1, minimum.  Apple makes Reset and Restore (two processes) painless with these Help articles:  "Use iTunes to restore your iOS device to factory settings," and "Back up and restore your iPhone, iPad, or iPod Touch using iCloud or iTunes."  The Restore took a while to synch my photos and iTunes libraries.

  • Create a fast refresh materialized view with partitioned primary index

    Hi,
    I have a materialized view that is based on a table with primary key.
    I want to create a materialized view with a partitioned primary index . do you have any way of doing it?
    I tried to create a materialized view with rowid and then I created a partitioned primary index on it.
    It did not work as what I expected. I could not perform a fast refresh on it. the materialized view can only complete refresh
    thank you

    Hi,
    Here is some info from the Oracle Documentation.
    http://download.oracle.com/docs/cd/E11882_01/server.112/e10706/repmview.htm
    http://download.oracle.com/docs/cd/E11882_01/server.112/e10810/basicmv.htm
    Determining the Fast Refresh Capabilities of a Materialized View
    http://download.oracle.com/docs/cd/E11882_01/server.112/e10706/repmview.htm#BABEDIAH
    Regards,

  • Primary Index to create

    Hi Experts,
    In BW 3.5, in DB24 there's a missing_index warning for a transparent table /BIC/Fxxxx (xxxx is not the name of the infocube)
    In SE14, it says thaht this primary index /BIC/Fxxxx-0 must not be created on the database.
    First of all, what kind of table is this one ? /BIC/F : partitionned F-Fact table ?
    Second, should I create this index or not ?
    Regards,

    Hi again,
    Thanks for your answers.
    But so why brconnect sends a missing_index error, and only for this table ? There are other fact tables whitout primary index, and brconnect does not send any error.
    regards,

  • Creating Primary Key Index

    Hi,
    Iam trying to create a Primary Index on Large table and Iam using the Parallelism to make it fast. I have used the following script to create a Primary key Index.
    CREATE UNIQUE INDEX FLX_ARCH.FUND_SEC_RESULT_STAT_ARCH_PK
    ON FLX_ARCH.FUND_SEC_RESULT_STAT_ARCH
    (FUND_SEC_RESULT_MAP_INST,RATES_RULE_ID,BASE_LOCAL_FLAG,EFFECTIVE_DATE)
    TABLESPACE FLX_PERF_I
    parallel;
    Index created.
    Elapsed: 01:43:37.37
    Index got created in 1 hour 43 minutes. But when Iam adding constarint to those columns, it is taking almost 12hours with or without parallelism.
    ALTER TABLE FLX_ARCH.FUND_SEC_RESULT_STAT_ARCH
    ADD CONSTRAINT FUND_SEC_RESULT_STAT_ARCH_PK
    PRIMARY KEY
    (FUND_SEC_RESULT_MAP_INST,RATES_RULE_ID,BASE_LOCAL_FLAG,EFFECTIVE_DATE)
    parallel
    Would you please let me know what could be the problem and how can we add the Primary constraint fast.
    Your early response is much appreciated.
    Thanks and Regards
    Suresh. D

    Would you please let me know what could be the problem and You have presented no details that indicate any problem exists.
    how can we add the Primary constraint fast. Fast is relative.
    If I asked you to make my snail fast, could you do so?
    ALTER SESSION SET SQL_TRACE=TRUE;
    ALTER TABLE FLX_ARCH.FUND_SEC_RESULT_STAT_ARCH
    ADD CONSTRAINT FUND_SEC_RESULT_STAT_ARCH_PK
    PRIMARY KEY
    (FUND_SEC_RESULT_MAP_INST,RATES_RULE_ID,BASE_LOCAL_FLAG,EFFECTIVE_DATE)
    parallel
    ALTER SESSION SET SQL_TRACE=FALSE;
    now find the trace file within ./udump folder
    tkprof <trace_file.trc> trace_results.txt explain=<username>/<password>
    By inspecting the trace file you will be able to see where time is being spent.
    With this knowledge, you may or may not be able to determine the bottleneck & therefore improve it

  • Compute Statistics after create a index

    Good night, I tryed it before and got no help.
    So I'll try it again, because I didn't find it yet.
    I have to create some indexes in a production database.
    I have a doubt.
    Do I need to Compute Statistics after creating indexes?
    Or when I create they automatically are computed?
    Thanks for any help.
    The version I'm using is:
    Oracle Database 10g Release 10.2.0.5.0 - 64bit Production
    PL/SQL Release 10.2.0.5.0 - Production
    CORE     10.2.0.5.0     Production
    TNS for 64-bit Windows: Version 10.2.0.5.0 - Production
    NLSRTL Version 10.2.0.5.0 - Production

    Petri BR wrote:
    Good night, I tryed it before and got no help.
    So I'll try it again, because I didn't find it yet.
    I have to create some indexes in a production database.
    I have a doubt.
    Do I need to Compute Statistics after creating indexes?after CREATE INDEX what are the values for the index statistics?
    Or when I create they automatically are computed?Just look at the statistics values after issuing CREATE INDEX & answer the question yourself!

  • Procedure for creating an Index on SAP standard table

    Hi,
    I need to create an Index on a SAP standard table. Please help me out with the process of doing so.
    Regards,
    Sharadendu

    Please search the forum. You will find lots of related threads:
    how to create an index
    how to create secondary index

  • SAP hana Visualize Plan show Create Temp Index and it shows Unsupported Operator Order.  Why the hana build temp index and how to get rid of it ?

    When I see the Visualize Plan for SQLScript it show Create Temp Index at many Places.
    It also shows unsupported operator order CS_JOIN over CS_AGGREGATION.
    Why does hana create Temp Index and how to get rid of it.

    Just browsing for some more information on this operation, and it just happened that my favorite VizPlan expert has chimed in here
    I have a scenario where I am using a nested SQLScript based calc view. One view (child view) retrieves a relatively small data set, then the main CV takes that and joins with a bunch of other data from another schema.
    When I started out, just the nested portion of (child calc view), takes roughly 800ms to return it's resultset, for which I was pleased. Now, when I added some other joins onto that dataset in the main CV, I am seeing my runtime spike to approximately 16 seconds, with 15 seconds being spent on a "Create Temp Index" process that only results in 12 rows (from what I can see).
    Now the question becomes - since I have identified where 95+% of my runtime is being generated, how can I really pinpoint what operation is causing this? I was thinking the ID column (shown in the details) may offer some clues if one knew how to locate the details in a system table somewhere?
    Otherwise, all I can see in the VizPlan is that YES there is a long running component, but I have no idea what it might be, therefore can't do anything about it
    Regards,
    Justin

Maybe you are looking for