Indexes on Oracle Level Missing at Data Dic. (SAP) Level

Hello Masters,.
My Question is,  There are 3 standard tables in CRM server Say for example table A, table B, table C.
Table A has 2 index , B have 1 Index, C has 2 Index in Database level.
These Indexes are not available or somehow deleted from Data Dictionary (SAP) level.
Are these Indexes working  if i am writing ABAP Code Select statement or in any situitions...??

The one way that I can see to find that out quickly is to do a trace on a query performed against the table(s)... through ST05 and then drilling down into the Execution Plan in the Trace List..
Those indexes that are being used will be listed there.
Your queries will need to encourage use of these indexes, though, to avoid false negatives...
Regards,
Dushyant Shetty

Similar Messages

  • PowerPlant passive Oracle dblink to read data from SAP ECC?

    Hello
    We have a third party vendor named PowerPlant that is proposing to integrate with SAP ECC using a passive dblink to pull data into PowerPlant.  This is proposed when SAP ECC is running on an Oracle database.  Presumably this is a read-only connection to the SAP ECC Oracle database (at database level).  Has anyone used this and/or have recommnedations for or against?
    Thank you,
    Harold

    Hi Harold,
    Technically it is possible but, check your Oracle license with the local SAP office. As far as I know that SAP is not permitted 3rd party connections to the database.
    Best regards,
    Orkun Gedik

  • How to add an index to a materialized view in Data Modeler 3.3

    Hello everyone,
    I'm looking for a how-to to add an index to a materialized view in Data Modeler 3.3.0.747, as I coudn't find a way to do it so far.
    I looked here:
    Relational Model
    Physical Model
    Oracle 11g
    Materialized Views
    "my_mv_name"
    "INDEXES" IS NOT HERE IN THE TREE
    "Tables" does not include it either
    Thank you & Best regards,
    Blama

    Hi David,
    thanks a lot. I did so and it worked, but I found a minor bug while doing so:
    I marked the table as "Implement as Materialized View" and went to File->Export->DDL (for Oracle 11g).
    The generated code (I checked all options in "Drop Selection") includes a row:
    DROP MATERIALIZED VIEW mv_mymatview CASCADE CONSTRAINTS ;
    which produces a syntax error.
    Best regards,
    Blama

  • "Document name invalid.  The index.xml file is missing"

    I have OSX Yosemite 10.10 and am having a problem just today in Numbers.  A file that I have saved to my desktop and opened daily to input data since the first of November won't open.  It gives me a message that the file is invalid, the "index.xml file is missing"   Running Numbers '09 version 2.3   Have been run ing Yosemite 10.10 for quite some time.  How can I open/retrieve the file?

    I am having the same problem.  I am also running Yosemite 10.10.  The last I accessed the file was on November 13, 2014. I have been noticing that the layout /view of the file with the program has been a little different for about a week before then.  I have not upgraded to the newer version although it is constantly asking to upgrade. Still running Number '09, Version 2.3. I am very concerned as the file I am accessing contains critical records that go back for several years. I am not having problems opening .numbers files that are a couple of years old. Thanks for any help.

  • "Document name is invalid.  The index.xml file is missing

    I have OSX Yosemite 10.10 and am having a problem just today in Numbers.  A file that I have saved to my desktop and opened daily to input data since the first of November won't open.  It gives me a message that the file is invalid, the "index.xml file is missing"   Running Numbers '09 version 2.3   Have been run ing Yosemite 10.10 for quite some time.  How can I open/retrieve the file?

    An original Numbers ’09 v2.3 document will save as a compressed zip archive that has its extension renamed .numbers. It contains an index.xml file. If you double-click this Numbers ’09 document, it will open in Numbers ’09 v2.3. This is contrary to how Pages v5.5.1 handles Pages ’09 documents.
    You have to explicitly open a Numbers ’09 v2.3 document in Numbers v3.5. When you make a change, Numbers v3.5 will drop a sheet dialog that is very clear in its portent:
    Would you like to upgrade this Numbers '09 spreadsheet?
    Edit a Copy
    Cancel
    Upgrade
    When you choose either the Edit a Copy, or Upgrade — and save your changes, the resulting document is now in the new document architecture whose structure no longer uses the index.xml file. Double-clicking this new document format will now default to Numbers v3.5 on open, not Numbers '09 v2.3. You have to explicitly choose Open with Numbers (2.3) from the contextual menu to get the pop-op below because Pages v5.5.1, Numbers v3.5, and Keynote 6.5 documents are not backwards compatible.

  • Catalog error : Index of values is missing for characteristic PLANTu0092

    Error" Index of values is missing for characteristic " in mapping 
    Hi
    We are in SRM
    I was trying to transfer Master records to catalog using
    BAdi : BBP_CCM_CHAR_MAINT but I’m getting an error’ Index of values is missing for characteristic PLANT’.
    Plant has multiple value.
    I’ve following code in the method ENHANCE_CTY_DEF
    ls_char_definition-char_name = 'PLANT'.
    ls_char_definition-char_datatype = 'CHAR'.
    ls_char_definition-MULTIVAL_INDICATOR = 'X'.
    APPEND ls_char_definition TO et_char_definition.
    Following code in the method VALUATE_PRODUCT
    **"plant data
    v_index = 1.
      ls_char_valuation-char_name = 'PLANT'.
      LOOP at t_plant into WA_plant.
    ls_char_valuation-char_value = WA_PLANT-PLANT_1.
      IF  WA_plant-PLANT_1 is not initial.
       ls_char_valuation-MULTIVAL_INDEX = v_index.
       APPEND ls_char_valuation TO et_char_valuation.
       v_index = v_index + 1.
      ENDIF.
      ENDLOOP.
    Can anyone tell me how to resolve it. I’ll definitely reward points to any useful answer.
    -Joe

    Hi
    Have you raised any OSS message with SAP yet ? If no, raise one immediately.
    Meanwhile, try this ->
    <u>It seems the Following code in the method VALUATE_PRODUCT
    has some problem -></u>
    **"plant data
    v_index = 1.
    ls_char_valuation-char_name = 'PLANT'.
    LOOP at t_plant into WA_plant.
    ls_char_valuation-char_value = WA_PLANT-PLANT_1.
    IF WA_plant-PLANT_1 is not initial.
    ls_char_valuation-MULTIVAL_INDEX = v_index.
    APPEND ls_char_valuation TO et_char_valuation.
    v_index = v_index + 1.
    ENDIF.
    ENDLOOP.
    Try changing the code - like this ->
    **"plant data
    LOOP at t_plant into WA_plant.
    v_index = sy-tabix.
    ls_char_valuation-char_name = 'PLANT'.
    ls_char_valuation-char_value = WA_PLANT-PLANT_1.
    IF WA_plant-PLANT_1 is not initial.
    ls_char_valuation-MULTIVAL_INDEX = v_index.
    APPEND ls_char_valuation TO et_char_valuation.
    ENDIF.
    ENDLOOP.
    Some SAP OSS notes to refer ->
    <b>Note 850613 - Maintaining the BBP_CCM_CHAR_MAINT BAdI using Customizing
    Note 875039 - BBP_CCM_CHAR_MAINT: Valuating predefined characteristics
    Note 724097 - Search of the comp. in case of structured characteristics</b>
    Do let me know.
    Regards
    - Atul

  • (R11) HOW TO FIX MISSING CM DATA IN AR_PAYMENT_SCHEDULES_ALL TABLE

    제품 : FIN_AR
    작성날짜 : 2004-11-09
    (R11) HOW TO FIX MISSING CM DATA IN AR_PAYMENT_SCHEDULES_ALL TABLE
    ==================================================================
    PURPOSE
    Credit Memo data가
    AR_PAYMENT_SCHEDULES_ALL TABLE과 AR_RECEIVABLE_APPLICATIONS_ALL 에 생성
    이 되어 있지 않는 경우 datafix script 으로 해결할 수 있도록 한다.
    Problem Description
    주로 타 시스템에서 autoinvoice 를 통해 넘어온 cm data에서 자주 발생하는 문제로 credit transaction 화면을 통해 보았을때 100% 로 applied 된 credit memo 임에도 applied 된 transaction의 balance에 전혀 영향을 주지 않는다.
    cm_info_11.sql 을 수행해 보면 ar_payment_schedules_all table과 ar_receivable_applications_all table에 해당 데이타가 존재하지 않음을 알수 있다.
    Workaround
    Solution Description
    1. AR Support Script Page (http://eastapps.us.oracle.com/appsar/SQL/SQL.htm) 에서 trx.sql과 pay.sql을 download 받는다. 이때 홈페이지 상의 90 Days, 150 Days 등은 transaction date scope을 의미한다. 즉 현재일로 부터 90일전까지의 transaction인 경우는 trx.sql을 150 일전까지의 transaction인 경우는 trx150.sql 을 download 받아야 한다.
    2. trx.sql 을 수행한다.
    3. trx.sql 수행시 생성되는 trx_<trx_id>.log 를 trx.sql 로 rename하여 다시 수행한다.
    4. 정상적으로 수행되었는지 확인 후 pay.sql을 수행한다.
    5. pay_<ps_id>.log 를 pay.sql로 rename한 후 수행한다.
    Reference Documents
    BUG 1401610

    'ORA-29278: SMTP transient error: 421 Service not available' --> "Oracle had a problem communicating with the SMTP server"
    Did you check:
    How I can resolve the error " 421 Service not available"
    Also look at document id 604763.1
    Edited by: MccM on Feb 23, 2010 11:38 AM

  • How to implement "Filtered Indexes" in Oracle?

    I know that function-based index can (kind-of) create index for only a portion of all the rows in a table,
    create index idx_abc_01 on t_abc (case when end_dt > '01-APR-2008' then end_dt else null end);
    /* to use the index, you have to use the same express in the filter clause */
    select * from t_abc where xyz in ('O', 'S', 'X') and
         case when end_dt > '01-APR-2008' then end_dt else null end
              between '12-JUL-2008' and '15-JUL-2008';But this approach is very inflexible because the expression in the where clause has to exactly match the index creation, otherwise that filtered index will not be used.
    MS SQL Server 2008 has a new feature called - Filtered Indexes
    CREATE NONCLUSTERED INDEX fidx_abc_01
        ON t_abc ( EndDate )
    WHERE EndDate > '20040401';
    select * from t_abc where EndDate between '20050121' and '20050130';[url http://msdn.microsoft.com/en-us/library/cc280372.aspx]Filtered Indexes can be very helpful in both OLTP and DW. I'm wondering how to get the same thing done in Oracle.
    Any suggestion beyond [url http://erturkdiriksoy.wordpress.com/2008/06/30/filtered-indexes-on-oracle/]Dunyada?

    In 11.2, if you're building a data warehouse (and assuming that implies you have the partitioning option), you can declare that certain partitions of a local index are unusable which prevents Oracle from building them. If T_ABC were partitioned on END_DT, that would allow arbitrary queries where Oracle could perform partition pruning to make use of the index if it was available for those partitions without the need for an exact expression match.
    Beyond that, if you are creating a function-based index, you would often want to create (or modify) the view that your application queries to specify the condition (i.e. a VW_RECENT_ABC that only has post-April, 2008 data) so that queries didn't have to specify the condition exactly.
    Justin

  • Using Oracle XE as a Data Source for Powerpivot

    I will be migrating my database from MS Access 2007 (*.accdb) to an Oracle XE. Any good links or papers on this or possibly any past experience with this (powerpivot connection)? Just trying to make sure that the connector is already within the Powerpivot download.
    Using Powerpivot v2.0
    For more info about powerpivot:
    www.powerpivot.com

    Hello,
    I'm not sure this is an XE-specific question. However, there's not much examples outside, probably because it's not a very common solution. I'd say it should work, as PowerPivot offers OLE DB and ODBC connectors and XE comes with appropriate ODBC drivers installed. So, as long as your XE is on the same machine as your PowerPivot, you don't have to care much about that. Otherwhise you'll probably have to install the ODBC driver, e.g. from [url http://www.oracle.com/technetwork/developer-tools/visual-studio/downloads/index.html]Oracle Data Access Components (ODAC).
    Then create your ODBC data source, test it, and link it in PowerPivot.
    -Udo

  • Oracle Development Survey on Data Warehouses: How Does Yours Compare?

    At the start of most data warehouse projects, or even during a project, I am sure you as customers try to find answers to the following questions to help you plan and manage your environments:
    * Where can I find trend and comparison information to help me plan for future growth of my data warehouse?
    * How many cpu's do other customers use per terabyte?
    * How many partitions are typically used in large tables? How many indexes?
    * How much should I allocate for memory for buffer cache?
    * How does my warehouse compare to others of similar and larger scale?
    The data warehouse development team, here at Oracle, would like to help provide answers to these questions. However, to do this we need your help. If you have an existing data warehouse environment, we would like to obtain more technical information about your environment(s) by running a simple measurement script and returning the output files to us, here at Oracle. This will allow our developers to provide comprehensive documents that explain best practices and get a better understanding of which features our customers use the most. This will also allow you as Customers, to benchmark your environments compared to other customers’ environments.
    From a Company perspective we are also interested to get feedback on features we have added to the database, are these features used, how are they used etc. For example we are keen to understand:
    * Which initialization parameters are most frequently used at what values?
    * How many Oracle data warehouses run on RAC? on single nodes?
    * Is there a trend one-way or the other, especially as data volumes increase?
    * Does this change with newer releases of the database?
    All results from these scripts will be held confidential. No customers will be mentioned by name; only summaries and trends will be reported (e.g., “X percent of tables are partitioned and Y percent are indexed in data warehouses that are Z terabytes and larger in size.” or “X percent of Oracle9i and Y percent of Oracle10g data warehouses surveyed run RAC”). Results will be written up as a summarized report. Every participating customer will receive a copy of the report.
    Terabyte and larger DW are the primary interest, but information on any data warehouse environment is useful. We would like to have as many customers as possible submit results, ideally by the end of this week. However, this will be an on going process so regular feedback after this week is extremely useful.
    To help our developers and product management team please download and run the DW measurement script kit from OTN which is available from the following link:
    http://www.oracle.com/technology/products/bi/db/10g/dw_survey_0206.html
    Please return the script outputs using the link shown on the above web page, see the FAQ section, or alternatively mail them directly to me: [email protected].

    969224 wrote:
    Hi Guys, just a quick question. when we have a primary key on 4 coloumns and we have, say 20 million rows and we want to add one extra row. How does oracle check whether the data on the primary key is unique to the record being added compared to the 20 million rows. Does it actually compare the record being added to all the rows present in the table?
    Edited by: 969224 on May 10, 2013 8:14 AMNot the whole row, it compares the 4 columns in the INDEX against the 4 columns in the new row.

  • Webcast : Sun Oracle Database Machine for Data Warehousing  -Sep 30 noon ET

    Sun Oracle Database Machine for Data Warehousing
    Jean Pierre Dijcks - Data Warehousing Product Mgmt, Oracle
    https://conference.oracle.com/imtapp/app/cmn_jm_hub.uix?mID=158101510
    On September 15 Oracle announced the second generation of its Database Machine, making an already strong data warehousing product significantly stronger. The new version runs on Sun hardware and offers important new features. Available in full rack, half rack, quarter rack, and basic unit configurations, the Sun Oracle Database Machine can add value at many data warehouse size levels.
    The Sun Oracle Database Machine runs on Oracle Database 11g Release 2 and has new features such as:
    Smart Flash Cache memory for ultra-fast IO - Reaches 50GB/second on a full rack system (not even counting gains from compression)
    Exadata Hybrid Columnar Compression - Maximizes data capacity and reduces scan times: think 500GB/second IO
    Offloaded Data Mining Scoring - Moves CPU-intensive operations from database servers to Exadata storage servers
    In-Memory Parallel Execution - Caches full tables in memory across nodes: foundation of new TPC-H world record
    There is plenty more we have not listed above, so come to this TechCast and learn about this major new product!
    Audio Dial-In: 888 967 2253 Audio Meeting ID: 572994 Audio Meeting Passcode: 334451
    Web Conference: https://conference.oracle.com/imtapp/app/cmn_jm_hub.uix?mID=158101510
    Compatibility Check: If you have not used Oracle's web conference system before, please ensure your system
    compatibility by going to https://conference.oracle.com/imtapp/app/nuf_sys.uix

    Is there any way, one could get this webcast to watch it offline?
    regards

  • Migration data from SAP DB2 to Oracle 10g

    Hi , I am assigned to migrate data from SAP system which is using DB2 to Oracle 10g.I not very familiar with SAP. I hope someone can help me., especially for those familiar in SAP.
    Please help me.
    thanks
    jebatco

    Hello,
    just migrating a DB2 database to Oracle 10g might be an easy task. The Oracle Migration Workbench is the tool for such a migration:
    http://www.oracle.com/technology/tech/migration/workbench/index.html
    I have no idea about SAP, and that might complicate the picture. But there exist specialists for this task:
    Oracle Expertise in the SAP environment
    The Solution Center SAP Support and Service – located in Walldorf – offers SAP
    customers the following services:
    • Advanced Customer Services (ACS)
    • Performance Analysis and Tuning
    • Development of concepts for Backup/Restore/Recovery, and High Availability,
    Administration
    • Security concepts
    • Optimizing of ABAP/4 programs (performance improvement)
    • Migration service for customers, who want to use Oracle as the database for SAP
    applications (from Informix, MaxDB, DB2, or SQL Server to Oracle).
    • Migration services from “Oracle to Oracle” (e.g. Tru64 to HP_UX)
    • Integration-Products and –Services
    • Oracle Database: The Database of Choice for Deploying SAP Solutions
    This is taken from http://www.oracle.com/newsletters/sap/docs/ora4sap-db-of-choice.090213.pdf
    Best regards
    Wolfgang

  • Oracle OLAP as OBIEE Data Source

    I've got a couple of questions regarding the use of Oracle OLAP (Analytic Workspace/Cube) as an OBIEE data source.
    First: As a general rule when creating a dimension, we create a total roll-up for the dimension i.e. "Total Product", "Total Geog", "Total Customer" etc... Generally, I don't create a total roll-up for time dimensions. When importing metadata from OLAP to OBIEE, OBIEE creates a "Total" level for all dimensions. Now, I understand why OBIEE does that; to support queries that might exclude one or more dimensions. My question is: what is the best method/procedure to deal with the extra "Total" level?
    Second: I would appreciate it if someone could explain this error for me: [nQSError: 59137] Filter level YEAR is below the projected level Total on dimension CMP_TIME while an externally aggregated measure is present. (HY000). I understand the words, but have no clue what OBIEE is trying to tell me. This error pops up constantly and I see no rhyme or reason that would cause it. the specific case above occurred when I clicked on the sort icon for a measure included in a report.
    Thanks,

    Mark,
    Thanks for the reply. However, I'm not sure I made myself clear. I have created a "Product" dimension in AWM (Analytic Workspace Manager) in the following structure: Product -> Product Line -> Total Product. Withing the context this hierarchy, "Total Product" is the "Grand Total" Level. When this data is imported into OBIEE using "Oracle OLAP" as a data source, the Product hierarchy is created in the Physical Layer as an "Oracle OLAP Dimension". In the BMM Layer, the hierarchy is structured as: Product -> Product Line -> Total Product -> Total. There are now two "Total" Levels. Naturally only one, the OBIEE generated Total, is defined as "Grand Total". The only child of the Total level is Total Product. I have two hierarchy levels that are the same. So, do we need both? should we keep both? Should a dimension defined within AWM for use in OBIEE NOT include a total level? It's not really a problem, it just doesn't seem to make any sense to have TWO total levels within a hierarchy.
    On the second issue, I wish I could provide some detail, but I'm really not sure how I'd do that. That's why I asked for the meaning of the error. What is OBIEE telling me that I'm doing wrong. All I really did was import the metadata, drag it to the BMM Layer, deleted some of the hierarchy level keys, renamed some columns and dragged the stuff over to the Presentation Layer. So, it's pretty much drag-and-drop.
    Another example of the error: We have a Category Dimension (Sub Category -> Category -> Category Group -> Model -> All Categories -> Total) and I want to see the top 10 values of a measure by Category by Model. In an Analysis, adding the Model column works fine, just not the best visualization. Move the Model column to "Sections" and all works; move the Model column to Pivot Table Prompts and it errors. Obviously, I'm asking OBIEE to do something it doesn't want to do, so I'm looking for the root cause of the error.
    Thanks,

  • Normal index in oracle.

    i have a city_tbl with columns ( city_cd, city_name,state_cd) in oracle 11g release2
    city_cd is a primary key.
    i ran the query
    select city_name from city_tbl where state_cd = 'AL'
    expalin plan shows full table scan
    now i create normal index on state_cd column
    after running the query still explain plan shows full table scan.
    after creating the index why there is full table scan ?

    Hi,
    it looks like the optimizer is doing the right thing by ignoring your index and doing the full table scan. Full table scan is a perfectly valid way of retrieving data and in many cases is more efficient than an index access, because a) it uses multiblock reads b) doesn't have to go through index blocks first before reading the data.
    In your case, the optimizer estimates that the cost of reading the table via a FTS is equivalent to 34 single-block reads. Index access in your case is more expensive, because while it's pretty cheap to acquire rowids of desired rows (only 2 reads are needed), the desired table rows are scattered around 115 blocks, so Oracle would have to make 117 reads to retrieve the data via an index.
    Of course, this is only true if the optimizer is right in its assumptions about table size, index structure, predicate selectivity and clustering factor. If it's not, then you need to find where exactly the optimizer is wrong and correct it (e.g. by collecting a histogram on the column in WHERE clause).
    Best regards,
    Nikolay

  • Oracle Pkg missing?? Pls Help!!

    I have installed oracle on my pc(win2000 os) with version below:
         Personal Oracle8i Release 8.1.7.0.0 - Production
         With the Partitioning option
         JServer Release 8.1.7.0.0 - Production
    Here is few lines from file TNSNAMES.ORA
    TNSNAMES.ORA Configuration File:E:\oracle\ora81\NETWORK\ADMIN\tnsnames.ora     
    Generated by Oracle Enterprise Manager V2                         
    Date..........: Sat May 11 17:22:02 CDT 2002                         
    LINK2HR2 =                                             
    (DESCRIPTION =                                        
    (ADDRESS_LIST =                                        
    (ADDRESS = (PROTOCOL = TCP)(HOST = pria)(PORT = 1521))               
    (CONNECT_DATA = (SERVICE_NAME = link2hr2))                         
    I have installed VisualAge for Java 4.0 with JDK 1.2.2 that is offered for free by IBM.
    I did the following sqlj setup in VAJ workbench:
    1. workbench--->Tools--->SQLJ--->Properties
    Perform online semantics checking is checked
    JDBC Driver:     oracle.jdbc.driver.OracleDriver
    Default URL:     jdbc:oracle:thin:@pria:1521:link2hr2
    User:          hrdb     
    Password:     ????????
    2. Window--->Options--->Resources. I have the following in Workspace classpath I typed the following:
    "E:\oracle\ora81\jdbc\lib\classes12.zip;" w/o quotes.
    3. I have many Projects in my workspace along with "Java class libraries", "SQLJ Runtime Libraries".
    My Problem is:
    When I try to load and register JDBC driver with statement below:
    DriverManager.registerDriver(new oracle.jdbc.driver.OracleDriver());
    I get compile error saying "Type named oracle.jdbc.driver.OracleDriver is not defined."
    My questions are:
    Do I need to have any "oracle" Projects in my workspace. If so Where do I get it from??.
    I think I am missing some oracle classes.
    P L E A S E H E L P !!

    Try a Class.forName("oracle.jdbc.driver.OracleDriver");

Maybe you are looking for

  • Bug report: vpn (ipsec) interface number in snmp always change

    Hi, this is a bug report for RV082 hardware version 3 and 4, firmware version 1.x, 2.x and 4.x (all latest versions). I hope someone from cisco/belkin reads it. Summary: The snmp interface number of a VPN Tunnel change when the VPN tunnel disconnect

  • Another iTunes Won't run question

    Yesterday I tried to open iTunes to listen to my music and nothing-it wouldn't launch. So I uninstalled everything per Apple support, reinstalled iTunes and QT and still it will not launch or open. The hard drive spins as if to access the program the

  • Learning XI

    HI Friends.. I had been working as an abap developer for the past couple of years to one of the top most company in U.S. Right now iam planning to shift my career into SAP XI. I Found this forum really useful to get started with SAP XI. Though i get

  • Fresh Install 10.4.11 -Library 'Preferences' Folder

    Hi Gang I'm no novice to reformatting/installing new system drives Just wanna make sure I have this straight; The Library 'Preferences' Folder, obviously contains data relevant to each application that was installed. It was recommended I copy this EN

  • Deadlock when inserting lob

    I am getting several deadlocks. I have three client machines connecting via JDBC to our Oracle 10g server. Each of these clients insert about 10,000 rows a minute into a specific table. This table is defined as follows: create table contextdata ( acc