Regarding ODS and Physical Tables

Hi,
can you tell me the difference between ODS and a physical table created using transaction SE11?
I am writing a program to update this table not by any update rules, it is just updating the table using modify statement.
suggest me which one to use a physical table or ODS?
Thanks,
Vijaya

Hi Vijaya,
You want to insert the records into the table/ODS using some ABAP...(not through update rules)..
In such a case you can do it either way...but going for an ODS will give an added advantage. You will be able to report on the data in ODS.
Hope it helps..
Regards,
Amol )

Similar Messages

  • Regarding ska1 and skb1 tables

    Hi ,
             I got the requirement to extract the data from ska1 and skb1 tables, is their any standard datasource maintaining for that or else shall i extract based on generic extraction on view. Is the table which i mentioned is correct or not
    This was the requirement which am having
    The data which am having in excel is maintained in Budget in the year and budget in the month and previous. And parallel in the same data am having Actuals for the month and Actuals for the previous month.

    Hi,
    These tables are master data for G/L accounts; there are no figures stored in these tables.
    Regards,
    Eli

  • Question regarding Classic and Advance table

    Hi,
    I have classic and advance table. I am populating with VO that brings 100 record. When i render page i see last record being displayed. How can i show my first record by defult on both tables?

    By default the first record will be displayed, if you are seeing the last one instead, you probably did a vo.last() in the code.
    Thanks
    Tapash

  • Regarding index and anla table

    Hi
    i have code as below and i have Select option input
    for fields bukrs, anln1 and anln2.
    SELECT bukrs      anln1     anln2     anlkl txt50 ktogr invnrz sernr deakt               
        FROM anla                    
        INTO TABLE o_anla                    
       WHERE bukrs IN wa_ex_bukrs.
       if sy-subrc = 0.
       DELETE o_anla WHERE anln2 NOT IN wa_ex_anln2.
       DELETE o_anla WHERE anln1 NOT IN wa_ex_anln1.
       endif.
    Does the below code change is more faster than above code.?
    SELECT bukrs anln1 anln2 anlkl txt50 ktogr invnrz sernr deakt               
        FROM anla                    
        INTO TABLE o_anla                    
       WHERE bukrs IN wa_ex_bukrs
       and anln1 in wa_ex_anln1
       and anln2 in wa_ex_anln2.
    I feel bukrs and anln1 and anln2 are primary key fields and could increase performance better.
    However there is no std index available for this 3 fields. do you think creating secondary index
    for this 3 field can increase performance.
    Pls confirm.
    Regards

    Hi,
    If bukrs, anln1 and anln2 makes the primary key then you are hitting the table with primary index itself(which is always a better option for good performance). Your 2nd query would always do better.  No need to create secondary index in this case.
    Regards,
    Ravi

  • Logical and physical table - help

    How to solve problem to connect dimension ''persons'' twice on the same mesure. It has to be connected twice on two different foreign keys...

    I have natural physical connection with dimension ''persons'' with mesure. but i cannot duplicate table persons on physical layer two times, so I did it on business layer adding new logical column ''persons_diff'' and I connected that (with new foreign key) column with the same mesure ''on business layer'', but it doesn't give me expected results...
    Please help.
    Ed

  • Query regarding passivation and PS_TXN tables

    Hi All ,
    I am working on a read only Dashboard UI where the DB user has only read privileges.
    The jbo.server.internal_connection uses the same DB connection to create PS_TXN & PS_TXN_SEQ tables which fails for obvious reasons & I get the error -
    "Couldnot create persistence table PS_TXN_seq".
    I have disabled passivation at all the VO levels and also disabled jbo.isSupportsPassivation to be false at the AMLocal level , but still I am getting this error.
    I have also increased the initial AM pool size in bc4j.xcfg & Connection Pool at the weblogic server level to avoid snapshots been written to this table.
    Is there any way I can prevent any interaction with this table as the client too is not interested in any kind of passivation to happen for the time being.
    Thanks

    Thanks for your reply Chris.
    Chris Muir wrote:
    You might be taking the wrong approach to solving this. Rather than disabling the AM pooling (which btw is not supported by Oracle) to the database, instead you can get ADF to passivate to file or memory of the app server.I am confused as to what exactly jbo.ampool.issupportspassivation = false does then ? I read on one of the blogs that its a viable use case for Programatic VOs ?
    Also regarding passivating to file system as per http://download.oracle.com/docs/cd/E12839_01/web.1111/b31974/bcstatemgmt.htm#ADFFD1307 , its not really a recommended approach.. hence was not going for that.
    Can you please throw some more light ?
    Thanks

  • Three questions regarding DB_KEEP_CACHE_SIZE and caching tables.

    Folks,
    In my Oracle 10g db, which I got in legacy. It has the init.ora parameter DB_KEEP_CACHE_SIZE parameter configured to 4GB in size.
    Also there are bunch of tables that were created with CACHE turned on for them.
    By querying dba_tables table , with CACHE='Y', I can see the name of these tables.
    With time, some of these tables have grown in size (no. of rows) and also some of these tables are not required to be cached any longer.
    So here is my first question
    1) Is there a query I can run , to find out , what tables are currently in the DB_KEEP_CACHE_SIZE.
    2) Also how can I find out if my DB_KEEP_CACHE_SIZE is adqueataly sized or needs to be increased in size,as some of these
    tables have grown in size.
    Third question
    I know for fact, that there are 2 tables that do not need to be cached any longer.
    So how do I make sure they do not occupy space in the DB_KEEP_CACHE_POOL.
    I tried, alter table <table_name> nocache; statement
    Now the cache column value for these in dba_tables is 'N', but if I query the dba_segments tables, the BUFFER_POOL column for them still has value of 'KEEP'.
    After altering these tables to nocache, I did bounce my database.
    Again, So how do I make sure these tables which are not required to be cached any longer, do not occupy space in the DB_KEEP_CACHE_SIZE.
    Would very much appreciate your help.
    Regards
    Ashish

    Hello,
    1) Is there a query I can run , to find out , what tables are currently in the DB_KEEP_CACHE_SIZE:You may try this query:
    select owner, segment_name, segment_type, buffer_pool
    from dba_segments
    where buffer_pool = 'KEEP'
    order by owner, segment_name;
    2) Also how can I find out if my DB_KEEP_CACHE_SIZE is adqueataly sized or needs to be increased in size,as some of these tables have grown in size.You may try to get the total size of the Segments using the KEEP BUFFER:
    select sum(bytes)/(1024*10124) "Mo"
    from dba_segments
    where buffer_pool = 'KEEP';To be sure that all the blocks of these segments (Table / Index) won't be often aged out from the KEEP BUFFER, the total size given by the above query should be less than the size of your KEEP BUFFER.
    I know for fact, that there are 2 tables that do not need to be cached any longer.
    So how do I make sure they do not occupy space in the DB_KEEP_CACHE_POOL.You just have to execute the following statement:
    ALTER TABLE <owner>.<table> STORAGE(BUFFER_POOL DEFAULT);Hope this help.
    Best regards,
    Jean-Valentin

  • Regarding BP  and Premise tables

    Hi,
    what are the tables that will be updated when we create Business partner and Premise.
    Please Reply
    Regards,

    BP details should be there in BUT000 table.
    Address Number should be there at BUT020
    address details will be there at ADRC
    telephone details will be there at ADR6
    Please press BUT* in se11 u wiill get the exact description also.
    premise :  EVBS,EVBST.....
    Please Allot points
    Regards,
    Shiv.

  • Improve Performance of Dimension and Fact table

    Hi All,
    Can any one explain me the steps how to improve performance of Dimension and Fact table.
    Thanks in advace....
    redd

    Hi!
    There is much to be said about performance in general, but I will try to answer your specific question regarding fact and dimension tables.
    First of all try to compress as many requests as possible in the fact table and do that regularily.
    Partition your compressed fact table physically based on for example 0CALMONTH. In the infocube maintenance, in the Extras menu, choose partitioning.
    Partition your cube logically into several smaller cubes based on for example 0CALYEAR. Combine the cubes with a multiprovider.
    Use constants on infocube level (Extras->Structure Specific Infoobject properties) and/or restrictions on specific cubes in your multiprovider queries if needed.
    Create aggregates of subsets of your characteristics based on your query design. Use the debug option in RSRT to investigate which objects you need to include.
    To investigate the size of the dimension tables, first use the test in transaction RSRV (Database Information about InfoProvider Tables). It will tell you the relative sizes of your dimensions in comparison to your fact table. Then go to transaction DB02 and conduct a detailed analysis on the large dimension tables. You can choose "table columns" in the detailed analysis screen to see the number of distinct values in each column (characteristic). You also need to understand the "business logic" behind these objects. The ones that have low cardinality, that is relate to each other shoule be located together. With this information at hand you can understand which objects contribute the most to the size of the dimension and separate the dimension.
    Use line item dimension where applicable, but use the "high cardinality" option with extreme care.
    Generate database statistics regularily using process chains or (if you use Oracle) schedule BRCONNECT runs using transaction DB13.
    Good luck!
    Kind Regards
    Andreas

  • Duplicate table  and  Alias table in the physical layer

    Hi,
    I want a clear information regarding the duplicate tables in the physical layer of the rpd.
    I mean what is the purpose of the duplicate tables in the physical layer.........
    thanks,
    chinna.

    what do you mean by duplicate of physical table? is it Alias table?
    if then,
    -> aliases are created to break the circular joins (which are not supported by obiee)
    -> and, as per apps standards no operation can be done on base tables which are imported directly to physical layer. Only aliases are joined and will be used in BMM layer
    -> useful to re-use the table more than one time , best example: W_DAY_D in apps rpd
    Edited by: Kishore Guggilla on Sep 23, 2010 10:41 PM

  • Alias Tables and Duplicate tables in Physical Layer

    Hi all,
    I have a small doubt in physical layer. We have two options when we right click table, 'Alias' and 'Duplicate'. To create another instance of a table, can we use any one of these. Does it make any difference
    Thanks & Regards,
    Jagi

    Hi,
    If u duplicate any table the join conditions will carry with respective tables , same name appending (TableName#1)
    In case of alias u need to specify joins conditions explicitly
    We can use same table with different constrains to the same table
    Aliases used, if you need to replicate a physical table and not create multiple tables with the same data
    If you delete any column in the Original table automatically deleted in alias
    table and Its not reflected in Duplicate table.
    Steps:
    Right Click the original table. Choose New Object and choose Alias.
    The mistake most people make is try to bring in the same table again and rename it or choose Duplicate by right clicking the original table.
    Thanks,
    Balaa...

  • Physical table list against Cluster Table CDPOS and PCDPOS

    Hello experts,
    For function customized requirement, we need to know the physical table
    list against Cluster Table CDPOS and PCDPOS and EDID4, just like
    Cluster Table BSEG contain with six physical tables
    BSAD/BSID/BSAS/BSIS/BSAK/BSIK. Also we want to know if there is any
    general way to find out the physical table list for any cluster table.
    My question is:
    1. How can I find all the transparent table for Cluster Table CDPOS?
    just like Cluster Table BSEG has transparent tables of
    BSAD/BSID/BSAS/BSIS/BSAK/BSIK.
    2. How can I find all the transparent table for Cluster Table PCDPOS?
    3. How can I find all the transparent table for Cluster Table EDID4?
    4. Additionally,I want to know if there is any
    general way to find out all the transparent tables for an specific
    cluster table.
    Many thanks.

    Hello,
    simply look in transaction SE11.
    Example:
    1.  SE11 -> Table CDPOS - Display. On Tab 'Delivery and Maintenance' you'll find Pool/Cluster 'CDCLS'.
    2. SE11 -> Table CDCLS -> Display. On next screen position on CDCLS-> Where-used-list -> Tables -> you'll find tables CDPOS and PCDPOS.
    Same thing with EDID4    -> EDI40 ...
    Regards Wolfgang

  • Table to know the last loaded date for a ODS  and InfoObject?

    Table to know the last loaded date  and time for a ODS  and InfoObject?

    Hi Muruganand,
        The only way is through code ... there is no direct and simple method ... what you need to do is find hte highest Timestamp value for DP_NR = 000001. Then get the request field (RNR) ... this is the last request. Since you have to do it for 300 cubes ... I suggest you do some coding .. I can guide you through if you need assistance.
    Best regards,
    Kazmi

  • Dimension table and fact table exists data physically

    Hi experts,
    can anyone plz tell me weather dimension table and fact table exists data physically or not/

    Hi..Sudheer
    SAPu2019s BW is based on "Enhanced Star schema" or "Info Cubes" database design.This database design has a central database table, known as u2018Fact Tableu2019 which is surrounded by associated dimension tables.
    Fact table is surrounded by dimensional tables. Fact table is usually very large, that means it contains
    millions to billions of records.
    These dimension tables doesn't contain data  it contain references to the pointer tables that point to the master data tables which in turn contain Master data objects such as customer, material and destination country stored in BW as Info objects. An InfoObjects can contain single field definitions such as transaction data or complex Customer Master Data that hold attributes, hierarchy and customer texts that are stored in their own tables.
    SID is surrogate ID generated by the system. The SID tables are created when we create a master data IO. In SAP BW star schema, the distinction is made between two self contained areas: Infocube & master data tables/SID tables.
    The master data doesn't reside in the satr schema but resides in separate tables which are shared across all the star schemas in SAP BW. A numer ID is generated which connects the dimension tables of the infocube to that of the master data tables.
    The dimension tables contain the dim ID and SID of a particular IO. Using this SID the attributes and texts of an master data Io is accessed.
    The SID table is connected to the associated master data tables via teh char key.
    Fact table(Transaction data,DIM ID)<>Dimention Table(SID and Dim ID)<->Masterdata table(SID,IO)
    Thanks,
    Abha

  • Expand the Warehouse tables X adding physical tables and joins

    I was questioned about these options:
    a) Add physical tables (from other DBs) and modifying the physical model with new joins + modifying logical model to include new columns
    b) Expand warehouse tables to include new columns in the tables. Doing a complementar ETL in order to feed the additional columns with data. Logical model would be updated but the idea in this option is to avoid joins in the physical layer.
    My understanding is that option a) despite the joins in the physical layer would be a better strategy.
    Pls. I'd appreciate any comments on the performance side or in the amount of effort to create the complementar ETL...
    Txs.
    Antonio

    Hi Lombo,
    I am not sure about what you mean with option a. In my understanding, you are asking for a comparison between:
    a) adding data from an additional source in the RPD:
    This means that you have to create an additional data source in the physical layer. I do not think you can create relations between physical tables sourced from different data sources in your physical layer. This means that the data is related in the logical layer. By doing this, the BI Server has to join the data from both sources for each front-end request. Also, it will have to perform the last aggregations instead of shipping it to the database. This is a big performance hit in terms of response time.
    However! It can be much and much faster in terms of development effort required to support this. If it's only a limited number of columns and tables being added, this may be a valid option. You can also use this approach to prototype option b.
    b) adding data from an additional source using ETL:
    Quite some work in terms of development effort. However, all complexity and performance hits are moved to the bottom of the stack and process: database and ETL. You will need to create an additional data source in the DAC, additional custom folder(s) in Informatica and the ETL logic to update the existing tables and/or load new tables.
    Additionaly, you need to expand the RPD with the new columns / tables, just as you would do in option a. However, now it will be sourced from the same datasource: less work for the BI Server to deliver the dataset to satisfy the request.
    So basically, I would consider:
    - How does the system currently perform: response times in the front end ( bad -> go for option b )
    - How many columns will be added and how often will they be used in the front-end
    - Informatica / DAC / OBI expertise available
    - How long do you have to facilitate reporting on data from the 2nd source
    Good luck!

Maybe you are looking for

  • Ideal Settings For HD Playback Of Timeline?

    Any suggestions for high def playback of the timeline?  Seems a bit choppy.  The timeline exports the files just fine however. ??  I have Win 7 64 bit i7 980X@ 3.33 6 core 12 GB RAM 1600DDR Radeon 5850 GPU. Is there a BIOS setting that could be tweak

  • Add recurring entries for days of the week

    Hello, I'm trying to setup a daily schedule for my dog walking business. I'd like to have each dog's name automaticaly listed under the day of the week which they are to walk on. For example: "Scarlett" walks on Tuesdays, Wednesdays, and Fridays. If

  • Performance tuning Issue

    Hi folks, I having a problem with performance tuning ... Below is a sample query SELECT /*+ PARALLEL (K 4) */ DISTINCT ltrim(rtrim(ibc_item)), substr(IBC_BUSINESS_CLASS, 1,1) FROM AAA K WHERE ltrim(rtrim(ibc_item)) NOT IN select /*+ PARALLEL (II 4) *

  • Alert Category - Runtime Work Bench

    Hi Experts, Can any please let me know how to logon to Runtime Workbench (RWB) to create an alert rule...as I'm new to Alert Management? Thanks in advance, Regards Vinod

  • Nuance Dragon PC and/or Mac on Intel Macs???

    The Nuance advisory relative to system requirements is that Dragon Dictate works on any Mac with an Intel processor.  A rep I spoke with at Nuance, assured me that this is the case and applies to booting into Windows on a Mac to use Dragon Naturally.