Any table for the data in ODS, PSA, INFOCUBE

Hi All,
      We are loading data to PSA and then to ODS and then to INFOCUBE , in such condition is there any table where we can check the number of data that can be loaded in each run into  PSA and then to ODS and then to INFOCUBE .
   When u check the data in the PSA you will see (Suppose):
                    Processing          PSA
                    Update Mode       Full update
                    Records Check    32 Records Received
Now here it is showing the number of data loaded to PSA then there must be one table where it is loading the data ..
Do anyone have the idea  about the table.
Thanks,
Gautam

Hi,
You can check the data in the PSA table and Active table of ODS.
PB

Similar Messages

  • Shortdump problem for loadinf data from ODS to InfoCube

    hai
    im trying to load the data from ODS to InfoCube.But i got the following error like below
    Short dump in the Warehouse
    Diagnosis
    The data update was not completed. A short dump has probably been logged in BW providing information about the error.
    <b>System response
    "Caller 70" is missing.
    Further analysis:
    Search in the BW short dump overview for the short dump belonging to the request. Pay attention to the correct time and date on the selection screen.
    You get a short dump list using the Wizard or via the menu path "Environment -> Short dump -> In the Data Warehouse".
    Error correction:
    Follow the instructions in the short dump.</b>
    I looked at the shortdump.But it says that there is no shortdump for that particular date selection.
    pls tell me wht i have to do
    i ll assing the points
    bye
    rizwan

    Hi Rizwan,
    Why does the error occurs ?
    • This error normally occurs whenever BW encounters error and is not able to classify them. There could be multiple reasons for the same
    o Whenever we are loading the Master Data for the first time, it creates SID’s. If system is unable to create SID’s for the records in the Data packet, we can get this error message.
    o If the Indexes of the cube are not deleted, then it may happen that the system may give the caller 70 error.
    o Whenever we are trying to load the Transactional data which has master data as one of the Characteristics and the value does not exist in Master Data table we get this error. System can have difficultly in creating SIDs for the Master Data and also load the transactional data.
    o If ODS activation is taking place and at the same time there is another ODS activation running parallel then in that case it may happen that the system may classify the error as caller 70. As there were no processes free for that ODS Activation.
    o It also occurs whenever there is a Read/Write occurring in the Active Data Table of ODS. For example if activation is happening for an ODS and at the same time the data loading is also taking place to the same ODS, then system may classify the error as caller 70.
    o It is a system error which can be seen under the “Status” tab in the Job Over View.
    What happens when this error occurs ?
    • The exact error message is “System response "Caller 70" is missing”.
    • It may happen that it may also log a short dump in the system. It can be checked at "Environment -> Short dump -> In the Data Warehouse".
    What can be the possible actions to be carried out ?
    • If the Master Data is getting loaded for the first time then in that case we can reduce the Data Package size and load the Info Package. Processing sometimes is based on the size of Data Package. Hence we can reduce the data package size and then reload the data again. We can also try to split the data load into different data loads
    • If the error occurs in the cube load then we can try to delete the indexes of the cube and then reload the data again.
    • If we are trying to load the Transactional and Master Data together and this error occurs then we can reduce the size of the Data Package and try reloading, as system may be finding it difficult to create SID’s and load data at the same time. Or we can load the Master Data first and then load Tranactional Data
    • If the error is happening while ODS activation cause of no processes free, or available for processing the ODS activation, then we can define processes in the T Code RSCUSTA2.
    • If error is occurring due to Read/Write in ODS then we need to make changes in the schedule time of the data loading.
    • Once we are sure that the data has not been extracted completely, we can then go ahead and delete the red request from the manage tab in the InfoProvider. Re-trigger the InfoPackage again.
    • Monitor the load for successful completion, and complete the further loads if any in the Process Chain.
    (From Re: caller 70 missing).
    Also check links:
    Caller 70 is missing
    Re: Deadlock - error
    "Caller 70 Missing" Error
    Caller 70 missing.
    Bye
    Dinesh

  • Error while loading the data from ODS to InfoCube

    hai
    Im trying to load the data from ODS to InfoCube for particular year .
    But it says that there is a source system problem .
    why it is like that .
    pls tell me
    i ll assing the points
    rizwan

    Hi Rizwan,
    you didn't mention the error message in details. there could be a few places to be checked:
    - check if BW itself source system is active and in tact and reactivate if necessary
    - check if update rule is active and reactivate if necessary
    - check if ODS is active and reactivate if necessary
    Regards,
    Lilly

  • How to Read the internal table for the data download from the spool

    HI all,
    I have one issue regarding the spool ,we are getting the correct output as per requirement of  user but when we send the same to the user in pdf format they did notget the same.
    they are telling that the due date is missing from the pdf.
    Please advice me how to track the internal  table for the spool data converted intopdf in a readable format.
    the FM used for the above task is : 
    call function 'CONVERT_OTFSPOOLJOB_2_PDF'
    Please reply if any one worked on the same.
    Thanks in advance.
    Gaurav,

    Hi Wang,
    Please let me know how you solved your question.
    Points will be rewarded.
    Thanks,
    Arun.

  • Is there any certificate for the data mining or OLAP ?

    Which certificate is that?

    user8117130 wrote:
    Which certificate is that?If you can't find it on the following page, it doesn't exist yet,
    http://education.oracle.com/pls/web_prod-plq-dad/db_pages.getpage?page_id=141
    HTH
    Aman....

  • Tables for Contract data

    Hi all,
    if you know any tables for contract data(For Contract report),
    please tell me.....
    Thanks,
    kishore.

    VBAK-contract header
    VBAP-contract item
    VEDA-changed contracts
    KONV-condition records
    MARA-material related to contract

  • The table for storing data for infocube and ODS

    Hi all:
        could you please tell me how to find the table for storing data for infocube and ODS?
    thank you very much!

    Hi Jingying Sony,
    To find tables for any infoprvider go to SE11.
    In database table field enter the following
    Cube -
    Has fact table and dimension table
    For customized cube - ie cube names not starting with ' 0 '
    Uncompressed Fact table - /BIC/F<infocubename>
    Compressed fact table - /BIC/E<infocubename>
    Dimension table - /BIC/D<infocubename>
    For standard cube - ie cube names  starting with ' 0 '
    Uncompressed Fact table - /BI0/F<infocubename>
    Compressed fact table - /BI0/E<infocubename>
    Dimension table - /BI0/D<infocubename>
    Click on display.
    For DSO,
    For standard DSO active table- /BI0/A<DSO name>00.
    You use 40 for new table.
    Click on display.
    For customized DSO use- /BIC/A<DSO name>00.
    An easier way is in the database table field, write the name of the cube/DSO preceeded and followed by ' * ' sign. Then press F4 . It shall give you the names of the available table for that info provider.
    Double click on the name and choose display.
    Hope this helps,
    Best regards,
    Sunmit.

  • Is there any way for updating wrong data in ODS and infocube?

    Hi all
        Is there any way for correcting wrong data in ODS and infocube?

    You could do a selective deletion, by specifying the filter values, so that you only delete a subset of the entire data. You could then do a re-load for the same selection so that there is no duplication of records.
    This is on the assumption, that your incorrect data is due to some incorrect update or transfer rules, which you have now corrected (before the re-load).
    mate, you have mentioned that You could then do a re-load for the same selection .
    could you please tell me how to do a re-load ?   thank you very much

  • How many years we can store the data in ODS as well as psa data ?

    Hi guru's
    how many years we can store the data in ODS? can u give particular years ?
    how many days we can store the data in psa ? could u tell me step by step procedure ?
    Thanks
    Reddy prabha

    Hi Prabhakarreddy,
    As many days as you like:
    In the PSA you can keep data as long as you like but you will need more and more disk-space. So its advisable to clean up from time to time.
    Example:
    PSA's from full loads Clean up in a weeks time,
    PSA's from Deltas: Clean up after you are sure data is correct. (e.g. a month or 90 days)
    Remember that you cannot do a reconstruction from dataloads of which you have deleted the PSA. If some thing is wrong you will have to do an Init again!
    In ODS's you must keep data in order to keep your delta loads to cubes in order. Removing data will force you to do a new initial load.
    Sometime companies delete old data form ODS's that is older than a certain period (e.g. more than 3 years old) by running a selective deletion. You will have to do an init again. But loads and activation may run faster and of course you gain on disk space.
    Udo

  • CLIENTThe maintenance tables for the logical files; may we use any client

    Please  see this link http://help.sap.com/saphelp_47x200/helpdata/en/9f/db95e635c111d1829f0000e829fbfe/content.htm
    See the import parameter
    CLIENT     The maintenance tables for the logical files and paths are client-dependent. Therefore, the desired client can be imported. The current client is stored in the system field SY-MANDT.
    Does it mean that we may use any client  even if our client number is different?

    I put it back once more

  • In which tables are the data for the aplication KSB1?

    I'm getting data from Qlik View with a BI tool called Qlik View.
    I need help to find in which tables are the data for KSB1 aplication.
    I've found something in COBK and COEP but there's more table involved since I didn't find all data.
    Somente can help me please?
    Regards.
    Robson

    I ran a trace and found some tables that may help you.
    COVP(View)
    CSKT
    CSLT
    CSKS
    CSKU
    AUFK
    Thanks
    Rathish

  • Index used or not for selecting data from ODS in a start routine

    Dear friends,
    In the start routine of the update rules to a cube, I am reading some data of an ODS in to an internal table .
    The ODS is indexed. But, I am not sure if the index is at all used in the Select statement (that gets the data from ODS to the internal table in the start routine) while loading data to the cube.
    Any help is highly appreciated.
    regards,
    atlaj

    Hi Atlaj
    You can findout this is display execution plan for SQL statement in DB02.
    Goto DB02. and under diagnostic, you find explain. Select that and enter your query. Make sure that everything here is in capital format. Below is a sample query which I have entered.
    SELECT "CRM_SALORG" "SALESORG" FROM "/BI0/QORGUNIT"
    WHERE "SALESORG" = ? AND "OBJVERS" = ? AND "DATETO" >= ?
    AND "DATEFROM" <= ?
    The select parameters should be inside qoutes and in caps and even the from table. Once you enter your query in this format, click on explain. It will show the index scan if the index is present. the output for my query will be something like
    0 SELECT STATEMENT ( Estimated Costs =  1,348E+01 [timerons] )
            1 (COOR) RETURN
                2 (    0) TQ
                    3 (    0) FETCH /BI0/QORGUNIT
                        4 (    0) IXSCAN /BI0/QORGUNIT~Z1
    Where my last statement (line 4) is showing index scan and the name of index read is Z1.
    Hope this helps.
    Please let me know if you have any problems entering the query in the specified format and u get any error.
    Regards
    Sriram

  • CLOSING STOCK QTY FOR THE DATE

    Hi
    Can any body tell, In which table i can find the closing stock qty for the particular date and for the particular material without including intransit stock qty.
    Regards
    Mani.

    Dear Santhosh,
    Yes, we tried to take the last MBLNR(Value) of the date from MKPF and Try to use the LBKUM Field from MSEG table. But the LBKUM value is with intransit stock.
    So , we failed in that logic to bring the warehouse stock for the date.
    Any other suggestions pls.
    Regards
    Mani.

  • Tables for Credit data

    Hi all,
    please tell me the tables  for Credit data.
    Thanks,
    kishore.

    1) Check to see if there are any entries in table S066 and S067. These SIS tables hold the credit data for sales order value and delivery order value.
    hope it helps

  • When target DB is down in which table is the data from source is stored, also where are the error messages stored in ODI

    When target DB is down in which table is the data from source is stored, also where are the error messages stored in ODI( I am not getting any error message in E$_TARGET_ANI_TEST).
    When i am running the interface i am getting the below error against the errored step
    "ORA-01045: user ABC lacks CREATE SESSION privilege; logon denied."
    Only E$_TARGET_ANI_TEST  is created with no data. No such tables like C$_0TARGET_ANI_TEST, I$_TARGET_ANI_TEST are created and also data is not inserted in the target table TARGET_ANI_TEST.

    Hi,
    I have checked that only E$ table is created. C$ and I$ table are not created ( I have selected my target schema as the part for the staging table).
    All the parameters for dropping the tables are selected as "<default>:false".
    I am importing the following KMs with the following parameters:
    1) CKM Oracle
    DROP_ERROR_TABLE
    :false
    DROP_CHECK_TABLE
    :false
    CREATE_ERROR_INDEX
    :true
    COMPATIBLE
    :9
    VALIDATE
    :false
    ENABLE_EDITION_SUPPORT
    :false
    UPGRADE_ERROR_TABLE
    :false
    2) LKM SQL to SQL
    DELETE_TEMPORARY_OBJECTS
    :true
    3) IKM SQL Incremental Update
    INSERT
    :true
    UPDATE
    :true
    COMMIT
    :true
    SYNC_JRN_DELETE
    :true
    FLOW_CONTROL
    :true
    RECYCLE_ERRORS
    :false
    STATIC_CONTROL
    :false
    TRUNCATE
    :false
    DELETE_ALL
    :false
    CREATE_TARG_TABLE
    :false
    DELETE_TEMPORARY_OBJECTS
    :true 

Maybe you are looking for

  • Adobe PDF form / submit sutton does nothing in Safari 5.1.2

    Hi, We display embedded Acrobat PDF forms containing a submit button in Safari. If I press "submit" nothing happens, this worked some Safari versions before. My Adobe Reader is: Adobe Reader X, Version 10.1.1 My Safari version is 5.1.2, I removed and

  • Flash Builder 4.5.1 Error when trying to run

    Hi, Since I installed 4.5.1 of Flash Builder (OS/X 10.6.7) I get the following error when I try to run anything: An internal error occurred during: "Launching New_configuration". java.lang.NullPointerException I created a new project to make sure it

  • I can not make 16bit tiff file out

    Hello. This is kamecame from japan. I am seeing this US board because Japanese discussion board is not active. I have a queation about shake. I hope to convert Quick Time(motion jpg)to Tiff file format(16bit LZW). I made a success of 8bit Tiff file o

  • Actiontec MI424-WR Rev. E and Windows 8

    Will this router I have funtion with W8. Thanks

  • Is there a way to import pictures into LR in original numerical order? .

    I am new to LR. Imported few hundreds of pictures taken by two different cameras and stored in numerical order in a folder. LR4 rearranged them in some mysterious order that makes no sense to me. I tried to re-sort them in numerical order, but there