Transactional data is loaded in a infocube

Hi Experts,
How to identify how much transactional data is loaded in a infocube?  Please recommend the best way to quantify this. 
For any suitable answare will be awarded with suitable points.
Thanks!!

CUBE -> Right Click -> Manage-> Requests Tab- > View the loaded Request
See two columns .
Transfered Record and Added number of Records.
Based on Update Rules, Added Record Count might be <= Transfered Rec Count.
Good Luck

Similar Messages

  • Can we load Transaction data without loading Master data, explain  w.r.tSID

    Can we load Transaction data without loading Master data, if so can you explain me how and when Surrogate Ids/ Dimension Ids  gets generated.

    Hi,
    We can load the transaction data without loading master data.
    But for loading performance issues it is recommended not to load the transaction data first.
    Every infoobject in the BW is identified by SID, that means which we are loading the transaction data, considering a scenario for an infocube, the system will identify the infoobject by existing sid against that infoobject in the bw and correspondingly create DIM id's.
    Now consider the secnario where master data is not loaded and transaction data is being loaded.
    Now the system should have sid to identify and infoobject and to create DIM id ( which is the combination of sid's of the characterisc in it)
    And as the sid's are not yet created i mean as there is no referece for the master data, the system will first create the sid for that infoobject it means it will create a copy of the master data in that infoobject nd then create the dim ids based on it and then load the data.
    Here the additional task of creating sids are also being done during loading the transaction data which hinders the loading performance
    I hope this is clear5 now
    Janardhan Kumar

  • What if i load transaction data without loading master data

    Hello experts,
    What are the consequences if i load transaction data without loading master data?Are there any other factors except the load performance because of SID generations etc and inconsistencies.
    <b>What kind of potential inconsistencies will occur?</b>
    Problem here is:
    when the transaction load starts a new master data such as employee(x) would be have been created in R/3 which does not exist in BW and hence transaction load fails.
    Thanks and Regards
    Uma Srinivasa rao

    Hi Rao,
    In case you load the master data after loading the transcation data and if there is any look up at the master data in teh update rules, then you can delete and reconstruct the requests in the ODS/Cube so that the latest master data is pulled in in the data target.
    Make sure you do the Apply hier/attr change before doing the delete and reconstruct.
    Bye
    Dinesh

  • Can transaction data be loade into info object

    Hi Gurus
    Can a Transaction data be loaded into info objects. Appreciate if some one give a simple Definition of transaction data.
    GSR

    Hi,
    You can probably do that but why would you want to do it? Transaction data is generally required for querying purposes and queries run best on multidimensional cube structure, not a flat structure.

  • Error in Transaction Data - Full Load

    Hello All,
        This is the current scenario that I am working on:
    There is a process chain which has two transaction data load (FULL LOADS) processes to the same cube.In the process monitor everything seems okay (data loads seem fine) but overall status for both loads failed due to 'Error in source system/extractor' and it says 'error in data selection'.
    Processing is set to data targets only.
    On doing a manage on the cube, I found 3 old requests that were red and NOT set to QM status red. So I set them to QM status red and Deleted them and the difference I saw was that the subsequent requests became available for Reporting.
    Now this data load which is a full load takes for ever - I dont even know why I do not see a initialize delta update option there - can Anyone tell me why I dont see that.
    And, coming to the main question, how do I get the process chain completed - will I have to repeat the data loads or what options do I have to have a succesfully running process chain or at least these 2 full loads of transaction data.
    Thank you - points will be assigned for helpful answers
    - DB
    Edited by: Darshana on Jun 6, 2008 12:01 AM
    Edited by: Darshana on Jun 6, 2008 12:05 AM

    One interesting discovery I just found in R/3, was this job log with respect to the above process chain:
    it says that the job was cancelled in R/3 because the material ledger currencies were changed.
    the process chain is for inventory management and the data load process that get cancelled are for  the job gets cancelled in the source system:
    1. Material Valuation: period ending inventories
    2. Material Valuation: prices
    The performance assistant says this but I am not sure how far can I work on the R/3 side to rectify this:
    Material ledger currencies were changed
    Diagnosis
    The currencies currently set for the material ledger and the currency types set for valuation area 6205 differ from those set at conversion of the data (production startup).
    System Response
    The system does not allow you to post transactions after changing the currency settings to ensure consistency.
    Procedure
    Replace the current settings with the those entered at production
    start-up.
    If you wish to change the currency settings, you must use programs to convert data from the old to the new currencies.
    Inform your system administrator
    Anyone knowledgable in this area please give your inputs.
    - DB

  • Set up table and transaction data delta load init

    Dear experts,
    I am following a document left by my predecessor on "post processing activities" for BI prod post go-live.
    And for transaction data it says to :
    1. execute init with data transfer infopack manually
    2. execute delta update infopack manually
    3. go to ECC : delete setup data (lbwg)
    4. setup the tables for  (OLIxxBW)
    5. then finally activate Process chains.
    Is this sequence correct?
    Shouldn't I first delete the setup table, setup the tables and then execute the infopackages??
    Regards,
    Alice

    Hi,
    The sequence is correct ,and after this you need to run the job for the collective run in SAP ,which is based on your delta queue.
    i.e if you choose direct delta it is not required ,but if you use 'queued delta' then you need to run the collective run to move the delta from the extraction queue to delta queue .This job should run before the delta load.
       It is better to fill the set-up table during downtime .ie.no documents posted during the set-up table run. But if you dont have an option you can fill the set-up table during operation hours ,but need fo follow certain steps to get all the delta data. You may need to use 'Init without delta transfer' or early delta initialization in the infopackage .You need ODS as a data target to achieve this . Please search SDN ,you will get more information about this.
    Thanks.

  • Deleting master data after loading transactional data using flat file

    Dear All,
    I have loaded transaction data  into an infocube using a flat file . While loading DTP i have checked the  option "load  transactional data with out master data exists" . So transactional data is loaded even if no master data is there in BW.
    While loading the flat file, I made a mistake for DIVISION Characteristic  where original master data value is '04000' , but i loaded the transactional data with value '4000' .Then i later realized after seeing the data from the infocube and deleted the request. then i reloaded data with value '04000'. Till now every thing is fine.
    But when I see the master data for DIVISION , i can see a new entry  with value '4000'.
    My question is how to delete the entry value '4000' from DIVISION. I tried deleting manually this entry from 'maintaining masterdata' , but it is not allowing me to do so .
    I have also checked if any transactional data exists for that value '4000' , as i said earlier I have deleted the transactional data with that values. even tried to delete the entries from the master data table, but i donot see a option to delete entries there.
    Please suggest me on this.
    Regards,
    Veera

    Hi,
    Goto RSA1 right click on the Info object and select Delete Master data. This will delete the master data unused existing in the table.
    If this master data is not used any where else just delete the master data completely with SID option.
    If even this doesnt work you can delete the complete table entire in SE14. But this will wipe out the entire table. Be sure if you wanna do this.
    Hope this helps
    Akhan.

  • This year master data-previous year transaction data load

    hi guys,
    I support some infocubes where financial data is stored and reported on.our source systems are financial databases.
    now we load transaction data every month-end throughout a year.we load master data(full load) from flatfiles whenever there are changes.whenever we change master data,we include these changes in transaction data infopackage selections,so the transaction data related to new master data will also be loaded.
    now my situation is:
    Now this is June,2010.My boss said I have to reload transaction data for 2009 as there was some changes in rates on which the data was calulated in 2009.Some new rates came in,so they re-calculated 2009 transaction data and sent it to us.Now we have to reload this 2009 data and delete OLD 2009 data from cubes.
    My question---As I am going to re-load 2009 transaction data...but the master data present in system is up-to-date..i.e, June 2010..(its not the same master data of 2009....there were few additions and renamings until today.)...so what do we need to do in this type of situations?
    Also,as I said above,whenever we have changes in master data,we change transaction data infopackage selections so they load transaction data related to this new master data also...so it means Infopackage selections of today donot reflect the infopackage selections of 2009....but we are still trying to reload 2009 transaction data with todays infopackage??
    How do we need to do in these type of situations?Do we just go ahead and load 2009 data with 2010 master data and 2010 infopackage selections........?
    Thanks alot for your patience....
    Rgds.

    we are actually having time-dependent master data....one version of master data for each year ...like for 2005..we have verion 2....for 2006..we have version 3....master data version and to which year they are linked to is maintained in a customised table....
    so coming back to my previous post...
    for 2009...we have version 4....and transaction data was loaded then.....
    for 2010..we have version 5...and transaction data will be loaded at the end of every month...and also we have decided to reload 2009 data because the the rates on which 2009 data was calculated,changed and so our business recalculated that data and sent to us and we have to reload 2009 data and delete OLD 2009 data....
    now my questions...if we re-load 2009 data,today master data is of 2010(version 5)....this is meant for 2010 transaction data loadings....Doesnt it create inconsistencies....or am I thinking too much...is just to go ahead and load....?
    Also the infopackage selections today donot match with infopackage selections of 2009 as there were additions in master data and so we included in selections....Do we need to change back the infopackage selections to 2009 until 2009 re-load is over?
    Hope I explained it better...............
    Thanks alot.

  • Transactional Data Load without Respective Master Data in Place

    Hi Folks,
    Could anyone let me know if i could load transactional data without loading the respective master data. If yes, well could you give me a scenario when you do it. Also could you let me know the option of how to load transactional data without master data in place.
    Best Regards,
    Prithvi

    Hi,
    You can load transaction data without loading master data.
    You have the option in the info package to load data even if no master data exists..
    You may use it in cases where your transfer rules/update rules or routines donot refer to master data tables to populate transaction data..
    Ashish

  • Can i load transaction data thro flat files without loading masters?

    Hi all,
    is it possible to load history data using flat files in sales cube without having loaded any masters?
    thanks

    Hi
    You can load the transaction data without loading the master data. You have to check the option in update rules (update also if no master data exists).
    In reporting you will not be able to see the attribute and text data. Its always adivceable to load master data first and run apply hierarchy/attribute change and load the transaction data.
    AHP: Nice to see you after a long time
    REGards
    Rak

  • Relation between master and transaction data

    Hi experts,
    I have a basic question how the relation between masterdata and transction data is maintained in R/3. As I know first we extract master data and later transaction data in a typical R/3 based scenario.
    1.How does BW knows that , the transaction data are  related masterdata in the same sense as in R/3 ?
    I am involved in an enhancement of R/3 extractor. I need to derive 3 more fields and map it to info objects in BW. Cld any one explain me ( with an example plz) how these alignments to be made. How can I establish relation b/w derived fields and transaction data during extractions ?
    regs
    D Bret

    Hi Dave.
    Let's see if I can get all of your questions answered.  First, when Transactional Data is loaded, BW checks to see if a SID exists for each characterisatic field & value in the load.  If one doesn't, it either creates a SID (Surregate ID) in the master data tables (default action) or the load fails (if the enforece referential intgegrity flag is set).  It's better to load master data first in either case as it will improve the performance of the transactional load.
    In terms of R/3, the data has similar relationships as in BW.  There are master data tables and transactional data tables.  The difference is that while R/3 uses ~1 master data table per field, BW uses more.  Same thing for transactional data.
    As for your enhancement, if everything you need for the derivation of the extra field is in R/3, then you can enhance the extractor in RSA6 and do the mapping in R/3.  Once you replicate the DataSource to BW and reconnect the transformation & update rules, the load should work.  If you can give a bit more information on what sort of enhancement you are attempting, I can give a better example.
    Cheers,
    Adam

  • BPC 7.5 NW: Loading transaction data from infocube

    Hello, I am trying to load transaction data from an infocube into a BPC application using a package based on /CPMB/LOAD_INFOPROVIDER.
    The master data (cost center and cost element) are already loaded. As they are compounded, we have added in the key the CO area (eg ID=0CO_AREA+ID)
    I am facing an issue that the system ads leading 0 in front of the values of those master data comming from the cube.
    So the data looks like CO_AREA00000000000cost_center.
    As the data where correctly loaded in the dimensions (CO_AREAcost_center), the loading fails.
    How can we remove those 'O'?
    Thanks!

    Hi Anton,
    We use this in most of our projects; since auditing is one of the important requirements.
    You should enable the auditing for selective categories only. With each transaction data entered, it will keep a log, which can increase the size a lot, and hence impacting the performance.
    If you are enabling the log, you should frequently purge the audit data as well. However, you should take a backup of this and then purge it.
    Hope this helps.

  • Problem in loading of flat file transaction data into BI7

    Hi Experts,
    I am new to BI7, got in problem in creation of Data source.
    I have fields in flat file are:
    CUSTID,SREPID,MATID,PRICE PER UNIT,UNIT,QUANTITY,TRANS.DATE
    one of the row under the above fields is
    C100,SR100,MAT01, 50,CS,5,19991005
    Created infoobjects are:
    IO_CUSID
    IO_SRID
    IO_MATID
    IO_PRC
    IO_QTY
    IO_REV
    0CALDAY
    IO_QTY was created with Unit/currency '0UNIT'
    In creation of Data source, under field tab system shown extra CS field, I did confuse that which infoobject should be map to this field. At the same time i created IO_QTY with 0UNIT so when i entered IO_QTY under infoobjects column, system automatically created 0UNIT infoobject. I didn't completly understand why one of the unit (CS) sat under the fields column. could you clarify me please.
    And also could you please tell me the procedure of creating transfer routine for sales revenue as we didn't have this object in flat file (SS). In earlier version of BW(Addison Wesley Step by Step book) i did this at infosource level. Here i am loading the data without infosource so just i would like to know is it possible to create transfer routine at DS level or do i need to create Infosource in this case. If it is the case also could you please make me understand with procedure to create transfer routine for sales revenue.
    Your solutions will be more appreciated with points.
    Thank you,
    Shri

    Hi,
    1.Open Administrator Workbench: Modeling, from the menu or using the transaction RSA1
    2.Go to Source Systems (File)and Create a new System.
    3. Double click on the datasources, create data source
    4.Entitle the Datasource, choose Transaction Data as Data Type Datasource
    5. In next screen, Go to Extraction , Activate the data source
    6.Go to Preview. Press Read Preview Data.
    7.Data will be loaded: Save and activate the datasource
    8.Go to InfoProvider, go to Info area and create Infocube
    9.Entitle the infocube, press create
    10.Display all info objects: . Choose your info objects , Move the characteristics and key figure to the infocube by drag and drop adn activate
    11.Go to InfoCube, create transformation
    12.Click on the ‘Show Navigator’ button. Match the fields in navigator. Save and activate.
    13.go to info provider, your info cube, create data transfer process
    14.Choose: Extraction Mode: full
    Update: Valid records update, Reporting possible (Request green)
    Activate, execute
    14.Double click on the data source, Read Preview Data
    15.Create an InfoPackage by right clicking on the datasource. Execute the load into PSA.
    16.Verify data was loaded into PSA before proceeding with the next step.
    17. Execute the Data Transfer Process
    18.Monitor the results (Menu: GoTo: DTP Monitor)
    19.Try to display the data at the BW Frontend.
    Thanks,
    Sankar M

  • BPC:NW - Best practices to load Transaction data from ECC to BW

    I have a very basic question for loading GL transaction data into BPC for variety of purposes, would be great if you can point me towards best practices/standard ways of making such interfaces.
    1. For Planning
    When we are doing the planning for cost center expenses and need to make variance reports against the budgets, what would be the source Infocube/DSO for loading the data from ECC via BW, if the application is -
    YTD entry mode:
    Periodic entry mode:
    What difference it makes to use 0FI_GL_12 data source or using 0FIGL_C10 cube or 0FLGL_O14 or 0FIGL_D40 DSOs.
    Based on the data entry mode of planning application, what is the best way to make use of 0balance or debit_credit key figures on the BI side.
    2. For consolidation:
    Since we need to have trading partner, what are the best practices for loading the actual data from ECC.
    What are the typical mappings to be maintained for movement type with flow dimensions.
    I have seen multiple threads with different responses but I am looking for the best practices and what scenarios you are using to load such transactions from OLTP system. I will really appreciate if you can provide some functional insight in such scenarios.
    Thanks in advance.
    -SM

    For - planning , please take a look at SAP Extended Financial Planning rapid-deployment solution:  G/L Financial Planning module.   This RDS captures best practice integration from BPC 10 NW to SAP G/L.  This RDS (including content and documentation) is free to licensed customers of SAP BPC.   This RDS leverages the 0FIGL_C10 cube mentioned above.
      https://service.sap.com/public/rds-epm-planning
    For consolidation, please take a look at SAP Financial Close & Disclosure Management rapid-deployment solution.   This RDS captures best practice integration from BPC 10 NW to SAP G/L.  This RDS (including content and documentation) is also free to licensed customers of SAP BPC.
    https://service.sap.com/public/rds-epm-fcdm
    Note:  You will require an SAP ServiceMarketplace ID (S-ID) to download the underlying SAP RDS content and documentation.
    The documentation of RDS will discuss the how/why of best practice integration.  You can also contact me direct at [email protected] for consultation.
    We are also in the process of rolling out the updated 2015 free training on these two RDS.  Please register at this link and you will be sent an invite.
    https://www.surveymonkey.com/s/878J92K
    If the link is inactive at some point after this post, please contact [email protected]

  • Problem in loading transactional data from to 0MKT_DSO1(ods) to 0MKTG_C01

    Hi,
    I am trying to load lead transaction data to the standard Crm lead management cube from ODS.There is a problem while loading transaction data from 0MKT_DSO1(ods) to the infocube 0MKTG_C01 as the field 0STATECSYS2(CRM STATUS)  is set to 10 in ods -meaning incorrect transaction. This feild is not there in the infocube.
    There is a routine in the cube that deletes data records with (0statecsys2) set to 10.
    THIS field is not coming in the transaction.
    so, where can i see the master data in crm source system? and why is that feild getting set to 10 ?
    thanks in advance!

    Thanks for the reply..
    I have checked the Fact table which shows
    1. packet Dimension
    2. Time dimension
    3. Unit dimension.
    I have kept the 0CALDAY as the time characteristics.
    Sample data i have loaded from ODS to Cube.
    Sample data in ODS.
    Sales order No_____0CALDAY_____AMOUNT
    800001___________12/02/2009____15
    I have loaded this data in Cube with Full Upload.
    Data in Cube.
    Sales order No_____0CALDAY_____AMOUNT
    800001___________12/02/2009____15
    Again i am loading the same data to cube
    Data in cube after loading.
    Sales order No_____0CALDAY_____AMOUNT
    800001___________12/02/2009____15
    800001___________12/02/2009____15
    The data is duplicated and it is not cumulating.
    Am i missing anything on this.
    Pls help..
    Thanks,
    Siva.

Maybe you are looking for

  • Wrong creation of legacy asset in AS91

    Hi ALL,     I had create an asset using tcode AS91 and the system assigned main asset number 50000001 automatically to it. Later on, we discovered that we do not want this asset migrated to SAP. (Not yet post to asset reconciliation account in GL) Th

  • Help? Adobe Acrobat 11 Standard - just purchased for windows 7 now I can't open pdf files or attach?

    Help!!! I just puchased the latest Adobe Acrobat XI Standard. I put in the Serial Key number but my computer tells me that a qualifying product could not be found on my computer so I can't use the adobe product. So now I can't open any files or attac

  • Save and load jtree

    Hi, I have a jtree which I would like to save and then load and display in a JPanel. Just for info, the nodes of the tree are instances of a user object i have created. Now I have looked into two ways of doing this: 1) Serialize the actual JTree obje

  • Laptop to Tv Screen

    Is there any way i can plug my Laptop into my Tv Screen?

  • How to associate jnlp content with javaws?

    When I follow a link to a Java WebStart .jnlp file it opens as an XML file and not useing WebStart (javaws). It is not obvious to me how to make the association. I'm currently working with Ubuntu 10.10 but it may be a more general problem.