Procedures for implementing a snapshot scenario with custom DataSources

Hi Gurus,
I have checked the How To paper ([How to Handle Inventory Management Scenarios in BW (NW2004)|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/f83be790-0201-0010-4fb0-98bd7c01e328]). However, only SAP standard BW objects are mentioned in the paper e.g. InfoCube (0IC_C03), Material Stock InfoSource (2LIS_03_BX), Material movements IS (2LIS_03_BF) and Revaluations IS (0LIS_03_UM).
On the contrary, I need to handle custom DataSources for the Snapshot scenario. Are there any differences in the implementation methodology? Which additional aspects should I take into consideration? For example, the load sequence, delta type, etc.
Could you please list out the step-by-step procedures for such an implementation?
Thanks in advance!
Regards,
Meng

Hi Meng,
You can approach this in two ways.
1) If the volume of data is not much, you can derive the balance at query level, as follows.
User enters the date, based on this restrict your key figure to display all values less than this date.
2) If the volume of data is high, then you will have issues with performance if you are calculating the balance in the front end. In this case, you can model this with 'Non cumulative' key figure.  Again there are 2 ways of approaching this back end solution based on the volume of data. ( Say in one case you have 2 years of history in your DSO and in the second case, you have  5 years of history ).
A) For example, If there are only 2 years of history
Create a non cumulative Key figure 'ZBALANCE' with inflow and outflow, in a cube.
Map this to your credit and debit as + and - respectively and map the calender day to posting date.
Just initialise the dataload with data transfer and start loading the delta as normal.
You will be able to see the balances for each and every calday in your reporting.
This approach is straight forward and simple.
Compress the cube for getting the better performance.
B) If there are 5 years of history and you are not interested in loading all the 5 years data in getting the balance
Here you want to have the initial balance, continue delta and would like to load 2 years of history.
The cube and non cumulative KF are created as mentioned above.
For generating initial balance, you have to create another DSO without calander day and ZBalance mapped to credits and debits in additive mode. Load your DSO data into this new DSO to generate initial balance. This balance will be loaded to your cube as initial balance. ( Like 2LIS_03_BX ).
You have to compress this request with marker update ( Must ).
Load your historical data for 2 years from the original DSO. Compress without marker update ( Must ).
initialise without data transfer from DSO to cube and load deltas normally.
Compress the delta requests normally for performance reasons.
Please read the 'Inventory document' in detail.
Please let me know, if any of the information is still not clear.
Thanks,
Krishnan

Similar Messages

  • Implementing an EDI scenario with XI

    Hi,
    I need to implement an EDI scenario with a non SAP system. I would like to know what are the "components" that I should have in order to acomplish the task ?
    (how is it different from an ALE scenario)
    I'm new to EDI ! (so take is slow...)
      Thanks, Nir.

    Hi,
    I need to implement an EDI scenario with a non SAP system. I would like to know what are the "components" that I should have in order to acomplish the task ?
    (how is it different from an ALE scenario)
    I'm new to EDI ! (so take is slow...)
      Thanks, Nir.

  • Design Approach for 1:N Multimapping scenario with SAP ECC Receiver

    Hi Experts,
    I am trying to find the best approach to implement the following scenario. its desribed as follows:
    Legacy Database ->XI ->  SAP scenario.
    1. Pick records from database table with status ='n'
    2. The records picked can have one or more RefNos i.e. if 2000 records are picked 1500 can have RefNo :1111 and rest 500 RefNo: 2222
    3. 1st condition is to split the records into multiple messages if reference nos are different, so for scenario in point 2 it will be 2 messages. Secondly if the no of records are more than 1000 then split further into 2 message. So 1500 Ref No would be split into messages with 1000 and 500 records. Hence we get 3 messages. I know something similar was achieved in this blog:
    </people/claus.wallacher/blog/2006/06/29/message-splitting-using-the-graphical-mapping-tool
    4. Once the data is forwarded to ECC and its successfully updated, we need to update the status of the database table on the sender side to 'y';
    So the point where i am not very clear as of now is,
    1. What alternative I choose for PI-> ECC call? Its a 1:N multimapping split scenario and as far as I know the call has to go via AE, so is RFC the only option or is there a way to have a call with Proxy/IDOC? Or will there be an option to limit such a scenario at the JDBC Adapter level so that split is not needed. I am not sure if that can be achieved and is a better solution.
    2. If I go for either of the approaches mentioned above, what is the best way to achieve point 4, i.e. update the database table at the sender. As JDBC Adapter would be async and hence we either use a Aysnc-Sync bridge(for proxy/rfc receiver) or some other option like triggering a outbound interface from SAP side with the update data?
    Hope that experts can provide the inputs on the best way forward.Let me know if anything is missing from scenario details perspective.
    Best Regards,
    Pratik

    Pratik,
    2. If I go for either of the approaches mentioned above, what is the best way to achieve point 4, i.e. update the database table at the sender. As JDBC Adapter would be async and hence we either use a Aysnc-Sync bridge(for proxy/rfc receiver) or some other option like triggering a outbound interface from SAP side with the update data?
    Use the solution # 2, make it like this:-
    DB (async) -> PI -> Proxy (sync) -> PI -> DB (async)
    Just out of curiosity, why you breaking the records into 1000's in PI? Why not you select only 1000 records when polling the DB? That will help in improving the overall performance.
    Regards,
    Neetesh

  • Complete procedure for implementing sms facility in SAP

    Hi All,
    Can you please tell me ? how can we enable SMS facility ( SAP to mobile ) .
    I know some settings thru tcode SCOT, but its still not working..  what will be the complete procedure for this?
    do we need to implement new SMS server for it.. or what new things need to do by me?
    plz tell me.. i know this que asked several times in SDN but i want more precise information on this...
    Thankss a lott in advance,
    Shailesh

    Hello,
    Check the below link, might help.
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/b03d0983-bda1-2b10-c09c-e93bb9956055&overridelayout=true
    http://sappandit.wordpress.com/2009/05/03/sms-setting-in-sap-scot-configuration-for-sms/
    Regards,
    Yoganand.V
    Edited by: Yoganand Vedagiri on Dec 11, 2009 2:02 PM

  • Procedures for implementing Process chain

    Hai,
    Can any help me to know the step by step procedure to implement process chain
    in BW.
    Regards,
    Giri

    hi,
    refer to this link.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/8f/c08b3baaa59649e10000000a11402f/frameset.htm
    hope it helps,
    regards,
    Parth.

  • Extract Structure not Being transported with custom Datasource

    I'm having trouble transporting a custom datasource from BW 3.5 dev to QAS.  The datasource is coming through and showing up in RSA6 but when I click on it I get the following error:
    "The extract structure ZOXDBW0123 of the DataSource ZRSPCTRIGTEXT is invalid
    Message no. R8444
    Diagnosis
    An invalid extract structure has been assigned to the DataSource. An extract structure has to be active in the DDIC. It cannot be a view, since the customer is not able to add append fields for filling in the customer exits to a view."
    Any ideas?
    Thanks,
    Luke

    Thanks for your input.  We managed to get through one of the errors on about the 8th try to transport everything together.  It might of had something to do with the way things were being grouped in the packages.
    So now were just getting an RJ044 Error: "ZRSPCTRIGTEXT: Only tables and views are permitted for extraction". 
    I think it may be some kind of program error after all, cause our basis consultant took a look and couldn't figure it out at all either.  The transport says that everything is going through, but the view were extracting from is still not showing up in QAS.  He suggested we blow everything away and try building the datasource again from scratch.  So if we can't figure it out today we may try that route tomorrow. 
    Anyway, thanks again for all the help.

  • Master Data With Custom Datasource.

    Hi,
    When we create the custom datasource ( Transactional ) and load it to the BW, it has  master data field Company code,Cost Centre.
    Doubt is when the company code, Cost Centre are coming from the transactional Custom data source(with data), do we need to still load the Master data from the Master data data sources .
    What are points to be considered and known?
    Please provide your valuable inputs to understand this condition.

    Hi Adhvirao,
    As suman said there is master data sources like customer master, vendor master...
    check there is SAP predefined master datasources like 0CUSTOMER_ATTR..
    First you have to load master data and the only transaction data.. then only you can get rid of data mismatches in the report.
    Otherwise there will SID collusion, then it might be chaces you won't get the proper result in reports.
    Regards,
    Ranganath.
    Edited by: Ranganath on Jul 27, 2011 1:01 PM

  • Stored procedures for read only snapshots

    Hi
    I have to convert some tables into read only snapshots as per client's requirements,
    There are some stored procedures which are based on these tables which basically fetch values from these tables, how do I use these existing procedures so that they can be used even after converting the tables into snapshots ?

    >
    Sean Byrne wrote:
    > I have a calculated date shared variable that I would like to use as a parameter. This date is not contained in the database. The user would put in 12/31/09, and only records that are before this date would show up. I want to use a stored procedure to achieve this result. Does this calculated date have to be written to a table for this to work? The tables are controlled by our ERP on demand vendor, which restricts us from writing anything to their tables. They allow us to use Crystal Reports to retrieve data but no modifications of their tables. One way around this restriction is to create another database on the server to pull data from the ERP database. Any advice would be appreciated? Can I write stored procedures within Crystal Reports?
    I think you are just trying to query your database with a date parameter correct?  In your case I doubt you want to use a "stored procedure", but a command object that holds a simply sql query.  Are you familiar with SQL?  When you create your report, you will be asked to select your datasource. Once your database credentials are entered you should see rows that says "Add Command, Tables, Views, etc".  Select Add Command and you will be presented with a window in which to enter your SQL command.  On the right of that box you will set your parameter. Create your "date" parameter.
    Your query will look something like this
    Select * from table.table where date <
    You can add your dataparam by double clicking it after you created it.
    HTH
    Chad

  • Procedure for creating Recording in  LSMW with header data and line item

    Hello Friends,
    I have to do the Initial Upload of Model Service specification using LSMW ( Recording).
    Transaction : ML10
    The problem i am facing is that i have been given 2 files one with HEADER data and the other with LINE ITEM(Service data).
    Can we use the RECORDING method to upload 2 files?
    If not possible how do we proceed.
    Thanks and regards
    Ashish Naik.

    Step-by-Step Procedure:
    Details of the BAPI used in this scenario:
    Business Object: BUS2012
    Method: CreateFromData
    Details of Message Type and Basic IDoc Type:
    Message Type: PORDCR
    Basic IDoc Type: PORDCR02
    Letu2019s have a look at the BAPI first, before proceeding to the LSMW:
    1.     Go to Transaction BAPI
    2.     Click on Search Button
    3.     Enter the value u201CBUS2012u201D and select u201CObj.type(Technical Object Nameu201D
    4.     Press ENTER
    5.     Following screen appears:
    6.     On the left side of the screen, Expand the u201CPurchaseOrderu201D.
    7.     Select u201CPurchaseOrderu201D and double-click on the same for details.
    Building LSMW using BAPI:
    1. Go to Transaction LSMW.
    2. Enter the Project, Subproject and Object information and click on CREATE.
    3. Enter the descriptions for Project, Subproject and Object as prompted.
    4. Now select Settings à IDoc Inbound Processing
    5. u201CIDoc Inbound Processingu201D screen appears. Enter the required details as shown below:
    6. Click on u201CActivate IDoc Inbound Processingu201D.
    7. Click on u201CYesu201D when prompted for u201CActivate IDoc Inbound?u201D
    8. Hit on u201CBacku201D to return to the main screen.
    9. Click on Continue (F8). Following Screen appears:
    10. Select the Step 1 u201CMaintain Object Attributesu201D and select u201CExecuteu201D.
    11. Select the radio button u201CBusiness Object Methodu201D and enter the following details:
    Business Object: BUS2012
    Method: CreateFromData
    Hit ENTER
    12. Save and click on BACK button. Following information message is displayed.
    13. Now select step 2 u201CMaintain Source Structuresu201D and click u201CExecuteu201D.
    14. In this step, we need to maintain the source structure. In our example, lets consider the example of a file with 2 structures Head and Item data as shown below:
    Click on Create and name the source structure as HEADERDATA. Now select HEADERDATA and click on u201CCreateu201D again to create the child structure. Following popup appears:
    Select u201CLower Levelu201D and click on Continue. Enter the Item data structure name.
    Click Save and hit BACK button to go to the main screen.
    15. Select step 3 u201CMaintain Source Fieldsu201D and hit execute.
    16. Enter the fields as shown below:
    17. Click SAVE and return to main screen.
    18. Select step 4 u201CMaintain Structure Relationsu201D and click Execute.
    Select E1PORDCR and click on CREATE RelationShip. Following screen appears:
    Select HEADERDATA and hit ENTER
    Similarly do the same for the structure E1BPEKKOA, E1BPEKPOC and E1BPPEKET.
    Click Save and return to main screen.
    19. Select the step u201CMaintain Field Mapping and Conversion Rulesu201D and click on execute. Maintain the Field Mapping as seen below:
    20. Select step 7 u201CMaintain Source Filesu201D and provide the link for the test file created. (Create a test file with the same structure as defined earlier).
    Save and return to main screen.
    21. Select the step u201CAssign Filesu201D and click on Execute.
    Assign the file provided to the source structure. Here the same file is provided for both the structures.
    Save and return to the main screen.
    22. Select the step u201CRead Datau201D and click on Execute.
    Click on Execute.
    Return to the main screen.
    23. Select the step u201CDisplay read datau201D and click on execute.
    Click on the structure name to get the field level values.
    24. Return to main screen and now select u201CConvert Datau201D.
    25. Return to the main screen and select u201CDisplay Converted datau201D.
    26. Return to main screen and select u201CStart IDoc generationu201D.
    27. Now select the step u201CStart IDoc Processingu201D on the main screen.
    28. Return to main screen and click on u201CCreate IDoc overviewu201D. Here the data record and status records of the IDoc could be viewed:
    Check this. This may help

  • Message flow in SAP PI 7.31 JAVA only for ABAP Proxy synchronous scenario with Oracle DB

    Dear Experts,
    I am working on a synchronous scenario Sender ABAP Proxy <===> SAP PI 7.31 JAVA only <===> Oracle Database.I have successfully
    configured all the proxy configurations as well both Sender SOAP and JDBC receiver channels are showing no error when pinged from
    Cc monitoring in NWA.
    1.During testing at Tx SPROXY , I am triggering a message and after 5 minutes I receive an PARSING error message with 2 messages in ECC
    local IE i.e. Tx SXMB_MONI.
    Note : In NWA , I can see only 1 message that too in cancelled status.
    2.Seeing the trace of the messages I assume that the messages have not moved from sender ECC system. I am still confused with flow of message
    from the Message Audit log. Audit log of error messages.Could any one share the full Audit log of a successful message generated in SAP PI 7.31 JAVA only.
    3.I observed that the QIN Scheduler status in ECC is in INACTIVE status. There is no queue id for 2 messages generated in ECC. Does
    it should in ACTIVE status in Tx SMQR..
    Regards
    Rebecca

    Hi Thomas,
    Valuable inputs..
    1. Since I don't have any audit log for successful message, would you be able to share referring to the
    below screenshot whether the message was put into receiver JDBC adapter and JDBC adapter tried to retrieve the Select response from Oracle Database.
    2.I have done some setting in Advance mode of the receiver JDBC to increase the time out and other parameters as per SAP note but still the issue is same. If you perhaps have some experience in such scenarios, please share some input so that I can ask the Oracle Database administrator exactly what is required from our side.
    3.Regarding the slow response from Oracle , there was some suggestion that the datatype of the request and response should match the column name of the Oracle database. I did that.. Do you have any more suggestion that may be hampering the generation of the response message from the Oracle like SAP note or document.
    Regards
    Rebecca

  • Best practices for a development/production scenario with ORACLE PORTAL 10G

    Hi all,
    we'd like to know what is the best approach for maintaining a dual development/production portal scenario. Specially important is the process of moving from dev. to prod. and what it implies in terms of portal availability in the production environment.
    I suppose the best policy to achieve this is to have two portal instances and move content via transport sets. Am I right? Is there any specific documentation about dev/prod scenarios? Can anybody help with some experiences? We are a little afraid regarding transport sets, as we have heard some horror stories about them...
    Thanks in advance and have a nice day.

    It would be ok for a pair of pages and a template.
    I meant transport sets failed for moving an entire pagegroup (about 100 pages, 1Gb of documents).
    But if your need only deals with a few pages, I therefore would direclty developp on the production system : make a copy of the page, work on it, then change links.
    Regards

  • Character Encoding for IDOC to JMS scenario with foreign characters

    Dear Experts,
    The scenario is desribed as follows:
    Issue Description:
    There is an IDOC which is created after extracting data from different countries (but only one country at a time). So, for instance first time the data is picked in Greek and Latin and corresponding IDOC is created and sent to PI, the next time plain English and sent to PI and next Chinese and so on. As of now every time this IDOC reaches PI ,it comes with UTF-8 character encoding as seen in the IDOC XML.
    I am converting this IDOC XML into single string flat file (currently taking the default encoding UTF-8) and sending it to receiver JMS Queue (MQ Series). Now when this data is picked up from the end recepient from the corresponding queue in MQ Series, they see ? wherever there is a Greek/latin characters (may be because that should be having a different encoding like ISO-8859_7). This is causing issues at their end.
    My Understanding
    SAP system should trigger the IDOC with the right code page i.e if the IDOC is sent with Greek/Latin code page should be ISO-8859_7, if this same IDOC is sent with Chinese characters the corresponding code page else UTF-8 or default code page.
    Once this is sent correctly from SAP, Java Mapping should have to use the correct code page when righting the bytes to outputstream and then we would also need to set the right code page as JMS Header before putting the message in the JMS queue so that receiver can interpret it.
    Queries:
    1. Is my approach for the scenario correct, if not please guide me to the right approach.
    2. Does SAP support different code page being picked for the same IDOC based on different data set. If so how is it achieved.
    3. What is the JMS Header property to set the right code page. I think there should be some JMS Header defined by MQ Series for Character Encoding which I should be setting correctly) I find that there is a property to set the CCSID in JMS Receiver Adapter but that only refers to Non-ASCII names and doesn't refer to the payload content.
    I would appreciate if anybody can give me pointers on how to resolve this issue.
    Thanks,
    Pratik

    Hi Pratik,
         http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/502991a2-45d9-2910-d99f-8aba5d79fb42?quicklink=index&overridelayout=true
    This link might help.
    regards
    Anupam

  • Tomahawk for jsf 1.2 problem with customizing treetable component

    hi,
    i've tried many times bu i've got no clue how to set custom icons in the tree. what i want to achieve is to display the server's local file system. however this component doesn't allow to explicit specify what the node type is (a directory or a file). as a result when it finds an empty directory entry, no icon is displayed. i googled a bit and i guess the project's documentation is not sufficient. any help appreciated. ofc if someone knows a better method (or a framework) to do the job, please let me know. thanks
    cheers

    r035198x wrote:
    You could store the value in the session and remove it when you need to use it.Thanks, r035198x. I tried that before and users from other Java forums suggest that I shouldn't do it this way as I am adding more codes than necessary to pass technically a value from one bean to another. Hence, I decided to follow their advice and went and created a generic bean that can be referenced by other beans for pages calculation. Some users also suggest using Dependency Injection (available for JSF 1.2) to try to accomplish the result I want. I've also read about this method in JSF 1.2 spec - but without much example, I'm pretty much in the dark as to how this work. Unfortunately, I've only been on JSF for the last 4 weeks so I'm still a newbie - please excuse any "dumb" questions I have asked on this forum so far...
    What I currently has working for me is moving the method to calculate the pages from mybean1 to mybean2. This means all I have to do now for the other beans (11 remaining) is passed a few parameters to mybean2's method that calculate pages.
    Edited by: icepax on 12/11/2009 18:27

  • Step by step procedure for implementing User auditing in Solaris 10

    we have no.of users who use either SSH terminals and Java Desktop Environment through Xbrowser through network.
    We would like to know what are all the files being accessed by logged in users and what are all the changes being done, if any....
    Earlier, we have enabled "script" command to get invoked automatically in ~/.bash_profile. At this stage, users are not able to get Java Desktop through xbrowser. Hence, dropped this option.
    Please suggest to get rid of this script problem or some better solution through Audit logging.
    Thanks in advance.

    In my experience, its very rare for people to write step-by-step lists here, this is a forum which is excellent for general questions and technical problems, but, since most people who answer posts here do it in their spare time, its rare with detailed instructions of things which is rather well documented in the manual..
    Having said that, the things you asks for is actually documented, have a look at:
    http://docs.sun.com/app/docs/doc/816-4557/audittm-1?l=en&q=audit&a=view
    .7/M.

  • BAPI for creation of Credit Note with custom pricing condition

    I need to create a Credit Note with reference Invoice via program.  So, I am planning to use BAPI. Could you please suggest appropriate BAPI?  I need to send a pricing condition that is set manually in the credit note as well.  So, I will needto pass pricing condition and value.  Thanks.

    Hello,
    Why you have to create the credit note using BAPI w.r.t. invoice ? Can you brief the senario .
    The simple way is create a credit memo request in VA01 w.r.t. invoice number. It will have the block . Remove the block and Create credit memo in VF01 .
    Thanks
    RK

Maybe you are looking for

  • Problem with inbound ichats

    We are having a problem with inbound video ichats to our domain. It doesn't happen all the time but it happens enough to where it is a problem. The users can type back and forth, but when they go to ichat it fails. Here is the error log: Date/Time: 2

  • How to customize SORT event in the View Form ?

    I'm using Jdevelper 9.0.3 with BC4J+UIXML, and I need customize the View Object Query, including "bind variables" as: SELECT A.COD_CLIENTE, A.NUM_TESTE, A.NOM_TESTE FROM TESTE A WHERE A.COD_CLIENTE = :1 I create the java class for prepare assign valu

  • Connection Issues in the Connection Manager

    Good Morning Friends, I hope i will get some answer or some directions here . I have been trying to find a solution for months. I have two servers TESTSQL(2008R2) PRODSQL(2008R2) . On these twos servers i have a package that loads data from source da

  • ASSIGN field-symbol doesn't throw exception?

    if <fs> is already assigned, and then assign it to a non-exist column. it seems that <fs> is still assign to the original column, no exception raised.     i've no idea how could i detect the failure of assgin and raise exception by myself?     thanks

  • Can we extend our Illustrator CS6 trial?

    We downloaded a 30-day trial of Adobe Illustrator CS6 and then our computer needed repair, so 3 of the 4 weeks of the trial, we were without the computer!  Can we extend our trial?