Best Practice Process Chains

Hi All,
      What is the Best Practice Process chains recommended from SAP mainly for FICO related stuff.
       Do anyone have the structure of FI staging ,reporting layer Process chains step by step.
Thanks in Advance
Regards
con
Edited by: con on May 21, 2009 10:01 PM

Hello con ,
two thing to follow..
1)have a Change run Step Once master data load completes for all the infoobjects which you have loaded.
The Change Run activates the master data and adapts all aggregates for newly loaded master data and hierarchies.
2)If you want to replace existing data in the DataTarget completely, first delete the data in DataTarget(also in PSA if present )and load afterwards. This will help to improve on timing parameters.
Hope this helps you!!!
- Nandita

Similar Messages

  • Need best practices PROCESS to schedule RMAN Backups.

    Hi All,
    I would like a suggestion on the following for RMAN Backups:
    Details:
    Database: - 11gR2, Size 3TB on ASM - DW database.
    Like suggestions on:
    1) What kind of backups to schedule - Incremental along with Block backups?
    2) Size required to allocate for the Backup Space - Can it be ASM or Disk Space?
    3) Anything else - please suggest.
    Thank you.

    For that size, you might do weekly L0 and daily L1 (differential) backups.  Try to do L1 (cumulative) backups if possible.
    Getting the total sizes of Segments (from DBA_SEGMENTS) will give you an approximation for the minimum size.  (Tables/Indexes that have been dropped and are no longer in DBA_SEGMENTS wouldn't appear but the underlying blocks for them if unused by other Tables/Indexes would still be backed up as they'd have been formatted).
    You could run the Backup as a COMPRESSED BACKUPSET to reduce the size.
    Whether you want ASM or FileSystem depends on your comfort level.  If you run backups to ASM, you can only use RMAN to backup from ASM to Tape.  However, backups on FileSystem can be copied to tape using any other method (tape backup utility).  Also consider if you want to replicate/copy the backups to other servers and what methods you'd have available.
    Hemant K Chitale

  • Best Practice for Expired updates cleanup in SCCM 2012 SP1 R2

    Hello,
    I am looking for assistance in finding a best practice method for dealing with expired updates in SCCM SP1 R2. I have read a blog post: http://blogs.technet.com/b/configmgrteam/archive/2012/04/12/software-update-content-cleanup-in-system-center-2012-configuration-manager.aspx
    I have been led to believe there may be a better method, or a more up to date best practice process in dealing with expired updates.
    On one side I was hoping to keep a software update group intact, to have a history of what was deployed, but also wanting to keep things clean and avoid issues down the road as i used to in 2007 with expired updates.
    Any assistance would be greatly appreciated!
    Thanks,
    Sean

    The best idea is still to remove expired updates from software update groups. The process describes in that post is still how it works. That also means that if you don't remove the expired updates from your software update groups the expired updates will
    still show...
    To automatically remove the expired updates from a software update group, have a look at this script:
    http://www.scconfigmgr.com/2014/11/18/remove-expired-and-superseded-updates-from-a-software-update-group-with-powershell/
    My Blog: http://www.petervanderwoude.nl/
    Follow me on twitter: pvanderwoude

  • Best practices with LDIF Development for RBAC?

    I'm currently working on enforcing RBAC (Role Based Access controls) in OID that may be subject to change every few months. What I've currently been doing is writing LDIF files to make changes to the existing RBAC once the changes have been finalized.
    Unfortunately, now we have ended up with a growing list of LDIF files that must be run in sequential order if we were to build a new environment. Any defects or development errors that slip through developer unit testing must be handled in the same manner.
    What is the best practice process for performing this type of development? Would it make more sense to have one LDIF file that removes all of the RBAC enforcement (via ldapmodify -c), and then a separate file that will install the latest and most up to date version? I've also considered just using one LDIF file, appending any updates to the end of it and using the ldapmodify command with the -c parameter

    With regard to the 29.97/30 thing, you'll find that video people are idiosyncratically imprecise about that. We say 60 when we mean 59.94, we say 30 when we mean 29.97 and we say 24 when we mean 23.976.
    We're quirky.
    Whenever somebody says one of those nice, round numbers, you can assume they're really talking about the corresponding ugly fraction.
    Unless they're film people, in which case +24 means 24, dangit.+

  • APO - Aggregate Forecasting and Best Practices

    I am interested in options to save time and improve accuracy of forecasts in APO Demand Planning.   Can anyone help me by recommending best practices, processes and design that they have seen work well?
    We currently forecast at the Product level (detailed).  We are considering changing that to the Product Family level.   If you have done this, please reply.

    Hello Dan -
    Doing it at the Product level is very detailed (but it depends on the number of SKU's Available and that are to be forecasted).
    Here for me on my project we have about sample size 5000 finished goods to start with and forecasting at that minute level wont help. For me i have classified a product group level where i have linked all the similar FGs. tht way when you are working with the product group you are able to be a little from assertive in your forecasting. After that you can use proportional factors that will help you in allocating the necessary forecast down to product level. This way you are happy as well as the management is happy (high level report) as they dont have to go thru all the data that is not even necessary for them to see.
    Hope this helps.
    Regards,
    Suresh Garg

  • Best Practices for Loading Master Data via a Process Chain

    Currently, we load attributes, text, and hierarchies before loading the transactional data.  We have one meta chain.  To load the master data it is taking more than 2 hours.  Most of the master data is full loads.  We've noticed that a lot of the master data, especially text, has not changed or changed very little since we implemented 18 months ago.  Is there a precedence or best practice to follow such as do we remove these processes from the chain?  If so, how often should it be run?  We would really like to reduce the amount of the master data loading time.  Is there any documentation that I can refer to?  What are other organizations doing to reduce the amount of time to load master data?
    Thanks!
    Debby

    Hi Debby,
    I assume you're loading Master Data from a BI system? The forum here are related to SAP NetWeaver MDM, so maybe you should ask this question in a BI forum?
    Nevertheless, if your data isn't changed this much, maybe you could use a delta mechanism for extraction? This would send only the changed records and not all the unchanged all the time. But this depends on your master data and of course on your extractors.
    Cheers
    Michael

  • Import of Process Chains - What is the best practice

    Hi Colleagues,
    Wish to check on what is considered as a best practice:
    1. Move Process chains to the Production Environment through Transports
    Else
    2. Create them in the production system directly (obviously after having tested the scenario in other envrionments)
    Also would be glad to know if you have some links to posts which have discussions on the errors encountered in transports of process chains.
    Good Day and regards!
    Uma Shankar

    Hi Srinivas,
    Thanks a lot for the reply and the time taken to update this post.
    I will have to try this suggestion and shall update this post. Am now away from the project site, however since i was trying to help a colleague, will forward this reply to him to try it out and seek his observation.
    However a couple of things:
    1. This problem arose when the process chain was transported from Development to the Quality system.
    2. Though the error came up in the transport but when we were trying load the data into the Quality system using this process chain we were able to do so.
    So looks like the error had no impact on the performance of either the info-pack or the process chain. However we were sceptical on the our decision to import this process chain to the Production due to this error.
    However having seen hyour reply i will suggest my colleague to try it out (assuming that this error is still a pending issue) and shall update this post for the benefit of all.
    Good Day and regards!
    Uma Shankar

  • Best practices to modify process chain

    Hi,
    What is the best practice to modify the  process chain whether in in the production or transport back to the dev.
    Does Query performing tuning setting should be done in production like read modes,cache settings.
    Thanks
    nikhil

    Hi Nikhil,
    The best practice to modify the process chains will be making the change in Development system and transporting to Quality and Production. But if you are making a simple change like changing the scheduling time ( change of scheduled date and time by editing the variant) you can do it directly in Production. But if you are adding new steps to process chain, it is better to make change in Dev and move it to Prod.
    Also once the change reach production you may need to open the chain in edit mode and activate it manually.
    It is a common issue for the process chains containing delta DTP, that the chain will be saved in modified (M)  version and not in active version by transports.
    The query read mode and cache settings, you can do it in Development and move it to production.
    But the pre-filling of cache using broadcaster settings can be done directly in production.
    Also creating of cube aggregates to improve query can be done directly in production rather than creating in development and transporting to Prod.
    Again, it depends on project, as in some projects they make changes directly in Prod, while in some projects changes in Prod are not allowed. I think it is better always start the changes in dev and move it to P.
    Hope it helps,
    Thanks,
    Vinod-

  • User wants to schedule the process chain - what is the best practice

    I posted earler and got the anwser.
    Create a custom ABAP program that calls the RSPC_API_CHAIN_START Function Module, with a selection for the Process Chain where your user can enter the Meta Chain name to start the process.
    To make it simpler for the user to execute this, create a custom tcode for them in SE93, so that they logon to your BW environment, call up that tcode, enter the Meta Chain name and execute. With this, they wouldn't have to have access to RSPC.
    I did the above and the user not able to execute it since he doesn't have access to the cube/dso to update the cube.
    To overcome this I wrote another abap program and used the
    SUBMIt program user ALEREMOTE
    and there is a problem.I called the process chain start program using ALEREMOTE .
    Now the user  needs access to authorization object S_BTCH_NAM and authorization field BTCUNAME with value ALEREMOTE.
    So again due to SOX the basis doesn't want to give the necessary access.
    What is the best solution.
    How in general if the user needs to execute the process chain how they do it.
    Please let me know. Thanks for your help in advance.
    Senthil

    Hi Senthil
    In general production scenerio, we do not run process chain manually. To execute a process chain you should have necessary authorization. As you have assigned alertmote in the start variant, any user other thatn that will not be able to trigger the same.
    Instead of alertmote , use the nornal start variant to trigger the process chain and give the user the normal authorization. Whenever user wants to run the infopackage manually or via process chain they should have the auth to view the infoprovider content as well....

  • Process chain best practice - attribute change runs

    When building attribute change runs into Process Chains I see that there are potentially two main options to consider:
    1. One to One mapping.  Incorporate a separate attribute change run object type after each InfoObject has been loaded.
    2. Many to One mapping.  Build the process chain so that one single attribute change run occurs after all the InfoObjects have been loaded. 
    I would be interested to see how other people have incorporated their attribute loading into Process Chains.
    any info appreciated.
    Mike

    We run the attribute change run twice in our daily process chain, here's our order:
    - Load all master data
    - Attrib change run
    - Load all tx data
    - Run ABAP programs that generate master data
    - Attrib change run
    I doubt there will be much of a performance difference if you do the change run after each master data load, but it makes the process chain a lot easier to read.

  • Deadline Branche in Correlation Process - Best Practice

    Hello,
    I have an integration process with a correlation - there is a asynchronous send step which activates a correlation and afterwards an asynchronous receive step that uses that correlation.
    Furthermore I have a deadline branch to cancel the process after 24 hours.
    My question now is:
    There could be (rare) cases where a message arrives later than 24 hours, so according to my understanding the received message will block the inbound queue as no active correlation can be found anymore. Is this correct? How can I avoid this situation, I guess a blocked queue would also block other messages that are sent to the integration process?
    What would be best practice to handle such a scenario? I could leave the process intance open for 1 month, however this might have a significant impact on system performance.....
    Thank you for your advice.

    There could be (rare) cases where a message arrives later than 24 hours, so according to my understanding the received
    essage will block the inbound queue as no active correlation can be found anymore
    No correlation found error will occur only when the BPM instance is running and the message tries to enter into the relevant receive step (not the first one)
    However when you say the process is cancelled you need not worry about the message going into the queue and blocking the BPM queue.
    Regards,
    Abhishek.

  • Best practice for handling errors in EOIO processing on AEX?

    Hi,
    I'm looing for resources and information describing the best practice on how to handle processing errors of asynchronous integrations with Exactly-Once-In-Order QoS on an AEX (7.31) Java only installation. Most information I've found so far are describing a the monitoring and restart jobs on AS ABAP.
    Situation to solve:
    multiple different SOAP messages are integrated using one queue with an RFC receiver. On error the message status is set to Holding and all following messages are Waiting. Currently I need to manually watch over the processing, delete the message with status holding and restart the waiting ones.
    Seems like I can setup component based message alerting to trigger an email or whatever alert. I still need to decide on how to handle the error and resolve it (ie. delete the errornous message, correct the data at sender and trigger another update). I still need to manually find the oldest entry with status waiting and restart it. I've found a restart job in Background jobs in configuration and monitoring home, but it can be only scheduled in intervals of 1 or more hours.
    Is there something better?
    Thank you.
    Best regards,
    Nikolaus

    Hi Nikolaus -
    AFAIK - For EOIO, you have to cancel the failed message and then process the next message in the sequence manually..
    Restart job only works the messages which are in error state but not in holding state.. So you have to manually push the message... So there is no other alternative.
    But it should not be that difficult to identify the messages in a sequence..
    How to deal with stuck EOIO messages in the XI ... | SCN
    Though it is for older version, it should be the same.. you should be able to select the additional columns such as sequence ID from the settings..

  • What is the best practice to display info of completed task in process flow

    Hi all,
    I'm starting to study BPM modeling with CE7.1 EHP1. Thanks to the tutorial and example on SDN site and I can easily build my own process in NWDS and deploy to server, start it, finish it.
    I like the new runtime which can show a BPMN diagram to the processors. However, I can't find a way to let the follow up processor to review the task result completed in previous step. I'm more familiar with Guided Procedure, and know there is "Display Callable Object" which can used to show some info of a completed task when the processor/owner/admin/overseer click on a completed task.  Where is the feature in BPM ? What is the best practice to show such task information in BPM environment.
    For example, A multiple level approval process, the higher level approver need to know the comment written by the previous approver. Can he read this information from process flow ?
    I think it is very important feature for a BPM platform. In Guided Procedure, such requirement can be done with Display Callable Object + View Permission, and you just need some coding for the UI. If BPM is superior to GP, I think there must be a way to achieve this, I just do not know how ?
    Can anyone shed me some light on it ?

    Oliver,
    Thanks for your quick reply.
    Yes, Notes and Attachment CAN BE USED for the purpose. But I'm still looking for a more elegant solution.
    With the solution of using Notes/Attachment, the processor need to give input at two places : the task UI and Note/Attach , with similar or same data. It is really annoying.
    Is there any SAP BPM real-world deployment ? None of customer has the requirement ?

  • Best practice to process inbound soapenc:Array

    Hi, I'm working my way through a real world example where I've sent a keyword search query to Amazon and it return 10 results. This works great. Now I want to loop thru the Array and put the returned values into a table in the database. I can't seen to get down to the element level via XQuery to assign the returned values to my local variables. The XML returned is defined in the wdsl as:
         <xsd:complexType name="DetailsArray">
         <xsd:complexContent>
         <xsd:restriction base="soapenc:Array">
    <xsd:attribute ref="soapenc:arrayType" wsdl:arrayType="typens:Details[]"/>
    </xsd:restriction>
    </xsd:complexContent>
    </xsd:complexType>
         <xsd:complexType name="Details">
    What's the best practice for walking thru this incoming XML Array?
    Thanks...Matt

    Hello again!
    Thanks for quick reply!
    The instance state says : closed.completed.
    Perhaps the problem is not the insert into db, but the transformation activity?
    This is what the audit trail (raw xml) looks like:
    <?xml version="1.0" encoding="UTF-8" ?>
    - <audit-trail>
    - <event sid="0" cat="2" type="2" n="0" date="2008-06-24T10:16:21.875+02:00">
    - <message>
    - <![CDATA[ New instance of BPEL process "BPELProcess3" initiated (# "150001").
      ]]>
    </message>
    </event>
    - <event sid="BpPrc0.1" cat="1" type="2" label="process" n="1" date="2008-06-24T10:16:21.875+02:00" psid="0">
    - <message>
    - <![CDATA[ _cr_
      ]]>
    </message>
    </event>
    - <event sid="BpTry0.2" cat="1" type="2" n="2" date="2008-06-24T10:16:21.890+02:00" psid="BpPrc0.1">
    - <message>
    - <![CDATA[ _cr_
      ]]>
    </message>
    </event>
    - <event sid="BpSeq0.3" cat="1" type="2" label="sequence" n="3" date="2008-06-24T10:16:21.890+02:00" psid="BpTry0.2">
    - <message>
    - <![CDATA[ _cr_
      ]]>
    </message>
    </event>
    - <event sid="BpSeq0.3" cat="2" type="2" wikey="150001-BpRcv0-BpSeq0.3-1" n="4" date="2008-06-24T10:16:21.921+02:00">
    - <message>
    - <![CDATA[ Received "inputVariable" call from partner "client"
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <inputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="payload"><ns1:BPELProcess3ProcessRequest xmlns:ns1="http://xmlns.oracle.com/BPELProcess3">
                <ns1:stations>18700</ns1:stations>
                <ns1:username/>
            </ns1:BPELProcess3ProcessRequest>
    </part></inputVariable>
      ]]>
    </details>
    </event>
    - <event to="GetStationProperties_getStationsProperties_InputVariable" sid="BpSeq0.3" cat="2" type="1" wikey="150001-BpAss0-BpSeq0.3-2" n="5" date="2008-06-24T10:16:21.953+02:00">
    - <message>
    - <![CDATA[ Updated variable "GetStationProperties_getStationsProperties_InputVariable"
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <GetStationProperties_getStationsProperties_InputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="stations"><stations>18700</stations>
    </part><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="username"><username xmlns="" xmlns:def="http://www.w3.org/2001/XMLSchema" xsi:type="def:string" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"/>
    </part></GetStationProperties_getStationsProperties_InputVariable>
      ]]>
    </details>
    </event>
    - <event to="GetStationProperties_getStationsProperties_InputVariable" sid="BpSeq0.3" cat="2" type="1" wikey="150001-BpAss0-BpSeq0.3-2" n="6" date="2008-06-24T10:16:21.953+02:00">
    - <message>
    - <![CDATA[ Updated variable "GetStationProperties_getStationsProperties_InputVariable"
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <GetStationProperties_getStationsProperties_InputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="stations"><stations>18700</stations>
    </part><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="username"><username/>
    </part></GetStationProperties_getStationsProperties_InputVariable>
      ]]>
    </details>
    </event>
    - <event sid="BpSeq0.3" cat="2" type="2" wikey="150001-BpInv0-BpSeq0.3-3" partnerWSDL="MetDataService2Ref.wsdl" n="7" date="2008-06-24T10:16:22.968+02:00">
    - <message>
    - <![CDATA[ Invoked 2-way operation "getStationsProperties" on partner "MetDataService".
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <messages><GetStationProperties_getStationsProperties_InputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="stations"><stations>18700</stations>
    </part><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="username"><username/>
    </part></GetStationProperties_getStationsProperties_InputVariable><GetStationProperties_getStationsProperties_OutputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="return"><return xmlns:ns2="http://schemas.xmlsoap.org/soap/encoding/" xsi:type="ns2:Array" xmlns:ns3="http://no.met.metdata/IMetDataService.xsd" ns2:arrayType="ns3:no_met_metdata_StationProperties[1]" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
    <item xsi:type="ns3:no_met_metdata_StationProperties">
    <toDay xsi:type="xsd:int">0</toDay>
    <toMonth xsi:type="xsd:int">0</toMonth>
    <fromYear xsi:type="xsd:int">1937</fromYear>
    <municipalityNo xsi:type="xsd:int">301</municipalityNo>
    <amsl xsi:type="xsd:int">94</amsl>
    <latDec xsi:type="xsd:double">59.9427</latDec>
    <lonDec xsi:type="xsd:double">10.7207</lonDec>
    <toYear xsi:type="xsd:int">0</toYear>
    <department xsi:type="xsd:string">OSLO</department>
    <fromMonth xsi:type="xsd:int">2</fromMonth>
    <stnr xsi:type="xsd:int">18700</stnr>
    <wmoNo xsi:type="xsd:int">492</wmoNo>
    <latLonFmt xsi:type="xsd:string">decimal_degrees</latLonFmt>
    <name xsi:type="xsd:string">OSLO - BLINDERN</name>
    <fromDay xsi:type="xsd:int">25</fromDay>
    </item>
    </return>
    </part></GetStationProperties_getStationsProperties_OutputVariable></messages>
    ]]>
    </details>
    </event>
    - <event to="StoreStationProperties_insert_InputVariable" sid="BpSeq0.3" cat="2" type="1" wikey="150001-BpAss1-BpSeq0.3-4" n="8" date="2008-06-24T10:16:23.000+02:00">
    - <message>
    - <![CDATA[ Updated variable "StoreStationProperties_insert_InputVariable"
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <StoreStationProperties_insert_InputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="EklimaStationsTblCollection"><EklimaStationsTblCollection xmlns:ns0="http://xmlns.oracle.com/pcbpel/adapter/db/top/db" xmlns="http://xmlns.oracle.com/pcbpel/adapter/db/top/db"/>
    </part></StoreStationProperties_insert_InputVariable>
      ]]>
    </details>
    </event>
    - <event sid="BpSeq0.3" cat="2" type="2" wikey="150001-BpInv1-BpSeq0.3-5" partnerWSDL="db.wsdl" n="9" date="2008-06-24T10:16:27.312+02:00">
    - <message>
    - <![CDATA[ Invoked 1-way operation "insert" on partner "db".
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <StoreStationProperties_insert_InputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="EklimaStationsTblCollection"><EklimaStationsTblCollection xmlns:ns0="http://xmlns.oracle.com/pcbpel/adapter/db/top/db" xmlns="http://xmlns.oracle.com/pcbpel/adapter/db/top/db"/>
    </part></StoreStationProperties_insert_InputVariable>
      ]]>
    </details>
    </event>
    - <event sid="BpSeq0.3" cat="2" type="2" wikey="150001-BpRpl0-BpSeq0.3-6" n="10" date="2008-06-24T10:16:27.312+02:00">
    - <message>
    - <![CDATA[ Reply to partner "client".
      ]]>
    </message>
    - <details>
    - <![CDATA[
    <outputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="payload"><BPELProcess3ProcessResponse xmlns="http://xmlns.oracle.com/BPELProcess3"/>
    </part></outputVariable>
      ]]>
    </details>
    </event>
    - <event sid="BpSeq0.3" cat="1" type="2" n="11" date="2008-06-24T10:16:27.312+02:00">
    - <message>
    - <![CDATA[ _cl_
      ]]>
    </message>
    </event>
    - <event sid="BpPrc0.1" cat="1" type="2" n="12" date="2008-06-24T10:16:27.312+02:00">
    - <message>
    - <![CDATA[ _cl_
      ]]>
    </message>
    </event>
    - <event sid="BpPrc0.1" cat="2" type="2" n="13" date="2008-06-24T10:16:27.328+02:00">
    - <message>
    - <![CDATA[ BPEL process instance "150001" completed
      ]]>
    </message>
    </event>
    </audit-trail>
    The webservice-method I'm calling in this case is getStationsProperties.
    Kind regards
    Jørn Eirik

  • Best Practice for sugar refinery process

    hello, my company need to deploy a new business concerning raw sugar refinery.
    so we need to analyze the business requirements and propose a process for refinery management  .
    step 1: arrival of goods in docks
    step 2: raw sugar need to be charged in our stock ( quantity and value ) but is not  our property
    step 3: goods need to be delivered to our plant ( we pay the transport as service for our business partner )
    step 4: goods need to be verified ( for quality and quantity ) and accepted by operators
    step 5: goods are processed in a refinery plant, we need to verify timing, costs, quantity and human resources employed ( for costs remittance and transfer  )
    step 6: sugar is delivered to other industrial plants, warehouse and finally sold ( but is not our property ), for us it's like a refinery service.
    step 7: we need to trace production lot from raw sugar arrival to the docks up to step 6 .
    step 8: inventory and  maintenance costs need to be traced because our profit is a part of this refinery service reduced by costs incurred
    any suggestions to find the right best practice ?
    I'm not a skilled BPS, I was looking for oil refinery but  is not the same process, so.. what can i look for?
    TNks

    Hi Kumar,
    In order to have Consigment in SAP u need to have master data such as material master, vendor master and purchase inforecord of consignment type. U have to enter the item category k when u enter PO. The goods receipt post in vendor consignment stock will be non valuated.
    1. The intial steps starts with raising purchase order for the consignment item
    2. The vendor recieves the purchase order.
    3. GR happens for the consignment material.
    4. Stocks are recieved and placed under consignment stock.
    5. When ever we issue to prodn or if we transfer post(using mov 411) from consignment to own stock then liability occurs.
    6. Finally comes the settlement using mrko. You settle the amount for the goods which was consumed during a specific period.
    regards
    Anand.C

Maybe you are looking for