Best Practice for sugar refinery process

hello, my company need to deploy a new business concerning raw sugar refinery.
so we need to analyze the business requirements and propose a process for refinery management  .
step 1: arrival of goods in docks
step 2: raw sugar need to be charged in our stock ( quantity and value ) but is not  our property
step 3: goods need to be delivered to our plant ( we pay the transport as service for our business partner )
step 4: goods need to be verified ( for quality and quantity ) and accepted by operators
step 5: goods are processed in a refinery plant, we need to verify timing, costs, quantity and human resources employed ( for costs remittance and transfer  )
step 6: sugar is delivered to other industrial plants, warehouse and finally sold ( but is not our property ), for us it's like a refinery service.
step 7: we need to trace production lot from raw sugar arrival to the docks up to step 6 .
step 8: inventory and  maintenance costs need to be traced because our profit is a part of this refinery service reduced by costs incurred
any suggestions to find the right best practice ?
I'm not a skilled BPS, I was looking for oil refinery but  is not the same process, so.. what can i look for?
TNks

Hi Kumar,
In order to have Consigment in SAP u need to have master data such as material master, vendor master and purchase inforecord of consignment type. U have to enter the item category k when u enter PO. The goods receipt post in vendor consignment stock will be non valuated.
1. The intial steps starts with raising purchase order for the consignment item
2. The vendor recieves the purchase order.
3. GR happens for the consignment material.
4. Stocks are recieved and placed under consignment stock.
5. When ever we issue to prodn or if we transfer post(using mov 411) from consignment to own stock then liability occurs.
6. Finally comes the settlement using mrko. You settle the amount for the goods which was consumed during a specific period.
regards
Anand.C

Similar Messages

  • Best practice for increasing image processing speed

    Are there best practises for efficient image processing so that the overall perfomance speed will be improved? I have a need to do near real time image processing (threshold, filtering, particle analysis/cleaning and measurements) at 10fps. So far I am not satisfied with my cycle time so I am wondering if there are documented ways to speed up performance.
    Solved!
    Go to Solution.

    Hi,
    Navigates to the "Vision and Motion" palette, then "Vision Utilities" and "Image Management" palette, you will find ther the "IMAQ multi-core options" vi.
    You will be able to set or get the number of cores used by the IMAQ functions.
    Regards

  • Best practice for identifying BPEL process boundary

    Hi
    Want to know if there is any best practices / guidelines available to identify how many BPEL processes are required.. I assume once we do a process decomposition we might arrive at an abstract business process boundary.
    However, I also think those processes can be further broken down while actually implementing them using BPEL PM?
    Is there any guidelines that I can use or follow to specify how many BPEL process to define
    Thanks
    Prashanth Rao

    Hi Kumar,
    In order to have Consigment in SAP u need to have master data such as material master, vendor master and purchase inforecord of consignment type. U have to enter the item category k when u enter PO. The goods receipt post in vendor consignment stock will be non valuated.
    1. The intial steps starts with raising purchase order for the consignment item
    2. The vendor recieves the purchase order.
    3. GR happens for the consignment material.
    4. Stocks are recieved and placed under consignment stock.
    5. When ever we issue to prodn or if we transfer post(using mov 411) from consignment to own stock then liability occurs.
    6. Finally comes the settlement using mrko. You settle the amount for the goods which was consumed during a specific period.
    regards
    Anand.C

  • Best practice for credit card processing??

    Hello,
       Our application processes credit card transactions. For whatever reason recently double charges have been happening despite our clients stating they are not submitting the data twice. It kinda doesn't really matter what the cause is as we need a solid server side solution to never let double transactions occur. I have thought of a few methods but I'd very much appreciate someone else's opinion/experience on this. Whatcha think! Many thanks.
    We are using coldfusion 8 with SQL Server 2008.
    Patrick

    My suggestion is to 1) pick a reliable gateway and 2) pick a gateway that allows for pre-settlement auditing.
    Double posting is a fact of life when two applications interface to each other. You want a reliable gateway to minimize the occurrence. Because double postings can occasionally occur, I always recommend pre-settlement auditing to make sure application A balances with application B.
    A gotcha in the payments industry is when you double post, you get dinged a discount rate and transaction fees -- something like 2.00% + $0.35 as an example. If the double post makes it through interchange, and you need to refund one of the transactions back to the customer, you don't get the original fees back and you will most likely be dinged additional transaction fees (the $0.35 in sample rate). Using my rate example, if your average ticket is $10, your average processing fee is $0.55 ($10 x 2% + $0.35) so a refund will cost you the $0.90 (the original fee that does not get refunded plus another $0.35 for the refund). Now if your average tick is $100, now your approaching $3 for each double posting. If you pre-audit and stop the double posts before interchange, it'll only cost you the single $0.35 fixed fee for the double auth.
    Now some API's are better than others and some can stop double postings all together. I still recommend the pre-settlement auditing if notithing else, for volume trending and fraud prevention. Again, a reliable gateway should have a solid API that can prevent double-postings.

  • Best Practice for Automotive

    Dear All:
    Who has the package of best practice for Automotive? Specially for product traceability and external procurement! Please share with me, thanks in advance.
    Good day
    Tao

    Hi,
    I only find the best practice for automotive external process, you can refer to the following address:
    http://help.sap.com/content/bestpractices/industry/index.htm
    But not find it for product traceability, hope it's useful to you!

  • Suggest the Best Practice for Procurement of Commodities like crude, copper

    Dear Gurus,
    Please suggest the best practice for following business Process.
    My client is having the procurement need for a comodity whose prices are fluctuating. Say Crude Oil . The prices are changing every day. Now The client would like to pay to vendor on day of good receipt. But it may be possible that the price on GR day is much higher. How to control these kind of procurement. Presently I have activated the price variance through invoice posting but this is not working.
    Can you suggest best practice for  procurement of comodities.
    Thank you for your consistant support.
    Regards
    Vinod Kakade
    Edited by: vinodkakade on Jul 14, 2011 2:34 PM

    Hi Vinod,
    Would you know the price by the time the GR is to happen, in this case you can ask the vendor to send the confirmations, just prior to that with the correct price.
    Please refer this link
    Change in PO Price after goods receipts and goods issue
    though the thread is marked unanswered.
    Regards
    Shailesh

  • Can anyone recommend tips and best practices for FrameMaker-to-RoboHelp migration ?

    Hi. I'm planning a migration from FM (unstructured) to RH. I'd appreciate any tips and best practices for the migration process. (Note that at the moment I plan to import the FM documents into, not link to them from, RH.)
    For example, my current FM files are presently not optimally "chunked", so that autoconverting FM file sections (based on, say, Header 1 paragraph layout) won't always result in an optimal topic set. I'm thinking of going through the FM docs and inserting dummy paragraphs with a tag somethike like "topic_break", placed in more appropriate locations that the existing headers. Then, during import to RH, I'd use the topic_break paragraph to demark the topics. Is this a good technique? Beyond paragraph-based import delineation, do you know of any guidelines for redrafting FM chapter file content into RH topics?
    Also, are there any considerations/gotchas in the areas of text review workflow, multiple authoring, etc. after the migration? (I've not managed an ongoing RH doc project before, so any advice would be greatly appreciated.
    Thanks in advance!
    -Kurt
    BTW, the main reason for the migration: Info is presently scattered in various (and way to many) PDF files. There's no global index. I'd like to make a RoboHelp HTML interface (probably WebHelp layout) so it can be a one-stop documentation shop for users.

    Jeff
    Fm may produce better output for your requirements but for many what Rh produces works just fine. My recent finding re Word converting images to JPG before import will mean a better experience for many.
    Once Rh is set up, and it's not difficult, for many its printed documents will do the job. I would say try it and then judge.
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • SAP Best Practice for Chemicals in an oil refinery

    Does anyone have experience in implementing the SAP Best Practice for Chemicals in an oil refinery that will also use IS-OIL?
    What would be the pros and cons?
    What is the implementation price of considering a best practice solution for a large complex organization?
    Oded Dagan
    Oil Refineries LTD.
    Israel

    Hi Oded,
    I don't know about any Best Practice Chemicals implementation in an refinery so far.
    But you can use Best Practice Chemicals with SAP IS-Oil for the non IS-Oil functionality as well.
    Best Practice Chemicals gives you benefits within standard business processes, but for the
    IS-Oil business processes you have to consider the traditon implementation methods.
    Best Practice Chemicals gives you a broad business process set, usually used in Chemical
    corporation. If you can cover 50% of your needed business processes out of Best Practice Chemicals you save approx 50% implementation time. It is not only implemenentation you save a lot in documentation and training material as well. Most of our Best Practice Chemicals implemenations used
    60-80% out of Best Practice Chemicals. At a large corporation the percentage of standard ERP processes is normally smaller, because of other additional needed SAP solutions e.g. APO, SRM, CRM etc.
    regards Rainer

  • Best Practice for the Vendor Consignment Process

    Hiii ,,,
    Can anybody have the best practices for the Vendor consignment process???
    Please Send me the document.
    Please explain me the Consignment process in SAP??
    Thanx & Regards,
    Kumar Rayudu

    Hi Kumar,
    In order to have Consigment in SAP u need to have master data such as material master, vendor master and purchase inforecord of consignment type. U have to enter the item category k when u enter PO. The goods receipt post in vendor consignment stock will be non valuated.
    1. The intial steps starts with raising purchase order for the consignment item
    2. The vendor recieves the purchase order.
    3. GR happens for the consignment material.
    4. Stocks are recieved and placed under consignment stock.
    5. When ever we issue to prodn or if we transfer post(using mov 411) from consignment to own stock then liability occurs.
    6. Finally comes the settlement using mrko. You settle the amount for the goods which was consumed during a specific period.
    regards
    Anand.C

  • Best practice for optimizing processing of big XMLs?

    All,
    What is the best practice when dealing with large XML files ... (say couple of MBs).
    Instead of having to read the file from the file system everytime a static method is run, what would be the best way in which the program reads the file once and then keeps it in memory. So the next time it would not have to read and parse it all over again?
    Currently my code just read the file in the static method like ...
    public static String doOperation(String path,...) throws Exception
    try{           
    String masterFile = path+"configfile.xml";
    Document theFile = (Document)getDocument(masterFile);
    Element root = theFile.getDocumentElement();
    NodeList nl = root.getChildNodes();
    // ... operations on file
    Optimization tips and tricks most appreciated :-)
    Thanks,
    David

    The best practice for multi-megabyte XML files is not to have them at all.
    However if you must, presumably you don't need all of the information in your XML, repeatedly. Or do you? If you need a little bit of it here, then another little bit of it there, then yet another little bit of it later, then you shouldn't have stored your data in one big XML.
    Sorry if that sounds unhelpful, but I'm having trouble imagining a scenario when you need all the data in an XML document repeatedly. Perhaps you could expand on your design?
    PC²

  • Best-practice for Catalog Views ? :|

    Hello community,
    A best practice question:
    The situtation: I have several product categories (110), several items in those categories (4000) and 300 end-users.    I would like to know which is the best practice for segment the catalog.   I mean, some users should only see categories 10,20 & 30.  Other users only category 80, etc.    The problem is how can I implement this ?
    My first idea is:
    1. Create 110 Procurement Catalogs (1 for every prod.category).   Each catalog should contain only its product category.
    2. Assign in my Org Model, in a user-level all the "catalogs" that the user should access.
    Do you have any idea in order to improve this ?
    Saludos desde Mexico,
    Diego

    Hi,
    Your way of doing will work, but you'll get maintenance issues (to many catalogs, and catalog link to maintain for each user).
    The other way is to built your views in CCM, and assign these views to the users, either on the roles (PFCG) or on the user (SU01). The problem is that with CCM 1.0 this is limitated, cause you'll have to assign one by one the items to each view (no dynamic or mass processes), it has been enhanced in CCM 2.0.
    My advice:
    -Challenge your customer about views, and try to limit the number of views, with for example strategic and non strategic
    -With CCM 1.0 stick to the procurement catalogs, or implement BADIs to assign items to the views (I experienced it, it works, but is quite difficult), but with a limitated number of views
    Good luck.
    Vadim

  • Best Practice for Securing Web Services in the BPEL Workflow

    What is the best practice for securing web services which are part of a larger service (a business process) and are defined through BPEL?
    They are all deployed on the same oracle application server.
    Defining agent for each?
    Gateway for all?
    BPEL security extension?
    The top level service that is defined as business process is secure itself through OWSM and username and passwords, but what is the best practice for security establishment for each low level services?
    Regards
    Farbod

    It doesnt matter whether the service is invoked as part of your larger process or not, if it is performing any business critical operation then it should be secured.
    The idea of SOA / designing services is to have the services available so that it can be orchestrated as part of any other business process.
    Today you may have secured your parent services and tomorrow you could come up with a new service which may use one of the existing lower level services.
    If all the services are in one Application server you can make the configuration/development environment lot easier by securing them using the Gateway.
    Typical probelm with any gateway architecture is that the service is available without any security enforcement when accessed directly.
    You can enforce rules at your network layer to allow access to the App server only from Gateway.
    When you have the liberty to use OWSM or any other WS-Security products, i would stay away from any extensions. Two things to consider
    The next BPEL developer in your project may not be aware of Security extensions
    Centralizing Security enforcement will make your development and security operations as loosely coupled and addresses scalability.
    Thanks
    Ram

  • Best practices for ARM - please help!!!

    Hi all,
    Can you please help with any pointers / links to documents describing best practices for "who should be creating" the GRC request in below workflow of ARM in GRC 10.0??
    Create GRC request -> role approver -> risk manager -> security team
    options are : end user / Manager / Functional super users / security team.
    End user and manager not possible- we can not train so many people. Functional team is refusing since its a lot of work. Please help me with pointers to any best practices documents.
    Thanks!!!!

    In this case, I recommend proposing that the department managers create GRC Access Requests.  In order for the managers to comprehend the new process, you should create a separate "Role Catalog" that describes what abilities each role enables.  This Role Catalog needs to be taught to the department Managers, and they need to fully understand what tcodes and abilities are inside of each role.  From your workflow design, it looks like Role Owners should be brought into these workshops.
    You might consider a Role Catalog that the manager could filter on and make selections from.  For example, an AP manager could select "Accounts Payable" roles, and then choose from a smaller list of AP-related roles.  You could map business functions or tasks to specific technical roles.  The design flaw here, of course, is the way your technical roles have been designed.
    The point being, GRC AC 10 is not business-user friendly, so using an intuitive "Role Catalog" really helps the managers understand which technical roles they should be selecting in GRC ARs.  They can use this catalog to spit out a list of technical role names that they can then search for within the GRC Access Request.
    At all costs, avoid having end-users create ARs.  They usually select the wrong access, and the process then becomes very long and drawn out because the role owners or security stages need to mix and match the access after the fact.  You should choose a Requestor who has the highest chance of requesting the correct access.  This is usually the user's Manager, but you need to propose this solution in a way that won't scare off the manager - at the end of the day, they do NOT want to take on more work.
    If you are using SAP HR, then you can attempt HR Triggers for New User Access Requests, which automatically fill out and submit the GRC AR upon a specific HR action (New Hire, or Termination).  I do not recommend going down this path, however.  It is very confusing, time consuming, and difficult to integrate properly.
    Good luck!
    -Ken

  • Workflow not completed, is this best practice for PR?

    Hi SAP Workflow experts,
    I am new in workflow and now responsible to support existing PR release workflow.
    The workflow is quite simple and straightforward but the issue here is the workflow seems like will never be completed.
    If the user released the PR, the next activity is Requisition released that using task TS20000162.
    This will send work item to user (pr creator) sap inbox which when they double click it will complete the workflow.
    The thing here is, in our organization, user does not access SAP inbox hence there are thousands of work item that has not been completed. (our procurement system started since 2009).
    Our PR creator will receive notification of the PR approval to theirs outlook mail handled by a program that is scheduled every 5 minutes.
    Since the documentation is not clear enough, i can't digest why the implementer used this approach.
    May I know whether this is the best practice for PR workflow or not?
    Now my idea is to modify the send email program to complete the workitem after the email being sent to user outlook mail.
    Not sure whether it is common or not though in workflow world.
    Any help is deeply appreciated.
    Thank you.

    Hello,
    "This will send work item to user (pr creator) sap inbox which when they double click it will complete the workflow."
    It sounds liek they are sending a workitem where an email would be enough. By completing the workitem they are simply acknowledging that they have received notification of the completion of the PR.
    "Our PR creator will receive notification of the PR approval to theirs outlook mail handled by a program that is scheduled every 5 minutes."
    I hope (and assume) that they only receive the email once.
    I would change the workflow to send an email (SendMail step) to the initiator instead of the workitem. That is normally what happens. Either that or there is no email at all - some businesses only send an email if something goes wrong. Of course, the business has to agree to this change.
    Having that final workitem adds nothing to the process. Replace it with an email.
    regards
    Rick Bakker
    hanabi technology

  • JSF - Best Practice For Using Managed Bean

    I want to discuss what is the best practice for managed bean usage, especially using session scope or request scope to build database driven pages
    ---- Session Bean ----
    - In the book Core Java Server Faces, the author mentioned that most of the cases session bean should be used, unless the processing is passed on to other handler. Since JSF can store the state on client side, i think storing everything in session is not a big memory concern. (can some expert confirm this is true?) Session objects are easy to manage and states can be shared across the pages. It can make programming easy.
    In the case of a page binded to a resultset, the bean usually helds a java.util.List object for the result, which is intialized in the constructor by query the database first. However, this approach has a problem: when user navigates to other page and comes back, the data is not refreshed. You can of course solve the problem by issuing query everytime in your getXXX method. But you need to be very careful that you don't bind this XXX property too many times. In the case of querying in getXXX, setXXX is also tricky as you don't have a member to set. You usually don't want to persist the resultset changes in the setXXX as the changes may not be final, in stead, you want to handle in the actionlistener (like a save(actionevent)).
    I would glad to see your thought on this.
    --- Request Bean ---
    request bean is initialized everytime a reuqest is made. It sometimes drove me nuts because JSF seems not to be every consistent in updating model values. Suppose you have a page showing parent-children a list of records from database, and you also allow user to change directly on the children. if I hbind the parent to a bean called #{Parent} and you bind the children to ADF table (value="#{Parent.children}" var="rowValue". If I set Parent as a request scope, the setChildren method is never called when I submit the form. Not sure if this is just for ADF or it is JSF problem. But if you change the bean to session scope, everything works fine.
    I believe JSF doesn't update the bindings for all component attributes. It only update the input component value binding. Some one please verify this is true.
    In many cases, i found request bean is very hard to work with if there are lots of updates. (I have lots of trouble with update the binding value for rendered attributes).
    However, request bean is working fine for read only pages and simple binded forms. It definitely frees up memory quicker than session bean.
    ----- any comments or opinions are welcome!!! ------

    I think it should be either Option 2 or Option 3.
    Option 2 would be necessary if the bean data depends on some request parameters.
    (Example: Getting customer bean for a particular customer id)
    Otherwise Option 3 seems the reasonable approach.
    But, I am also pondering on this issue. The above are just my initial thoughts.

Maybe you are looking for

  • *Connection problems Sap A Java Application*

    Hi Gurus, I have a problem. I'm riding the part of the JCO SERVER for SAP from sending information to a Java application. I created the RFC destination of the type TCP / IP and I have registered my JAVA application in the visual administrator. This a

  • Debug console

    I am having trouble with my iphone the internet 'safari' connection is extremely slow and I keep recieving error messages and messages to debug console. any suggestions

  • Photos not staying after rendering

    I had this problem before, after installing Leopard. Got the machine fixed and doubled the RAM. Now I have a video of the holidays to edit and it STILL won't keep photos. They render and the panel turns black, the video stops when it gets to the phot

  • How to determine if a management exception is a warning or error

    I'm using MBeans to deploy an application. The 'task' MBean indicates via a getError call that a weblogic.management.ManagementException was thrown. In this particular case the detail message says "..[Management:149315]No server ...". In the message

  • "Could not initialize Photoshop because the disk is nor available"

    I'm trying to install Photoshop CS4 for the very first time onto a Mac Powerbook G4, MacOS 10.4.11.  The installation succeeded, but when I launch Photoshop CS4, I get the "Could not open a scratch file because the disk is not available" error follow