How to configure LoadRunner 8.0 to test EP 7 ESS ?

Hi all:
    We have EP 7 with ESS/MSS will go live on July 1st.
    I would like to perfom some stress testing to simulate 1000 users to login in EP 7 aix .
    How to use Load Runner 8.0 to perform test ? Which protocal can we use ?

Hi,
Some hints:
- The best is to use Web/HTTP protocol with URL based recording
- Scripting important points:
  - Give special attention to all correlations
  - Add content check for all requests giving you relevant information! (all interaction with the Portal and/or iView) this is the only way to be sure you are getting the expected answer.
- For WebDynpro iView interaction, give a special attention to eventQueue request parameter, as this is were the interaction with the WebDynpro iView is describ. Be also careful about how this parameter is encoded. And add a content check for all request having such eventQueue parameter!
- Use SMD and Wily Introscope for J2EE and EP monitoring
Hope this help,
Sylvain

Similar Messages

  • How to configure applet/servlet in Tomcat?

    I'm building a web application which is going to be deployed in Tomcat. I have an applet(called MYAPPLET here) which "calls" a servlet(called MYSERVLET here). I cannot figure out how to configure the whole thing.
    /test is my docbase. In /test I made a directory "applet classes" and a directory "servlet". The applet tag I use in the HTML file looks like this:
    <applet
    code="MYAPPLET.class"
    codebase="http://localhost:8080/test/applet classes/"
    <!-- Tell the applet where its peer is located -->
    <param
    name="servlet"
    value="http://localhost:8080/test/servlet/MYSERVLET" />
    </applet>
    Now, when opening the page, the applet can be found but Tomcat gives an error 404 saying that it can not find the servlet. On the other hand, If I make an entry in my web.xml (with url mapping
    <url-pattern>/test </url-pattern> or <url-pattern>/test/servlet </url-pattern>, depending on where I put MYSERVLET)
    for the servlet mentioned above and I put the MYSERVLET in either the WEB-INF/classes directory or in /test/servlet Tomcat gives an error 403. It refuses to connect to the servlet!
    Please can anyone help me. THanx in advance
    debeumers

    You are configuring Tomcat and if you have further questions then post new questions in the forum "Java Servlet Technology".
    Given that your servlet url is "localhost:8080/test/servlet/MyServlet"
    This assumes -
    1) You have created a web application called "test". You have a directory called "test" under- $CATALINA_HOME/webapps
    2) Your html files, jsp files, and subdirectory "applet-classes" (Note - do Not use spaces when naming directories) exist in this top level directory of your web app. If you choose to place your applet classes in this subdir then the codebase can be "applet-classes" - omitting the "/" means this directory is located relative to the current directory of the html file. (No jar files? No archive param then..)
    3) You have a class file called "MyServlet" located - $CATALINA_HOME/webapps/test/WEB-INF/classes
    Some versions of Tomcat require you to uncomment out the area of the Mapping for the Invoker Servlet element in $CAT_HOME/conf/web.xml.
    <servlet-name>invoker</servlet-name>
    <url-mapping>/servlet/*</url-mapping>
    Once this is done, you need not create a directory called "servlet" anywhere in your structure. It is the url mapping.
    You will want to apply url-patterns and mappings to your application specific web.xml (This is located in - $C_H/webapps/test/WEB-INF) when it comes time to deploy the app, but that is not required for testing and development. You need not make changes to the app specific web.xml at this time just to get the servlet to run.
    And finally I suggest you read a book on this subject...

  • How to configure soap service?

    Hi
    I have installed the oracle9ias, and read all meassages about soap in the otn forums, but I do not know how to configure the soap service and test it.
    Anyone help me? thanks

    I have resovled the problem.
    first you should get the soap.ear file.
    then deploy it in the OC$J_home instance.
    url bind is /soap
    after deployment, you can test it.

  • How to configure snmp on loopback interface, on vrf TEST, in ASR 9000 series

    Hi there!
    How to configure snmp v2, with community string, for ASR 9000 series, in interface loopback 1, that has vrf TEST
    Regards!

    oh that is somewhat vague, but let me try...
    in order to enable snmp access to an interface in a vrf you need to enable MPP (management plane protection) in that vrf.
    Something like this:
    control-plane
     management-plane
      inband
       interface Loopback100
        allow all
    if you have this and it doesn't work, a config snippet and some debugs (like debug snmp packet) will help to define what to do next.
    regards
    xander

  • What is load balancer in fms cluster and how to configure it and testing fms clustering

    Hey guys,
                I am new to fms and clustering the fms server in origin-edge mode.But i am confused actually what is load balancer and how to configure it .
    After configuring , how to test and confirm that fms clustering is working.Is there any one can help me, plz....................
                                                                                                   thanks in Advance........

    Flash Media Server does not come bundled with a load balancer.  In theory, one deploys FMS in an edge/origin configuration like this
    1 Origin ----> N Edges ------> Customer Traffic
    The idea here is that N edges shoulder the load for most customers with a single origin coordinating resources like VOD content or script access.  The issue that remains is "how do I spread customers over the N edges" and that's where load balancing comes in.  Most customers use a prepackaged load balancing solution - as an Adobe staffer I can't recommend any of them over the other, but you should be able to get some recommendations here along with just googling for it.  The short answer contained here though is, you're going to have to find one outside of FMS if that's needed for your setup.
    Asa

  • How to configure, push and test Idoc from XI to R/3

    Assuming mapping between imported Idoc and some file structure is fine ,I have to configure XI and R/3 for pushing Idocs from Xi to r/3. I have to make sure they talk.
    -I have done this long time ago for pushing Idocs from R/3 to XI but I wonder if it is same procedure or it is different ?
    -Can someone tell me  - how to configure both for this purpose partner profiles , logical systems etc ?
    -And also after configurations , how I can check if Idocs is received from Xi and the entire process is successful?
    -does it make difference the type of Idocs during the configurations?
    -Are the test for success same for all type Idocs?
    Thanks all !
    Jon

    Here are the steps:
    ===========================
    First - Maintain the Sender R/3 System:
    SM59 : Create a RFC destination to XI
    WE21 : Create a TRFC Port ->Specify the RFC Destination Created
    BD54 : Create a Logical System for the Idoc Receiver
    WE20 : Create Partner Profile ->Maintain Outbound and the Inbound Parameters
    Second - Log on to XI System:
    SM59 : RFC Destination for Sender System
    IDX1 : Create the port to get Idoc Metadata from Sender System ( The Port Name must match the port name in the idoc header - Usually in format SAP<System ID>. eg. SAPID1 [Optional Step. Not mandatory]
    IDX2 : Maintain the Idoc Metadata. This is needed only by XI, and not by other SAP systems. IDX2 is needed because XI needs to construct IDoc-XML from the IDoc. No other SAP system needs to do that.
    To Enable Acknowledgement:
    SXMB_ADM ->Integration Engine Configuration ->Specific Configuration ->Add New entry -> Select parameters as:
    Category: RUNTIME
    Parameters: ACK_SYSTEM_FAILURE
    Current Value: 1
    Additional References
    Ale Configuration for pushing Idocs form SAP to XI
    /people/swaroopa.vishwanath/blog/2007/01/22/ale-configuration-for-pushing-idocs-from-sap-to-xi

  • How to configure Test System

    Hello Guys,
    I want to some guidelines to install and configure the SAP NWCE as Test System. Any link or documentation would be encouraged.
    Regards,
    Naeem

    Hi Naeem,
    The SCAfiles for the ESR comes along with the CE package, you need to deploy those SCA files on NWDS in order to access ESR which is a part of the CE installation process, though I undestand your point that it is not required unless you need ESR/SR. The NWDS accesses ESR and SR to deploy the EAR files for the service Interfaces created on ESR/SR on the Java server from where the services would be accessed in run time. I think you don't need the ESR/SR in the test system unless it would be used by the PI system as well (if you have one in your landscape), as NWDS accesses the ESR once to get the service interface and then deploy it on Java server. and the deploed EAR files can be transported to the Test environment. having said that, you should have ESR/SR in  your landscape as you would like the DEV, QA, PRD systems in sync.
    On your other question, NWDS accesses the ESR/SR to get the service interfaces and create ear files for the interfaces and deploy them on the Java server and used at runtime.
    Hope it helps.

  • Don't know which technology to utilize or how to configure ASA5505

    I have an ASA5505.  Currently, it is using static NAT on several ports to forward traffic to several devices inside my network.  It is a pain not only to configure but from the end user side.
    The issue I am having is the applicatoins I am using to access the devices become a mess with dual configurations, one for when I am connected to the internal network and one for when I am away from the office and accessing from the internet.  For example, I have 2 Cisco VC240 IP Cameras behind the ASA5505.  One is set use port 9091 and the other 9092.  When I am inside the office, I access them via http://10.1.2.215:9091 and http://10.1.2.216:9092.  But when I am away from the office, I have to have another configuration in an Android app to use them, http://external_ASA_IP:9091 and 9092 and then NAT 9091 to the object for Camera1 and 9092 for Camera2.  This is only one scenario.  I also have a UC320W that I would like to put an IP phone at home and it sounds like AnyConnect is the only way to do this.
    It sounds like to me that if I use some type of VPN, I can access the same devices using the same IP whether internal or external with the external connection using the VPN to tunnel the IP to the local network.  There seems to be quite a few ways to do this with an ASA 5505.
    AnyConnect seems like the way to go but after reading Cisco documentation, it requires your Android device to be root'd if it is not a particular Samsung model.  If I understand correctly, root'ing your phone voids the warranty.  I know it is common practice but would think Cisco would have a better solution as I am sure Cisco would not want another manufacturer telling their customers to void the warranty on their Cisco equipment in order to get it to work.
    I believe I can just use IPSEC and use the native VPN of the Android OS and also tunnel L2TP as the Android supports IPSEC-PSK/L2TP or IPSEC-CRT/L2TP.  But will either of these will support the IP phone to the UC320W?
    A friend also told me to use NginX to proxy URL's so the URL http://www.fqdn.com/camera1 gets proxy'd to the internal IP of Camera1 and http://www.fqdn.com/camera2 gets proxy'd to Camera2.  He says I should be able to store a cookie on the phone and let the phone authenticate to the camera and if the phone cannot, the proxy can authenticate internally to the IP camera over SSL.
    I don't know anymore, I am so confused and just want to simplify my life as I am just a small business with me and a couple other employees but I have full-time job and it is not IT/Network Technician, it is only CTO/CEO/CIO/CFO.  I don't have hours upon hours to set this up and test and I don't have hours upon hours to manage it.  I just need to simplify this and have so that it is a set-it-and-forget-it for 6 months to 1 year and re-evaluate or update.  So, if someone suggests IPSEC, I would not know how to configure anyway and you should expect another post.  The same for AnyConnect or any of the other suggestions.
    Thanks in advance for any advice.

    Hi!
    1. Set Calculation Mode property of ITEM_5 to Formula.
    Formula property:
    nvl(:Block_Name.ITEM_1, 0) + nvl(:Block_Name.ITEM_2, 0) + nvl(:Block_Name.ITEM_3, 0) + nvl(:Block_Name.ITEM_4, 0)
    OR
    Function_Name(Param_1,... Param_N);
    Have in view of, that the ITEM_5 data will not be saved in DataBase.
    2. When-Validate-Item trigger is usfull when is necessary to store calculated item data in DataBase.
    Rename you Post-Query trigger to When-Validate-Item.
    Modify trigger: Store calculation result in the variable.
    (Don't forget to round variable value!)
    Then compare it with ITEM_5. If they are different - :ITEM_5 := var_name.
    I prefer the first method.

  • How to configure ACS 5.2 for policy condition on TACACS+ Service

    In https://supportforums.cisco.com/message/3953175#3953175 thread, I was able to get the ACS 5.2 work with SRX for both SSH CLI and J-Web TACACS+ accounts. However, I found the behavior is different on our production environment. I found our ACS 5.2 was configured authorization rule with condition "TACACS+ Service" = "junos-exec". I don't know how to configure this on my ACS 5.2 Please guide me how to configure this.
    I found there was NO TACACS+ "Authorization Request" when access via J-Web in our production SRX and ACS. However, there were TACACS+ "Authorzation Request" when access via J-Web in our production SRX and ACS. The difference between my lab ACS and production ACS is the authorization rule condition. In my condition, I configure with all "SRX" Device Type. but in our production ACS 5.2, it was configure to TACACS+ Service=junos-exec. so I like to test it in our lab to find out the difference. Thanks.

    I would suggest you to go through the below two link.
    http://www.cisco.com/en/US/docs/net_mgmt/cisco_secure_access_control_system/5.2/migration/guide/Migration_Configure.html
    http://www.cisco.com/en/US/docs/net_mgmt/cisco_secure_access_control_system/5.2/user/guide/common_scenarios.html

  • How to configure Solaris 10 IPMP for Oracle VDI 3.3.2

    Hi,
    Does anyone have an indication on how to configure Solaris 10 IPMP in a manner that supports Oracle VDI?
    We have setup two servers with 2 test addresses on physical and 1 logical for the hostname of the box, but when we configure VDI the VDI database does not come up on the 2nd box. We have also tried configuring IPMP without test addresses, but it doesn't make any difference - the DB still doesn't go into the up state after configuration on the 2nd server.
    Solaris 10 u9 with patches
    Two physical NICs on management VLAN via 2 switches
    Two physical NIcs on VDI VLAN via 2 switches
    VDI 3.3.2
    All hostnames are in DNS and resolve for short name, FQDN and also reverse IP lookup
    The is a proposal pdf on Oracle website that mentions IPMP, so someone has done it. Just could do with a hint on how it was done so that it works.
    ( http://www.oracle.com/us/technologies/virtualization/vdi-design-proposal-1401195.pdf )
    Thanks
    Paul

    OK, updating this with what was causing the issue.
    Not an IPMP problem this was a DNS problem. The VDI servers have access to two DNS environments, so had a DNS search path that had the domain where the VMs where going to go and another for management of the box with two DNS servers listed in /etc/resolv.conf.
    We configured VDI using the FQDN for the management DNS, however the vda-config script takes the hostname of the box adds the first DNS domain from /etc/resolv.conf search entry and configures using (it seams to ignore what you put into the vda-config). When the config script comes to configuring MySQL because the names didn't match (FQDN it created by adding hostname to the 1st entry in DNS search path & FQDN that you put into vda-config command), it decides that it is configuring a "Client MySQL" instance rather than a "Slave MySQL" instance, which means the VDA DB never comes up in the out of vda-center status.
    So the fix is to:
    1) configure IPMP without test addresses (so traffic comes out of the correct IP and can be reverse looked up in DNS by the other host)
    2) if you have multiple DNS search entries, configure VDI using the first entry in your search path as its FQDN
    Paul

  • [Forum FAQ] How to configure a Data Driven Subscription which get multi-value parameters from one column of a database table?

    Introduction
    In SQL Server Reporting Services, we can define a mapping between the fields that are returned in the query to specific delivery options and to report parameters in a data-driven subscription.
    For a report with a parameter (such as YEAR) that allow multiple values, when creating a data-driven subscription, how can we pass a record like below to show correct data (data for year 2012, 2013 and 2014).
    EmailAddress                             Parameter                      
    Comment
    [email protected]              2012,2013,2014               NULL
    In this article, I will demonstrate how to configure a Data Driven Subscription which get multi-value parameters from one column of a database table
    Workaround
    Generally, if we pass the “Parameter” column to report directly in the step 5 when creating data-driven subscription.
    The value “2012,2013,2014” will be regarded as a single value, Reporting Services will use “2012,2013,2014” to filter data. However, there are no any records that YEAR filed equal to “2012,2013,2014”, and we will get an error when the subscription executed
    on the log. (C:\Program Files\Microsoft SQL Server\MSRS10_50.MSSQLSERVER\Reporting Services\LogFiles)
    Microsoft.ReportingServices.Diagnostics.Utilities.InvalidReportParameterException: Default value or value provided for the report parameter 'Name' is not a valid value.
    This means that there is no such a value on parameter’s available value list, this is an invalid parameter value. If we change the parameter records like below.
    EmailAddress                        Parameter             Comment
    [email protected]         2012                     NULL
    [email protected]         2013                     NULL
    [email protected]         2014                     NULL
    In this case, Reporting Services will generate 3 reports for one data-driven subscription. Each report for only one year which cannot fit the requirement obviously.
    Currently, there is no a solution to solve this issue. The workaround for it is that create two report, one is used for view report for end users, another one is used for create data-driven subscription.
    On the report that used create data-driven subscription, uncheck “Allow multiple values” option for the parameter, do not specify and available values and default values for this parameter. Then change the Filter
    From
    Expression:[ParameterName]
    Operator   :In
    Value         :[@ParameterName]
    To
    Expression:[ParameterName]
    Operator   :In
    Value         :Split(Parameters!ParameterName.Value,",")
    In this case, we can specify a value like "2012,2013,2014" from database to the data-driven subscription.
    Applies to
    Microsoft SQL Server 2005
    Microsoft SQL Server 2008
    Microsoft SQL Server 2008 R2
    Microsoft SQL Server 2012
    Please click to vote if the post helps you. This can be beneficial to other community members reading the thread.

    For every Auftrag, there are multiple Position entries.
    Rest of the blocks don't seems to have any relation.
    So you can check this code to see how internal table lt_str is built whose first 3 fields have data contained in Auftrag, and next 3 fields have Position data. The structure is flat, assuming that every Position record is related to preceding Auftrag.
    Try out this snippet.
    DATA lt_data TYPE TABLE OF string.
    DATA lv_data TYPE string.
    CALL METHOD cl_gui_frontend_services=>gui_upload
      EXPORTING
        filename = 'C:\temp\test.txt'
      CHANGING
        data_tab = lt_data
      EXCEPTIONS
        OTHERS   = 19.
    CHECK sy-subrc EQ 0.
    TYPES:
    BEGIN OF ty_str,
      a1 TYPE string,
      a2 TYPE string,
      a3 TYPE string,
      p1 TYPE string,
      p2 TYPE string,
      p3 TYPE string,
    END OF ty_str.
    DATA: lt_str TYPE TABLE OF ty_str,
          ls_str TYPE ty_str,
          lv_block TYPE string,
          lv_flag TYPE boolean.
    LOOP AT lt_data INTO lv_data.
      CASE lv_data.
        WHEN '[Version]' OR '[StdSatz]' OR '[Arbeitstag]' OR '[Pecunia]'
             OR '[Mita]' OR '[Kunde]' OR '[Auftrag]' OR '[Position]'.
          lv_block = lv_data.
          lv_flag = abap_false.
        WHEN OTHERS.
          lv_flag = abap_true.
      ENDCASE.
      CHECK lv_flag EQ abap_true.
      CASE lv_block.
        WHEN '[Auftrag]'.
          SPLIT lv_data AT ';' INTO ls_str-a1 ls_str-a2 ls_str-a3.
        WHEN '[Position]'.
          SPLIT lv_data AT ';' INTO ls_str-p1 ls_str-p2 ls_str-p3.
          APPEND ls_str TO lt_str.
      ENDCASE.
    ENDLOOP.

  • How to configure Email output in SD

    Dear friends Pls sent me details of how to configure email output in SD module.
    <REMOVED>
    Thanks & Regards,
    <REMOVED>

    Hi
    Idoc Def :Standard SAP format for electronic data interchange between systems (Intermediate Document). Different message types (such as delivery confirmations or purchase orders) normally represent different specific formats, the IDoc types. However, multiple message types with related content can be assigned to one IDoc type: For example, the IDoc type ORDERS01 transfers the "logical" message types ORDERS (purchase order) and ORDRSP (order confirmation).
    Idocs are of two types basic type and the extension type .we need to configure the system settings for this process .
    IDOC (Intermediate Document) - A data holder.
    IDOC is divided in to three parts.
    Control Record
    Data Record
    Status Record
    Control record (Table: EDIDC):
    - Every IDOC has only one Control Record
    - Each Control Record contains header information like:
    o IDOC Number
    o Direction of IDOC: Inbound or Outbound
    o Date and Time of creation of IDOC
    o Date and time when the IDOC was last modified.
    o Message Type of IDOC
    o IDOC type and extension of IDOC
    o Sender and Receiver Partner
    Data record – (Table: EDID4):
    - Data Record contains Data to be processed.
    - Every IDOC has one data record with multiple segments in hierarchy.
    - Segments and Hierarchy of Segments are defined by IDOC Type and Extension.
    - IDOC created has to strictly follow the hierarchy; else IDOC fails with Syntax error.
    - Segments which are repetitive have qualifiers attached to it
    Status record – (Table: EDIDS):
    - Status Record describes the status of IDOC.
    - Each IDOC contains one status Record with multiple status information.
    - Status at each level is appended to IDOC. E.g. When IDOC is created in SAP, Status is “This IDoc has been generated through a test transaction”, When the IDOC is added to system it is “IDOC added”, “IDOC ready to be transferred to Application”……
    - Status should always be read bottom-up. Status at the top is the latest status.
    - Some Example of Status Records:
    o Inbound:
    § 53 - IDOC successfully posted
    § 51 – IDOC Failed
    § 64 - IDOC ready to be transferred to Application
    o
    Outbound:
    § 30 – IDOC ready for Dispatch
    § 03 – IDOC passed to port OK
    § 12 – IDOC Dispatched
    § 16 – Functional Acknowledgement Positive
    § 17 – functional Acknowledgement Negative
    IDoc Type: Defines the segments and hierarchy of segments
    o Transaction Code:
    § WE30 – To create, change or display the IDOC type and the extension.
    § WE31 – To create the Segment
    - IDOC type defines the segments to be used in the IDOC.
    - It also defines the hierarchy and syntax of the segments.
    - IDOC extension is nothing but to add segments to standard IDOC types.
    - Transaction WE31 allows you to create segments.
    - Program RSEIDOC3 documents the use of each IDOC type.
    Segments:
    Attributes of a Segment:
    - Mandatory Segment: If checked, this segment should always exist in the IDOC.
    - Minimum Number:
    - Maximum Number: Maximum number of times this segment can be repeated in IDOC. -
    Parent Segment: Parent of this segment
    - Hierarchy level: Level of hierarchy.
    Segment Definition (WE31):
    Messsage Type: Defines the type of data in the IDOC
    o Transaction Code:
    § WE81 – To create, change or display the Message type and the extension.
    § WE82 – Using this transaction you can link Message Type, IDOC Type, IDOC Extension and version.
    - Message type identifies the type of data IDOC holds. E.g. Orders (ORDERS), Delivery (DESADV), Invoice (INVOICE). It also defines what needs to be done with the data in the IDOC, in case of Inbound IDOC, and which data to be extracted in case of Outbound IDOC.
    - Message Type is linked to a process code, which in turn is linked to a Function Module. This function module extracts from or posts data to SAP depending on direction of IDOC.
    - Relation between Message Type, IDOC type and IDOC extension needs to define. Without this relation Message type or IDOC type cannot be used.
    Message Type Create, Change or Display (WE81):
    Setup link between Message Type, IDOC Type, IDOC Extension and Version (WE82)
    Process Code: Function Module is linked to a process code. This function module in executed for inbound or outbound IDOC.
    o Transaction Codes:
    § WE41 – Outbound Process Code
    § WE42 – Inbound Process Code
    - Process codes are linked to a Function Module.
    - Relationship is Message Type is linked to a Process Code which is linked to a Function Module.
    - In case if you are using a stand alone code to trigger an IDOC, you need not define a process code.
    RFC Destination: System definition of destination.
    o Transaction Code: SM59
    - RFC destination identifies the destination of IDOC.
    - In case of ALE:
    o In ALE the communication mode is IDOC to IDOC, hence the type used is R/3 Connections.
    o It is the destination SAP system which will receive the IDOC.
    o In RFC destination you define the destination SAP system details like System, Login and Password.
    Go thr below links:
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCMIDALEIO/BCMIDALEIO.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCMIDALEPRO/BCMIDALEPRO.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/CABFAALEQS/CABFAALEQS.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCSRVEDISC/CAEDISCAP_STC.pdf
    Sail

  • How to configure quality process with batch management.

    Dear Friends,
                we have a scenario,our client receive goods with 101movement type  into quality inspection stock.
    once stock is passed quality check stock,then we will take into unrestricted stock with 321 movement type.
    failed in quality test will  be rejected.please give me suggestion how to map the scenario.while i am testing
    the process in my sandbox,its asking batch number.its not allowing me to post unrestricted stock with 321
    movement type.how to make batch as optional in this scenario?if batch is mainted for the material,how to
    configure the process.we are not using QM module.
    please suggest me.
    Regards,
    Varun

    In Batch management config in Logistic general, Check for the movement type 101 & 321 whether batch creation has been activated to Manual, Automatic, or no creation. In your case it looks like for mvt 321 batch creation is set to manual & for 101 no creation.
    if batch creation is set to manual, enter a manual batch no & proceed.
    But to cross verify wheter 101 has generated batch or not please check the stocks in MMBE.
    Hope this helps.
    Regards
    Mithun

  • How to configure to generate two accounting documents for Mvt 645

    Dear Experts,
    I read all documents related to STO one-step procedure (mvt 645). They said that the system generate two accounting documents for 2 company codes (receing and issuing company codes) after goods issue reference to DO.
    However, after configuration one step procedure and test my system, the system only generated one accounting document for receiving company code, not for issuing company code. In Material document, I saw 2 Mvt 645 and 101. In the lines having Mvt 645, I saw G/L account of cost-of-goods-sold.
    Could you please to show me how to generate two accounting documnets automatically after goods issue referent to DO with Mvt 645?
    Thank you so much,
    Best regards,
    Anh Duong

    My configuration steps are:
    Step 1: preparation:
          create customer and vendor master
          extend sales & purchasing views for material master
    Step 2: Assign customer number and the organizational units (sales organization, distribution channel, and division) to the Supplying plant and Receiving plant
    IMG --> Materials Management --> Purchasing --> Purchase Order --> Set up Stock Transport Order --> Define Shipping Data for Plants
    Step 3: tick on column "One step" for the line of supplying, receiving plant, and PO type cross transport order
    IMG --> Materials Management --> Purchasing --> Purchase Order --> Set up Stock Transport Order --> Assign Document Type, One-Step Procudure, Underdelivery Tolerance
    Step 4: assign delivery type NLCC for PO type cross transport order
    IMG --> Materials Management --> Purchasing --> Purchase Order --> Set up Stock Transport Order --> Assign Delivery Type and Checking Rule
    These above steps are in MM module.
    Please help me,
    Thank you,
    Anh Duong

  • How to Configure logical path and how to retrieve physical path for an IDOC

    Hi ,
    I want to configure logical path for an IDOC,
    and then I need to retrieve Physical path from this logical path
    using a FM (eg. FILE_GET_NAME).
    How to configure the logical path for an idoc ?
    Regards ,
    Harshit Rungta

    Using Logical Files in ABAP Programs
    To create a physical file name from a logical file name in your ABAP programs, use the function module FILE_GET_NAME. To insert the function module call in your program, choose Edit ® Insert statement from the ABAP Editor screen. A dialog box appears. Select Call Function and enter FILE_GET_NAME. The parameters of this function module are listed below.
    Import parameters
    Parameters
    Function
    CLIENT
    The maintenance tables for the logical files and paths are client-dependent. Therefore, the desired client can be imported. The current client is stored in the system field SY-MANDT.
    LOGICAL_FILENAME
    Enter the logical file name in upper case letters that you want to convert.
    OPERATING_SYSTEM
    You can import any operating system that is contained in the list in Transaction SF04 (see Assigning Operating Systems to Syntax Groups). The physical file name will be created according to the syntax group to which the operating system is linked. The default parameter is the value of the system field
    SY-OPSYS.
    PARAMETER_1
    PARAMETER_2
    If you specify these import parameters, the reserved words  in the physical path names will be replaced by the imported values.
    USE_PRESENTATION
    _SERVER
    With this flag you can decide whether to import the operating system of the presentation server instead of the operating system imported by the parameter OPERATING_SYSTEM.
    WITH_FILE_EXTENSION
    If you set this flag unequal to SPACE, the file format defined for the logical file name is appended to the physical file name.
    Export Parameters
    Parameters
    Function
    EMERGENCY_FLAG
    If this parameter is unequal to SPACE, no physical name is defined in the logical path. An emergency physical name was created from table FILENAME and profile parameter DIR_GLOBAL.
    FILE_FORMAT
    This parameter is the file format defined for the logical file name. You can use this parameter, for example, to decide in which mode the file should be opened.
    FILE_NAME
    This parameter is the physical file name that you can use with the ABAP statements for working with files.
    Exception Parameters
    Parameters
    Function
    FILE_NOT_FOUND
    This exception is raised if the logical file is not defined.
    OTHERS
    This exception is raised if other errors occur.
    Suppose the logical file MYTEMP and the logical path TMP_SUB are defined as in the preceding topics and we have the following program:
    DATA: FLAG,
          FORMAT(3),
          FNAME(60).
    WRITE SY-OPSYS.
    CALL FUNCTION 'FILE_GET_NAME'
         EXPORTING
              LOGICAL_FILENAME        = 'MYTEMP'
              OPERATING_SYSTEM        = SY-OPSYS
              PARAMETER_1             = '01'
         IMPORTING
              EMERGENCY_FLAG          = FLAG
              FILE_FORMAT             = FORMAT
              FILE_NAME               = FNAME
         EXCEPTIONS
              FILE_NOT_FOUND          = 1
              OTHERS                  = 2.
    IF SY-SUBRC = 0.
      WRITE: /  'Flag      :', FLAG,
             / 'Format    :', FORMAT,
             / 'Phys. Name:', FNAME.
    ENDIF.
    The output appears as follows:
    HP-UX
    FLAG :
    FORMAT : BIN
    Phys. Name: /tmp/TEST01
    In this example, the R/3 System is running under the operating system HP-UX, which is member of the syntax group UNIX. The logical file name MYTEMP with the logical path TMP_SUB is converted into a physical file name /tmp/TEST01 as defined for the syntax group UNIX. The field FNAME can be used in the further flow of the program to work with file TEST01 in directory /tmp.
    Suppose we have a logical file name EMPTY with the physical file name TEST, connected to a logical path that has no specification of a physical path. If you replace the EXPORTING parameter 'MYTEMP' with 'EMPTY' in the above example, the output appears as follows:
    HP-UX
    FLAG : X
    FORMAT :
    Phys. Name: /usr/sap/S11/SYS/global/TEST
    The system created an emergency file name, whose path depends on the installation of the current R/3 System.

Maybe you are looking for