What are the backgroundjobs we schedule in grc

hello experts,
                        pls tell me what is the requirement of scheduling  background jobs in GRC
sanajy

sanajy,
I would kindly ask you to either visit an Access Control training or at least read the manual - all your questions so far are extremely basic ones, which can not easily be answered out of context. You're much better off gaining some deeper knowledge first and then coming back to ask questions.
Frank.

Similar Messages

  • How to schedule a load job from bi to r3 daily ? What are the available opt

    I think we can create a PC for the same so should we code an ABAP routine with include to do the process. So can it be scheduled daily, what are the other option in BI7 other than PC ?

    Hi Manohar,
    To make the data source Real-Time enabled you need to run a small program which updates the REALTIME flag in ROOSOURCE table. Program is as below.
    report z_rda_flag.
    tables : roosource.
    parameters : p_oltpsr like roosource-oltpsource.
    update roosource set realtime = 'X'
    where oltpsource = p_oltpsr
    and objvers = 'A'.
    Enter the data source name in the selection screen and run the program.
    Now go back to SE16 and see the contents of 0FI_GL_4 inROOSOURCE.
    You can see that the program has updated REALTIME flag for the datasource 0FI_GL_4.
    hope this answer solves ur issue
    u can do this only for an ODs and not for a cube. will let u know if there are any other options for a cube.
    assign values to the answers as its the way of thanking in SDN.
    regards,
    Ranjith

  • What are  the roles & responsibilities  of GRC/SECURTIY CONSULTANT IN IMPLE

    HELLO EXPERTS,
                                  CAN U PLS TELL WHAT ARE THE ROLES & RESPONSIBILITIES OF GRC/SECURITY CONSULTANT IN IMPLEMENTATION OF GRC TOOL.
    REGARDS,
    SANJAY

    Hi Sanjay,
    The role of GRC & Securtiy Consultant is basicly the implementation of GRC Application in Customor envoriment. In most of cases he gives technical support for installation and configuration. What we are really do is following..
    1. Preperation of System Landscape Concept --> Provide Hardware and Software spesifications to customer
    2. Installation of Software Components --> This contains the installation of NW and GRC Components and also the installation of SAP Backend Components like Real Time Agents and the post-Installation configurations
    3. Post-Installation and Basic Configuration --> Create Connectors and additional basic configurations
    4. Hosting of Workshops for Risk Creation and Ruleset customizing
    5. Implementation of Custom defined Rules in the Application
    6. Test of Applications
    7. Hostin of Workshops for Approval Methodology and Workflow Use Cases
    8. Hostin of Workshops for Superuser Role Usage and Management of Superuser
    9. Role Change and Creation Workshops in GRC ERM.
    10 Documentation and Deliver of Custom Concepts for Compliant User & Role Administration
    11. Handling of Error and Trobles.
    this are my daily business...
    Best Regards.

  • What are the main transaction codes in XI?

    What are the main transaction codes in XI?

    SAP XI TRANSACTION CODES
    ALRTCATDEF- Alerts Configuration.
    ALRTDISP - Display of Alerts Created
    SXMB_IFR -> Start Integration Builder
    SXMB_MONI -> Integration Engine - Monitoring
    SXI_MONITOR -> XI: Message Monitoring
    SXI_CACHE -> To Access IS runtime cache
    SXI_SUPPORT -> Test all the Repository and Directory Objects
    SXI_CACHE -> XI data directory cacheidx1 -> idoc adapter
    IDX2 -> idoc metadata
    IDX5 -> monitor idoc adapter
    SM21 -> log di sistema
    ST22 -> dump abap (analisi di eccezioni)
    SMQ1 -> messages inbound e outbound queue
    SMQ2 -> messages inbound e outbound queue
    SMICM -> J2EE administration
    ST06 -> stato della macchina cpu memory filesystem machine status - cpu, memory and file system.
    ST03 -> workload.
    SCC4 -> visualizza mandanti del sistema
    WE20 -> partner profiles
    sale - ale settings
    RZ10 - edit system profile
    SM59 ->mantain rfc destinations
    BD64 -> mantain distribution model (trasformazioni)
    SU01 -> users
    AL08 -> list all logged users(user login logon)
    SE10 -> Change Request
    SE09 -> Change Request
    WE05 -> All idocs
    WE19 -> IDoc Test Tool
    WE21 -> port definition
    SE11 -> Data dictionary
    SM21 -> log sistema xi
    se37 -> mantain funcion modules
    SICF -> http server configuration
    SMGW -> trace, alzare livello di trace.
    BD13 ->
    BD64 -> modelli di distribuzione
    PFCG -> Roles
    tabella TSTC -> sap transactions codes
    tabella TSTCT -> transaction descriptions
    STMS -> transports management
    SPAM -> apply ABAP support packages
    SPAU -> manage objects after apply support packages
    SE01 -> manage change requests
    SLDCHECK -> Test SLD Connection
    SLDAPICUST-> SLD API Customizing
    SXMB_ADM -> Integration Engine - Administration
    SXMB_MONI_BPE -> Process Engine - Monitoring
    SE38 -> ABAP Editor
    SE11 -> ABAP Dictionary
    ST22 -> ABAP dump analysis
    SPROXY-> ABAP Proxy Generation
    SE80 -> Object Navigator
    ABAPDOCU -> ABAP Documentation and Examples
    SE24-> Class Builder
    SM21-> Online System Log Analysis
    SMQ1-> qRFC Monitor (Outbound Queue)
    SMQ2-> qRFC Monitor (Inbound Queue)
    RZ70-> SLD Administration
    SM58-> Asynchronous RFC Error Log
    SM59-> RFC Destinations (Display/Maintain)
    SMICM-> ICM Monitor
    WE60-> Documentation for IDoc types
    BD87-> Status Monitor for ALE Messages
    IDX1-> Port Maintenance in IDoc Adapter
    IDX2-> Meta Data Overview in IDoc Adapter
    WE05-> IDoc Lists
    WE02-> Display IDoc
    WE19-> Test tool
    WE09-> Search for IDocs by Content
    WE20-> Partner Profiles
    WE21-> Port definition in XI
    SE16-> Data Browser
    SE93-> Maintain Transaction Codes
    SM30-> Call View Maintenance
    SU01-> User Maintenance
    SM02-> System Messages
    BD54 -Logical System Creation.
    SWXF_PBUILDER -> for Detail BPM Process
    SMQS - > to register the destination in QOUT scheduler
    WEOUTQUEUE - > to start the queue processing
    SMQR - > to register the queue
    IDXPW - > to activate the IDOC message package
    IDXP - > to monitor the message packages.
    SWF_XI_CUSTOMIZING -> transaction to check prerequisites for integration processes.
    http://www.erpgenie.com/abaptips/content/view/452/62/

  • What are the importent transaction codes in edi

    what are the importent transaction codes in edi

    hi,
    =>sale - ALE Customizing
    =>bale - Distribution Administration (ALE)
    =>wedi - IDOC Type and EDI Basis
    =>shdb - BDC (Transaction Recorder)
    =>lsmw - LSMW
    =>bapi - BAPI
    =>spau - Display Modified DE Objects
    =>spdd - Display Modified DDIC Objects
    =>sara - central idoc archiving
    =>stms - Transport Management System
    =>wedi - IDOC Type and EDI Basis (IDOC Configuration)
    =>bd21 - Analyze change pointers - create IDOC from change pointer
    =>bd10 - Send Material Master
    =>bd12 - Send Customer Master
    =>bd14 - Send Vendor Master
    =>bd50 - Activate Change Pointer for Message Type
    =>bd55 - Conversion rule user exit (Link conversion rule user exit to the different system \ partner combinations)
    =>bd54 - Maintain Logical Systems
    =>bd71 - Distribute Customer Model
    =>bd87 - Process Inbound IDOCs
    =>bd88 - Process Outbound IDOCs
    =>bdm2 - Cross-system IDOC Reporting
    =>bdm7 - ALE Audit - statistical analyses
    =>bdm8 - ALE Audit - sending confirmations
    =>bd61 - Activate Change Pointer generally
    =>bmv0 - Direct Input Data Transfer Administration
    =>rz04 - Operation Mode
    =>rz12 - RFC Server Group Maintenance
    =>sar3 - Archiving object customizing
    =>s001 - SAPoffice Inbox
    =>s010 - Standards Texts (SAP Script)
    =>se01 - Transport Organizer (Extended View - Performing and managing the transport of development object across different systems)
    =>se09 - Transport (Workbench) Organizer (Controlling and keeping track of Development work and Development Objects)
    =>se78 - Administration of Form Graphics (Import Graphics into SAP System)
    =>se84 - ABAP Repository Information (search for SAP objects)
    =>su01 - User Maintenance (Security)
    =>su02 - Maintain Authorization Profiles
    =>su03 - Maintain Authorizations
    =>sm36 - Background Job Scheduling
    =>sm37 - Background Job Monitoring (display batch jobs)
    =>sm31 - Table Maintenance
    =>sm30 - Calling View Maintenance
    =>sm35 - Batch Input Monitoring (Error Logs)
    =>sm50 - Process Overview
    =>sm58 - Transactional RFC Monitoring
    =>sm59 - Maintain RFC Destinations (Define RFC Configuration)
    =>snro - Number Range Objects (for IDOCs)
    =>we02 - Display IDOC
    =>we05 - IDOC lists
    =>we07 - IDOC statistics
    =>we12 - Inbound processing of outbound file - convert outbound to inbound file
    =>we14 - Process (dispatch) IDOCs through Port - RSEOUT00
    =>we15 - Outbound IDOC from NAST - message control
    =>we16 - Inbound File
    =>we18 - Generate status file
    =>we19 - Test tool
    =>we20 - Add new Message Type to Partner Profile
    =>we21 - Maintain Port Definition
    =>we30 - IDOC Type development
    =>we31 - IDOC Segment development
    =>we41 - Process Code Outbound
    =>we42 - Process Code Inbound
    =>we46 - IDoc Administration
    =>we47 - Status Code Maintenance
    =>we57 - Assign function module to logical message and IDoc type
    =>we63 - IDOC Type for Parser Output
    =>we82 - Assign IDOC to Message Type
    =>we60 - IDOC Documentation - IDOC Types
    =>we61 - IDOC Documentation - IDOC Record Types
    =>we81 - Create new IDOC Message Types
    =>we82 - Assign IDOC to Message Type
    =>pa20 - Display HR Master Data
    =>pa30 - Maintain HR Master Data
    =>bd59 - Create and assign Filter Object Type for Message Type
    =>bd64 - Maintenance of ALE or any other Distribution Model
    Hope this helps, Do reward.

  • What are  the tables and fields  in fico & sd module

    what are  the tables and fields  in fico & sd module

    Hi,
    <b> Sales and Distribution:</b>
    Customers KNA1 General Data
    KNB1 Customer Master – Co. Code Data (payment method, reconciliation acct)
    KNB4 Customer Payment History
    KNB5 Customer Master – Dunning info
    KNBK Customer Master Bank Data
    KNKA Customer Master Credit Mgmt.
    KNKK Customer Master Credit Control Area Data (credit limits)
    KNVV Sales Area Data (terms, order probability)
    KNVI Customer Master Tax Indicator
    KNVP Partner Function key
    KNVD Output type
    KNVS Customer Master Ship Data
    KLPA Customer/Vendor Link
    Sales Documents VBAKUK VBAK + VBUK
    VBUK Header Status and Administrative Data
    VBAK Sales Document – Header Data
    VBKD Sales Document – Business Data
    VBUP Item Status VBAP Sales Document – Item Data
    VBPA Partners
    VBFA Document Flow
    VBEP Sales Document Schedule Line
    VBBE Sales Requirements: Individual Records
    SD Delivery DocumeLIPS Delivery Document item data, includes referencing PO
    LIKP Delivery Document Header data
    Billing Document VBRK Billing Document Header
    VBRP Billing Document Item
    SD Shipping Unit VEKP Shipping Unit Item (Content)
    VEPO Shipping Unit Header
    <b>Financial Accounting</b>
    FBAS             Financial Accounting “Basis”
    BKPF             Accounting Document Header              BUKRS / BELNR / GJAHR
    BSEG             Accounting Document Segment             BUKRS / BELNR / GJAHR / BUZEI
    BSIP             Index for Vendor Validation of Double   BUKRS / LIFNR / WAERS / BLDAT /
                     Documents                               XBLNR / WRBTR / BELNR / GJAHR / BUZEI
    BVOR             Inter Company Posting Procedure         BVORG / BUKRS / GJAHR / BELNR
    EBKPF            Accounting Document Header (docs from   GLSBK / BELNR / GJHAR / GLEBK
                     External Systems)
    FRUN             Run Date of a Program                   PRGID
    KLPA             Customer / Vendor Linking               NKULI / NBUKR / NKOAR / PNTYP
                                                             / VKULI / VBUKR / VKOAR
    KNB4             Customer Payment History                KUNNR / BUKRS
    KNB5             Customer Master Dunning Data            KUNNR / BUKRS / MABER
    KNBK             Customer Master Bank Details            KUNNR / BANKS / BANKL / BANKN
    KNC1             Customer Master Transaction Figures     KUNNR / BUKRS / GJHAR
    KNC3             Customer Master Special GL Transactions KUNNR / BUKRS / GJAHR / SHBKZ
                     Figures
    LFB5             Vendor Master Dunning Data              LIFNR / BUKRS / MABER
    LFBK             Vendor Master Bank Details              LIFNR / BANKS / BANKL / BANKN
    LFC1             Vendor Master Transaction Figures       LIFNR / BUKRS / GJHAR
    LFC3             Vendor Master Special GL Transactions   LIFNR / BUKRS / GJHAR / SHBKZ
                     Figures
    VBKPF            Document Header for Document Parking    AUSBK / BUKRS / BELNR / GJHAR
    FBASCORE         Financial Accounting General Services “Basis”
    KNB1             Customer Master (Company Code)          KUNNR / BUKRS
    LFA1             Vendor Master (General Section)         LIFNR
    LFB1             Vendor Master (company Code Section)    LIFNR / BUKRS
    SKA1             G/L Account Master (Chart of Accounts)  KTOPL / SAKNR
    SKAT             G/L Account Master (Chart of Accounts – SPRAS / KTOPL / SAKNR
                     Description)
    MAHNS            Accounts Blocked by Dunning Selection   KOART / BUKRS / KONKO / MABER
    MHNK             Dunning Data (Account Entries)          LAUFD / LAUFI / KOART / BUKRS /
                                                             KUNNR / LIFNR / CPDKY / SKNRZE /
                                                             SMABER / SMAHSK / BUSAB
    FI-GL-GL (FBS)   General Ledger Accounting: Basic Functions- G/L Accounts
    SKAS             G/L Account Master (Chart of Accounts – SPRAS / KTOPL / SAKNR / SCHLW
                     Key Word list)
    SKB1             G/L Account Master (Company Code)       BUKRS / SAKNR
    FI-GL-GL (FBSC)  General Ledger Accounting: Basic
                     Functions - R/3 Customizing for G/L Accounts
    FIGLREP          Settings for G/L Posting Reports        MANDT
    TSAKR            Create G/L account with reference       BUKRS / SAKNR
    FI-GL-GL (FFE)   General Ledger Accounting: Basic
                     Functions - Fast Data Entry
    KOMU             Account Assignment Templates for G/L    KMNAM / KMZEI
                     Account items
    FI-AR-AR (FBD)   Accounts Receivable: Basic Functions - Customers
    KNKA             Customer Master Credit Management :     KUNNR
                     Central Data
    KNKK             Customer Master Credit Management :     KUNNR / KKBER
                     Control Area Data
    KNKKF1           Credit Management : FI Status data      LOGSYS / KUNNR / KKBER / REGUL
    RFRR             Accounting Data – A/R and A/P           RELID / SRTFD / SRTF2
                     Information System
    FI-BL-PT         Bank Accounting: Payment (BFIBL_CHECK_D)  Transactions – General Sections
    PAYR             Payment Medium File                     ZBUKR / HBKID / HKTID / RZAWE /
                                                             CHECT
    PCEC             Pre-numbered Check                      ZBUKR / HBKID / HKTID / STAPL
    FI-BL-PT-AP(FMZA)Bank Accounting: Payment Transactions – Automatic Payments
    F111G            Global Settings for Payment Program for MANDT
                     Payment Requests
    FDZA             Cash Management Line Items in Payment   KEYNO
                     Requests
    PAYRQ            Payment Requests                        KEYNO
    Hope this solves your concern...
    Regards,
    Habeeb
    Assign points if helpful

  • Migrate from taxinj to taxinn  what are the precautions to take, what will

    CAN ANYONE HELP ME, FOR MY CLINENT THE REQUIREMENT IS TO CHANGE FROM  EXISTING TAXINJ PROCEDURE TO TAXINN, THE ARE CHANGING FROM TAXINJ TO TAXINN  ..BCAZ IN THE CURRENT TAXINJ PROCEDURE THEY ARE MAINTAINING 300 TAX CODES FOR 20 COMPANYCODES AND 190 PLANTS, 250 LOCATIONS. 1230 ENDUSERS.
    SO THEY ARE COMING ACROSS DAY TO DAY PROBLEM.
    BOFORE GOING TO MIGRATE FROM TAXINJ TO TAXINN WHAT ARE THE PRECAUTIONS AND IMPLICATIONS NEED TO BE TAKE INTO CONSIDERATIONS.
    FROM BUSINESS POINT VIEW, WITHOUT DISTURBING MASTER DATA,
    CAN ANYONE HELP URGENT.
    THANKS IN ADVANCE
    CHEERS
    SRINIVAS.ANGAJALA

    hi GK
              When u delete the request in ODS or CUBE u want to delete the available data on those targets.
    For deleting the datas in the targets u just right click on the ods or cube and give delete data so that if there is a data in the target it gets deleted. this is the main step u want to do when u reconstrcut the request and after doing this u can direclty schedule data to the target.
    In targets after completing the data transfer process activate it to get that data in report also. dont forget to do this
    In psa level there is no problem u can go and directly delete the request.
    Assign points and gain some points in SDN
    cheers
    bye

  • What are the Calculations of Compression days in MRP?

    Hi,
    Can anyone please explain me with example that what are the calculations for compression days calculated for planned orders?? I am using MRP planning.
    Thanks and Regards
    Farhan

    HI,
    compression days are the days by which lead time of your buy items gets compressed to fulfill the demand due dates.
    For ex,  Finished Good is having Due date ( as scheduled ship date/request date in SO)   on 15th of Feb, but the buy item used in this assembly has lead time of 20 days.
    Consider the plan is running today i.e. on 6th Feb, so buy item will be actually available  on 6+20=26th of  Feb,
    But your demand due date is 15th Feb, which needs to be met, so planning engine will suggest an exception for this buy item as 'Item with compression days'
    by compressing its lead time as 26th Feb-15th Feb=11days as compression days.
    Please mark this post as correct or helpful it serves your purpose.
    Thanks,
    Avinash

  • What are the major process to transfer the data from legacy to sap system.

    What are the major process to transfer the data from legacy to sap system using BDC at Real Time only?

    hi,
    BATCH DATA COMMUNICATION
    main methods are:
    1. SESSION METHOD
    2. CALL TRANSACTION
    3. DIRECT INPUT
    Advantages offered by BATCH INPUT method:
    1. Can process large data volumes in batch.
    2. Can be planned and submitted in the background.
    3. No manual interaction is required when data is transferred.
    4. Data integrity is maintained as whatever data is transferred to the table is through transaction. Hence batch input data is submitted to all the checks and validations.
    To implement one of the supported data transfers, you must often write the program that exports the data from your non-SAP system. This program, known as a “data transfer” program must map the data from the external system into the data structure required by the SAP batch input program.
    The batch input program must build all of the input to execute the SAP transaction.
    Two main steps are required:
    • To build an internal table containing every screen and every field to be filled in during the execution of an SAP transaction.
    • To pass the table to SAP for processing.
    Prerequisite for Data Transfer Program
    Writing a Data Transfer Program involves following prerequisites:
    Analyzing data from local file
    Analyzing transaction
    Analyzing transaction involves following steps:
    • The transaction code, if you do not already know it.
    • Which fields require input i.e., mandatory.
    • Which fields can you allow to default to standard values.
    • The names, types, and lengths of the fields that are used by a transaction.
    • Screen number and Name of module pool program behind a particular transaction.
    To analyze a transaction::
    • Start the transaction by menu or by entering the transaction code in the command box.
    (You can determine the transaction name by choosing System – Status.)
    • Step through the transaction, entering the data will be required for processing your batch input data.
    • On each screen, note the program name and screen (dynpro) number.
    (dynpro = dyn + pro. Dyn = screen, pro = number)
    • Display these by choosing System – Status. The relevant fields are Program (dynpro) and Dynpro number. If pop-up windows occur during execution, you can get the program name and screen number by pressing F1 on any field or button on the screen.
    The technical info pop-up shows not only the field information but also the program and screen.
    • For each field, check box, and radio button on each screen, press F1 (help) and then choose Technical Info.
    Note the following information:
    - The field name for batch input, which you’ll find in its own box.
    - The length and data type of the field. You can display this information by double clicking on the Data Element field.
    • Find out the identification code for each function (button or menu) that you must execute to process the batch-input data (or to go to new screen).
    Place the cursor on the button or menu entry while holding down the left mouse button. Then press F1.
    In the pop-up window that follows, choose Technical info and note the code that is shown in the Function field.
    You can also run any function that is assigned to a function key by way of the function key number. To display the list of available function keys, click on the right mouse button. Note the key number that is assigned to the functions you want to run.
    Once you have program name, screen number, field name (screen field name), you can start writing.
    DATA TRANSFER program.
    Declaring internal table
    First Integral Table similar to structure like local file.
    Declaring internal table like BDCDATA
    The data from internal table is not transferred directly to database table, it has to go through transaction. You need to pass data to particular screen and to particular screen-field. Data is passed to transaction in particular format, hence there is a need for batch input structure.
    The batch input structure stores the data that is to be entered into SAP system and the actions that are necessary to process the data. The batch input structure is used by all of the batch input methods. You can use the same structure for all types of batch input, regardless of whether you are creating a session in the batch input queue or using CALL TRANSACTION.
    This structure is BDCDATA, which can contain the batch input data for only a single run of a transaction. The typical processing loop in a program is as follows:
    • Create a BDCDATA structure
    • Write the structure out to a session or process it with CALL TRANSACTION USING; and then
    • Create a BDCDATA structure for the next transaction that is to be processed.
    Within a BDCDATA structure, organize the data of screens in a transaction. Each screen that is processed in the course of a transaction must be identified with a BDCDATA record. This record uses the Program, Dynpro, and Dynbegin fields of the structure.
    The screen identifier record is followed by a separate BDCDATA record for each value, to be entered into a field. These records use the FNAM and FVAL fields of the BDCDATA structure. Values to be entered in a field can be any of the following:
    • Data that is entered into screen fields.
    • Function codes that are entered into the command field. Such function codes execute functions in a transaction, such as Save or Enter.
    The BDCDATA structure contains the following fields:
    • PROGRAM: Name of module pool program associated with the screen. Set this field only for the first record for the screen.
    • DYNPRO: Screen Number. Set this field only in the first record for the screen.
    • DYNBEGIN: Indicates the first record for the screen. Set this field to X, only for the first record for the screen. (Reset to ‘ ‘ (blank) for all other records.)
    • FNAM: Field Name. The FNAM field is not case-sensitive.
    • FVAL: Value for the field named in FNAM. The FVAL field is case-sensitive. Values assigned to this field are always padded on the right, if they are less than 132 characters. Values must be in character format.
    Transferring data from local file to internal table
    Data is uploaded to internal table by UPLOAD of WS_UPLOAD function.
    Population of BDCDATA
    For each record of internal table, you need to populate Internal table, which is similar to BDCDATA structure.
    All these five initial steps are necessary for any type of BDC interface.
    DATA TRANSFER program can call SESSION METHOD or CALL TRANSACTION. The initial steps for both the methods are same.
    First step for both the methods is to upload the data to internal table. From Internal Table, the data is transferred to database table by two ways i.e., Session method and Call transaction.
    SESSION METHOD
    About Session method
    In this method you transfer data from internal table to database table through sessions.
    In this method, an ABAP/4 program reads the external data that is to be entered in the SAP System and stores the data in session. A session stores the actions that are required to enter your data using normal SAP transaction i.e., Data is transferred to session which in turn transfers data to database table.
    Session is intermediate step between internal table and database table. Data along with its action is stored in session i.e., data for screen fields, to which screen it is passed, the program name behind it, and how the next screen is processed.
    When the program has finished generating the session, you can run the session to execute the SAP transactions in it. You can either explicitly start and monitor a session or have the session run in the background processing system.
    Unless session is processed, the data is not transferred to database table.
    BDC_OPEN_GROUP
    You create the session through program by BDC_OPEN_GROUP function.
    Parameters to this function are:
    • User Name: User name
    • Group: Name of the session
    • Lock Date: The date on which you want to process the session.
    • Keep: This parameter is passed as ‘X’ when you want to retain session after
    processing it or ‘ ‘ to delete it after processing.
    BDC_INSERT
    This function creates the session & data is transferred to Session.
    Parameters to this function are:
    • Tcode: Transaction Name
    • Dynprotab: BDC Data
    BDC_CLOSE_GROUP
    This function closes the BDC Group. No Parameters.
    Some additional information for session processing
    When the session is generated using the KEEP option within the BDC_OPEN_GROUP, the system always keeps the sessions in the queue, whether it has been processed successfully or not.
    However, if the session is processed, you have to delete it manually. When session processing is completed successfully while KEEP option was not set, it will be removed automatically from the session queue. Log is not removed for that session.
    If the batch-input session is terminated with errors, then it appears in the list of INCORRECT session and it can be processed again. To correct incorrect session, you can analyze the session. The Analysis function allows to determine which screen and value has produced the error. If you find small errors in data, you can correct them interactively, otherwise you need to modify batch input program, which has generated the session or many times even the data file.
    CALL TRANSACTION
    About CALL TRANSACTION
    A technique similar to SESSION method, while batch input is a two-step procedure, Call Transaction does both steps online, one after the other. In this method, you call a transaction from your program by
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>
    Messages into <MSGTAB>.
    Parameter – 1 is transaction code.
    Parameter – 2 is name of BDCTAB table.
    Parameter – 3 here you are specifying mode in which you execute transaction
    A is all screen mode. All the screen of transaction are displayed.
    N is no screen mode. No screen is displayed when you execute the transaction.
    E is error screen. Only those screens are displayed wherein you have error record.
    Parameter – 4 here you are specifying update type by which database table is updated.
    S is for Synchronous update in which if you change data of one table then all the related Tables gets updated. And sy-subrc is returned i.e., sy-subrc is returned for once and all.
    A is for Asynchronous update. When you change data of one table, the sy-subrc is returned. And then updating of other affected tables takes place. So if system fails to update other tables, still sy-subrc returned is 0 (i.e., when first table gets updated).
    Parameter – 5 when you update database table, operation is either successful or unsuccessful or operation is successful with some warning. These messages are stored in internal table, which you specify along with MESSAGE statement. This internal table should be declared like BDCMSGCOLL, a structure available in ABAP/4. It contains the following fields:
    1. Tcode: Transaction code
    2. Dyname: Batch point module name
    3. Dynumb: Batch input Dyn number
    4. Msgtyp: Batch input message type (A/E/W/I/S)
    5. Msgspra: Batch input Lang, id of message
    6. Msgid: Message id
    7. MsgvN: Message variables (N = 1 - 4)
    For each entry, which is updated in database, table message is available in BDCMSGCOLL. As BDCMSGCOLL is structure, you need to declare a internal table which can contain multiple records (unlike structure).
    Steps for CALL TRANSACTION method
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. UPLOAD or WS_UPLOAD function to upload the data from local file to itab. (Considering file is local file)
    4. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>.
    Refresh BDCTAB.
    Endloop.
    (To populate BDCTAB, You need to transfer each and every field)
    The major differences between Session method and Call transaction are as follows:
    SESSION METHOD CALL TRANSACTION
    1. Data is not updated in database table unless Session is processed. Immediate updation in database table.
    2. No sy-subrc is returned. Sy-subrc is returned.
    3. Error log is created for error records. Errors need to be handled explicitly
    4. Updation in database table is always synchronous Updation in database table can be synchronous Or Asynchronous.
    Error Handling in CALL TRANSACTION
    When Session Method updates the records in database table, error records are stored in the log file. In Call transaction there is no such log file available and error record is lost unless handled. Usually you need to give report of all the error records i.e., records which are not inserted or updated in the database table. This can be done by the following method:
    Steps for the error handling in CALL TRANSACTION
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. Internal table BDCMSG like BDCMSGCOLL
    4. Internal table similar to Ist internal table
    (Third and fourth steps are for error handling)
    5. UPLOAD or WS_UPLOAD function to upload the data from the local file to itab. (Considering file is local file)
    6. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tr.code> using <Bdctab>
    Mode <A/N/E>
    Update <S/A>
    Messages <BDCMSG>.
    Perform check.
    Refresh BDCTAB.
    Endloop.
    7 Form check.
    IF sy-subrc <> 0. (Call transaction returns the sy-subrc if updating is not successful).
    Call function Format_message.
    (This function is called to store the message given by system and to display it along with record)
    Append itab2.
    Display the record and message.
    DIRECT INPUT
    About Direct Input
    In contrast to batch input, this technique does not create sessions, but stores the data directly. It does not simulate the online transaction. To enter the data into the corresponding database tables directly, the system calls a number of function modules that execute any necessary checks. In case of errors, the direct input technique provides a restart mechanism. However, to be able to activate the restart mechanism, direct input programs must be executed in the background only. Direct input checks the data thoroughly and then updates the database directly.
    You can start a Direct Input program in two ways;
    Start the program directly
    This is the quickest way to see if the program works with your flat file. This option is possible with all direct input programs. If the program ends abnormally, you will not have any logs telling you what has or has not been posted. To minimize the chance of this happening, always use the check file option for the first run with your flat file. This allows you to detect format errors before transfer.
    Starting the program via the DI administration transaction
    This transaction restarts the processing, if the data transfer program aborts. Since DI document are immediately posted into the SAP D/B, the restart option prevents the duplicate document posting that occurs during a program restart (i.e., without adjusting your flat file).
    Direct input is usually done for standard data like material master, FI accounting document, SD sales order and Classification for which SAP has provided standard programs.
    First time you work with the Direct Input administration program, you will need to do some preparation before you can transfer data:
    - Create variant
    - Define job
    - Start job
    - Restart job
    Common batch input errors
    - The batch input BDCDATA structure tries to assign values to fields which do not exist in the current transaction screen.
    - The screen in the BDCDATA structure does not match the right sequence, or an intermediate screen is missing.
    - On exceptional occasions, the logic flow of batch input session does not exactly match that of manual online processing. Testing the sessions online can discover by this.
    - The BDCDATA structure contains fields, which are longer than the actual definition.
    - Authorization problems.
    RECORDING A BATCH INPUT
    A B recording allows you to record a R/3 transaction and generate a program that contains all screens and field information in the required BDC-DATA format.
    You can either use SHDB transaction for recording or
    SYSTEM ? SERVICES ? BATCH INPUT ? EDIT
    And from here click recording.
    Enter name for the recording.
    (Dates are optional)
    Click recording.
    Enter transaction code.
    Enter.
    Click Save button.
    You finally come to a screen where, you have all the information for each screen including BDC_OKCODE.
    • Click Get Transaction.
    • Return to BI.
    • Click overview.
    • Position the cursor on the just recorded entry and click generate program.
    • Enter program name.
    • Click enter
    The program is generated for the particular transaction.
    BACKGROUND PROCESSING
    Need for Background processing
    When a large volume of data is involved, usually all batch inputs are done in background.
    The R/3 system includes functions that allow users to work non-interactively or offline. The background processing systems handle these functions.
    Non-interactively means that instead of executing the ABAP/4 programs and waiting for an answer, user can submit those programs for execution at a more convenient planned time.
    There are several reasons to submit programs for background execution.
    • The maximum time allowed for online execution should not exceed 300 seconds. User gets TIMEOUT error and an aborted transaction, if time for execution exceeds 300 seconds. To avoid these types of error, you can submit jobs for background processing.
    • You can use the system while your program is executing.
    This does not mean that interactive or online work is not useful. Both type of processing have their own purposes. Online work is the most common one entering business data, displaying information, printing small reports, managing the system and so on. Background jobs are mainly used for the following tasks; to process large amount of data, to execute periodic jobs without human intervention, to run program at a more convenient, planned time other than during normal working hours i.e., Nights or weekends.
    The transaction for background processing is SM36.
    Or
    Tools ? Administration ? Jobs ? Define jobs
    Or
    System ? services ? Jobs
    Components of the background jobs
    A job in Background processing is a series of steps that can be scheduled and step is a program for background processing.
    • Job name. Define the name of assigned to the job. It identifies the job. You can specify up to 32 characters for the name.
    • Job class. Indicates the type of background processing priority assigned to the job.
    The job class determines the priority of a job. The background system admits three types of job classes: A B & C, which correspond to job priority.
    • Job steps. Parameters to be passed for this screen are as follows:
    Program name.
    Variant if it is report program
    Start criteria for the job: Option available for this are as follows:
    Immediate - allows you to start a job immediately.
    Date/Time - allows you to start a job at a specific name.
    After job - you can start a job after a particular job.
    After event - allows you to start a job after a particular event.
    At operation mode - allows you to start a job when the system switches to a particular operation mode.
    Defining Background jobs
    It is two step process: Firstly, you define the job and then release it.
    When users define a job and save it, they are actually scheduling the report i.e., specifying the job components, the steps, the start time.
    When users schedule program for background processing, they are instructing the system to execute an ABAP/4 report or an external program in the background. Scheduled jobs are not executed until they are released. When jobs are released, they are sent for execution to the background processing system at the specified start time. Both scheduling and releasing of jobs require authorizations.
    HANDLING OF POP UP SCREEN IN BDC
    Many times in transaction pop up screen appears and for this screen you don’t pass any record but some indication to system telling it to proceed further. For example: The following screen
    To handle such screen, system has provided a variable called BDC_CURSOR. You pass this variable to BDCDATA and process the screen.
    Usually such screen appears in many transactions, in this case you are just passing information, that YES you want to save the information, that means YES should be clicked. So you are transferring this information to BDCDATA i.e., field name of YES which is usually SPOT_OPTION. Instead of BDC_OKCODE, you are passing BDC_CURSOR.
    BDC_CURSOR is also used to place cursor on particular field.
    A simple transaction where you are entering customer number on first screen and on next screen data is displayed for the particular customer number. Field, which we are changing here, are name and city. When you click on save, the changed record gets saved.
    Prerequisite to write this BDC interface as indicated earlier is:
    1. To find screen number
    2. To find screen field names, type of the field and length of the field.
    3. To find BDC_OKCODE for each screen
    4. Create flat file.
    generally  Batch Input usually are used to transfer large amount of data. For example you are implementing a new SAP project, and of course you will need some data transfer from legacy system to SAP system.
    CALL TRANSACTION is used especially for integration actions between two SAP systems or between different modules. Users sometimes wish to do something like that click a button or an item then SAP would inserts or changes data automatically. Here CALL TRANSACTION should be considered.
    2. Transfer data for multiple transactions usually the Batch Input method is used.
    check these sites for step by step process:
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    Check these link:
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://www.sap-img.com/abap/question-about-bdc-program.htm
    http://www.itcserver.com/blog/2006/06/30/batch-input-vs-call-transaction/
    http://www.planetsap.com/bdc_main_page.htm
    call Transaction or session method ?
    null

  • What are the ODM components available in SNC?

    Hi experts,
    Could you please explain on
    1. what are the ODM components available with SNC an their use?
    2. Settings that required to activate/generate/regenerate ODM?
    Thanks in advance and eagerly looking for your great words.
    Regards
    Vinoth

    Hi Rohan
    The following are the order components available with SNC
    1HDR,1KNB,1NOT,1PAC,1PRA,1PRC,1PRT,1SCN 1SUB 1PRT 1SCN 1SUB 1TST 2IT2 2IT3 2NOT 2PRA 4NOT 4SDL 4TST 7MP1 7MP2 7MP3 7REF 7RQT BWHG BWIG BWLG DLH1 DLI1 DLI2 DLI3 DLI4 DLI5 DLI6 DLI8 DRIT DRSL FTD1 FTDH FTU1 FTUH HEXC HHU1 HHU2 HHU3 HHU4 HMSG HPRT HSTT HTRA ICAR IEXC IPRD ISTO ISTT IVH1 IVI1 IVI2 KNBR OECN OECP OEHD OEIL OEIR
    OEIT OESL PIHD PIND  PIOR RWAA RWAT RWCG  RWCP RWDS RWFA  RWGN RWHN RWIB RWIN RWIR RWIV RWLC RWNS RWOH RWOI RWOP RWP1 RWP2  RWPA RWPG RWPH RWPN RWPP RWPR  RWPV RWPW  RWRC RWRS  RWSN RWTS  SCNR
    TLBH
    Depending on the data you exchange between the two systems, you need to activate the necessary components.
    These components form the basic framework where the data in your order document gets stored in the respective components
    For example:A double-level order document, use the components 1HDR (header) and 2IT2 (item). For a triple-level order document, use 1HDR (header), 2IT3 (item) and 4SDL (schedule line).
    As mentioned by Nikhil and Sandeep, the above notes help you, in activating the necessary ODM's
    Kindly read the help document available is SAP IMG for further understanding.
    Best Regards
    Vinod

  • What are the settings for datasource and infopackage for flat file loading

    hI
    Im trying to load the data from flat file to DSO . can anyone tel me what are the settings for datasource and infopackage for flat file loading .
    pls let me know
    regards
    kumar

    Loading of transaction data in BI 7.0:step by step guide on how to load data from a flatfile into the BI 7 system
    Uploading of Transaction data
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( Transaction data )
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to create ODS( Data store object ) or Cube.
    • Specify name fro the ODS or cube and click create
    • From the template window select the required characteristics and key figures and drag and drop it into the DATA FIELD and KEY FIELDS
    • Click Activate.
    • Right click on ODS or Cube and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.
    4. Monitor
    Right Click data targets and select manage and in contents tab select contents to view the loaded data. There are two tables in ODS new table and active table to load data from new table to active table you have to activate after selecting the loaded data . Alternatively monitor icon can be used.
    Loading of master data in BI 7.0:
    For Uploading of master data in BI 7.0
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( master data attributes, text, hierarchies)
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to select Insert Characteristics as info provider
    • Select required info object ( Ex : Employee ID)
    • Under that info object select attributes
    • Right click on attributes and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.

  • What are the settings required to maintain consignment storgae location

    Hi, what are the setting required to maintain new consignmnet storgae location for a customer.
    Thank you.

    Here I am giving you over all idea of Consignment:
    Consignment goods are goods which are stored at the customer location but which are owned by your company. The customer is not obliged to pay for these goods until they remove them from consignment stock. Otherwise, the customer can usually return consignment goods which are not required.
    Consignment processing offers the participating partners several advantages. Customers store the consignment goods at their own warehouses. The customer can access the goods in the consignment warehouse at any time. They are only billed for the goods when they are removed from the warehouse and only for the actual quantity taken.
    consignment stock must be
    Managed separately from the rest of your stock so that you know exactly what stock is stored at the customer location
    Managed separately for each customer
    In inventory management, the consignment stock is managed as special stock in your inventory and is assigned to specific customers. This enables you to keep track of returnable packaging stock by customer.
    If the consignment stocks are not managed by the sold-to party but by a central office, you can use the partner function for special stock partner.
    Prerequisites
    Special Stock Partner
    The special stock partner has been defined for carrying out consignment stock processing by means of a third party rather than the customer. This means you can manage all special stock under one partner. It makes sense to use the special stock partner if your customer is using decentralized order processing but manages consignment stock centrally.
    Inventory Management Using the Special Stock Partner
    If you want to process your consignment goods using a special stock partner, proceed as follows:
    Create a customer master record for the special stock partner (account groups 0001 and DEBI are defined for this purpose in the standard system).
    Enter the special stock partner in the relevant customer master record on the partner screen using the partner function SB.
    When you create an order for this customer, the system automatically proposes the partner function SB in the document header and the document items. Consignment goods which have been entered in a consignment fill-up are always posted to the stock of the special stock partner when goods issue is carried out. You can also enter the partner function SB manually in the document header or in the relevant items if a special stock partner is required and is not proposed from the customer master record. If a special stock partner does not exist in the document header, inventory management is carried out using the sold-to party.
    Process Flow
    There are four main transactions for processing consignment stock in the R/3 System, all of which support separate management of stock:
    Creating a Consignment Fill-Up
    Creating Consignment Issue
    Creating a Consignment Pick-Up
    Displaying Consignment Returns
    Consignment fill-up
    Consignment fill up is used to supplement the customer’s consignment stock.
    Goods issue of the appropriate stock is posted from the unrestricted-use stock to consignment stock (special stock). The goods remain in the possession of the vendor.
    When you ship consignment stock to the customer, you record the transaction by creating a consignment fill-up order (order type KB). As a result, the system carries out the following actions:
    If special stock does not yet exist in your inventory for the customer (or special stock partner), the system creates it when goods issue is posted
    The relevant quantity is removed from regular inventory in your plant and is added to the special stock for the customer. The total valuated stock for the plant remains the same.
    The transaction is not relevant for pricing since the consignment stock remains the property of your company.
    Consignment pick-up
    Consignment issue enables the customer to take consignment goods from the special stock for their use or to sell.
    Consignment issue involves removing the goods from the special stock and making it the property of the customer.
    When the customer removes consignment stock to use or sell, you record the transaction in the system by creating a consignment issue order (order type KE). As a result, the system carries out the following actions:
    When goods issue is posted, the relevant quantity is deducted from both the customer’s special stock and your own total valuated stock.
    The transaction is relevant for pricing since the goods now become the property of the customer.
    Consignment issue
    Any consignment goods stored at the customer’s warehouse that haven’t been used can be reposted to your company’s warehouse with a consignment pick-up.
    If the customer returns consignment stock to you, you record the transaction in the system by creating a consignment pick-up order (order type KA). As a result, the system carries out the following actions:
    When goods issue is posted, the relevant quantity is deducted from the customer’s special stock and is added back into your regular stock at the plant where the goods are returned. Your total valuated stock remains the same since the returned stock was regarded as part of your own inventory even while it was at the customer’s premises.
    This transaction is not relevant for billing.
    Consignment return
    Consignment returns are used for when your customer wants to return goods to the consignment stock.
    If the customer wishes to claim on consignment goods which have already been issued, you can record this transaction by creating a consignment return order (order type KR). As a result, the system carries out the following actions:
    When goods issue is posted, the relevant quantity is added to the customer’s special stock at the plant where the goods are returned
    Since the ownership of the goods is passed from the customer back to your company, the transaction is relevant for billing. In this case, the customer receives a credit memo for the returned goods.
    Depending on the settings in Customizing for Sales, you can create consignment returns with direct reference to a consignment issue. The consignment issue appears in the document flow for the consignment return. This function is not supported in the standard version of the SAP R/3 System.
    When you create consignment returns in the standard version of the SAP R/3 System, the system automatically sets a billing block. To credit a return, you must first approve the request for a credit memo by removing the billing block in the return header. You can control the automatic setting of a billing block in Customizing for Sales.
    The following table provides an overview of the order types used in the standard system for processing consignments:
    Transaction
    Sales order type
    Consignment fill-up
    KB
    Consignment issue
    KE
    Consignment pick-up
    KA
    Consignment return
    KR
    Pricing and Availability Check
    Whether pricing is carried out for the individual transactions depends on the item category. The schedule line category controls the availability check and the transfer of requirements in each of the transactions. The schedule line category itself is determined by the item category and the MRP type of the material. The order types KE and KA check against the consignment stock for the customer involved and order type KB checks against your own plant stock. Requirement records for consignment stock are stored as individual requirements, independently of the setting in the material master

  • What are the different ways to upload file data to SAP? Please help

    Hi Experts,
       I have to transfer huge file data (few lakhs records) to SAP business transaction. What are the different ways to do the same? I have heard of BDC and LSMW. But are there any more options?
    Which option is best suited for huge file data?
    Is LSMW an old technology and SAP will not support it any more?
    Kindly answer my queries at the earliest.
    I will be greatful to you if you can help me.
    Thanks
    Gopal

    for uplodig data to non sap we have 2 methodes
    i) if u know bapi u will use lasm
    2) bdc
    but u mentioned so many records isthere
    best thing is u will uplode all record sto al11 using XI interface
    then u have to write bdc / lsmw  program
    beter to go for lsmw before that u will find bapi
    if u will unable to find bapi
    u have to create bapi and use it in lasmw
    ofter that u have schedule the lsmw program as a bockground
    then u have to create a job for it
    and release from sm 37
    then u have to moniter through bd87
    if u want to go through i will help u.
    if it is usefull to u pls give points
    Saimedha

  • What are the different transaction codes for PGI?

    1. What are the different transaction codes for PGI?
    2. What transaction code should I use for a Group Delivery?
    Regards,
    Nazim.

    Hi
    Go to VL02 and you can do the PGI.
    Regfarding Group Delivery, you can create a batch job and link the program to the batch job so that the job picks up all the orders and create deliveries.
    Creating a Background Job Using the Job Wizard
    •From the main menu select System>Services>Jobs>Define Job (transaction SM36).
    •Press the Job Wizard button and step through the screens as follows:
    •1:Continue
    •2: Enter a name for your job. Continue.
    •3: Continue with ABAP Program Step selected
    •4: Enter the ABAP Program Name and variant name.
    Press the Print Parameters button, specify the required printer and set the "Time of Print" to "Send to Print Spooler for now".
    Press green tick
    Continue
    •5: If more reports need to be added to the job tick the checkbox, press continue and repeat screens 3 and 4.
    •6: Select how the job is to scheduled (eg Date and Time)
    •7: Enter the required scheduling information ( eg Date and Time) and if required tick "Periodic Jobs"
    •8: If "Periodic Job" was selected select the frequency (eg Monthly)
    •9: Check the job details and press Complete
    Hope the above info helps.
    Reward if useful
    Regards

  • What are the User Exits for Sales Order creation process?

    Hi,
    what are the User Exits for Sales Order creation process? how can I find them?
    thanks in advance,
    will reward,
    Mindaugas

    Please check this info:
    User Exits In Sales Document Processing
    This IMG step describes additional installation-specific processing in sales document processing. In particular, the required INCLUDES and user exits are described.
    Involved program components
    System modifications for sales document processing affect different areas. Depending on the modification, you make the changes in the program components provided:
    MV45ATZZ
    For entering metadata for sales document processing. User-specific metadata must start with "ZZ".
    MV45AOZZ
    For entering additional installation-specific modules for sales document processing which are called up by the screen and run under PBO (Process Before Output) prior to output of the screen. The modules must start with "ZZ".
    MV45AIZZ
    For entering additional installation-specific modules for sales document processing. These are called up by the screen and run under PAI (Process After Input) after data input (for example, data validation). The modules must start with "ZZ".
    MV45AFZZ and MV45EFZ1
    For entering installation-specific FORM routines and for using user exits, which may be required and can be used if necessary. These program components are called up by the modules in MV45AOZZ or MV45AIZZ.
    User exits in the program MV45AFZZ
    The user exits which you can use for modifications in sales document processing are listed below.
    USEREXIT_DELETE_DOCUMENT
    This user exit can be used for deleting data which was stored in a separate table during sales document creation, for example, if the sales document is deleted.
    For example, if an additional table is filled with the name of the person in charge (ERNAM) during order entry, this data can also be deleted after the sales order has been deleted.
    The user exit is called up at the end of the FORM routine BELEG_LOESCHEN shortly before the routine BELEG_SICHERN.
    USEREXIT_FIELD_MODIFICATION
    This user exit can be used to modify the attributes of the screen fields.
    To do this, the screen fields are allocated to so-called modification groups 1 - 4 and can be edited together during a modification in ABAP. If a field has no field name, it cannot be allocated to a group.
    The usage of the field groups (modification group 1-4) is as follows:
    Modification group 1: Automatic modification with transaction MFAW
    Modification group 2: It contains 'LOO' for step loop fields
    Modification group 3: For modifications which depend on check tables or on other fixed information
    Modification group 4: is not used
    The FORM routine is called up for every field of a screen. If you require changes to be made, you must make them in this user exit.
    This FORM routine is called up by the module FELDAUSWAHL.
    See the Screen Painter manual for further information on structuring the interface.
    USEREXIT_MOVE_FIELD_TO_VBAK
    Use this user exit to assign values to new fields at sales document header level. It is described in the section "Transfer of the customer master fields into the sales document".
    The user exit is called up at the end of the FORM routine VBAK_FUELLEN.
    USEREXIT_MOVE_FIELD_TO_VBAP
    Use this user exit to assign values to new fields at sales document item level. It is described in the section "Copy customer master fields into the sales document".
    The user exit is called up at the end of the FORM routine VBAP_FUELLEN.
    USEREXIT_MOVE_FIELD_TO_VBEP
    Use this user exit to assign values to new fields at the level of the sales document schedule lines.
    The user exit is called up at the end of the FORM routine VBEP_FUELLEN.
    USEREXIT_MOVE_FIELD_TO_VBKD
    Use this user exit to assign values to new fields for business data of the sales document. It is described in the section "Copy customer master fields into sales document".
    The user exit is called up at the end of the FORM routine VBKD_FUELLEN.
    USEREXIT_NUMBER_RANGE
    Use this user exit to define the number ranges for internal document number assignment depending on the required fields. For example, if you want to define the number range depending on the sales organization (VKORG) or on the selling company (VKBUR), use this user exit.
    The user exit is called up in the FORM routine BELEG_SICHERN.
    USEREXIT_PRICING_PREPARE_TKOMK
    Use this user exit if you want to include and assign a value to an additional header field in the communication structure KOMK taken as a basis for pricing.
    USEREXIT_PRICING_PREPARE_TKOMP
    Use this user exit if you want to include or assign a value to an additional item field in the communication structure KOMP taken as a basis for pricing.
    USEREXIT_READ_DOCUMENT
    You use this user exit if further additional tables are to be read when importing TA01 or TA02.
    The user exit is called up at the end of the FORM routine BELEG_LESEN.
    USEREXIT_SAVE_DOCUMENT
    Use this user exit to fill user-specific statistics update tables.
    The user exit is called up by the FORM routine BELEG-SICHERN before the COMMIT command.
    Note
    If a standard field is changed, the field r185d-dataloss is set to X. The system queries this indicator at the beginning of the safety routine. This is why this indicator must also be set during the maintenance of user-specific tables that are also to be saved.
    USEREXIT_SAVE_DOCUMENT_PREPARE
    Use this user exit to make certain changes or checks immediately before saving a document. It is the last possibility for changing or checking a document before posting.
    The user exit is carried out at the beginning of the FORM routine BELEG_SICHERN.
    User exits in the program MV45AFZA
    USEREXIT_MOVE_FIELD_TO_KOMKD
    Use this user exit to include or assign values to additional header fields in the communication structure KOMKD taken as a basis for the material determination. This is described in detail in the section "New fields for material determination".
    USEREXIT_MOVE_FIELD_TO_KOMPD
    Use this user exit to include or assign values to additional item fields in the communication structure KOMPD taken as a basis for the material determination. This is described in detail in the section "New fields for material determination".
    USEREXIT_MOVE_FIELD_TO_KOMKG
    Use this user exit to include or assign values to additional fields in the communication structure KOMKG taken as a basis for material determination and material listing. This is described in detail in the section "New fields for listing/exclusion".
    USEREXIT_MOVE_FIELD_TO_KOMPG
    Use this user exit to include or assign values to additional fields in the communication structure KOMPG taken as a basis for material determination and material listung. This is described in detail in the section "New fields for listing/exclusion".
    USEREXIT_REFRESH_DOCUMENT
    With this user exit, you can reset certain customer-specific fields as soon as processing of a sales document is finished and before the following document is edited.
    For example, if the credit limit of the sold-to party is read during document processing, in each case it must be reset again before processing the next document so that the credit limit is not used for the sold-to party of the following document.
    The user exit is executed when a document is saved if you leave the processing of a document with F3 or F15.
    The user exit is called up at the end of the FORM routine BELEG_INITIALISIEREN.
    User-Exits in program MV45AFZB
    USEREXIT_CHECK_XVBAP_FOR_DELET
    In this user exit, you can enter additional data for deletion of an item. If the criteria are met, the item is not deleted (unlike in the standard system).
    USEREXIT_CHECK_XVBEP_FOR_DELET
    In this user exit, you can enter additional data for deletion of a schedule line. If the criteria are met, the schedule line is not deleted (unlike in the standard system).
    USEREXIT_CHECK_VBAK
    This user exit can be used to carry out additional checks (e.g. for completion) in the document header. The system could, for example, check whether certain shipping conditions are allowed for a particular customer group.
    USEREXIT_CHECK_VBAP
    This user exit can be used to carry out additional checks (e.g. for completion) at item level.
    USEREXIT_CHECK_VBKD
    The user exit can be used to carry out additional checks (e.g. for completion) on the business data in the order.
    USEREXIT_CHECK_VBEP
    This user exit can be use to carry out additional checks (e.g. for completion) on the schedule line. During BOM explosion, for example, you may want certain fields to be copied from the main item to the sub-items (as for billing block in the standard system).
    USEREXIT_CHECK_VBSN
    You can use this user exit to carry out additional checks (e.g. for completion) on the serial number.
    USEREXIT_CHECK_XVBSN_FOR_DELET In this user exit, you can enter additional criteria for deletion of the serial number. If the criteria are met, the serial number is not deleted (unlike in the standard system).
    USEREXIT_FILL_VBAP_FROM_HVBAP
    You can use this user exit to fill additional fields in the sub-item with data from the main item.
    USEREXIT_MOVE_FIELD_TO_TVCOM_H
    You can use this user exit to influence text determination for header texts. For example, you can include new fields for text determination or fill fields that already exist with a new value.
    USEREXIT_MOVE_FIELD_TO_TVCOM_I
    You can use this user exit to influence text determination for item texts. For example, you can include new fields for text determination or fill fields that already exist with a new value.
    User-Exits for product allocation:
    The following user exits all apply to structure COBL, in which the data for account determination is copied to item level.
    USEREXIT_MOVE_FIELD_TO_COBL
    Option to include new fields in structure COBL.
    USEREXIT_COBL_RECEIVE_VBAK
    Option to assign values from the document header to the new fields.
    USEREXIT_COBL_RECEIVE_VBAP
    Option to supply values from the item to the new fields.
    USEREXIT_COBL_SEND_ITEM
    A changed field can be copied from the structure into the item. You could use the user exit to display a certain field in the account assignment block (see also MV45AFZB).
    USEREXIT_COBL_SEND_HEADER
    A changed field can be copied from the structure to the header (see source text MV45AFZB)
    USEREXIT_SOURCE_DETERMINATION
    You can use this user exit to determine which plant will be used for the delivery. In the standard system, the delivering plant is copied from the customer master or the customer-material info record. If you want to use a different rule, then you must enter it in this user exit.
    USEREXIT_MOVE_FIELD_TO_ME_REQ
    With this user exit you can include additional fields for the following fields:
    EBAN (purchase requisition)
    EBKN (purchase requisition-account assignment)
    USEREXIT_GET_FIELD_FROM_SDCOM
    Option to include new fields for the variant configuration. Fields that are included in structure SDCOM can be processed and then returned to the order.
    USEREXIT_MOVE_WORKAREA_TO_SDWA
    You can use this user exit to format additional work areas for the variant configuration. You will find notes on the user exit in MV45AFZB.
    User-Exits for first data transfer:
    The following user exits can only be used for the first data transfer.
    Note
    Only use the user exits if the names/fields do NOT have the same name.
    USEREXIT_MOVE_FIELD_TO_VBAKKOM
    Option to include additional fields in structure VBAKKOM (communiction fields for maintaining the sales document header)
    USEREXIT_MOVE_FIELD_TO_VBAPKOM
    Option to include additional fields in structure VBAPKOM (communication fields for maintaining a sales item)
    USEREXIT_MOVE_FIELD_TO_VBEPKOM
    Option to include additional fields in structure VBEPKOM (communication fields for maintaining a sales document schedule line)
    USEREXIT_MOVE_FIELD_TO_VBSN
    You can use this user exit to include fields in structure VBSN (scheduling agreement-related change status).
    USEREXIT_MOVE_FIELD_TO_KOMKH
    You can use this user exit to include new fields for batch determination (document header).
    USEREXIT_MOVE_FIELD_TO_KOMPH
    You can use this user exit to include new fields for batch determination (document item).
    USEREXIT_CUST_MATERIAL_READ
    You can use this user exit to set another customer number in the customer material info record (e.g. with a customer hierarchy)
    USEREXIT_NEW_PRICING_VBAP
    Option for entry of preconditions for carrying out pricing again (e.g. changes made to a certain item field could be used as the precondition for pricing to be carried out again). Further information in MV45AFZB.
    USEREXIT_NEW_PRICING_VBKD
    Option for entry of preconditions for carrying out pricing again (e.g. changes to the customer group or price group could be set as the preconditions for the system to carry out pricing again). Further information in MV45AFZB.
    User-Exits in Program MV45AFZD
    USEREXIT_CONFIG_DATE_EXPLOSION
    The BOM is exploded in the order with the entry date. You can use this user exit to determine which data should be used to explode the BOM (explosion with required delivery date, for example).
    User exits in the program FV45EFZ1
    USEREXIT_CHANGE_SALES_ORDER
    In the standard SAP R/3 System, the quantity and confirmed date of the sales document schedule line is changed automatically if a purchase requisition is allocated, and it or the sales document is changed (for example, quantity, date).
    If you want to change this configuration in the standard system, you can define certain requirements in order to protect your sales orders from being changed automatically. Use this user exit for this purpose. Decide at this point whether the schedule lines are to be changed.
    User-Exits in Program RV45PFZA
    USEREXIT_SET_STATUS_VBUK
    In this user exit you can you can store a specification for the reserve fields in VBUK (header status). Reserve field UVK01 could, for example, be used for an additional order status (as for rejections status, etc.).
    The following workareas are available for this user exit:
    VBUK (header status)
    FXVBUP (item status)
    FXVBUV (Incompletion)
    USEREXIT_SET_STATUS_VBUP
    In this user exit you can you can store a specification for the reserve fields for VBUP (item status).
    The following workareas are available for this user exit:
    FXVBAP (Item data)
    FXVBAPF (Dynamic part of order item flow)
    FXVBUV (Incompletion)
    USEREXIT_STATUS_VBUK_INVOICE
    You can use this user exit to influence billing status at header level.
    User exits in the screens
    Additional header data is on screen SAPMV45A 0309, additional item data on screen SAPMV45A 0459. These screens contain the Include screens SAPMV45A 8309 or SAPMV45A 8459 as user exits.
    Fields which are also to be included in the sales document for a specific installation should be included on the Include screens for maintaining. If an application-specific check module is needed for the fields, this can be included in the Include MV45AIZZ. The module is called up in the processing logic of the Include screens.
    For field transports, you do not have to make changes or adjustments.
    Example
    A new field, VBAK-ZZKUN, should be included in table VBAK.
    If the check is defined via the Dictionary (fixed values or check table) the field must be included with the fullscreen editor in the Include screen SAPMV45A 8309. In this case, no change has to be made to the processing logic.
    User Exits in Program MV45AFZ4
    USEREXIT_MOVE_FIELD_TO_KOMK
    You can use this user exit to add or edit additional header fields in the communication structure - KOMK- for free goods determination. For more information, see the New Fields for Free Goods Determination IMG activity.
    USEREXIT_MOVE_FIELD_TO_KOMP
    You can use this user exit to add or edit additional item fields in the communication structure KOMP for free goods determination. For more information see the New Fields for Free Goods Determination IMG activity.
    User Exits in the SAPFV45PF0E and SAPFV45PF0C Programs
    EXIT_SAPFV45P_001
    You can use this user exit to decide whether intercompany billing data is used in the profitability segment for cross-company code sales, or whether the data comes from external billing (external customer, sales data from the selling company code.
    Regards
    Eswar

Maybe you are looking for