Process of Querying the Backend similar to Toad

Hi
I would like to  know the process of querying the oracle database similar to Toad in fusion environment.
Could you please show some pointers .
Thanks
Santosh 

Hi.
Assuming this is part of your access (i.e. not SaaS), then I'd recommend SQL Developer.
Read more here
Kind regards
Richard
FA Dev Relations

Similar Messages

  • Processed Dimension IDs not saving in the backend server

    Hi Experts,
    After successfully processing dimensions, the newly added dimension IDs or the changed dimensions are not reflecting in the backend BI server. And naturally not in the Office client.
    I have checked logs in SLG1, couldnt find issue. Checked dumps, nothing particular found.
    Checked the .NET server as well, no problems.
    We are on SAP BPC 7.5 NW SP09.
    Any idea why this could be happening?
    Best Regards,
    Sayan

    Hi Sabine,
    I checked the log files earlier itself on the server but they didnot help. Last updates in the files were from November.
    What i have found today is, after i tried to process the dimensions for mutiple times, they finally got processed in the backend as well. I was checking in SM66, every time processes were triggered for BPC_ADMIN in the backend, when i started processing.
    But when i went to SLG1, there were entries with grey status (not processed) for same master data object and subobject 'master data update'. Since they are dialog process i could not any further entry from where i could find why they were not processing.
    Any idea? Thanks again for your patience with the responses
    Best Regards,
    Sayan

  • How to find the backend  SQL query of the JSP page in OIC

    Does anybody how the best way to find the backend SQL QUERY of OIV JSP page?

    How To Generate Trace Files in in HTML/JSP (using Profile Option)
    •     • Note: This requires proper responsibility to set SQL Initialization statement using Profile option.      
         Step 1.     Login to the desired Form application.     
         Step 2.     Select +Profile >> System ('Find System Profile Values' screen will pop up)     
         Step 3.     Check 'User' and Type in the Username (in which the account for that user will be trace)     
         Step 4.     Type 'Initialization%' in the Profile box and Hit 'Find' (Click here for preview.)     
         Step 5.     In the User box, type the following statement and Hit 'Save' (Click here for preview)
         BEGIN FND_CTL.FND_SESS_CTL('','','TRUE','TRUE','','ALTER SESSION SET TRACEFILE_IDENTIFIER = TESTING MAX_DUMP_FILE_SIZE = 5000000 EVENTS ='||''''||' 10046 TRACE NAME CONTEXT FOREVER, LEVEL 12'||'''');END;     
         Note:     specify any name you like to identify your trace, in this case, testing is the end name on the trace. You can also specify the amount of data allowable to be in the trace, in this case, 5000000 is the amount set. Make sure you hit 'Save' afterwards.[Quotes in the statement are all 'Single' quotes.]
              specifying TRACEFILE_IDENTIFIER value is mandatory when setting up the trace using the above profile option value
         Step 6.     Login to HTML / JSP page with username/password and start your flow. (Everything you do once login to HTML / JSP will get trace.)     
         Step 7.     Logout of HTML / JSP application once you completed with your flow.      
         Step 8.     Go back to the Profile option in the Form application and delete the Initialization SQL statement, and Hit 'Save'.     
         Step 9.     Log in to the database server or login server and retrieve your trace file.
         Identify and retrieve the trace file using the tracefile_identifier specified in Step 5.
         In this case the tracefile_identifier is “TESTING”. (Click here for Trace file locations) *     
         Note:     If you need to regenerate your trace or tracing a new flow, then repeat Step 1 to Step 8. To avoid self-confusion, choose a different name for your trace identifier everytime you set to trace.     
         Step 10.     See TKPROF section on how to format trace file into readable text.
         Trace Options Definition
         No Trace          Tracing is not activated
         Activities will not get traced.
         Regular Trace
         (Level 1)          Contains SQL, execution statistics, and execution plan.
         Provides execution path, row counts as well as produces smallest flat file.
         Trace with Binds
         (Level 4)          Regular Trace plus value supplied to SQL statement via local variables.
         Trace with Waits
         (Level 8)          Regular Trace plus database operation timings that the SQL waited to have done in order to complete, i.e. disk access.
         Trace with Binds and Waits
         (Level 12)          Regular trace with both waits and binds information.
         Contains the most complete information and will produce the largest trace file.
    ****Send me an email to [email protected],I will share the document with you.

  • Any way to find out the backend still in Process for the record after saved

    once after we create a record and save here we are using many procs and functions from backend, after the record saved is there any way we will know still the backend is still in process for that record to insert into concerned tables for that module?
    we are getting an error after the first record is saved for some reason the form is not getting clearing after the record is saved there is clear_form code is also there in my save button.
    this is my prod problem please do help me out..
    thanks in advance..

    The Commit will not return control to the form until it is finished. So unless you are starting asynchronous operations with things like DBMS_JOB it is not going to be a timing problem.

  • How to limit the no. of the backend's DIA processes with XI proxies?

    When connecting a SAP backend to XI using a RFC adapter,  the number of  the backend's dialog processes is easily limited in the RFC adapter configuration by setting the 'maximum connections' parameter. Is there any obvious way to limit the backend's DIA processes when using the XI proxy technique, instead? Thanks in advance.

    If you are not using the rfc adapter..then the default connection pool is maintained.
    Regards
    Ravi Raman

  • Is there any way to prevent the OS from querying the Superdrive when start

    Just a random question. Whenever my MacBook Pro starts up (either from sleep or a complete/fresh start, the OS queries the drive slot to see if there is a disc present. While this may be normal behavior, it seems to slow down the start up process. (I rarely ever have a disc in the drive). Just curious if this, in fact, normal, or if there's something awry.

    Is there any way to prevent the OS from querying the Superdrive when start
    No
    he OS queries the drive slot to see if there is a disc present.
    How can you tell? Based on the noise it makes?

  • Error while reading the PO in the Backend system. Inform system admin

    Hi All,
    We are having a peculiar issue of 'Error while reading the PO in the Backend system. Inform system admin'.
    The P.O is in ordered status in SRM but the same is not getting transferred to backend ECC system.
    No error messages or logs in RZ20, SLG1 any where.
    All programmes like BBP_GET_STATUS_2 and CLEAN_REQREQ_UP are running fine.
    Tried pushing the P.Os manually using function module (BBP_PD_PO_TRANSFER_EXEC_V2) to backend ECC.
    It was working fine till a week ago and suddenly this problem is coming.
    We had implemented few OSS notes suggested by SAP for the issue of 'shopping carts appearing in sourcing cockpit even after P.O creation' in both development and test system.
    Now this issue is coming up in test system where as development system is working fine.
    Please let us know where to look and how to resolve this issue.
    A quick response would be highly appreciated.
    Regards,
    Teja

    I am facing the same issue with one PO in the Production system.
    SRM 5.0 , R/3 4.6C Extended classic scenario.
    I checked the status of other PO's created today. I see them in R/3. There is one PO which was created a week back which shows up as "ordered" in SRM but the PO is missing in R/3. When clicked on the details on the web, system throws the error
    Error while reading the PO in the Backend system. Inform system admin.
    Message no. BBP_CF010
    I checked RZ20, SLG1 no errors were found. I checked RFC connection, it was working fine too.
    I tried pushing the PO using the FM BBP_PD_PO_TRANSFER_EXEC, it did not solve the problem.
    In SRM WEBGUI Process PO - Item data -->follow on documents --> PO status is shown as Archived.
    Any inputs would be greatly appreciated. Please throw some light on this issue.
    Krishna

  • Does each SAP transaction iview create a log on to the backend ECC?

    Hi,
    If I put multiple SAP transaction iviews in a role, each pointing to a different SAP transaction in the same backend system, does the launch of each iview create a new logon in the backend system when using SAP logon tickets?  If so, is it possible to make it only launch a new session if one logon has already been created?
    Regards
    Cindy

    Hello Cindy,
    launching of diffrent Transactio Iview form same back end is quite similar as loged in to backend with Proper credentials and opening diffrent transaction's by useing
    /oxx(Transaction Name)
    Hope it will add to ur knowladge.
    Regards.
    Soni Vinit

  • Columns are ordered differently at DSO and at the backend database. Why?

    Hello,
    I've created a standard DSO with key and data fields.
    If I watch this table in SE16 I see it in one order but when I go to the backend database (Oracle in our case) directly and query this table I see different columns order.
    Any idea why?

    Anyone...?

  • Matrix Report 's Query Output is not as Toad...!!!!!!

    Hello,
    i have the following Query
    SELECT ALL
    FINANCIAL_YEAR_MONTHS.month_desc||' '|| substr( FINANCIAL_YEAR_MONTHS.month_code,4,4)  months,
    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE,
    CDE_MATERIALS.MATERIAL_DESC, CDE_MATERIALS.PARENT_MATERIAL,
    PLN_PLAN_DISTRIBUTION_WAY.EXPECTED_QUANTITY,
    PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE,
    PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE
    FROM
    PLN_PLAN_DISTRIBUTION_WAY,
    PLN_PLAN_PRODUCTS,
    CDE_MATERIALS,
    FINANCIAL_YEAR_MONTHS
    WHERE
    (PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = PLN_PLAN_PRODUCTS.MATERIAL_CODE)
    AND (    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = CDE_MATERIALS.MATERIAL_CODE   )
    AND (PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE = PLN_PLAN_PRODUCTS.GROUP_CODE)
    AND ( PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE= PLN_PLAN_PRODUCTS.FIN_YEAR_CODE )
    AND (  PLN_PLAN_DISTRIBUTION_WAY.MONTH_CODE = 
    to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2)))
    AND( to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2)) >= 7 )
    AND (PLN_PLAN_DISTRIBUTION_WAY.DISTRIBUTION_WAY = :P_DISTRIBUTION_WAY )
    AND (PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE = :P_FIN_YEAR_CODE)
    UNION ALL
    SELECT ALL
    FINANCIAL_YEAR_MONTHS.month_desc||' '|| substr( FINANCIAL_YEAR_MONTHS.month_code,4,4)  months,
    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE,
    CDE_MATERIALS.MATERIAL_DESC, CDE_MATERIALS.PARENT_MATERIAL,
    PLN_PLAN_DISTRIBUTION_WAY.EXPECTED_QUANTITY,
    PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE,
    PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE
    FROM
    PLN_PLAN_DISTRIBUTION_WAY,
    PLN_PLAN_PRODUCTS,
    CDE_MATERIALS,
    FINANCIAL_YEAR_MONTHS
    WHERE
    (PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = PLN_PLAN_PRODUCTS.MATERIAL_CODE)
    AND (    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = CDE_MATERIALS.MATERIAL_CODE   )
    AND (PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE = PLN_PLAN_PRODUCTS.GROUP_CODE)
    AND ( PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE= PLN_PLAN_PRODUCTS.FIN_YEAR_CODE )
    AND (  PLN_PLAN_DISTRIBUTION_WAY.MONTH_CODE = 
    to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2)))
    AND to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2))<= 6 
    AND (PLN_PLAN_DISTRIBUTION_WAY.DISTRIBUTION_WAY = :P_DISTRIBUTION_WAY )
    AND (PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE = :P_FIN_YEAR_CODE) the Report is a Matrix Report with Group Material_names as Rows and financial Months as Columns and the expected_Quantity as cells of that report matrix.
    Well,the query runs just fine in Toad as
    Months Material_code Material_Desc ..etc.
    July 2007 1 abc
    July 2007 2 def ...etc.
    June 2008 1 abc ..etc.
    June 2008 2 def ...etc.
    But in a report the output of months are a mess not ordered as a fin_year months..
    i tried to change the design of the report from Matrix group to Matrix got same result...
    Any Suggestions pls.
    Regards,
    Abdetu..

    This the Query with order by which doesn't return the required order..
    SELECT ALL
    FINANCIAL_YEAR_MONTHS.month_desc||' '|| substr( FINANCIAL_YEAR_MONTHS.month_code,4,4)  months,
    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE,
    CDE_MATERIALS.MATERIAL_DESC, CDE_MATERIALS.PARENT_MATERIAL,
    PLN_PLAN_DISTRIBUTION_WAY.EXPECTED_QUANTITY,
    PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE,
    PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE
    FROM
    PLN_PLAN_DISTRIBUTION_WAY,
    PLN_PLAN_PRODUCTS,
    CDE_MATERIALS,
    FINANCIAL_YEAR_MONTHS
    WHERE
    (PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = PLN_PLAN_PRODUCTS.MATERIAL_CODE)
    AND (    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = CDE_MATERIALS.MATERIAL_CODE   )
    AND (PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE = PLN_PLAN_PRODUCTS.GROUP_CODE)
    AND ( PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE= PLN_PLAN_PRODUCTS.FIN_YEAR_CODE )
    AND (  PLN_PLAN_DISTRIBUTION_WAY.MONTH_CODE = 
    to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2)))
    AND( to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2)) >= 7 )
    AND (PLN_PLAN_DISTRIBUTION_WAY.DISTRIBUTION_WAY = :P_DISTRIBUTION_WAY )
    AND (PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE = :P_FIN_YEAR_CODE)
    UNION ALL
    SELECT ALL
    FINANCIAL_YEAR_MONTHS.month_desc||' '|| substr( FINANCIAL_YEAR_MONTHS.month_code,4,4)  months,
    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE,
    CDE_MATERIALS.MATERIAL_DESC, CDE_MATERIALS.PARENT_MATERIAL,
    PLN_PLAN_DISTRIBUTION_WAY.EXPECTED_QUANTITY,
    PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE,
    PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE
    FROM
    PLN_PLAN_DISTRIBUTION_WAY,
    PLN_PLAN_PRODUCTS,
    CDE_MATERIALS,
    FINANCIAL_YEAR_MONTHS
    WHERE
    (PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = PLN_PLAN_PRODUCTS.MATERIAL_CODE)
    AND (    PLN_PLAN_DISTRIBUTION_WAY.MATERIAL_CODE = CDE_MATERIALS.MATERIAL_CODE   )
    AND (PLN_PLAN_DISTRIBUTION_WAY.GROUP_CODE = PLN_PLAN_PRODUCTS.GROUP_CODE)
    AND ( PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE= PLN_PLAN_PRODUCTS.FIN_YEAR_CODE )
    AND (  PLN_PLAN_DISTRIBUTION_WAY.MONTH_CODE = 
    to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2)))
    AND to_number(substr( FINANCIAL_YEAR_MONTHS.month_code,1,2))<= 6 
    AND (PLN_PLAN_DISTRIBUTION_WAY.DISTRIBUTION_WAY = :P_DISTRIBUTION_WAY )
    AND (PLN_PLAN_DISTRIBUTION_WAY.FIN_YEAR_CODE = :P_FIN_YEAR_CODE)
    ORDER BY months

  • Java.sql.SqlRecoverableException error while querying the cube

    Hi
    I get the following error when I try to query the cube from a Java program
    java.sql.SqlRecoverableException: no more data to read from the socket.
    This error occurs sometimes and sometimes it does not occur.
    We observed that if many people simultaneously try to query the cube the error crops up.
    Is this a bug or is there a solution or a method to deal with this sort of situation

    That sounds like the connection to the database is not there, or was dropped. Check that you are not maxed out on the number of processes.

  • How to find out number of background work processe available in the system?

    Hi All,
    I have a FM that triggers a background job using JOB_OPEN, ABAP_SUBMIT and JOB_CLOSE. But the problem is if there are no background work processes available at that moment, it does not return error. Not sure whether the job is even scheduled!
    Is there a way to find out if there are any background work processes available in the system? I know a FM SPBT_INITIALIZE which gives info about dialog work processes. Is there any other FM similar to this which gives info about background wp??
    Regards,
    Shailesh

    your Job will remain in a queue by message handler. once workprocess is free, your job gets scheduled
    by the way, you can use the functionality used in FM TH_DISPLAY_WORKPROCESS_LIST to read the workprocesses
    the command which gets you the workprocess is:
    CALL 'ThWpInfo' ID 'OPCODE' FIELD OPCODE_WP_LIST
                      ID 'TAB' FIELD LIST_STATE_WPLST-TABL
                      ID 'CPU' FIELD LIST_STATE_WPLST-CPU.
    check how its written in include LSDEBF01, form SET_LIST_WPLST
    Edited by: Soumyaprakash Mishra on Dec 12, 2011 12:17 PM

  • Goods Receipt not transferred to the Backend System R/3

    We are currently using BBPCRM4.0 & SAP R/3 4.7 (extended scenario)
    For a particular Purchase Order which has a material number, We cannot transfer GR to the Backend System R/3 successfully.(The shopping cart of unsuccessfully transferred GR is created by using material)
    When I check R/3 using WE05, I can find the message "The plant data of the material G09002202 is locked by the user MRP_RFC". This must be the reason why GR is not transferred successfully to the Backend System, but I do not know what this message mean.(As you see, MRP_RFC is the user name that connects SRM and R/3. + That moment the material is not locked by any user in R/3)
    If we try to create the GR document again in SRM, the message "Item is already being processed" is shown, so cannot make BUS2203 Docs.
    To sort out this, we remove the GR Transfer Error Information in
    BBP_DOCUMENT_TAB.
    After removing the information of this error in BBP_DOCUMENT_TAB, surprisingly we can make GR successfully.
    Weird!.
    Only for a particular PO which has a material, but not everytime.
    Sometimes this problem happens.
    At the first time, it doesn't work, but after removing the data from BBP_DOCUMENT_TAB, when we try GR again it works.
    Could you help me why this symptom happends..?
    Edited by: Hyangsook Moon on Feb 28, 2008 6:44 AM

    Hi Hyangsook Moon,
    The system is allowing  you to create Confirmation again in SRM because the status is not yet updated in SRM since the G.R is not created in the backend successfully.
    Since the confirmation is created in SRM and could not able to create G.R in R/3 through IDOC the system is giving you message as " Document is being processed".
    So the system is allowing you to create confirmation again once you delete the table entry.
    Hope this makes you more clear.Clarifications are welcome.
    Award points for helpful answers.
    Rgds,
    Teja

  • What are the major process to transfer the data from legacy to sap system.

    What are the major process to transfer the data from legacy to sap system using BDC at Real Time only?

    hi,
    BATCH DATA COMMUNICATION
    main methods are:
    1. SESSION METHOD
    2. CALL TRANSACTION
    3. DIRECT INPUT
    Advantages offered by BATCH INPUT method:
    1. Can process large data volumes in batch.
    2. Can be planned and submitted in the background.
    3. No manual interaction is required when data is transferred.
    4. Data integrity is maintained as whatever data is transferred to the table is through transaction. Hence batch input data is submitted to all the checks and validations.
    To implement one of the supported data transfers, you must often write the program that exports the data from your non-SAP system. This program, known as a “data transfer” program must map the data from the external system into the data structure required by the SAP batch input program.
    The batch input program must build all of the input to execute the SAP transaction.
    Two main steps are required:
    • To build an internal table containing every screen and every field to be filled in during the execution of an SAP transaction.
    • To pass the table to SAP for processing.
    Prerequisite for Data Transfer Program
    Writing a Data Transfer Program involves following prerequisites:
    Analyzing data from local file
    Analyzing transaction
    Analyzing transaction involves following steps:
    • The transaction code, if you do not already know it.
    • Which fields require input i.e., mandatory.
    • Which fields can you allow to default to standard values.
    • The names, types, and lengths of the fields that are used by a transaction.
    • Screen number and Name of module pool program behind a particular transaction.
    To analyze a transaction::
    • Start the transaction by menu or by entering the transaction code in the command box.
    (You can determine the transaction name by choosing System – Status.)
    • Step through the transaction, entering the data will be required for processing your batch input data.
    • On each screen, note the program name and screen (dynpro) number.
    (dynpro = dyn + pro. Dyn = screen, pro = number)
    • Display these by choosing System – Status. The relevant fields are Program (dynpro) and Dynpro number. If pop-up windows occur during execution, you can get the program name and screen number by pressing F1 on any field or button on the screen.
    The technical info pop-up shows not only the field information but also the program and screen.
    • For each field, check box, and radio button on each screen, press F1 (help) and then choose Technical Info.
    Note the following information:
    - The field name for batch input, which you’ll find in its own box.
    - The length and data type of the field. You can display this information by double clicking on the Data Element field.
    • Find out the identification code for each function (button or menu) that you must execute to process the batch-input data (or to go to new screen).
    Place the cursor on the button or menu entry while holding down the left mouse button. Then press F1.
    In the pop-up window that follows, choose Technical info and note the code that is shown in the Function field.
    You can also run any function that is assigned to a function key by way of the function key number. To display the list of available function keys, click on the right mouse button. Note the key number that is assigned to the functions you want to run.
    Once you have program name, screen number, field name (screen field name), you can start writing.
    DATA TRANSFER program.
    Declaring internal table
    First Integral Table similar to structure like local file.
    Declaring internal table like BDCDATA
    The data from internal table is not transferred directly to database table, it has to go through transaction. You need to pass data to particular screen and to particular screen-field. Data is passed to transaction in particular format, hence there is a need for batch input structure.
    The batch input structure stores the data that is to be entered into SAP system and the actions that are necessary to process the data. The batch input structure is used by all of the batch input methods. You can use the same structure for all types of batch input, regardless of whether you are creating a session in the batch input queue or using CALL TRANSACTION.
    This structure is BDCDATA, which can contain the batch input data for only a single run of a transaction. The typical processing loop in a program is as follows:
    • Create a BDCDATA structure
    • Write the structure out to a session or process it with CALL TRANSACTION USING; and then
    • Create a BDCDATA structure for the next transaction that is to be processed.
    Within a BDCDATA structure, organize the data of screens in a transaction. Each screen that is processed in the course of a transaction must be identified with a BDCDATA record. This record uses the Program, Dynpro, and Dynbegin fields of the structure.
    The screen identifier record is followed by a separate BDCDATA record for each value, to be entered into a field. These records use the FNAM and FVAL fields of the BDCDATA structure. Values to be entered in a field can be any of the following:
    • Data that is entered into screen fields.
    • Function codes that are entered into the command field. Such function codes execute functions in a transaction, such as Save or Enter.
    The BDCDATA structure contains the following fields:
    • PROGRAM: Name of module pool program associated with the screen. Set this field only for the first record for the screen.
    • DYNPRO: Screen Number. Set this field only in the first record for the screen.
    • DYNBEGIN: Indicates the first record for the screen. Set this field to X, only for the first record for the screen. (Reset to ‘ ‘ (blank) for all other records.)
    • FNAM: Field Name. The FNAM field is not case-sensitive.
    • FVAL: Value for the field named in FNAM. The FVAL field is case-sensitive. Values assigned to this field are always padded on the right, if they are less than 132 characters. Values must be in character format.
    Transferring data from local file to internal table
    Data is uploaded to internal table by UPLOAD of WS_UPLOAD function.
    Population of BDCDATA
    For each record of internal table, you need to populate Internal table, which is similar to BDCDATA structure.
    All these five initial steps are necessary for any type of BDC interface.
    DATA TRANSFER program can call SESSION METHOD or CALL TRANSACTION. The initial steps for both the methods are same.
    First step for both the methods is to upload the data to internal table. From Internal Table, the data is transferred to database table by two ways i.e., Session method and Call transaction.
    SESSION METHOD
    About Session method
    In this method you transfer data from internal table to database table through sessions.
    In this method, an ABAP/4 program reads the external data that is to be entered in the SAP System and stores the data in session. A session stores the actions that are required to enter your data using normal SAP transaction i.e., Data is transferred to session which in turn transfers data to database table.
    Session is intermediate step between internal table and database table. Data along with its action is stored in session i.e., data for screen fields, to which screen it is passed, the program name behind it, and how the next screen is processed.
    When the program has finished generating the session, you can run the session to execute the SAP transactions in it. You can either explicitly start and monitor a session or have the session run in the background processing system.
    Unless session is processed, the data is not transferred to database table.
    BDC_OPEN_GROUP
    You create the session through program by BDC_OPEN_GROUP function.
    Parameters to this function are:
    • User Name: User name
    • Group: Name of the session
    • Lock Date: The date on which you want to process the session.
    • Keep: This parameter is passed as ‘X’ when you want to retain session after
    processing it or ‘ ‘ to delete it after processing.
    BDC_INSERT
    This function creates the session & data is transferred to Session.
    Parameters to this function are:
    • Tcode: Transaction Name
    • Dynprotab: BDC Data
    BDC_CLOSE_GROUP
    This function closes the BDC Group. No Parameters.
    Some additional information for session processing
    When the session is generated using the KEEP option within the BDC_OPEN_GROUP, the system always keeps the sessions in the queue, whether it has been processed successfully or not.
    However, if the session is processed, you have to delete it manually. When session processing is completed successfully while KEEP option was not set, it will be removed automatically from the session queue. Log is not removed for that session.
    If the batch-input session is terminated with errors, then it appears in the list of INCORRECT session and it can be processed again. To correct incorrect session, you can analyze the session. The Analysis function allows to determine which screen and value has produced the error. If you find small errors in data, you can correct them interactively, otherwise you need to modify batch input program, which has generated the session or many times even the data file.
    CALL TRANSACTION
    About CALL TRANSACTION
    A technique similar to SESSION method, while batch input is a two-step procedure, Call Transaction does both steps online, one after the other. In this method, you call a transaction from your program by
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>
    Messages into <MSGTAB>.
    Parameter – 1 is transaction code.
    Parameter – 2 is name of BDCTAB table.
    Parameter – 3 here you are specifying mode in which you execute transaction
    A is all screen mode. All the screen of transaction are displayed.
    N is no screen mode. No screen is displayed when you execute the transaction.
    E is error screen. Only those screens are displayed wherein you have error record.
    Parameter – 4 here you are specifying update type by which database table is updated.
    S is for Synchronous update in which if you change data of one table then all the related Tables gets updated. And sy-subrc is returned i.e., sy-subrc is returned for once and all.
    A is for Asynchronous update. When you change data of one table, the sy-subrc is returned. And then updating of other affected tables takes place. So if system fails to update other tables, still sy-subrc returned is 0 (i.e., when first table gets updated).
    Parameter – 5 when you update database table, operation is either successful or unsuccessful or operation is successful with some warning. These messages are stored in internal table, which you specify along with MESSAGE statement. This internal table should be declared like BDCMSGCOLL, a structure available in ABAP/4. It contains the following fields:
    1. Tcode: Transaction code
    2. Dyname: Batch point module name
    3. Dynumb: Batch input Dyn number
    4. Msgtyp: Batch input message type (A/E/W/I/S)
    5. Msgspra: Batch input Lang, id of message
    6. Msgid: Message id
    7. MsgvN: Message variables (N = 1 - 4)
    For each entry, which is updated in database, table message is available in BDCMSGCOLL. As BDCMSGCOLL is structure, you need to declare a internal table which can contain multiple records (unlike structure).
    Steps for CALL TRANSACTION method
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. UPLOAD or WS_UPLOAD function to upload the data from local file to itab. (Considering file is local file)
    4. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>.
    Refresh BDCTAB.
    Endloop.
    (To populate BDCTAB, You need to transfer each and every field)
    The major differences between Session method and Call transaction are as follows:
    SESSION METHOD CALL TRANSACTION
    1. Data is not updated in database table unless Session is processed. Immediate updation in database table.
    2. No sy-subrc is returned. Sy-subrc is returned.
    3. Error log is created for error records. Errors need to be handled explicitly
    4. Updation in database table is always synchronous Updation in database table can be synchronous Or Asynchronous.
    Error Handling in CALL TRANSACTION
    When Session Method updates the records in database table, error records are stored in the log file. In Call transaction there is no such log file available and error record is lost unless handled. Usually you need to give report of all the error records i.e., records which are not inserted or updated in the database table. This can be done by the following method:
    Steps for the error handling in CALL TRANSACTION
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. Internal table BDCMSG like BDCMSGCOLL
    4. Internal table similar to Ist internal table
    (Third and fourth steps are for error handling)
    5. UPLOAD or WS_UPLOAD function to upload the data from the local file to itab. (Considering file is local file)
    6. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tr.code> using <Bdctab>
    Mode <A/N/E>
    Update <S/A>
    Messages <BDCMSG>.
    Perform check.
    Refresh BDCTAB.
    Endloop.
    7 Form check.
    IF sy-subrc <> 0. (Call transaction returns the sy-subrc if updating is not successful).
    Call function Format_message.
    (This function is called to store the message given by system and to display it along with record)
    Append itab2.
    Display the record and message.
    DIRECT INPUT
    About Direct Input
    In contrast to batch input, this technique does not create sessions, but stores the data directly. It does not simulate the online transaction. To enter the data into the corresponding database tables directly, the system calls a number of function modules that execute any necessary checks. In case of errors, the direct input technique provides a restart mechanism. However, to be able to activate the restart mechanism, direct input programs must be executed in the background only. Direct input checks the data thoroughly and then updates the database directly.
    You can start a Direct Input program in two ways;
    Start the program directly
    This is the quickest way to see if the program works with your flat file. This option is possible with all direct input programs. If the program ends abnormally, you will not have any logs telling you what has or has not been posted. To minimize the chance of this happening, always use the check file option for the first run with your flat file. This allows you to detect format errors before transfer.
    Starting the program via the DI administration transaction
    This transaction restarts the processing, if the data transfer program aborts. Since DI document are immediately posted into the SAP D/B, the restart option prevents the duplicate document posting that occurs during a program restart (i.e., without adjusting your flat file).
    Direct input is usually done for standard data like material master, FI accounting document, SD sales order and Classification for which SAP has provided standard programs.
    First time you work with the Direct Input administration program, you will need to do some preparation before you can transfer data:
    - Create variant
    - Define job
    - Start job
    - Restart job
    Common batch input errors
    - The batch input BDCDATA structure tries to assign values to fields which do not exist in the current transaction screen.
    - The screen in the BDCDATA structure does not match the right sequence, or an intermediate screen is missing.
    - On exceptional occasions, the logic flow of batch input session does not exactly match that of manual online processing. Testing the sessions online can discover by this.
    - The BDCDATA structure contains fields, which are longer than the actual definition.
    - Authorization problems.
    RECORDING A BATCH INPUT
    A B recording allows you to record a R/3 transaction and generate a program that contains all screens and field information in the required BDC-DATA format.
    You can either use SHDB transaction for recording or
    SYSTEM ? SERVICES ? BATCH INPUT ? EDIT
    And from here click recording.
    Enter name for the recording.
    (Dates are optional)
    Click recording.
    Enter transaction code.
    Enter.
    Click Save button.
    You finally come to a screen where, you have all the information for each screen including BDC_OKCODE.
    • Click Get Transaction.
    • Return to BI.
    • Click overview.
    • Position the cursor on the just recorded entry and click generate program.
    • Enter program name.
    • Click enter
    The program is generated for the particular transaction.
    BACKGROUND PROCESSING
    Need for Background processing
    When a large volume of data is involved, usually all batch inputs are done in background.
    The R/3 system includes functions that allow users to work non-interactively or offline. The background processing systems handle these functions.
    Non-interactively means that instead of executing the ABAP/4 programs and waiting for an answer, user can submit those programs for execution at a more convenient planned time.
    There are several reasons to submit programs for background execution.
    • The maximum time allowed for online execution should not exceed 300 seconds. User gets TIMEOUT error and an aborted transaction, if time for execution exceeds 300 seconds. To avoid these types of error, you can submit jobs for background processing.
    • You can use the system while your program is executing.
    This does not mean that interactive or online work is not useful. Both type of processing have their own purposes. Online work is the most common one entering business data, displaying information, printing small reports, managing the system and so on. Background jobs are mainly used for the following tasks; to process large amount of data, to execute periodic jobs without human intervention, to run program at a more convenient, planned time other than during normal working hours i.e., Nights or weekends.
    The transaction for background processing is SM36.
    Or
    Tools ? Administration ? Jobs ? Define jobs
    Or
    System ? services ? Jobs
    Components of the background jobs
    A job in Background processing is a series of steps that can be scheduled and step is a program for background processing.
    • Job name. Define the name of assigned to the job. It identifies the job. You can specify up to 32 characters for the name.
    • Job class. Indicates the type of background processing priority assigned to the job.
    The job class determines the priority of a job. The background system admits three types of job classes: A B & C, which correspond to job priority.
    • Job steps. Parameters to be passed for this screen are as follows:
    Program name.
    Variant if it is report program
    Start criteria for the job: Option available for this are as follows:
    Immediate - allows you to start a job immediately.
    Date/Time - allows you to start a job at a specific name.
    After job - you can start a job after a particular job.
    After event - allows you to start a job after a particular event.
    At operation mode - allows you to start a job when the system switches to a particular operation mode.
    Defining Background jobs
    It is two step process: Firstly, you define the job and then release it.
    When users define a job and save it, they are actually scheduling the report i.e., specifying the job components, the steps, the start time.
    When users schedule program for background processing, they are instructing the system to execute an ABAP/4 report or an external program in the background. Scheduled jobs are not executed until they are released. When jobs are released, they are sent for execution to the background processing system at the specified start time. Both scheduling and releasing of jobs require authorizations.
    HANDLING OF POP UP SCREEN IN BDC
    Many times in transaction pop up screen appears and for this screen you don’t pass any record but some indication to system telling it to proceed further. For example: The following screen
    To handle such screen, system has provided a variable called BDC_CURSOR. You pass this variable to BDCDATA and process the screen.
    Usually such screen appears in many transactions, in this case you are just passing information, that YES you want to save the information, that means YES should be clicked. So you are transferring this information to BDCDATA i.e., field name of YES which is usually SPOT_OPTION. Instead of BDC_OKCODE, you are passing BDC_CURSOR.
    BDC_CURSOR is also used to place cursor on particular field.
    A simple transaction where you are entering customer number on first screen and on next screen data is displayed for the particular customer number. Field, which we are changing here, are name and city. When you click on save, the changed record gets saved.
    Prerequisite to write this BDC interface as indicated earlier is:
    1. To find screen number
    2. To find screen field names, type of the field and length of the field.
    3. To find BDC_OKCODE for each screen
    4. Create flat file.
    generally  Batch Input usually are used to transfer large amount of data. For example you are implementing a new SAP project, and of course you will need some data transfer from legacy system to SAP system.
    CALL TRANSACTION is used especially for integration actions between two SAP systems or between different modules. Users sometimes wish to do something like that click a button or an item then SAP would inserts or changes data automatically. Here CALL TRANSACTION should be considered.
    2. Transfer data for multiple transactions usually the Batch Input method is used.
    check these sites for step by step process:
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    Check these link:
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://www.sap-img.com/abap/question-about-bdc-program.htm
    http://www.itcserver.com/blog/2006/06/30/batch-input-vs-call-transaction/
    http://www.planetsap.com/bdc_main_page.htm
    call Transaction or session method ?
    null

  • Auto Replace functionality similar to TOAD

    I've played with Raptor 804 for a few hours now. Only a couple things keeping me from being able to move (comfortably) to Raptor from Toad at this point.
    1) Bind Variables not remembering values if I run queries repetively. Other posts address that issue
    2) The other is Toad's Auto Replace feature. I have huge library built where I can type in a few letters and when I hit the space bar (or other set characters) the characters are replaced with something else. ie...I type "srf" and I get "select a.*, a.rowid from" or I type "hca" and I get the table name "ar.hz_cust_accounts" etc....
    Because I've got so many auto replace keystrokes saved I am able to type out ad-hoc queries in a hurry. I see Raptor has an Accelerators option under Preferences but that doesn't do quite what I want.
    3) Other very nice to have (because of the way I work) is the Named Sql options that is available in Toad. Basically instead of having a bunch of standard queries that I use all the time saved into files, I've got them saved in my Named Sql's in Toad so I can quicly pop them up (with a shortcut keystroke Ctrl-N) and then click on the one similar to the Sql History window. In conjunction with the bind variable functionality listed in item 1, it helps me investigate things very quickly when researching issues.
    Initial results on Raptor look good. I'm looking forward to it improving to the point where I can ditch Toad (which I've lost all faith in).

    1) try the new build it's in there
    2) This will have to wait till 1.1 however I agree for no other reason then fixing my select * form dual.
    3) Would key shortcuts and the ability to manage the snippets work ( again in 1.1)
    -kris

Maybe you are looking for