How to use the program "Generate test data for BAI bank statement" ?

Hi all,
I use the program RFEBKAT5 to create an Electronic Bank Statement file in BAI format ,but it doesn't generate the file .
I don't know how to generate the file.
Please help me !
Thanks all !

You need to suitably adjust your GL accounts for "Posting Offset Account" (Whisch is defaulted as ++++++++19" and Checkout/Funds Out/Funds IN GL accounts to your configuration.
Once you do this try generating the file again.

Similar Messages

  • Generate test data for BAI bank statement

    Hi all,
    I use the program RFEBKAT5 to create an Electronic Bank Statement file in BAI format ,but it doesn't generate the file .
    I don't know how to generate the file.
    Please help me !
    Thanks all !

    Hello,
    Did you click in PC Download checkbox?
    Regards
    Waza

  • How to return the newly generated sequence id for an INSERT statement

    A record is to be inserted into a table with a sequence for the primary key. The newly inserted sequence value is to returned on successful insertion. Is it possible to do all this in a single statement (say executeUpdate or any other) using java.sql.* ?
    E.g.: - A student record is to be inserted into the STUDENT table. There is a sequence (by name Student_ID_SEQ) on the primary key Student_ID. Student_ID_SEQ.nextval will generate the new sequence id which will be provided as input to the SQL statement (say statement.executeUpdate) along with other student attribute values. On insertion the created sequence id should be returned. And all this should happen in a single statement (single call to database). Stored Procedures can accomplish this. But is this feasible without the use of Stored Procedures?
    Thanks.

    a better aproach is to generate the auto key on the
    database side, not on the application side.That's his problem - since the database is supplying the key for the new record his application which executed the SQL has no way to identify the record that was just added. I just create the key on the app server and accept the likelihood of overlap (which is extremely small).
    Here is a more technical explanation:
    Table Person
       ID,
       Name,
       Phone Number,
       Age
    }The field ID is an autonumber, and all other fields are not unique.
    Now, when this code executes:
    PreparedStatement pst = conn.prepareStatement("Insert Into Person (Name, Phone Number, Age) Values ?, ?, ?");
    pst.setString(1, "John");
    pst.setString(2, "405-444-5555");
    pst.setInt(3, 44);
    pst.executeUpdate();How can the app determine the ID of the person just added since no query is possible which is guaranteed to select just the record that was inserted?
    Since I am generally against Stored Procedures I would develop a way to insure that your keys were unique and generate them inside the app server.

  • How to use the program called ZSAPLINK to import Function Name

    How to use the program called ZSAPLINK to import Function Name

    Not sure why you posted into this forum?
    Note the header for this forum:
    This forum is dedicated to all other development-related questions which are not directly addressed by other forums. This includes Business Objects SDKs, products, or technologies which do not fall under BusinessObjects Enterprise, BusinessObjects Edge, Crystal Reports Server, or Crystal Reports (for example Desktop Intelligence SDK, Universe Designer SDK, Portal Integration Kits, Java User Function Libraries, and other third party technologies or development languages).
    Ludek

  • How to use the program RHMOVE30

    Dear Gurus,
    One of our consultant has created organisation structure in one client of development system and wants to transport the same to another client in the same system and also to another system Quality.
    Came to know that by using program RHMOVE30 it can be done, but can anyone tell me how to use this program, I mean what are the inputs to be given and how to get the transport request.
    Regards,
    Sharath

    Hi Ali,
    Yes, normally transports will be moved from DEV to PRD and hope the same configuration is there in u r setup. But if you want to move the transport from PRD to DEV there is a way, but for this you may need to open the client restriction in u r production system (which is not recommended or u have to do this when no user is logged on to the production system).
    First make changes in the client settings for enabling transport request generation and execute the program RHMOVE30 and give u r inputs and generate a transport request. Once the request is generated it wil create two files(cofile and datafile folders) in trans folder (/usr/sap/<SID>/trans or <drive>:\usr\sap\<SID>\trans)
    Based on the request number copy the cofile and data file from production system and paste it in the same path in development system. Then login to development system from SAP and go to the import queue of development server and then choose EXTRAS->OTHER REQUEST->ADD and this will add that request to development queue and u can transport it normally.
    Hope this clarifies. Award points if helpful
    Regards,
    Sharath

  • How to change the HTML/XML test data to test my_abap_proxy in SE80?

    Hello,
    I am trying to test my_abap_proxy in SE80 by usng a HTML/XML file of my desk top, the test data is as below
      <LoadingDate>20110422</LoadingDate>
      <DocumentDate>20110422</DocumentDate>
      <SDDocumentCategory>G</SDDocumentCategory>
    so on.......
    Now, i want to change the Loading date in this test data, so, i tried after Loading file in Test Service Provider screen of my_abap_proxy of SE80, but i could not make change ANY data in this test data, pls. let me know how to change the data?
    Thank you

    Hello ABAP_SAP_ABAP ,
                                          Youneed to select the option "generate template data " and then on the displaying template there you will see the option for editing in the top left corner toolbar.
    Then after editing you can do a XMl syntax check and test your proxy.
    Hope this helps.
    Thanks,
    Greetson

  • How to find the programs created by date

    Hi all
    I would like to find the programs created by date and by report title. Is there any report to find programs by date or by report title..?
    Thanks
    Prasead K

    Hi ,
    Can anybody tell me how fo find out the date on which a Tcode was created?
    Thanks,
    Manjula.s

  • How to generate test data for all the tables in oracle

    I am planning to use plsql to generate the test data in all the tables in schema, schema name is given as input parameters, min records in master table, min records in child table. data should be consistent in the columns which are used for constraints i.e. using same column value..
    planning to implement something like
    execute sp_schema_data_gen (schemaname, minrecinmstrtbl, minrecsforchildtable);
    schemaname = owner,
    minrecinmstrtbl= minimum records to insert into each parent table,
    minrecsforchildtable = minimum records to enter into each child table of a each master table;
    all_tables where owner= schemaname;
    all_tab_columns and all_constrains - where owner =schemaname;
    using dbms_random pkg.
    is anyone have better idea to do this.. is this functionality already there in oracle db?

    Ah, damorgan, data, test data, metadata and table-driven processes. Love the stuff!
    There are two approaches you can take with this. I'll mention both and then ask which
    one you think you would find most useful for your requirements.
    One approach I would call the generic bottom-up approach which is the one I think you
    are referring to.
    This system is a generic test data generator. It isn't designed to generate data for any
    particular existing table or application but is the general case solution.
    Building on damorgan's advice define the basic hierarchy: table collection, tables, data; so start at the data level.
    1. Identify/document the data types that you need to support. Start small (NUMBER, VARCHAR2, DATE) and add as you go along
    2. For each data type identify the functionality and attributes that you need. For instance for VARCHAR2
    a. min length - the minimum length to generate
    b. max length - the maximum length
    c. prefix - a prefix for the generated data; e.g. for an address field you might want a 'add1' prefix
    d. suffix - a suffix for the generated data; see prefix
    e. whether to generate NULLs
    3. For NUMBER you will probably want at least precision and scale but might want minimum and maximum values or even min/max precision,
    min/max scale.
    4. store the attribute combinations in Oracle tables
    5. build functionality for each data type that can create the range and type of data that you need. These functions should take parameters that can be used to control the attributes and the amount of data generated.
    6. At the table level you will need business rules that control how the different columns of the table relate to each other. For example, for ADDRESS information your business rule might be that ADDRESS1, CITY, STATE, ZIP are required and ADDRESS2 is optional.
    7. Add table-level processes, driven by the saved metadata, that can generate data at the record level by leveraging the data type functionality you have built previously.
    8. Then add the metadata, business rules and functionality to control the TABLE-TO-TABLE relationships; that is, the data model. You need the same DETPNO values in the SCOTT.EMP table that exist in the SCOTT.DEPT table.
    The second approach I have used more often. I would it call the top-down approach and I use
    it when test data is needed for an existing system. The main use case here is to avoid
    having to copy production data to QA, TEST or DEV environments.
    QA people want to test with data that they are familiar with: names, companies, code values.
    I've found they aren't often fond of random character strings for names of things.
    The second approach I use for mature systems where there is already plenty of data to choose from.
    It involves selecting subsets of data from each of the existing tables and saving that data in a
    set of test tables. This data can then be used for regression testing and for automated unit testing of
    existing functionality and functionality that is being developed.
    QA can use data they are already familiar with and can test the application (GUI?) interface on that
    data to see if they get the expected changes.
    For each table to be tested (e.g. DEPT) I create two test system tables. A BEFORE table and an EXPECTED table.
    1. DEPT_TEST_BEFORE
         This table has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look BEFORE the
         test for that test case is performed.
         CREATE TABLE DEPT_TEST_BEFORE
         TESTCASE NUMBER,
         DEPTNO NUMBER(2),
         DNAME VARCHAR2(14 BYTE),
         LOC VARCHAR2(13 BYTE)
    2. DEPT_TEST_EXPECTED
         This table also has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look AFTER the
         test for that test case is performed.
    Each of these tables are a mirror image of the actual application table with one new column
    added that contains a value representing the TESTCASE_NUMBER.
    To create test case #3 identify or create the DEPT records you want to use for test case #3.
    Insert these records into DEPT_TEST_BEFORE:
         INSERT INTO DEPT_TEST_BEFORE
         SELECT 3, D.* FROM DEPT D where DEPNO = 20
    Insert records for test case #3 into DEPT_TEST_EXPECTED that show the rows as they should
    look after test #3 is run. For example, if test #3 creates one new record add all the
    records fro the BEFORE data set and add a new one for the new record.
    When you want to run TESTCASE_ONE the process is basically (ignore for this illustration that
    there is a foreign key betwee DEPT and EMP):
    1. delete the records from SCOTT.DEPT that correspond to test case #3 DEPT records.
              DELETE FROM DEPT
              WHERE DEPTNO IN (SELECT DEPTNO FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3);
    2. insert the test data set records for SCOTT.DEPT for test case #3.
              INSERT INTO DEPT
              SELECT DEPTNO, DNAME, LOC FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3;
    3 perform the test.
    4. compare the actual results with the expected results.
         This is done by a function that compares the records in DEPT with the records
         in DEPT_TEST_EXPECTED for test #3.
         I usually store these results in yet another table or just report them out.
    5. Report out the differences.
    This second approach uses data the users (QA) are already familiar with, is scaleable and
    is easy to add new data that meets business requirements.
    It is also easy to automatically generate the necessary tables and test setup/breakdown
    using a table-driven metadata approach. Adding a new test table is as easy as calling
    a stored procedure; the procedure can generate the DDL or create the actual tables needed
    for the BEFORE and AFTER snapshots.
    The main disadvantage is that existing data will almost never cover the corner cases.
    But you can add data for these. By corner cases I mean data that defines the limits
    for a data type: a VARCHAR2(30) name field should have at least one test record that
    has a name that is 30 characters long.
    Which of these approaches makes the most sense for you?

  • How to use the path generated by A star to guide the robot to move?

    I now know how to use A star path planning on Voronoi and I can run this program successfully.\  
    however, my problem now is how i can use the path data (generated by the A* path planning algorithm) to guide the robot to drive along this path?   well, i suppose i can get a series of dots (denoted by x and y coordinates) from the path, and use these dots to guide the robot ?  
    can anyone please help?  
     thank you very much
    Solved!
    Go to Solution.

    Ravens Fan wrote:
    You can't.
    My post above will point everyone to the other post so that we don't have people trying to assist you in two different places without knowledge of the other thread.
    thank you 

  • How to use the random generator in java

    hey peeps, this is the class in which i am trying to implement the random generator in;
    public class Matchlist
        private studentdetails sd = new studentdetails();
        /** matchList StringBuilder stores the match list in progress */
        private StringBuilder matchList = new StringBuilder();
        private int loop = 0;
        private int matches = 0;
        public Matchlist()
            sd.createstudentdetails();
        /** Method to create the actual match list, returns the list as a string */
        public String CreateMatch()
            int g;
            int y;
                for (g = 0; g < 4; g++)//g = game
                    for (y = 17; y > -1; y--)//y = green house
                        /** Check to see if the game is empty */
                        if (sd.getgh(y).getGame(g).equalsIgnoreCase(""))
                            for (int x = 0; x < 18; x++) //x = yellow house
                                if (sd.getyh(x).getGame(g).equalsIgnoreCase(""))
                                    if (sd.getgh(y).getC_lass() != sd.getyh(x).getC_lass())
                                        /** Check to see if the person has played the other person */
                                        if (sd.getgh(y).checkPlayed(sd.getyh(x).getName()) == false)
                                            /** Set the game to the name of the opponent played */
                                            sd.getyh(x).changeGame(g, sd.getgh(y).getName());
                                            sd.getgh(y).changeGame(g, sd.getyh(x).getName());
                                            /** Build the match list step by step using append with \n at the end to create a new line */
                                            matchList.append(sd.getyh(x).getName() + " vs " + sd.getgh(y).getName() + "\n");
                                            matches++;
                                            break;
                /** Convert the stringbuilder into an actual string, then return it */
                String completeMatchList = matchList.toString();
                System.out.println(matches);
                for (int i = 0; i <18; i++)
                    sd.getyh(i).getEmptyMatches();
                    sd.getgh(i).getEmptyMatches();
                return completeMatchList;
        }what i dont understand is how to implement it to pick my matches at random using the http://java.sun.com/j2se/1.4.2/docs/api/java/util/Random.html java tutorials from here
    regards

    How to use Random ?
    First you open API then you read it, then you use it.
    You mention you try to use it but i just see a horrible nested for for for if if if loop.
    Restructure code and question and maybe it makes more sense.
    Edited by: pgeuens on 10-mrt-2008 22:58

  • I have purchased the Adobe program that allows you to convert PDF documents. How do use the application?. Thanks for your help.

    I have purchased the Adobe program that allows you to convert PDF documents but I can't figure out how to convert the PDF to a Word document. Does anyone know how the do this?
    Thanks.
    Andrew

    Which of these things did you buy?
    - ExportPDF
    - PDF Pack
    - Acrobat Standard
    - Acrobat Pro
    - Acrobat Plus

  • Using FPGA program to collect data for long time without 'gap'

    Our data collection system has a NI-PCIe 7852R card. We want to collect data at up to 10 KHz for up to 10 to 30 minutes. The data amount is too large so we have to separated them into multiple Array and save in multiple files. We do not want to loose data during the saving time. Anyone has idea on how to do it? We would appreciate your help.
    My current method:
    1 using FIFO memory, in FPGA program loop , write data from AIs to FIFO memory, after certain amount of data is written, say 2000 of it, asks a IRQ.
    2. In Host program loop, wait for IRQ, once the IRQ comes, read in FIFO data, do some processing, put the data into prelocated array, once the prelocated array is full, save it to the disk.
    So far the method is not working.
    1st, there seems never a IRQ, the wait for IRQ (invoke method) never took any time.
    2nd, the FIFO reading in the host program loop seems takes no less time than the FPGA program write same amount of data into FIFO...
    SO I have extra questions:
    1. Does the 7852R cards and its new FPGA still support IRQ?
    2. What is the reading speed of FIFO read in Host program? I thought it should be much faster than the FPGA program writing it....
    Any help will be propitiated.
    Thanks a lot

    Here is the FPGA program block
    Part of Host program block inside the loop
    Part of Running results.
       1. Some data is missing in every FIFO reading in Host program
       2. The FIFO reading in Host program takes more time than the AIs=> FIFO in the FIFO program
       3. Of cause each host loop takes more time than it takes FPGA program to get same amount of data...
    Attachments:
    FPGA_AI_block.png ‏48 KB
    HOST_AI_inloop_block.png ‏31 KB
    Running result.png ‏124 KB

  • How to use BAPI_OBJCL* to set classification data for Techn. Objects - PM ?

    Good afternoon !
            I´m working in a project to integrate some legacy system with the ECC 6.0, and I need to do some activities like, create functional locations and equipments in SAP PM, I have done this creating Z RFC modules that call bapis like BAPI_FUNCLOC_CREATE and BAPI_EQUI_CREATE, but these BAPIs doesn´t fill classification data, so, in a first moment, I solved this using BDC, but I would like to solve these needs using standard BAPIs if possible.
            To solve the classification, I tried to do it using the BAPI_OBJCL* BAPIs(BAPI_OBJCL_GET_KEY_OF_OBJECT, BAPI_OBJCL_CREATE_KEY), but I didn´t find any example of using these bapis with technical objects, so I tried to test the BAPI_OBJCL_GET_KEY_OF_OBJECT in a first moment and, when I call it passing '003' as classtype(003 is the classtype for the functional locations), this bapi returns me "Class type 003 requires additional objects"(the same occurs if I try it with 002, for equipment). What means this message ? Need I to do some customization in img/spro, or cannot these bapis be used with PM technical objects ?
            Thank you for your attention,
            Wilson

    Hi all !
            I have found the solution for this problem, I found that I was trying to use the wrong functions for my case, so, to solve my needs here, I will use BAPI_OBJCL_CREATE to set the classification data for Functional Locations and Equipments, and I can use BAPI_OBJCL_GETCLASSES to list all classes assigned for a Technical Object, and BAPI_OBJCL_GETDETAIL to get all characteristics assigned to a Technical Object for a specific class.
            Thank you all for your attention !
            Wilson

  • How to use the dynamical or static variable for ESSBASE cube name?

    Hi Experts,
    When I import ESSBASE Cube into physical layer, the cube name from ESSBASE is created automatically, such as H_Sales.
    I want to use the the static or dynamical variable for replacing the external name. So I try to create the static variable in RPD,such as cubeName, and use the following code
    'VALUEOF(cubeName)' into the textbox of external name.
    However, when I view the report in answer, it will generate the error message: Database VALUEOF(cubeName) does not exist.
    Is it possible to implement this functionality?
    Thanks..

    Hi,
    use <%=odiRef.getSchemaName("D")%>
    D as parameter if it is the Data Schema or W if you need the schema from Work Schema
    Your command will be like:
    select <%=odiRef.getSchemaName("D")%>.GER_LOT_EXEC_ODI('Fluxo', 1, 'C') FROM DUAL
    Works?
    Cezar Santos
    http://odiexperts.com

  • How to use the first row function value for the rest of records

    Hi all,
    I am having a select statement like this. We are calling function enabled for each row. So it is taking
    time. As per our bussiness logic, we can use what the function 'enabled' will give for the first row to
    other rows also. so that we can avoid calling the function 'enabled' next row onwars.
    Before
    SELECT
    decode(enabled(col1, col2, col3),'TRUE','xxx', col4),
    decode(enabled(col1, col2, col3),'TRUE','xxx', col5),
    decode(enabled(col1, col2, col3),'TRUE','xxx', col6),
    decode(enabled(col1, col2, col3),'TRUE','xxx', col7),
    decode(enabled(col1, col2, col3),'TRUE','xxx', col8),
    from
    table1 t1, table2 t2 where t1.col1 = t2.col1
    After
    SELECT
    decode(enabled(col1, col2, col3),'TRUE','xxx', col4),
    decode('true/false','TRUE','xxx', col5), --Here 'true/false' is the first row value after calling the function 'enabled'
    decode('true/false','TRUE','xxx', col6),
    decode('true/false','TRUE','xxx', col7),
    decode('true/false','TRUE','xxx', col8),
    from
    table1 t1, table2 t2 where t1.col1 = t2.col1
    Any thoughts on this, how to implement this logic
    Thanks,
    Pal

    It's not clear where col1, col2, col3... etc. come from, so I've assumed they're in table1.
    with table1_X as (select enabled(col1, col2, col3) as enbl
                            ,col1, col2, col3, col4, col5, col6, col7, col8
                      from   table1)
    SELECT decode(t1.enbl,'TRUE','xxx', col4)
          ,decode(t1.enbl,'TRUE','xxx', col5)
          ,decode(t1.enbl,'TRUE','xxx', col6),
          ,decode(t1.enbl,'TRUE','xxx', col7),
          ,decode(t1.enbl,'TRUE','xxx', col8),
    from   table1_x t1 join table2 t2 on (t1.col1 = t2.col1)p.s. you'd be better leaving out the enabled function altogether and incorporating the logic of the function directly in the SQL, and then you would be avoiding any context switching which will clearly impact performance on any large amount of data.
    Edited by: BluShadow on 14-May-2013 08:26

Maybe you are looking for