Populating the test data in table of IDES ECC 5.0 in Oracle

Hi Guys,
I have installed IDES ECC 5.0 successfully without any errors. But I don't see the data in tables like PA0001 etc. Can some body give me the steps for populating the tables with test data. I was able to sign on using DDIC in client 000.
Thanks,

you are using wrong client, login into client 800. check tcode SCC4 to check which client you want to log into.

Similar Messages

  • Problem about implementation of a controller in main_vi based on the test data of sub_vi

    Hi Group:
     I am developing a measurement and control system based on the PCI-6040E and SCXI1001,1102, and 1160.
    But now, I have a question ,which is how to implement a controller in main_vi based on the test data of sub_vi.
    my  program is consist of main_vi and  several sub_vi,  the main_vi is used for controlling and give the operation signal to hardware system (relay). and sub_vi is used for sampling data.
    for example: the sub_vi collects data from the hardware and the control algorithm and output is implemented in the main_vi, so the question is how to transfer the test data from the sub_vi to main_vi synchronously.
    I try to set the output connecter, but the data just return when  sub_vi is closed.
    anybody have good idea about it.
    Thanks a lot

    Hi hanwei,
    I'm making the assumption that you are not using LabVIEW Real-Time. If you are, you should use RT FIFOs. Otherwise, follow MikeS81's advice.
    I would go a step further and say that most likely it would be a better design to remove the loop (while loop?) from the subVI and instead place the subVI within a while loop in your top-level VI. This way on each execution of your subVI it will quickly return your data. Then you can also perform all the queue and dequeue operations in the two seperate while loops in your top-level VI. This organization will be easier to read, document, and understand.
    Mike Lyons
    National Instruments
    http://www.ni.com/devzone

  • Error when reading the DDIC-data for table

    Hi Tobias
    We upgraded from DMIS 2011 SP04 to SP07. We are in development system.
    To test initial load, we stopped replication for an existing table and then started replication for the same table.
    Mass transfer id connects ECC system to BW on HANA system.
    Replication errored with the following message:
    Error when reading the ddic-data for table MARD (RFC destination DWACLNT010).
    Looking for your guidance.
    Kind Regards
    Kamaljit Vilkhoo

    Was able to solve using oss note 1972533.

  • Pre-Populating the header data in the survey attached to an activity in CRM

    Hello Gurus,
    I have a question regarding Pre-Populating the header data in the survey attached to an activity in CRM Mobile 4.0. That in can we define a Functional module for a PBO (Process Before Output) for a survey u201CZ_SVY_Templateu201D in CRM 4.0.
    So that when a Sales rep opens the Survey in CRM Mobile 4.0 the Organizationu2019s Header data should automatically be displayed in the u201CZ_SVY_Templateu201D, so that he just needs to fill out only the remaining additional blank fields in the survey.
    << Moderator message - Please do not offer points >>
    Thanks,
    Siddhu
    Edited by: Rob Burbank on Oct 8, 2010 2:16 PM

    Moderator message - Cross post locked
    Rob

  • How to find the deleted data in tables

    guys,
    how to find the deleted data in tables example: i want to see whether anyone deleted data in MB5B report tables like mbew, etc.,
    regards,

    Hi,
    MBEWH is actually the history table of MBEW. It will record all the changes. As I have told you earlier if you have deleted the record dirctly from the table then it will not come even in the table MBEWH
    That means no changes have been made.
    regards

  • Can we use the test data dictionery saved for a FM for any other FM

    Hi experts,
    i want to use the test data dictionary which i have maintained for a FM. now i want to use the same data for another FM.
    Can i use that saved dictionary. if yes then how.
    Points are sure.....
    Thanks
    Ashish

    Amit,
    Yes i can, but this fm is calling from some another  applications (front end) and that is a big scenario. and that will be really vary time consuming so i want to do in this way.
    If this is not possible then the last way i have to opt is to do excercise again .
    Thanks

  • Two iterators browsing the same data (a table and a tree) problem

    I have an ADF UIX application with a navigation tree and a central page with a table of the records of a Works view object.
    There is a hierarchical relationship of Works ->Subworks-> Tasks which is shown in the tree. In the central part of the page, there is a table of Works.
    I want
    1) the tree to show all the data, but only about 10 works at a time (using small arrows for scrolling up and down)
    2) the table to show only 10 or 20 records each time. Not the same with the tree.
    3) Both clicking on a link on the tree and selecting a work on the table will open the appropriate page.
    4) Opening the children of a node in the tree and clicking on the links being able to send me directly in the correct subwork or task page EVEN if I haven't visited the parent node (work) first.
    I cannot seem to be able to succed in the "even if" part of 4 and I can succeed in only any 2 of the first three.
    If the tree iterator and table iterator point to the same view object (work1 for example), it is impossible to show a different set of works in the tree and the table
    If I use different view objects, then, if the tree part is working, the table selection doesn't work. I imagined that's a problem with "SetCurrentRowWithIndex" and I tried to set up correctly a "SetCurrentRowWithKey" event, but I didn't know how to write the handler. (I use JHS event handling, by the way)
    I suppose the following has errors:
    <event name="setCurrentRowWithKey" source="GetaeergView1">
    <set target="${bindings.GetaeergView1Iterator}" property="currentRowKeyInRange" value="${ui:tableSelectedKey(uix, 'GetaeergView1')}"/>
    </event>
    Any suggestions?

    Noone has tried this again?

  • How to find out the aligning Data dictionary table for a structure.

    Hi
    As the table controls are associated with strucures, the data input goes to the data dictionary table aligned to that structure. Is there any way we can find out the table related to that particular structure ?
    Thanks,
    Dhareppa

    Hi,
    The structure and its fields may be associated with multiple database tables.
    You can try by where-used list of the structure/field and then you can look and determine the table by hit and try method.
    Also as suggested above you can try get the SQL trace and then ypu can look for table associated behind the structure/fields I believe this is the best way to find the DB table.
    Thanks,
    Ravi

  • To find the greatest date from table

    guys,
    I currently use this query to find the difference between sysdate and the oldest date in the table
    select (SYSDATE - cur_date) from comp_check
    the problem is it fives multiple records as output.
    i want to design a query that will give the difference between sysdate and the oldest date in the table(it should throw only one record as the result)
    how do i do that.
    thanks in advance

    SQL> select sysdate-min(created) from all_users
    SQL> /
    SYSDATE-MIN(CREATED)
              507,373866
    SQL> Nicolas.
    MIN usage for oldest. Correction after ennisb's remark, thanks.
    Message was edited by:
    N. Gasparotto

  • Prob in populating the current date in drop down

    I want to get the current date into a drop down list.
    There are three drop down.
    1.Month 2. Day 3. Year.
    If anytime the page loads there should be current date in the drop down list. and if any one wants to change he can change the date accordingly.
    Can any one help me in sorting out this prob.
    I can think of getting the current date in the drop down by keeping it in the 1st option of the select tag. But when you click on the list the current date is not at proper location compared to other numbers in the list.
    Please reply.
    Message was edited by:
    aniketh_parmar

    function addit(){
    var dd = new Date();
    var yearnow = dd.getFullYear(); //gets 4 digit year
    //use getDate() & getMonth() methods in seperate fns, and populate
    //date and months
    for (var i=yearnow; i <= yearnow + 25 ;++i){
    addOption(document.getElementById("year"), i, i);
    function addOption(selectbox,text,value)
    var optn = document.createElement("OPTION");
    optn.text = text;
    optn.value = value;
    selectbox.options.add(optn);
    }in your jsp:
    <body onload="addit();">
    <form>
    <select name="month">
    </select>
    <select name="day">
    </select>
    <select name="year">
    </select>
    Have a seperate fn, and using the above 3 methods get
    today's (current) month, day & year.
    then something like,
    document.getElementById("year").selectedIndex = currYear;
    document.getElementById("month").selectedIndex = currMonth; //0 is Jan and 11 is Dec
    document.getElementById("day").selectedIndex = currDate;

  • SAP IDES ECC 6.0 EHP7 Oracle Installation error (DODS.SQL is missing)

    Hello community,
    I am having a problem with the installation of the current IDES ERP 6.0 EHP7 Oracle for Linux x86_64 installation images.
    I am using the Software Provisioning Manager 1.0 SP04_9-20009701 and selected installation option "SAP Business Suite 7i 2013 Support Release 1" > "EHP7 for SAP ERP 6.0 ABAP Support Release 1" > "Oracle" > "SAP Systems" > "Application Server ABAP" > "Standard System" .
    I am installaing on SUSE Linux Enterprise Server 11 SP3 x86_64.
    The installation stops at the "Import ABAP" stage:
    Program "Migration Monitor" exits with error code 103. For details see log files(s) import_monitor.java.log, import_monitor.log).
    import_monitor.log
    TRACE: 2014-07-04 04:16:14 com.sap.inst.migmon.LoadTask processPackage
    Task file generation for 'SAPDODS' finished import package:
    /usr/sap/IQ4/SYS/exe/uc/linuxx86_64/R3load -ctf I /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/
    51048312_1_EHP7_FOR_SAP_ERP_6.0_Installation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DATA/SAPDODS.STR
    /tmp/sapinst_instdir/BS2013SR1/ERP607SR1/ORA/INSTALL/BS2013SR1/ERP607SR1/ORA/STD/ABAP/DDLORA_LRG.TPL SAPD
    OD
    S.TSK ORA -l SAPDODS.log
    TRACE: 2014-07-04 04:16:14 com.sap.inst.migmon.LoadTask processPackage
    Loading of 'SAPDODS' starting import package into database:
    /usr/sap/IQ4/SYS/exe/uc/linuxx86_64/R3load -i SAPDODS.cmd -dbcodepage 4103 -l SAPDODS.log -stop_on_error -
    loadprocedure fast
    TRACE: 2014-07-04 04:16:14 com.sap.inst.migmon.LoadTask processPackage
    Loading of 'SAPDDIM' starting import package into database:
    /usr/sap/IQ4/SYS/exe/uc/linuxx86_64/R3load -i SAPDDIM.cmd -dbcodepage 4103 -l SAPDDIM.log -stop_on_error -
    loadprocedure fast
    INFO: 2014-07-04 04:16:15 com.sap.inst.migmon.LoadTask run_report
    Monitor jobs: running 2, waiting 4, completed 157, failed 1, total 164  error processing of package SAPDOD
    S
    ERROR: 2014-07-04 04:16:15 com.sap.inst.migmon.LoadTask run
    Loading of 'SAPDODS' import package is interrupted with R3load error.
    Process '/usr/sap/IQ4/SYS/exe/uc/linuxx86_64/R3load -i SAPDODS.cmd -dbcodepage 4103 -l SAPDODS.log -stop_o
    n_error -loadprocedure fast' exited with return code 2.
    For mode details see 'SAPDODS.log' file.
    Standard error output:
    sapparam: sapargv(argc, argv) has not been called!
    sapparam(1c): No Profile used.
    sapparam: SAPSYSTEMNAME neither in Profile nor in Commandline
    SAPDODS.log
    (SQL) INFO: Searching for SQL file SQLFiles.LST
    (SQL) INFO: SQLFiles.LST not found
    (SQL) INFO: Searching for SQL file /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/51048312_1_EHP7
    _FOR_SAP_ERP_6.0_Installation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DB/SQLFiles.LST
    (SQL) INFO: found /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/51048312_1_EHP7_FOR_SAP_ERP_6.0_
    Installation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DB/SQLFiles.LST
    (SQL) INFO: Trying to open /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/51048312_1_EHP7_FOR_SAP
    _ERP_6.0_Installation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DB/SQLFiles.LST
    (SQL) INFO: /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/51048312_1_EHP7_FOR_SAP_ERP_6.0_Instal
    lation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DB/SQLFiles.LST opened
    (SQL) INFO: Searching for SQL file DODS.SQL
    (SQL) INFO: DODS.SQL not found
    (SQL) INFO: Searching for SQL file /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/51048312_1_EHP7
    _FOR_SAP_ERP_6.0_Installation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DB/ORA/DODS.SQL
    (SQL) INFO: /net/iqfs1/install/sap/EHP7_for_SAP_ERP_6.0_IDES_Oracle/51048312_1_EHP7_FOR_SAP_ERP_6.0_Instal
    lation_export_IDES-Unicode_Part1/DATA_UNITS/EXPORT1/DB/ORA/DODS.SQL not found
    (SQL) ERROR: SQL list was not built successfully
    (DDL) ERROR: check_sql_list() failed for /BI0/APERS_BOD00
    (DB) INFO: disconnected from DB
    /usr/sap/IQ4/SYS/exe/uc/linuxx86_64/R3load: job finished with 1 error(s)
    /usr/sap/IQ4/SYS/exe/uc/linuxx86_64/R3load: END OF LOG: 20140704041615
    I have searched for the file DODS.SQL on the IDES import media. It is missing.
    Could that be the cause of the error?

    Hi Daniel.
    It looks to be problem with export DVDs. Can you download the same again and re-extract it.
    Regards,
    Deepak Kori

  • Populating the cache with VPD tables fails

    Using the Add Tables wizard I am trying to add tables to an Oracle Database cache. The tables have had RLS policies applied using the DBMS_RLS package. The Add Tables wizard fails with an ORA-28112 error (failed to execute policy function).
    Are there any known problems with cacheing tables of a virtual private database?
    Of course we want to cache an entire table, not just a portion as defined by that table's security policy. We have a procedure call that can cause the security policy to return a null predicate (effectively turning off security for all tables). We have set up a logon trigger to run this procedure for the user that populates the cache (we have done this successfully for other users), but we still get the error.
    Any help?
    Cache system:
    SPARC/Solaris 8
    9iAS 1.0.2.1
    Oracle EE 8.1.7.0
    Origin database:
    SPARC/Solaris 8
    Oracle EE 8.1.7.0
    Thanks,
    Steve

    Steve,
    In my opinion you should look first into the trace files. They are generally located under
    USER_DUMP_DEST directory. I believe that this error normally generates a trace file. So in the trace file you can see that it was trying to do and what happened. Also this may be happening due to some permission problems as well e.g. logged user versus policy user etc.
    HTH
    Prakash
    null

  • BADI for populating the Header data in MIRO

    Dear Gurus,
    I have tried to find the BADI for MIRO transaction as I want to populate the RKBP-XBLNR (Reference), RKBP-BUPLA (Business Place), RBKP-SECCO (Section Code) automatically. But I am unable to find any BADI to do this.
    Kindly advice,
    Thanks,
    Sam

    Hi ,
    You can try Business Add In 'INVOICE_UPDATE'.
    But with some modifications as said in note 392342
    Check the below part
    Solution
    If the data changed in the BAdI is supposed to be returned to the MIRO,
    this can only be executed by a program modification of BAdI
    'INVOICE_UPDATE' and function module 'MRM_BADI_INVOICE_CHECK'. Contact
    the remote consulting for this.

  • How to generate test data for all the tables in oracle

    I am planning to use plsql to generate the test data in all the tables in schema, schema name is given as input parameters, min records in master table, min records in child table. data should be consistent in the columns which are used for constraints i.e. using same column value..
    planning to implement something like
    execute sp_schema_data_gen (schemaname, minrecinmstrtbl, minrecsforchildtable);
    schemaname = owner,
    minrecinmstrtbl= minimum records to insert into each parent table,
    minrecsforchildtable = minimum records to enter into each child table of a each master table;
    all_tables where owner= schemaname;
    all_tab_columns and all_constrains - where owner =schemaname;
    using dbms_random pkg.
    is anyone have better idea to do this.. is this functionality already there in oracle db?

    Ah, damorgan, data, test data, metadata and table-driven processes. Love the stuff!
    There are two approaches you can take with this. I'll mention both and then ask which
    one you think you would find most useful for your requirements.
    One approach I would call the generic bottom-up approach which is the one I think you
    are referring to.
    This system is a generic test data generator. It isn't designed to generate data for any
    particular existing table or application but is the general case solution.
    Building on damorgan's advice define the basic hierarchy: table collection, tables, data; so start at the data level.
    1. Identify/document the data types that you need to support. Start small (NUMBER, VARCHAR2, DATE) and add as you go along
    2. For each data type identify the functionality and attributes that you need. For instance for VARCHAR2
    a. min length - the minimum length to generate
    b. max length - the maximum length
    c. prefix - a prefix for the generated data; e.g. for an address field you might want a 'add1' prefix
    d. suffix - a suffix for the generated data; see prefix
    e. whether to generate NULLs
    3. For NUMBER you will probably want at least precision and scale but might want minimum and maximum values or even min/max precision,
    min/max scale.
    4. store the attribute combinations in Oracle tables
    5. build functionality for each data type that can create the range and type of data that you need. These functions should take parameters that can be used to control the attributes and the amount of data generated.
    6. At the table level you will need business rules that control how the different columns of the table relate to each other. For example, for ADDRESS information your business rule might be that ADDRESS1, CITY, STATE, ZIP are required and ADDRESS2 is optional.
    7. Add table-level processes, driven by the saved metadata, that can generate data at the record level by leveraging the data type functionality you have built previously.
    8. Then add the metadata, business rules and functionality to control the TABLE-TO-TABLE relationships; that is, the data model. You need the same DETPNO values in the SCOTT.EMP table that exist in the SCOTT.DEPT table.
    The second approach I have used more often. I would it call the top-down approach and I use
    it when test data is needed for an existing system. The main use case here is to avoid
    having to copy production data to QA, TEST or DEV environments.
    QA people want to test with data that they are familiar with: names, companies, code values.
    I've found they aren't often fond of random character strings for names of things.
    The second approach I use for mature systems where there is already plenty of data to choose from.
    It involves selecting subsets of data from each of the existing tables and saving that data in a
    set of test tables. This data can then be used for regression testing and for automated unit testing of
    existing functionality and functionality that is being developed.
    QA can use data they are already familiar with and can test the application (GUI?) interface on that
    data to see if they get the expected changes.
    For each table to be tested (e.g. DEPT) I create two test system tables. A BEFORE table and an EXPECTED table.
    1. DEPT_TEST_BEFORE
         This table has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look BEFORE the
         test for that test case is performed.
         CREATE TABLE DEPT_TEST_BEFORE
         TESTCASE NUMBER,
         DEPTNO NUMBER(2),
         DNAME VARCHAR2(14 BYTE),
         LOC VARCHAR2(13 BYTE)
    2. DEPT_TEST_EXPECTED
         This table also has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look AFTER the
         test for that test case is performed.
    Each of these tables are a mirror image of the actual application table with one new column
    added that contains a value representing the TESTCASE_NUMBER.
    To create test case #3 identify or create the DEPT records you want to use for test case #3.
    Insert these records into DEPT_TEST_BEFORE:
         INSERT INTO DEPT_TEST_BEFORE
         SELECT 3, D.* FROM DEPT D where DEPNO = 20
    Insert records for test case #3 into DEPT_TEST_EXPECTED that show the rows as they should
    look after test #3 is run. For example, if test #3 creates one new record add all the
    records fro the BEFORE data set and add a new one for the new record.
    When you want to run TESTCASE_ONE the process is basically (ignore for this illustration that
    there is a foreign key betwee DEPT and EMP):
    1. delete the records from SCOTT.DEPT that correspond to test case #3 DEPT records.
              DELETE FROM DEPT
              WHERE DEPTNO IN (SELECT DEPTNO FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3);
    2. insert the test data set records for SCOTT.DEPT for test case #3.
              INSERT INTO DEPT
              SELECT DEPTNO, DNAME, LOC FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3;
    3 perform the test.
    4. compare the actual results with the expected results.
         This is done by a function that compares the records in DEPT with the records
         in DEPT_TEST_EXPECTED for test #3.
         I usually store these results in yet another table or just report them out.
    5. Report out the differences.
    This second approach uses data the users (QA) are already familiar with, is scaleable and
    is easy to add new data that meets business requirements.
    It is also easy to automatically generate the necessary tables and test setup/breakdown
    using a table-driven metadata approach. Adding a new test table is as easy as calling
    a stored procedure; the procedure can generate the DDL or create the actual tables needed
    for the BEFORE and AFTER snapshots.
    The main disadvantage is that existing data will almost never cover the corner cases.
    But you can add data for these. By corner cases I mean data that defines the limits
    for a data type: a VARCHAR2(30) name field should have at least one test record that
    has a name that is 30 characters long.
    Which of these approaches makes the most sense for you?

  • Test Data Management Server - Databases/Tables

    Hello,
    We are currently preparing to do a pilot with the Test Data Management Server and were point some initial documents together indicating technical changes it will bring to our landscape. We have attend the TZTDM3 course and have read the Master Guide but haven't located the detailed information regarding what databases/tables are created when TDMS is installed. If anyone could point me at this documentation or provided the info it would be greatly appreciated.
    Thanks,
    J

    Thanks for response but I don't want move data warehouse databases. My idea is create another server and change the data warehouse managemet server role to this new server preserving the databases in the original server. Is possible do this?

Maybe you are looking for

  • N8 User - Launch OVI maps, the phones restart

    Dear All, Everytome i open OVI maps the phone restart, just got it about 2 weeks now Please help any one. Or NOKIA can answer

  • Relationship between ERPi metadata rules and Data load rules

    I am using version 11.1.1.3 to load EBS data to ERPi and then to FDM. I have created a Metadata rule where I have assigned a ledger from the Source Accounting Entities. I have also created a Data Load rule that references the same ledger as the metad

  • How to return stream from bpel process

    Hi, As per my requirement I need to generate a file it could be pdf /text and then send this file as output of my bpel process. I am not getting any document which provide me any solution. Please help me out ASAP. Thnaks, Ana

  • Bluetooth headset disconnects if not in iTunes

    I have a Creative WP-450 Bluetooth Headset. I have a MacBook Air Summer 2012. I connect my headset. iTunes pops up. All is fine as long as I listen to music in iTunes. IF I close iTunes and stream media throug my browser or do anything other than lis

  • Adobe PDF Link Helper add on IE 8

    Hello, we have a client who is having issues with their IE 8 browser - it freezes, crashes and seems to leave behind extract iexplore processes that don't close properly - searching online, I discovered that the add on "Adobe PDF Link Helper" can som