About important tables in FI

Hi Friends,
I would like to know something about the tables and the data flow. I've not been a part of customising a "Z" report as yet, and I'm sure I would have to do it any time.
Please suggest to me as to where do I need to start off and the importance of learning the table names pertaining to FI/CO. Important factors that I need to bear in my mind before I could attempt to do so.
What information I need to give to the ABAPer or how do I prepare a Functional(or Technical) spec.
Would highly appreciate if there are any links pertaining to this and also if anyone could share some tips.
Regards,
Soujanya.

http://www.abap.es/Descargas/TAB%20-%20Relacion%20de%20las%20tablas%20por%20modulos.PDF
you may please search in search engines with
sap tables .pdf
sap tables .doc
You should be able to identify the tables and field names to be populated.
Technical spec. is more of technical in nature with table and field names, whereas functional spec. is more of explaining the functionality required. A lay man should be able to understanding functionality when they read the FS. Whereas, Technical spec. if purely for technical purposes with lot of table and field names and other technical stuff.
Check on google for some sample documents.
Regards,
Ravi

Similar Messages

  • Question about temporary tables and imports.

    Hello everybody.
    Interesting one this. We have an application that uses global temporary tables. We refresh live to test using import of the application schema. I noticed when doing this yesterday that the temporary tables were being recreated in test as perminant rather than temporary.
    Is there a reason for this? Has anyone come across this before? Is there a way around it (apart from manual checking)?
    Many thanks
    Rup

    Could you specify how you found out that it is coming in as permanent?
    I believe exp/Imp will export and import it is Temporary table. I have just done a simple test to check and it works fine;
    Here is my test log:
    SQL> connect scott
    Enter password:
    Connected.
    SQL>SQL> CREATE GLOBAL TEMPORARY TABLE test_global
      2          (startdate DATE,
      3           enddate DATE,
      4           class CHAR(20))
      5        ON COMMIT DELETE ROWS;
    Table created.
    SQL>  select table_name,temporary from user_tables where table_name='TEST_GLOBAL';
    TABLE_NAME                     T
    TEST_GLOBAL                    Y
    SQL> select table_name,temporary from user_tables;
    TABLE_NAME                     T
    DEPT                           N
    EMP                            N
    BONUS                          N
    SALGRADE                       N
    TEST_GLOBAL                    Y
    SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - Production
    With the Partitioning, Oracle Label Security, OLAP and Data Mining Scoring Engine options
    C:\Documents\dbmsdirect\Testing\global>exp
    Export: Release 10.2.0.2.0 - Production on Wed Feb 20 12:18:45 2008
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Username: scott
    Password:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - Production
    With the Partitioning, Oracle Label Security, OLAP and Data Mining Scoring Engine options
    Enter array fetch buffer size: 4096 >
    Export file: EXPDAT.DMP > test_global
    (2)U(sers), or (3)T(ables): (2)U > t
    Export table data (yes/no): yes >
    Compress extents (yes/no): yes >
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    server uses AL32UTF8 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    Table(T) or Partition(T:P) to be exported: (RETURN to quit) > test_global
    . . exporting table                    TEST_GLOBAL
    Table(T) or Partition(T:P) to be exported: (RETURN to quit) >
    Export terminated successfully without warnings.
    C:\Documents\dbmsdirect\Testing\global>sqlplus /nolog
    SQL*Plus: Release 10.2.0.2.0 - Production on Wed Feb 20 12:19:50 2008
    Copyright (c) 1982, 2005, Oracle.  All Rights Reserved.
    SQL> connect scott
    Enter password:
    Connected.
    SQL> drop table test_global purge;
    Table dropped.
    SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - Production
    With the Partitioning, Oracle Label Security, OLAP and Data Mining Scoring Engine options
    C:\Documents\dbmsdirect\Testing\global>imp
    Import: Release 10.2.0.2.0 - Production on Wed Feb 20 12:20:19 2008
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Username: scott
    Password:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - Production
    With the Partitioning, Oracle Label Security, OLAP and Data Mining Scoring Engine options
    Import file: EXPDAT.DMP > test_global.DMP
    Enter insert buffer size (minimum is 8192) 30720>
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    import server uses AL32UTF8 character set (possible charset conversion)
    List contents of import file only (yes/no): no >
    Ignore create error due to object existence (yes/no): no >
    Import grants (yes/no): yes >
    Import table data (yes/no): yes >
    Import entire export file (yes/no): no > yes
    . importing SCOTT's objects into SCOTT
    . importing SCOTT's objects into SCOTT
    Import terminated successfully without warnings.
    C:\Documents\dbmsdirect\Testing\global>sqlplus /nolog
    SQL*Plus: Release 10.2.0.2.0 - Production on Wed Feb 20 12:22:44 2008
    Copyright (c) 1982, 2005, Oracle.  All Rights Reserved.
    SQL> connect scott
    Enter password:
    Connected.
    SQL>
    SQL> select table_name,temporary from user_tables;
    TABLE_NAME                     T
    DEPT                           N
    EMP                            N
    BONUS                          N
    SALGRADE                       N
    TEST_GLOBAL                    Y
    SQL> insert into TEST_GLOBAL values (sysdate,sysdate,'test1');
    1 row created.
    SQL> select * from TEST_GLOBAL;
    STARTDATE ENDDATE   CLASS
    20-FEB-08 20-FEB-08 test1
    SQL> commit;
    Commit complete.
    SQL> select * from TEST_GLOBAL;
    no rows selected
    SQL>  select table_name,temporary from user_tables where table_name='TEST_GLOBAL';
    TABLE_NAME                     T
    TEST_GLOBAL                    Y
    SQL>

  • OBIEE Error while importing table from database

    Hi
    I am getting the following error when i am trying to import table from database.
    [nQSError: 16001]ODBC error state: IM004 code:0 message:
    [Microsoft][ODBC Driver Manager] Driver`s SQLAllocHandle on SQL_HANDLE_ENV failed.
    Any idea y such error.
    Thanks and Regards,
    Andy

    Looks like an error in the ODBC driver, not OBIEE as such.
    Have you tried googling it?
    Can you post details about your OS and DB.

  • Exceptions thrown while importing tables from SQL Server 2000

    Hi all,
    <MyProblem:
    I cannot import the tables defined in my local SQL Server. An exception is thrown --oracle.jdeveloper.cm.ds.db.ValidationException: Columns of type varchar cannot have a size specified.
    />
    <What I have done:
    In Oracle JDev 10g, Created a new DB connection called "SQLServerConnection" to connect to the SQL Server and it works fine. I am able to see all tables by expanding the connection.
    During the creation of a new partner link named "DBAdapter", I use the adapter wizard to select DB Adapter and then select the "SQLServerConnection" DB connection I have created. I select Perform an Operation on a Table and then check Select. After I pressed OK button on the[i] Import Tables  dialog, a Error messege box emmerged indicating the Exception: Columns of type varchar cannot have a size specified.
    I try to change the data type of the columns of the tables to char, there is still similar exception (Columns of type char cannot have a size specified).
    />
    I got the stack trace from the prompt by using jdev.exe.
    <StackTrace:
    oracle.jdeveloper.cm.ds.db.ValidationException: Columns of type varchar cannot h
    ave a size specified.
    at oracle.jdeveloper.offlinedb.OfflineDBTableValidator.validateObject(Of
    flineDBTableValidator.java:116)
    at oracle.jdeveloper.cm.ds.db.AbstractDBObjectProvider.validateObject(Ab
    stractDBObjectProvider.java:601)
    at oracle.jdeveloper.offlinedb.OfflineDBObjectProvider.updateObject(Offl
    ineDBObjectProvider.java:952)
    at oracle.tip.tools.ide.pm.wizard.adapter.db.DBAdapterDescriptorGenerato
    r.generate(DBAdapterDescriptorGenerator.java:274)
    at oracle.tip.tools.ide.pm.wizard.adapter.db.DBAdapterSelectTablePage.ac
    tionPerformed(DBAdapterSelectTablePage.java:451)
    at javax.swing.AbstractButton.fireActionPerformed(AbstractButton.java:17
    86)
    at javax.swing.AbstractButton$ForwardActionEvents.actionPerformed(Abstra
    ctButton.java:1839)
    at javax.swing.DefaultButtonModel.fireActionPerformed(DefaultButtonModel
    .java:420)
    at javax.swing.DefaultButtonModel.setPressed(DefaultButtonModel.java:258
    at javax.swing.plaf.basic.BasicButtonListener.mouseReleased(BasicButtonL
    istener.java:245)
    at java.awt.Component.processMouseEvent(Component.java:5100)
    at java.awt.Component.processEvent(Component.java:4897)
    at java.awt.Container.processEvent(Container.java:1569)
    at java.awt.Component.dispatchEventImpl(Component.java:3615)
    at java.awt.Container.dispatchEventImpl(Container.java:1627)
    at java.awt.Component.dispatchEvent(Component.java:3477)
    at java.awt.LightweightDispatcher.retargetMouseEvent(Container.java:3483
    at java.awt.LightweightDispatcher.processMouseEvent(Container.java:3198)
    at java.awt.LightweightDispatcher.dispatchEvent(Container.java:3128)
    at java.awt.Container.dispatchEventImpl(Container.java:1613)
    at java.awt.Window.dispatchEventImpl(Window.java:1606)
    at java.awt.Component.dispatchEvent(Component.java:3477)
    at java.awt.EventQueue.dispatchEvent(EventQueue.java:456)
    at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchTh
    read.java:201)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThre
    ad.java:151)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThre
    ad.java:141)
    at java.awt.Dialog$1.run(Dialog.java:540)
    at java.awt.Dialog.show(Dialog.java:561)
    at java.awt.Component.show(Component.java:1133)
    at java.awt.Component.setVisible(Component.java:1088)
    at oracle.bali.ewt.wizard.WizardDialog.runDialog(Unknown Source)
    at oracle.tip.tools.ide.pm.wizard.adapter.techAdapterWizard.display(tech
    AdapterWizard.java:151)
    at oracle.tip.tools.ide.pm.bpelgraph.editors.PartnerLinkEditPage.handleA
    dapterWizard(PartnerLinkEditPage.java:311)
    at oracle.tip.tools.ide.pm.bpelgraph.editors.PartnerLinkEditPage.actionP
    erformed(PartnerLinkEditPage.java:893)
    at javax.swing.AbstractButton.fireActionPerformed(AbstractButton.java:17
    86)
    at javax.swing.AbstractButton$ForwardActionEvents.actionPerformed(Abstra
    ctButton.java:1839)
    at javax.swing.DefaultButtonModel.fireActionPerformed(DefaultButtonModel
    .java:420)
    at javax.swing.DefaultButtonModel.setPressed(DefaultButtonModel.java:258
    at javax.swing.plaf.basic.BasicButtonListener.mouseReleased(BasicButtonL
    istener.java:245)
    at java.awt.AWTEventMulticaster.mouseReleased(AWTEventMulticaster.java:2
    31)
    at java.awt.Component.processMouseEvent(Component.java:5100)
    at java.awt.Component.processEvent(Component.java:4897)
    at java.awt.Container.processEvent(Container.java:1569)
    at java.awt.Component.dispatchEventImpl(Component.java:3615)
    at java.awt.Container.dispatchEventImpl(Container.java:1627)
    at java.awt.Component.dispatchEvent(Component.java:3477)
    at java.awt.LightweightDispatcher.retargetMouseEvent(Container.java:3483
    at java.awt.LightweightDispatcher.processMouseEvent(Container.java:3198)
    at java.awt.LightweightDispatcher.dispatchEvent(Container.java:3128)
    at java.awt.Container.dispatchEventImpl(Container.java:1613)
    at java.awt.Window.dispatchEventImpl(Window.java:1606)
    at java.awt.Component.dispatchEvent(Component.java:3477)
    at java.awt.EventQueue.dispatchEvent(EventQueue.java:456)
    at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchTh
    read.java:201)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThre
    ad.java:151)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThre
    ad.java:141)
    at java.awt.Dialog$1.run(Dialog.java:540)
    at java.awt.Dialog.show(Dialog.java:561)
    at oracle.tip.tools.ide.pm.bpelgraph.editors.PartnerLinkDialog.create(Pa
    rtnerLinkDialog.java:90)
    at oracle.tip.tools.ide.pm.bpelgraph.collaxalaf.BPELGraphPane.createPart
    nerLink(BPELGraphPane.java:324)
    at oracle.tip.tools.ide.pm.bpelgraph.collaxalaf.BPELDNDDropHandler.handl
    eDrop(BPELDNDDropHandler.java:77)
    at oracle.tip.tools.ide.common.graph.DefaultGraphPaneDropTargetListener.
    routeToDropHandler(DefaultGraphPaneDropTargetListener.java:212)
    at oracle.tip.tools.ide.common.graph.DefaultGraphPaneDropTargetListener.
    handleDrop(DefaultGraphPaneDropTargetListener.java:200)
    at oracle.tip.tools.ide.pm.bpelgraph.collaxalaf.BPELGraphPaneDropTargetL
    istener.handleDrop(BPELGraphPaneDropTargetListener.java:75)
    at oracle.tip.tools.ide.common.graph.DefaultGraphPaneDropTargetListener.
    drop(DefaultGraphPaneDropTargetListener.java:260)
    at java.awt.dnd.DropTarget.drop(DropTarget.java:398)
    at sun.awt.dnd.SunDropTargetContextPeer.processDropMessage(SunDropTarget
    ContextPeer.java:542)
    at sun.awt.dnd.SunDropTargetContextPeer.access$800(SunDropTargetContextP
    eer.java:52)
    at sun.awt.dnd.SunDropTargetContextPeer$EventDispatcher.dispatchDropEven
    t(SunDropTargetContextPeer.java:805)
    at sun.awt.dnd.SunDropTargetContextPeer$EventDispatcher.dispatchEvent(Su
    nDropTargetContextPeer.java:743)
    at sun.awt.dnd.SunDropTargetEvent.dispatch(SunDropTargetEvent.java:29)
    at java.awt.Component.dispatchEventImpl(Component.java:3494)
    at java.awt.Container.dispatchEventImpl(Container.java:1627)
    at java.awt.Component.dispatchEvent(Component.java:3477)
    at java.awt.LightweightDispatcher.retargetMouseEvent(Container.java:3483
    at java.awt.LightweightDispatcher.processDropTargetEvent(Container.java:
    3269)
    at java.awt.LightweightDispatcher.dispatchEvent(Container.java:3123)
    at java.awt.Container.dispatchEventImpl(Container.java:1613)
    at java.awt.Window.dispatchEventImpl(Window.java:1606)
    at java.awt.Component.dispatchEvent(Component.java:3477)
    at java.awt.EventQueue.dispatchEvent(EventQueue.java:456)
    at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchTh
    read.java:201)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThre
    ad.java:151)
    at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:145)
    at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:137)
    at java.awt.EventDispatchThread.run(EventDispatchThread.java:100)
    />
    <Oracle:
    Oracle JDeveloper 10.1.2.0.0 (Build 1811)
    Oracle IDE     10.1.2.17.84
    Business Components Version     10.1.2.17.96
    UML Modelers Version     10.1.2.16.71
    Versioning Support     10.1.2.16.71
    WebDAV Support Version     10.1.2.16.71
    Struts Modeler Version     10.1.2.6.15
    BPEL Designer     10.1.2.0.0 (Build 050504)
    Designer Generators Framework     10.1.2.7.56
    ADF UIX     2.2.16
    java.version     1.4.2_06
    java.home     E:\OraBPELPM\jdk\jre
    java.vm.name     Java HotSpot(TM) Client VM
    java.vm.vendor     Sun Microsystems Inc.
    java.vm.version     1.4.2_06-b03
    user.language     zh
    user.region     <no value assigned>
    user.name     Administrator
    user.home     C:\Documents and Settings\Administrator
    user.dir     E:\OraBPELPM\integration\jdev\jdev\bin
    os.name     Windows 2003
    os.version     5.2
    os.arch     x86
    http.proxyHost     jproxy.nju.edu.cn
    http.proxyPort     8080
    />
    <SQL Server:
    Microsoft SQL Server 2000 with the Microsoft SQL Server 2000 Driver for JDBC Service Pack 3 installed
    />
    By the way, I have read the thread Thread: DB Adapter not importing SQL Svr tables and take some of their advises but the problem still exist.
    Any help would be greatly appreciated.
    Thanks!
    And Merry Christmas to all!

    Hi, ronen!
    Thank you.
    I followed the steps and the there were no exceptions then.
    However, new runtime error emerged.
    The BPEL PM Server cannot build Toplink Session. But I do not know why.
    The messeges from the "Start BPEL PM Server" console about this project (named "TryDB") is as following (DoSelect is the service name of the DB Adapter):
    <::Outbound><oracle.tip.adapter.db.ox.TopLinkXMLProjectInitializer initialize> I
    nitializing an existing toplink project for use by the database adapter.
    <::Outbound><oracle.tip.adapter.db.ox.TopLinkXMLProjectInitializer initialize> I
    nitializing an existing toplink project for use by the database adapter.
    <2006-01-10 16:03:02,656> <INFO> <default.collaxa.cube.compiler> &#27491;&#22312;&#39564;&#35777; "E:\Or
    aBPELPM\integration\orabpel\domains\default\tmp\.bpel_TryDB_1.0.jar\TryDB.bpel"
    <2006-01-10 16:03:05,156> <INFO> <default.collaxa.cube.engine.deployment> <CubeP
    rocessFactory::generateProcessClass>
    &#27969;&#31243; "TryDB" (&#20462;&#35746; "1.0") &#32534;&#35793;&#25104;&#21151;&#12290;
    <2006-01-10 16:03:05,765> <INFO> <default.collaxa.cube.engine.deployment> &#27969;&#31243; "
    TryDB" (&#20462;&#35746;&#26631;&#35760; "1.0") &#24050;&#25104;&#21151;&#21152;&#36733;&#12290;
    06/01/10 16:03:34 Tutalii: E:\OraBPELPM\integration\orabpel\lib\orabpel.jar arch
    ive
    <2006-01-10 16:03:34,500> <INFO> <default.collaxa.cube.ws> <AdapterFramework::Ou
    tbound> file:/E:/OraBPELPM/integration/orabpel/domains/default/tmp/.bpel_TryDB_1
    .0.jar/DoSelect.wsdl [ DoSelect_ptt::DoSelectSelect(DoSelectSelect_inparameters,
    AirlineTableCollection) ] - Using JCA Connection Pool - max size = <unbounded>
    <2006-01-10 16:03:35,093> <ERROR> <default.collaxa.cube.ws> <Database Adapter::O
    utbound> <oracle.tip.adapter.db.DBManagedConnectionFactory acquireClientSession>
    unable to build the TopLink Session
    <2006-01-10 16:03:35,125> <ERROR> <default.collaxa.cube.ws> <Database Adapter::O
    utbound> <oracle.tip.adapter.db.DBInteraction executeOutboundRead> unable to exe
    cute the NamedQuery: TryDB.AirlineTable.DoSelectSelect
    <2006-01-10 16:03:35,125> <ERROR> <default.collaxa.cube.ws> <AdapterFramework::O
    utbound> file:/E:/OraBPELPM/integration/orabpel/domains/default/tmp/.bpel_TryDB_
    1.0.jar/DoSelect.wsdl [ DoSelect_ptt::DoSelectSelect(DoSelectSelect_inparameters
    ,AirlineTableCollection) ] - Could not invoke operation 'DoSelectSelect' against
    the 'Database Adapter' due to:
    ORABPEL-11622
    &#26080;&#27861;&#21019;&#24314;/&#35775;&#38382; TopLink &#20250;&#35805;.
    &#27492;&#20250;&#35805;&#29992;&#20110;&#36830;&#25509;&#21040;&#25968;&#25454;&#23384;&#20648;&#12290;.
    &#26377;&#20851;&#29305;&#23450;&#24322;&#24120;&#38169;&#35823;, &#35831;&#21442;&#38405;&#26681;&#24322;&#24120;&#38169;&#35823;&#12290;&#24744;&#21487;&#33021;&#38656;&#35201;&#22312; j2ee/home/application-deploymen
    ts/default/DbAdapter/oc4j-ra.xml &#20013;&#37197;&#32622;&#36830;&#25509;&#35774;&#32622;&#24182;&#37325;&#26032;&#21551;&#21160;&#26381;&#21153;&#22120;&#12290;&#30001; $J2EE_HOME/a
    pplication-deployments/default/DbAdapter/oc4j-ra.xml &#23548;&#33268;&#12290;
    <2006-01-10 16:03:35,125> <INFO> <default.collaxa.cube.ws> <AdapterFramework::Ou
    tbound> Change logging level for Logger 'default.collaxa.cube.ws' to DEBUG to se
    e full error stack
    <2006-01-10 16:03:35,125> <ERROR> <default.collaxa.cube.ws> <AdapterFramework::O
    utbound> file:/E:/OraBPELPM/integration/orabpel/domains/default/tmp/.bpel_TryDB_
    1.0.jar/DoSelect.wsdl [ DoSelect_ptt::DoSelectSelect(DoSelectSelect_inparameters
    ,AirlineTableCollection) ] - Rolling back JCA LocalTransaction
    And in the Audit part of the BPEL Console, I got the following "details"
    <remoteFault xmlns="http://schemas.oracle.com/bpel/extension">
    <part name="code">
    <code>4003</code>
    </part>
    <part name="summary">
    <summary>file:/E:/OraBPELPM/integration/orabpel/domains/default/tmp/.bpel_TryDB_1.0.jar/DoSelect.wsdl [
    DoSelect_ptt::DoSelectSelect(DoSelectSelect_inparameters,AirlineTableCollection) ] - WSIF JCA Execute of operation
    'DoSelectSelect' failed due to: ????/?? TopLink ??. ?????????????. ; nested exception is: ORABPEL-11622 ????/??
    TopLink ??. ?????????????. ????????, ??????????????? j2ee/home/application-deployments/default/DbAdapter/oc4j-ra.xml
    ????????????????? $J2EE_HOME/application-deployments/default/DbAdapter/oc4j-ra.xml ???</summary>
    </part>
    <part name="detail">
    <detail>??????: ????????? [com.microsoft.jdbc.sqlserver.SQLServerDriver]?</detail>
    </part>
    </remoteFault>
    My oc4j-ra.xml has the following content (SQLServerConnection is the name of the connection of SQL Server database):
    <connector-factory location="eis/DB/SQLServerConnection" connector-name="Database Adapter">
         <config-property name="driverClassName" value="com.microsoft.jdbc.sqlserver.SQLServerDriver"/>
         <config-property name="connectionString" value="jdbc:microsoft:sqlserver://localhost:1433;SelectMethod=cursor;databasename=lxbpel"/>
         <config-property name="userName" value="bpel"/>
         <config-property name="password" value="bpel"/>
         <config-property name="minConnections" value="5"/>
         <config-property name="maxConnections" value="5"/>
         <config-property name="minReadConnections" value="1"/>
         <config-property name="maxReadConnections" value="1"/>
         <config-property name="usesExternalConnectionPooling" value="false"/>
         <config-property name="dataSourceName" value=""/>
         <config-property name="usesExternalTransactionController" value="false"/>
         <config-property name="platformClassName" value="oracle.toplink.internal.databaseaccess.SQLServerPlatform"/>
         <config-property name="usesNativeSequencing" value="true"/>
         <config-property name="sequencePreallocationSize" value="50"/>
         <config-property name="tableQualifier" value=""/>
    </connector-factory>
    I think this configuration in this oc4j-ra.xml is fine
    So please help me again. I owe you so much.
    Thank you!

  • Character problem in toad import table data from excel

    Hi everybody,
    I want to import data from an excel file to an Oracle table, so I'm using Toad's "Import Table Data" tool for this purpose.
    The problem is Oracle doesn't import non-english characters properly.
    My database is XE and character set is 'AL32UTF8'.
    I searched web, but didn't find the solution.
    Please help...

    Hi again,
    Thank you for your reply Srini, but it is not about Toad.
    Character encoding of Excel causes the problem.
    I exported an unicode encoded csv file from excel and tried to load data from that file. It worked.
    Thanks.

  • Question about import database pk, fk

    Hi
    I want to ask a question about imported database. I imported database and I noticed tables doesnt have Primary key or foreign key. They havent relation between eachs. Why can this occured ?
    is it about import or another something?

    yeah ı am surprised none of tables have pk or fk ..
    but ı want to show you http://i51.tinypic.com/2hedc9d.jpg
    what is blue shapes in there ?
    can be foerign keys ?

  • Need information about Internal Tables

    Hi Every one!
    I Need some information about Internal tables. Pls help be above the same.
    Thanks & with Regards,
    Chandra.

    Hi..,
    <b>
    Internal tables </b>
    Internal tables provide a means of taking data from a fixed structure and storing it in working memory in ABAP. The data is stored line by line in memory, and each line has the same structure. In ABAP, internal tables fulfill the function of arrays. Since they are dynamic data objects, they save the programmer the task of dynamic memory management in his or her programs. You should use internal tables whenever you want to process a dataset with a fixed structure within a program. A particularly important use for internal tables is for storing and formatting data from a database table within a program. They are also a good way of including very complicated data structures in an ABAP program.
    Like all elements in the ABAP type concept, internal tables can exist both as data types and as data objects A data type is the abstract description of an internal table, either in a program or centrally in the ABAP Dictionary, that you use to create a concrete data object. The data type is also an attribute of an existing data object.
    <b>Internal Tables as Data Types</b>
    Internal tables and structures are the two structured data types in ABAP. The data type of an internal table is fully specified by its line type, key, and table type.
    <b>Line type</b>
    The line type of an internal table can be any data type. The data type of an internal table is normally a structure. Each component of the structure is a column in the internal table. However, the line type may also be elementary or another internal table.
    <b>Key</b>
    The key identifies table rows. There are two kinds of key for internal tables - the standard key and a user-defined key. You can specify whether the key should be UNIQUE or NON-UNIQUE. Internal tables with a unique key cannot contain duplicate entries. The uniqueness depends on the table access method.
    If a table has a structured line type, its default key consists of all of its non-numerical columns that are not references or themselves internal tables. If a table has an elementary line type, the default key is the entire line. The default key of an internal table whose line type is an internal table, the default key is empty.
    The user-defined key can contain any columns of the internal table that are not references or themselves internal tables. Internal tables with a user-defined key are called key tables. When you define the key, the sequence of the key fields is significant. You should remember this, for example, if you intend to sort the table according to the key.
    <b>
    Table type</b>
    The table type determines how ABAP will access individual table entries. Internal tables can be divided into three types:
    <u>Standard tables</u> have an internal linear index. From a particular size upwards, the indexes of internal tables are administered as trees. In this case, the index administration overhead increases in logarithmic and not linear relation to the number of lines. The system can access records either by using the table index or the key. The response time for key access is proportional to the number of entries in the table. The key of a standard table is always non-unique. You cannot specify a unique key. This means that standard tables can always be filled very quickly, since the system does not have to check whether there are already existing entries.
    <u>
    Sorted tables</u> are always saved sorted by the key. They also have an internal index. The system can access records either by using the table index or the key. The response time for key access is logarithmically proportional to the number of table entries, since the system uses a binary search. The key of a sorted table can be either unique or non-unique. When you define the table, you must specify whether the key is to be unique or not. Standard tables and sorted tables are known generically as index tables.
    <u>
    Hashed tables</u> have no linear index. You can only access a hashed table using its key. The response time is independent of the number of table entries, and is constant, since the system access the table entries using a hash algorithm. The key of a hashed table must be unique. When you define the table, you must specify the key as UNIQUE.
    <b>
    Generic Internal Tables</b>
    Unlike other local data types in programs, you do not have to specify the data type of an internal table fully. Instead, you can specify a generic construction, that is, the key or key and line type of an internal table data type may remain unspecified. You can use generic internal tables to specify the types of field symbols and the interface parameters of procedures . You cannot use them to declare data objects.
    <b>Internal Tables as Dynamic Data Objects</b>
    Data objects that are defined either with the data type of an internal table, or directly as an internal table, are always fully defined in respect of their line type, key and access method. However, the number of lines is not fixed. Thus internal tables are dynamic data objects, since they can contain any number of lines of a particular type. The only restriction on the number of lines an internal table may contain are the limits of your system installation. The maximum memory that can be occupied by an internal table (including its internal administration) is 2 gigabytes. A more realistic figure is up to 500 megabytes. An additional restriction for hashed tables is that they may not contain more than 2 million entries. The line types of internal tables can be any ABAP data types - elementary, structured, or internal tables. The individual lines of an internal table are called table lines or table entries. Each component of a structured line is called a column in the internal table.
    <b>
    Choosing a Table Type</b>
    The table type (and particularly the access method) that you will use depends on how the typical internal table operations will be most frequently executed.
    <b>
    Standard tables</b>
    This is the most appropriate type if you are going to address the individual table entries using the index. Index access is the quickest possible access. You should fill a standard table by appending lines (ABAP APPEND statement), and read, modify and delete entries by specifying the index (INDEX option with the relevant ABAP command). The access time for a standard table increases in a linear relationship with the number of table entries. If you need key access, standard tables are particularly useful if you can fill and process the table in separate steps. For example, you could fill the table by appending entries, and then sort it. If you use the binary search option with key access, the response time is logarithmically proportional to the number of table entries.
    <b>Sorted tables</b>
    This is the most appropriate type if you need a table which is sorted as you fill it. You fill sorted tables using the INSERT statement. Entries are inserted according to the sort sequence defined through the table key. Any illegal entries are recognized as soon as you try to add them to the table. The response time for key access is logarithmically proportional to the number of table entries, since the system always uses a binary search. Sorted tables are particularly useful for partially sequential processing in a LOOP if you specify the beginning of the table key in the WHERE condition.
    <b>
    Hashed tables</b>
    This is the most appropriate type for any table where the main operation is key access. You cannot access a hashed table using its index. The response time for key access remains constant, regardless of the number of table entries. Like database tables, hashed tables always have a unique key. Hashed tables are useful if you want to construct and use an internal table which resembles a database table or for processing large amounts of data.
    regards,
    sai ramesh

  • Can't import excel file using Powerpivot Import Table Wizard

    I am new to the powerpivot in 2010 Excel.   I was trying to use the Powerpivot's import table wizard to import an Excel file into Powerpivot.  I was able to connect the target source file when I was in the wizard window testing the connection. 
    However, after I accepted to import the file and clicked finished, I couldn't see all the columns of the file were completely uploaded into the PowerPivot worksheet.  Also, the file has only 138 rows but it ends up over thousands row.  Can you please
    let me kow if this feature works at all or did I miss any steps?

    Hi MySueSue,
    Here are some tutorials about how to Import Excel Data into a PowerPivot Workbook.
    http://msdn.microsoft.com/en-us/library/gg399137(v=sql.110).aspx
    http://technet.microsoft.com/en-us/sqlserver/dn151367.aspx
    If there is anything I can do for you regarding this issue, feel free to post back.
    Best regards,
    Greta Ge
    TechNet Community Support
    It's recommended to download and install
    Configuration Analyzer Tool (OffCAT), which is developed by Microsoft Support teams. Once the tool is installed, you can run it at any time to scan for hundreds of known issues in Office
    programs.

  • What does a BI Developer need to know about SAP tables and why?

    Hi,
    i.
    As a BI Developer, to what extent do you need to know about SAP tables?
    e.g. if you work mainly in the functional areas LO and FI
    ii.
    For example, I know that in Finance, GL relates to the tables BSEG and BKPF.
    And if it is important to know about them, then which other tables need to be known and for what reasons?
    e.g. do I need to know all these tables: http://www.erpgenie.com/sap/abap/tables_fi.htm
    iii.
    I thought as a BI Developer, knowing that a particular datasource, e.g. 0FI_GL_4 contains line items is enough and you replicate it to BW and work with it. Why is there the need to know the tables from which this datasource gets is data?
    iv. What is the main difference between the FI datasources and the LO datasource?
    Thanks.

    Hi,
    i.  As a BI Developer, to what extent do you need to know about SAP tables?
    e.g. if you work mainly in the functional areas LO and FI
    Not required, but today business demands so you required this Knowelgde.
    ii.
    For example, I know that in Finance, GL relates to the tables BSEG and BKPF.
    And if it is important to know about them, then which other tables need to be known and for what reasons?
    e.g. do I need to know all these tables: http://www.erpgenie.com/sap/abap/tables_fi.htm
    We can't say yes or No, so based on your requirement you need to sit with Functional COnsultanst and the understand the Business Requirements and develop the OBjetcs.
    iii.
    I thought as a BI Developer, knowing that a particular datasource, e.g. 0FI_GL_4 contains line items is enough and you replicate it to BW and work with it. Why is there the need to know the tables from which this datasource gets is data?
    If you want to know the data flow and reconsilization then you need to know it else not required
    iv. What is the main difference between the FI datasources and the LO datasource?
    Yes, one main difference is There is no setup table concept for FI DataSource*
    Note: If you want to become a good BW consulatnt you must know about Tables for normal BW consultant it is not required.
    Thanks
    Reddy

  • Importing tables with OMB+

    Hello all,
    I would like to import some tables, but not their constraints and I have not been able to do this through the OWB GUI (version 10.2), so I thought maybe I could do it using OMB+
    Does anybody know whether it is possible to import tables in a module using OMB+?
    As far as I have understood, (and please, correct me if I'm wrong!) OMBIMPORT only imports .mdl files and OMBRETRIEVE only gives you information about the existing tables in the repository.
    Thanks a lot in advance,
    Nerina

    Yes, you can import directly from your metadata source using OMB+, but as with the gui I beleive that it will import the FULL definition of the table, and that will include the constraints.
    However, if you want to play with this: You need to first connect to your repository, CC to your project, and then give it your password to your defined metadata location for your module (assuming that you normally need to set this). I have a config file of all relevant info, so you will be seeing a whole lot of dollar signs in this cut-paste job...
    #Connect to Design Repository
    set print [OMBCONNECT $OWB_DEG_USER/$OWB_DEG_PASS@$OWB_DEG_HOST:$OWB_DEG_PORT:$OWB_DEG_SIT USE REPOSITORY '$OWB_DEG_REPOS']
    puts "$print"
    #CC to Project
    set print [OMBCC '$PROJECT_NAME_DEV']
    puts "$print"
    #Enter Password to allow imports and deployments
    #NOTE: This needs to be modified to use paramters for proper release management.
    set print [OMBALTER LOCATION '$METADATA_LOCATION' SET PROPERTIES (PASSWORD) VALUES ('$PASSWORD')]
    puts "$print"
    OMBCOMMIT
    OMBSAVEOK, now you need to create your transient import plan, and then import from metadata location for each table.
    Sooo. assuming a table called SOURCE_TABLE you would do something like:
    set err [catch {
       set print [eval OMBCREATE TRANSIENT IMPORT_ACTION_PLAN 'IMPORT_ACTION1' \
                         ADD ACTION 'IMPORT_ACTION' \
                         SET REF SOURCE TABLE 'SOURCE_TABLE' \
                         SET REF TARGET ORACLE_MODULE 'MY_MODULE_NAME']
       puts "$print"
       set print [OMBIMPORT FROM METADATA_LOCATION FOR IMPORT_ACTION_PLAN 'IMPORT_ACTION1']
        puts "$print"
    } errMsg]
    if {$err != 0} {
              puts "ERROR Importing: SOURCE_TABLE!!!"
               puts "ERRMSG: $errMsg"
    }You can use this to import tables, sequences, and views from the defined metadata location by changing the source type in the action plan. Of course, you will also want to delete the action plan once you are done with it.
    Cheers,
    Mike

  • Important tables in MM and PP

    Hi,
      Hi could any one tell me the MM and PP flow,important tables ,fields,and the linkage between them.
    Regards,
    Charumathi.B

    Hi Charumathi,
    Here is the MM Flow,
    A.)MM Process Flow
    The typical procurement cycle for a service or material consists of the following phases:
    1. Determination of Requirements
    Materials requirements are identified either in the user departments or via materials planning and control. (This can cover both MRP proper and the demand-based approach to inventory control. The regular checking of stock levels of materials defined by master records, use of the order-point method, and forecasting on the basis of past usage are important aspects of the latter.) You can enter purchase requisitions yourself, or they can be generated automatically by the materials planning and control system.
    2. Source Determination
    The Purchasing component helps you identify potential sources of supply based on past orders and existing longer-term purchase agreements. This speeds the process of creating requests for quotation (RFQs), which can be sent to vendors electronically via SAP EDI, if desired.
    3. Vendor Selection and Comparison of Quotations
    The system is capable of simulating pricing scenarios, allowing you to compare a number of different quotations. Rejection letters can be sent automatically.
    4. Purchase Order Processing
    The Purchasing system adopts information from the requisition and the quotation to help you create a purchase order. As with purchase requisitions, you can generate Pos yourself or have the system generate them automatically. Vendor scheduling agreements and contracts (in the SAP System, types of longer-term purchase agreement) are also supported.
    5. Purchase Order Follow-Up
    The system checks the reminder periods you have specified and - if necessary - automatically prints reminders or expediters at the predefined intervals. It also provides you with an up-to-date status of all purchase requisitions, quotations, and purchase orders.
    6. Goods Receiving and Inventory Management
    Goods Receiving personnel can confirm the receipt of goods simply by entering the Po number. By specifying permissible tolerances, buyers can limit over- and under deliveries of ordered goods.
    7. Invoice Verification
    The system supports the checking and matching of invoices. The accounts payable clerk is notified of quantity and price variances because the system has access to PO and goods receipt data. This speeds the process of auditing and clearing invoices for payment
    MM flow is processing between business organisation and Vendor:
    1...Purchase Requisition --- T.code ---ME51
    2...Source List -
    T.code-----ME41
    3...Request for Quation (RFQ)-T.codeME21N (To Vendor)
    4...Purchase Order -
    T.code -
    ME21N (To Vendor)
    5...Goods Receipt-------T.code....MIGO
    (Tables: MKPF, MSEG)
    6...Invoice Verification----T.code...MIRO
    (Tables: BKPF , BSEG)
    7...Vendor Payment----T.code...FB60 (FI - Account Payable)
    B.) And the Important tables in MM module as as follows
    EINA Purchasing Info Record- General Data
    EINE Purchasing Info Record- Purchasing Organization Data
    MAKT Material Descriptions
    MARA General Material Data
    MARC Plant Data for Material
    MARD Storage Location Data for Material
    MAST Material to BOM Link
    MBEW Material Valuation
    MKPF Header- Material Document
    MSEG Document Segment- Material
    MVER Material Consumption
    MVKE Sales Data for materials
    RKPF Document Header- Reservation
    T023 Mat. groups
    T024 Purchasing Groups
    T156 Movement Type
    T157H Help Texts for Movement Types
    MOFF Lists what views have not been created
    A501 Plant/Material
    EBAN Purchase Requisition
    EBKN Purchase Requisition Account Assignment
    EKAB Release Documentation
    EKBE History per Purchasing Document
    EKET Scheduling Agreement Schedule Lines
    EKKN Account Assignment in Purchasing Document
    EKKO Purchasing Document Header
    EKPO Purchasing Document Item
    IKPF Header- Physical Inventory Document
    ISEG Physical Inventory Document Items
    LFA1 Vendor Master (General section)
    LFB1 Vendor Master (Company Code)
    NRIV Number range intervals
    RESB Reservation/dependent requirements
    T161T Texts for Purchasing Document Types
    C.) and about the fields it depends on transactions
    D.) PP Process Flow
    The PP flow consists of planning and execution as described. 
    Here is the flow of PP in simple language:
    First of all demand is generated for the product to be produced through demand management. It calculated the quantities to produce and the time for the final assembly. 
    The data from demand management is transferred to MRP.   The data from demand mgt can come from Customer Sales Order (firm) or Planned Order (forecast).
    The MRP checks for the availability of material at various stages of BOM. In case material is not available MRP generated planned order and Purchase requisition for production of components in house and procurement from vendor respectively.   
    The planned orders are converted into production order by the planner and the purchase requisition are converted into purchase order by the purchaser.
    Production in charge converts the planned order to production order and carries out the production activities once the order is released for production. Production is carried out as per the operation steps provided in the routing. Work centers are also mentioned with each operation in the routings where these operations are to be carried out. 
    Once the production is completed production confirmation is done and goods movement (delivery created) takes place. 
    Capacity planning is yet another part of PP which planes the capacities for various work centers. MPS is another tool which is used to plan out the master products. If MPS is to be run the inputs are provided from demand management to MPS. And the out put from MPS is then feed into MRP.
    E.) PP Module important Tables
    Demand Management
    PBED          Independent Requirements Data
    PBIM          Independent Requirements by Material
    Repetitive Manufacturing
    SAFK         RS Header Master Data
    S025           LIS -- Run Schedule Quantities
    S026           LIS -- Material Usage
    S028           LIS -- Reporting Point Statistics
    CEZP          Reporting Point Document Logs
    CPZP          Reporting Points - Periodic Totals
    MRP Records
    MDKP        MRP Document Header Data
    MDTB         MRP Table Structure (no data)
    PLSC          Planning Scenario (Long-term Planning)
    MDFD        MRP Firming Dates
    MDVM       Planning File Entries
    S094           LIS -- Stock/Requirements Analysis
    Reservations
    RESB          Reservations/Dependent Requirements
    Planned Orders
    PLAF          Planned Orders
    Discrete Production
    AFKO         Order Header
    AFPO          Order Item Detail
    AFVC          Order Operations Detail
    AFFL           Order Sequence Details
    AFFH          Order PRT Assignment
    AFBP          Order Batch Print Requests
    AFRU          Order Completion Confirmations
    AFFW         Confirmations -- Goods Movements with Errors
    AFRC          Confirmations -- Incorrect Cost Calculations
    AFRD          Confirmations -- Defaults for Collective Confirmation
    AFRH          Confirmations -- Header Info for Confirmation Pool
    AFRV          Confirmation Pool
    AFWI          Confirmations -- Subsequently Posted Goods Movements
    Classification
    KLAH          Class Detail
    CABN          Characteristic Detail
    AUSP          Characteristic Values
    CAWN          Characteristic Values
    CAWNT         Characteristic Value Texts
    KSML          Characteristic Allocation to Class
    KSSK          Material Allocation to Class
    Regards,
    Suresh.S

  • How to export / import tables from UAT Environment to Devlopment  in oracle

    i wanted to share some points about export/import tables in oracle.
    we have two environments one is development and integrated user acceptance test (UAT).
    we have finalized table structures with data in UAT. i wanted to move from UAT to Development.
    i found some commends in google. I tried but I am not able to export and then its throwing some errors.
    i am explaining below what i want to do and what i want did.
    Please provide solution what i need to do.
    admin credentials in UAT
    username :venkat_admin
    password :venkat_admin
    UAT user/schema
    username : ods_inb_stg
    password : ods_inb_stg
    development environment credentials
    admin user name : dev_admin
    password : dev_admin
    Development user/schema
    username :ods_inb_dev
    password :ods_inb_dev
    Above mentioned schema names i want move data from UAT (ods_inb_stg) to development (ods_inb_dev)
    I logged in uat ods_inb_stg user and then
    I used this command
    Exp ods_inb_stg/ods_inb_stg file=party_pt_address.expdat tables=(party,party_address)
    log=party_pt_address.log;
    getting error
    Unknown command beginning "exp ods_inb_stg/..." - rest of line ignored.
    What should I do any one help.
    One more thing i have more than 100 tables how to export all tables and which is the easiest way and how to do it.
    Thanks in Advance
    Venkat

    Hi,
    based import and export we need to do.
    suppose we think we have created tables in production and data also populated. we need to export the tables with data. i am able to export the tables with data some tables or full database.
    suppose we have development environment. in development environment there is no tables. we can treat nothing in development. we have only username,password and hostname.
    when you import the tables production to development automatically create the table structures with data.
    where ever we wanted to import the tables no need to create table structures. thats why no need to think about table structures.

  • JDBC Lookup - Import table data from a different schema in same DB

    Hi XI Experts,
    We are facing an issue while importing a Database table into the external definition in PI 7.1.
    The details are as below:
    I have configured user 'A' in PI communication channel to access the database. But the table that I want to access is present in schema "B". Due to this, I am unable to view the table that I have to import in the list available.
    In other words, I am trying to access a table present in a different schema in the same database. Please note that my user has been given all the required permissions to access different schema. Even then, I am unable to access the table in different schema.
    Kindly provide your valuable suggestions as to how I can import table which is present in another schema but in the same Database.
    Regards,
    Subbu

    If you are using PI 7.1, then you can do JDBC Lookup to import JDBC meta data (table structures from DB). Configure a jdbc receiver communication channel where you specify username and password which has permission to access schema A and Schema B of database. Specify database name in the connection string. Then you might have access to import both schema.
    Please refer these links
    SAP PI 7.1 Mapping Enhancements Series: Graphical Support for JDBC and RFC Lookups
    How to use JDBC Lookup in PI 7.1 ?

  • 11gR2 Data Pump. Import table in one schema into a different schema

    I stady oracle. I export shema HR into file hrexport.dmp. When I import tables from this file I had trouble. I used Enterprise Manager:
    1. connected by user SYSTEM as NORMAL
    2. selected file, chosen import type - tables
    3. data from file was imported
    4. chosen tables which tables to import
    5. in the next step I try insert row in table Re-Map Shemas and edit cell Destination Shema, but in the list is only one name of shema - HR! Why?
    Edited by: alvahtin on 10.03.2013 6:11

    Paul M. wrote:
    Did you try using impdp command at OS prompt ?I don't know what this command use. I try:
    impdp system/oracle remap_schema=hr:inventory tables=employees, departments, locations directory=ORACLES_HOME dumpfile=hrexport.dmp logfile=hrimport.log
    But it finished with error:
    ORA-39166: Object SYSTEM.EMPLOYEES was not found.
    ORA-39166: Object SYSTEM.DEPARTMENTS was not found.
    ORA-39166: Object SYSTEM.LOCATIONS was not found.
    Job "SYSTEM"."SYS_IMPORT_TABLE_01" successfully completed at 18:43:19
    Job "SYSTEM"."SYS_IMPORT_TABLE_01" successfully completed at 18:43:19
    Why SYSTEM? I want import to INVENTORY from HR - replaced in the file hrexport.dmp

  • How can i import tables from a different schema into the existing relational model... to add these tables in the existing model? plss help

    how can i import tables from a different schema into the existing relational model... to add these tables in the existing relational/logical model? plss help
    note; I already have the relational/logical model ready from one schema... and I need to add few more tables to this relational/logical model
    can I import the same way as I did previously??
    but even if I do the same how can I add it in the model?? as the logical model has already been engineered..
    please help ...
    thanks

    Hi,
    Before you start, you should probably take a backup copy of your design (the .dmd file and associated folder), in case the update does not work out as you had hoped.
    You need to use Import > Data Dictionary again, to start the Data Dictionary Import Wizard.
    In step 1 use a suitable database connection that can access the relevant table definitions.
    In step 2 select the schema (or schemas) to import.  The "Import to" field in the lower left part of the main panel allows you to select which existing Relational Model to import into (or to specify that a new Relational Model is to be created).
    In step 3 select the tables to import.  (Note that if there are an Foreign Key constraints between the new tables and any tables you had previously imported, you should also include the previous tables, otherwise the Foreign Key constraints will not be imported.)
    After the import itself has completed, the "Compare Models" dialog is displayed.  This shows the differences between the model being imported and the previous state of the model, and allows you to select which changes are to be applied.
    Just selecting the Merge button should apply all the additions and changes in the new import.
    Having updated your Relational Model, you can then update your Logical Model.  To do this you repeat the "Engineer to Logical Model".  This displays the "Engineer to Logical Model" dialog, which shows the changes which will be applied to the Logical Model, and allows you to select which changes are to be applied.
    Just selecting the Engineer button should apply all the additions and changes.
    I hope this helps you achieve what you want.
    David

Maybe you are looking for