Issues while migrating data from a disk based table to a memory optimized table

Hi All,
I have a Disk based table with 400000000 rows in it, We are trying to convert it into a memory optimized table.
We have  already created a memory optimized table with similar structure and trying to import data into this mem optimized table using 'insert into' from the disk table.
I am trying to Migrate around 10000000 rows at a time,  but I am getting an error 'There is insufficient system memory in resource pool 'default' to run this query.' Altough we have 128 GB RAM on the server and SS is utilizing more than 120 GB RAM.
Altough the query has been cancelled.
Wanted to Know how could we migrate the table with the available RAM or do we have increase our RAM?
aa

Josh,
Microsoft's documentation on this subject isn't at its best right now (I believe there will be incremental improvements for better understanding), but here is what I read so far.
http://msdn.microsoft.com/en-us/library/dn133190.aspx
"A hash index consists of a collection of buckets organized in an array. A hash function maps
index keys to corresponding buckets in the hash index."
Judging by this statement, a hash index is a hash table just like the ones used as work tables for hash operators in queries (hash matching or grouping). Doesn't contain (or include) other columns, i. e. it doesnt store any data.
"Multiple index keys may be mapped to the same hash bucket."
This means there is some kind of mapping, but this is not explained in the article above. However...
http://msdn.microsoft.com/en-us/library/dn282389.aspx
"For each hash index in the table, each row has an 8-byte address pointer to the next row in the index. Since there are 4 indexes, each row will allocate 32 bytes for index pointers (an 8 byte pointer for each index)."
Each row (in the table) has a pointer (for every index, 1:1 ratio) that points to a row (also known as bucket) in the hash index. So that is how the aforementioned mapping works huh!
> What happens if you include a column in two or three different indexes, or is that not allowed?
My conclusion is that the hash indexes works the same way as a hash worktable, with the addition of the column in the base table that is added to store pointers to the hash index.
When you create a new index, even if you use the same column twice, a hash table is created, hash calculations are distinctly made for each key and stored on it, and while this is done, the column that is exclusively used for this new index is populated
with pointers to this index. You can add a given column to the set of keys of different hash indexes as many times as you want. Correct if i'm wrong, I'm also new on this subject :D

Similar Messages

  • Special character issue while loading data from SAP HR through VDS

    Hello,
    We have a special character issue, while loading data from SAP HR to IdM, using a VDS and following the standard documentation: http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e09fa547-f7c9-2b10-3d9e-da93fd15dca1?quicklink=index&overridelayout=true
    French accent like (é,à,è,ù), are correctly loaded but Turkish special ones (like : Ş, İ, ł ) are transformed into u201C#u201D in Idm.
    The question is : does someone know any special setting to do in the VDS or in IdM for special characters to solve this issue??
    Our SAP HR version is ECC6.0 (ABA/BASIS7.0 SP21, SAP_HR6.0 SP54) and we are using a VDS 7.1 SP5 and SAP NW IdM 7.1 SP5 Patch1 on oracle 10.2.
    Thanks

    We are importing directly to the HR staging area, using the transactions/programs "HRLDAP_MAP", "LDAP" and "/RPLDAP_EXTRACT", then we have a job which extract data from the staging area to a CSV file.
    So before the import, the character appears correctly in SAP HR, but by the time it comes through the VDS to the IDM's temporary table, it becomes "#".
    Yes, our data is coming from a Unicode system.
    So, could it be a java parameter to change / add in the VDS??
    Regards.

  • Deadlock issue while sending data from PI to JDBC !!!

    Hi All,
    We are getting below error while sending data from PI to JDBC (Database).
    Error - JDBC message processing failed; reason Error processing request in sax parser: Error when executing statement for table/stored proc. 'store_order_details' (structure 'Order_details_Statement'): com.microsoft.sqlserver.jdbc.SQLServerException: Transaction (Process ID 61) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction.
    This is happening when more than one PI message is trying to deliver data at the same time to database.
    Can anyone let us know the root cause of this issue and how we can rectify this.
    Is there some setting which we can use to do parallel processing as Delivering one message at a time to database is causing performance issues & taking lot of time.
    Thanks
    Neha Verma

    Hello Neha,
    Strange but can please get below information.
    Please check with the DB admin about if the user is getting locked or is there any hanging threads related to user.
    Also confirm with DB admin if the exclusive lock is on table or on the row when you try insertign or updating information.
    You can share the user from the receiver channel.
    Regards,
    Hiren A.

  • Facing Issue while Extracting Data from Essbase using CalcScript

    Hi All,
    we are getting this error while extracting data from Essbase by CalcScript using ODI.
    Does anyone have the answer
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 1, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Cannot calculate file. Essbase Error(1030009): Name too long (C:\NewFolder\ACAD) in ESSAPI function EssCalcFile
         at com.hyperion.odi.essbase.ODIEssbaseDataReader.getAppData(Unknown Source)
         at com.hyperion.odi.essbase.AbstractEssbaseReader.extract(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx5046.f$0(<string>:1)
         at org.python.pycode._pyx5046.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlC.treatTaskTrt(SnpSessTaskSqlC.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.k(e.java)
         at com.sunopsis.dwg.cmd.h.A(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Thread.java:595)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Cannot calculate file. Essbase Error(1030009): Name too long (C:\NewFolder\ACAD) in ESSAPI function EssCalcFile
         at com.hyperion.odi.essbase.wrapper.EssbaseCalcDataIterator.init(Unknown Source)
         ... 33 more
    Caused by: com.essbase.api.base.EssException: Cannot calculate file. Essbase Error(1030009): Name too long (C:\NewFolder\ACAD) in ESSAPI function EssCalcFile
         at com.essbase.server.framework.EssOrbPluginDirect.ex_olap(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect.essMainCalcFile(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMainMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod2(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect._invokeProtected(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin.essMainCalcFile(Unknown Source)
         at com.essbase.api.datasource.EssCube.calculate(Unknown Source)
         at com.hyperion.odi.essbase.wrapper.EssbaseApplication.executeCalculationScript(Unknown Source)
         ... 34 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Cannot calculate file. Essbase Error(1030009): Name too long (C:\NewFolder\ACAD) in ESSAPI function EssCalcFile
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlC.treatTaskTrt(SnpSessTaskSqlC.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.k(e.java)
         at com.sunopsis.dwg.cmd.h.A(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Thread.java:595)

    Hi Now i am getting this error.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 1, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Cannot calculate file. Essbase Error(1030214): User [admin] cannot access calc script: ACAD
         at com.hyperion.odi.essbase.ODIEssbaseDataReader.getAppData(Unknown Source)
         at com.hyperion.odi.essbase.AbstractEssbaseReader.extract(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx5054.f$0(<string>:1)
         at org.python.pycode._pyx5054.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlC.treatTaskTrt(SnpSessTaskSqlC.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.k(e.java)
         at com.sunopsis.dwg.cmd.h.A(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Thread.java:595)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Cannot calculate file. Essbase Error(1030214): User [admin] cannot access calc script: ACAD
         at com.hyperion.odi.essbase.wrapper.EssbaseCalcDataIterator.init(Unknown Source)
         ... 33 more
    Caused by: com.essbase.api.base.EssException: Cannot calculate file. Essbase Error(1030214): User [admin] cannot access calc script: ACAD
         at com.essbase.server.framework.EssOrbPluginDirect.ex_olap(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect.essMainCalcFile(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMainMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod2(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect._invokeProtected(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin.essMainCalcFile(Unknown Source)
         at com.essbase.api.datasource.EssCube.calculate(Unknown Source)
         at com.hyperion.odi.essbase.wrapper.EssbaseApplication.executeCalculationScript(Unknown Source)
         ... 34 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Cannot calculate file. Essbase Error(1030214): User [admin] cannot access calc script: ACAD
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlC.treatTaskTrt(SnpSessTaskSqlC.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.k(e.java)
         at com.sunopsis.dwg.cmd.h.A(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Thread.java:595)

  • Issue while loading data from DSO to Cube

    Gurus,
    I'm facing a typical problem while loading Cube from a DSO. The load is based upon certain conditions.
    Though the data in DSO satisfies those condition, but the Cube is still not being loaded.
    All the loads are managed through Process Chains.
    Would there be any reason in specific/particular for this type of problem?
    Any pointers would be of greatest help !
    Regards,
    Yaseen & Soujanya

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Issue while uploading data from flatfile to Transaction(VK13).

    Hi All,
    I am facing an issue while uploading the data from flatfile to the transaction(vk13). My flat file is as shown below.
    SalesOrganization    DistributionChannel    Customer    Material    Releasestatus    Amount    Currency    Quantity    UOM    ValidFrom    ValidTo
    2000    01    0010029614    AT309739    A    20.00    USD    1    PC    09/11/2014    12/31/9999
    If I upload these data using the RV_CONDITION_COPY  FM it is succesfully uploading to the relevant tables(konh,konp) but in the tcode VK13  I am getting all values fine, except UOM. Instead of PC it is showning as ***. I did not understand why it is happening please give an idea to solve my issue.
    Regards,
    Chakradhar.

    Hi Raymond,
    Thanks for your reply.Yes,If I use CONVERSION_EXIT_CUNIT_INPUT in my program the issue is, Assume If the user is giving PC as value for UOM field in flat file and upload the flat file.It is successfully uploading the value PC to the UOM field in transaction VK13 but the in the database table(konp) it is showing the value as ST.
    Regards,
    Chakradhar.

  • Linking issues while Migrating code from SUN CC 4.3 to SUN CC 5.8

    Hello,
    I am migrating code written in SUN CC 4.3 to SUN CC 5.8
    I have compiled entire code with SUN CC 5.8 with -compact=4 option ,on machine with OS SunOS 5.8 Generic_Virtual sun4v sparc sun4v.
    during linking phase i am facing issues like
    (cd msggen; make CG="-fast -O3 -compat=4 -lC -features=extensions -i" all)
    /opt/SUNWspro/bin/CC -o msggen msggen.o licensemessage.o ../../../../libs/lib/libgeneral.a
    Undefined first referenced
    symbol in file
    cerr ../../../../libs/lib/libgeneral.a(string.o)
    unsafe_ostream::operator<<(long) ../../../../libs/lib/libgeneral.a(string.o)
    operator delete(void*) msggen.o
    operator new(unsigned int) licensemessage.o
    Iostream_init::Iostream_init(void) msggen.o
    Iostream_init::~Iostream_init(void) msggen.o
    istream::operator>>(char*) ../../../../libs/lib/libgeneral.a(string.o)
    setw(int) ../../../../libs/lib/libgeneral.a(string.o)
    unsafe_ostream::outstr(const char*, const char*) ../../../../libs/lib/libgeneral.a(string.o)
    exrethrow_q msggen.o
    ld: fatal: Symbol referencing errors. No output written to msggen
    I felt like these issues are due to the standard header files. I am including the headers which comes with SUN CC 5.8 and I am using Rogue Wave rwav/ver09.1.
    I am not getting any pointers to solve this issue,any pointer will be highly appriciated.
    Thanks in advance,
    Revathi R

    975741 wrote:
    Yes i did migration by following the Migration Guide, while Migration i ve come across situations where a class method is invisible at many situations like...
    C.h
    class C
    f1();
    B.h
    #include<C.h>
    class B
    f2();
    A.C
    #include<B.h>
    class A:B
    f1();-----> Error Not accessible so I changed to C::f1() then worked like this I ve came across many will it be a Problem, in migration Guide i haven't seen any of such.
    }When you post a code example, it is best to copy/paste from actual code that you verified shows the problem. This example is nowhere near valid for any compiler, even with your suggested change.
    - The class declarations are not terminated by a semicolon
    - You can't call C::f1() using that syntax outside of class C or a class derived from class C because it is not a static function, and requires an object. You could call it as c.f1() if you had a c object of type C, except that f1() would still not be accessible.
    The rules about accessibility have never changed. Please post an example that actually works with C++ 4.x or with C++ 5.x in -compat=4 mode, but that fails with -compat=5 mode.
    >
    Like this for operator overloading also , I haven't got any way to make function body available to function call, so i ve changed those functions to normal functions...I don't understand what you mean. An example would help.
    >
    what is a template cache?? How to delete it??
    The template cache is also known as the template repository. See section 3.3.5 Template Repository in the Migration Guide. To delete the cache (repository) just remove the entire cache directory.
    >
    >
    Another task of mine is, i am compiling a code written in SUN CC 5.5 to 5.8, would you suggest me the migration guide, I could get from 4 to 5 and 3 to 5 but
    not from 5.5 to 5.8, I am not sure about in which compiler code is written but, I could able to compile the code with out compact option with SUN CC 5.5, CC 5.5 and 5.8 are fully compatible. There should be no migration issues. The exceptions would be invalid code that compiled due to a bug in the earlier compiler but is correctly diagnosed as an error with the later compiler. In rare cases, the later compiler has a bug not present in the earlier compiler (a regression). When found, regressions are fixed in patches. If you have an Oracle service contract, you can download patches from My Oracle Support. Otherwise, get the most recent compiler, which should not have those regressions.
    but while trying to compile with 5.8 i am facing issues like
    1. typecastingI would have to see an example.
    2.virtual vector<int> getProcedureArgs() { return 0; } -----> Error: Cannot use int to initialize std::vector<int>
    The compiler error message is correct. The function returns a vector<int>, but you can't create a vector<int> from an integer 0. That is, the vector template has several constructors, but none that accept a single integer argument.
    At this point we have moved away from the original topic. If you have further questions, please start a new thread.

  • Issue while uploading data from R/3 production to BW production

    Hi,
    We are facing an issue as follows.
    We are trying to transfer master data from R/3 Production server to BW Production server.
    For the first time when we transfer data, upload is happening properly. But for the second upload - we dont get any error where as in the monitor data transfer shows 0 from 0 records and the status never changes.
    But if we restart the BW production server and then upload data, it works fine for the first upload but the same issue from second upload.
    Any suggestions please?
    Regards,
    Ram.

    Hi,
    Roberto - In ST22, there is no short dumps related to this issue. The issue is not specific for a single object. For eg. If i upload 0MATERIAL (after restarting the server) it works. Now I try to upload 0DIVISION (any object) it doesnot get uploaded.
    Sriram - We are trying Full Upload.
    Hari - I can see the records in RSA3 of the source system. The records are available and also they are extracted properly.
    Shreya - Basis people are also involved and trying to find out a solution.
    Manfred - Actually data is available. But the data doesnot get uploaded. So i guess changing the traffic lights settings might not help.
    Regards,
    Ram.

  • Performance issue while extracting data from non-APPS schema tables.

    Hi,
    I have developed an ODBC application to extract data from Oracle E-Business Suite 12 tables.
    e.g. I am trying to extract data from HZ_IMP_PARTIES_INT table of Receivables application(Table is in "AR" database schema) using this ODBC application.
    The performance of extraction (i.e. number of rows extracted per second) is very low if the table belongs to non-APPS schema. (e.g. HZ_IMP_PARTIES_INT belongs to "AR" database schema)
    Now if I create same table (HZ_IMP_PARTIES_INT) with same data in "APPS" schema, the performance of extraction improves a lot. (i.e. the number of rows extracted per second increases a lot) with the same ODBC application.
    The "APPS" user is used to connect to the database in both scenarios.
    Also note that my ODBC application creates multiple threads and each thread creates its own connection to the database. Each thread extract different data using SQL filter condition.
    So, my question is:
    Is APPS schema optimized for any data extraction?
    I will really appreciate any pointer on this.
    Thanks,
    Rohit.

    Hi,
    Is APPS schema optimized for any data extraction? I would say NO as data extraction performance should be the same for all schemas. Do you run "Gather Schema Statistics" concurrent program for ALL schemas on regular basis?
    Regards,
    Hussein

  • Performance issue while transferring data from one itab to another itab

    hi experts,
    i have stored all the general material details in one internal table and description, valuation details of the meterial is stored in another internal table which is of standard table. now i need to tranfer all the data from these two internal tables into one final internal table but it is taking lot of time as it has to transfer lacs of data.
    i have declared the output table as shown below
    DATA:
      t_output TYPE standard TABLE
               OF type_output
               INITIAL SIZE 0
               WITH HEADER LINE.
    (according to the standard i have to declare like this and the two internal tables are declared similar to the above one)
    could somebody suggest me how shall I procced through this...
    thanks in advance....
    ragerds,
    Deepu

    Have a look at the following article which you may find useful:
      <a href="https://www.sdn.sap.comhttp://www.sdn.sap.comhttp://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/40729734-3668-2910-deba-fa0e95e2c541">Improving performance in nested loops</a>
    good luck, damian

  • Performance issues while query data from a table having large records

    Hi all,
    I have a performance issues on the queries on mtl_transaction_accounts table which has around 48,000,000 rows. One of the query is as below
    SQL ID: 98pqcjwuhf0y6 Plan Hash: 3227911261
    SELECT SUM (B.BASE_TRANSACTION_VALUE)
    FROM
    MTL_TRANSACTION_ACCOUNTS B , MTL_PARAMETERS A  
    WHERE A.ORGANIZATION_ID =    B.ORGANIZATION_ID 
    AND A.ORGANIZATION_ID =  :b1 
    AND B.REFERENCE_ACCOUNT =    A.MATERIAL_ACCOUNT 
    AND B.TRANSACTION_DATE <=  LAST_DAY (TO_DATE (:b2 ,   'MON-YY' )  )  
    AND B.ACCOUNTING_LINE_TYPE !=  15  
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      3      0.02       0.05          0          0          0           0
    Fetch        3    134.74     722.82     847951    1003824          0           2
    total        7    134.76     722.87     847951    1003824          0           2
    Misses in library cache during parse: 1
    Misses in library cache during execute: 2
    Optimizer mode: ALL_ROWS
    Parsing user id: 193  (APPS)
    Number of plan statistics captured: 1
    Rows (1st) Rows (avg) Rows (max)  Row Source Operation
             1          1          1  SORT AGGREGATE (cr=469496 pr=397503 pw=0 time=237575841 us)
        788242     788242     788242   NESTED LOOPS  (cr=469496 pr=397503 pw=0 time=337519154 us cost=644 size=5920 card=160)
             1          1          1    TABLE ACCESS BY INDEX ROWID MTL_PARAMETERS (cr=2 pr=0 pw=0 time=59 us cost=1 size=10 card=1)
             1          1          1     INDEX UNIQUE SCAN MTL_PARAMETERS_U1 (cr=1 pr=0 pw=0 time=40 us cost=0 size=0 card=1)(object id 181399)
        788242     788242     788242    TABLE ACCESS BY INDEX ROWID MTL_TRANSACTION_ACCOUNTS (cr=469494 pr=397503 pw=0 time=336447304 us cost=643 size=4320 card=160)
       8704356    8704356    8704356     INDEX RANGE SCAN MTL_TRANSACTION_ACCOUNTS_N3 (cr=28826 pr=28826 pw=0 time=27109752 us cost=28 size=0 card=7316)(object id 181802)
    Rows     Execution Plan
          0  SELECT STATEMENT   MODE: ALL_ROWS
          1   SORT (AGGREGATE)
    788242    NESTED LOOPS
          1     TABLE ACCESS   MODE: ANALYZED (BY INDEX ROWID) OF
                    'MTL_PARAMETERS' (TABLE)
          1      INDEX   MODE: ANALYZED (UNIQUE SCAN) OF
                     'MTL_PARAMETERS_U1' (INDEX (UNIQUE))
    788242     TABLE ACCESS   MODE: ANALYZED (BY INDEX ROWID) OF
                    'MTL_TRANSACTION_ACCOUNTS' (TABLE)
    8704356      INDEX   MODE: ANALYZED (RANGE SCAN) OF
                     'MTL_TRANSACTION_ACCOUNTS_N3' (INDEX)
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      row cache lock                                 29        0.00          0.02
      SQL*Net message to client                       2        0.00          0.00
      db file sequential read                    847951        0.40        581.90
      latch: object queue header operation            3        0.00          0.00
      latch: gc element                              14        0.00          0.00
      gc cr grant 2-way                               3        0.00          0.00
      latch: gcs resource hash                        1        0.00          0.00
      SQL*Net message from client                     2        0.00          0.00
      gc current block 3-way                          1        0.00          0.00
    ********************************************************************************On a 5 node rac environment the program completes in 15 hours whereas on a single node environemnt the program completes in 2 hours.
    Is there any way I can improve the performance of this query?
    Regards
    Edited by: mhosur on Dec 10, 2012 2:41 AM
    Edited by: mhosur on Dec 10, 2012 2:59 AM
    Edited by: mhosur on Dec 11, 2012 10:32 PM

    CREATE INDEX mtl_transaction_accounts_n0
      ON mtl_transaction_accounts (
                                   transaction_date
                                 , organization_id
                                 , reference_account
                                 , accounting_line_type
    /:p

  • Semantic Partitioning Delta issue while load data from DSO to Cube -BW 7.3

    Hi All,
    We have created the semantic partion with the help of BADI to perform Partitions. Everthing looks good,
    first time i have loaded the data it is full update.
    Second time i init the data load. I pulled the delta  from DSO to Cube . DSO is standard where as Cube is partition with the help of semantic partition. What i can see is the records which are updated in Latest delta are shown into the report rest all are ignored by the system.
    I tried compression but still it did not worked.
    Can some has face this kind
    Thanks

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Issue while retrieving data from BAPI

    Problem:
    The initial exception that caused the request to fail, was:
    com.sap.dictionary.runtime.DdException: Wrong amount type (not decimal): Unit service cannot be instantiated
    at com.sap.dictionary.runtime.DdBroker.getUnitService(DdBroker.java:215)
    at com.sap.dictionary.runtime.DdBroker.getUnitService(DdBroker.java:233)
    at com.sap.tc.webdynpro.services.datatypes.core.DataTypeBroker.getUnitService(DataTypeBroker.java:337)
    at com.sap.tc.webdynpro.progmodel.context.DataAttributeInfo.initReferenceAttribute(DataAttributeInfo.java:342)
    at com.sap.tc.webdynpro.progmodel.context.NodeInfo.initStructureType(NodeInfo.java:708)
    ... 30 more
    I have a BAPI  - ZPM_CREATE_ORDER. I am trying to write a record to that BAPI. Did the Model Import from R/3[including the Commit ].
    The structures of BAPI are like this:
    Zpm_create_Order_Input
            Order         -  Zpm_Hdr
         Operations    -  [ Line Items structure ]
    I have created the mapping between Model & Custom Controller with Order & Operations. Did the mapping between the Custom Controller & the View Controller. Binded them to the UI elements from View Controller with an input form.
    But there are ABAP - decimal fields under Operations -  price & Amount  which are a Currency field and Quantity . So when I just try to run the application, I get the amount mismatch error. When I say run, I could n’t even see the GUI even. I get the 500 error with the above problem.
    I read in the forums that I need to declare as a Simple type and re cast the fields to right type and bind that structure back. How ever, I could n’t get the complete understanding of it.
    For a Retrieval Bapi with the exact same structure, the same issue if I map the Operations. How ever, if I skip the Operations and just map the Order i.e., header, then I could retrieve the records.
    Interestingly even if I just map a String(char) field under Operations structure I am getting teh same error.
    Any help is appreciated. Thanks in advance...

    Satyajit,
    You are right. One of the fields needed is Waers(currency) which I have mapped from the Model to Custom Controller. I think, that is one of the causes but there are other fields too (for instance, Decimal type).
    The surprising thing to me is that, even if I just map the character fields in that structure and not map the Decimal or Currency fields, then also I am getting the same error.
    Thanks for the Reply...
    --Vivek

  • While uploading data from local file

    Hi,
    While uploading data from local file having the header text into interal table using the GUI_UPLOAD function module,there is problem in upload because of header text in file, do we have any option in this FM itself to skip this header text and upload from sdecond line because as per our requirement we can't delete this header text from the file.

    Hi Manish,
    Do you have the problem while uploading? Is it giving any error?
    If not, if your uploading is successful into the internal table, then delete the first row, which is header, from the internal table using index eq 1.
    Regards,
    Chandra Sekhar

  • Raw Read Error Rate, Errors occurred while reading raw data from a disk

    hi
    when I turn my pavilion dv6 2170ee computer on screen shows below error.
    system diognostics (F2) does not any work
    and when my windows7 comming up this error was shown.
    I install Hard Disk Sentinel software and this program shows below:
    Failure Predicted - Attribute: 1 Raw Read Error Rate, Errors occurred while reading raw data from a disk. Indicate problem with the disk surface or the read/write heads.
    There are 28 weak sectors found on the disk surface. They may be remapped any time in the later use of the disk.
    26848 errors occured during data transfer. This may indicate problem of the device or with data/power cables. It is recommended to examine and replace the cables if possible.
    Replace hard disk immediately.
    how can I fix it?
    may drive format (Fdisk) and renewing windows can fix it?
    tanx and regards
    Mohammad

    Hello mohkhosh,
    I see that you're getting an error indicating possible hard driver failure, that you have this message from 3 different sources. I will try to help you with this.
    After reviewing everything  you're looking at needing to replace the hard drive. I would suggest contacting support to see what they can do for you.
    HP Technical Support, North America 1-800-474-6836. For all other regions click here.
    Thank you for posting on the HP Forums.
    I worked on behalf of HP.

Maybe you are looking for