SQL Memory issue

I am storing files in sql lite, but when i store large file 400+Mb system throw this exception
Error #1000: The system is out of memory.
    at flash.data::SQLStatement/internalExecute()
    at flash.data::SQLStatement/execute()
Please give suggestion
Thanks

One sql server box:
SYstem administrator pointed out: SQL server almost uses all memory and server memory usage almost hits 100%
How to fix this problem?
This is not a problem but normal behavior.SQL Server will use as much memory you provide to it.So its good to set proper value for max server memory setting in sp_configure.See example in below MS link . Set proper  memory usage for SQL Server leaving
enough RAM (4-5 G ) for OS.This will only include memory set for buffer pool there are memory allocations outside buffer pool which is directly done by OS
http://technet.microsoft.com/en-us/library/ms178067.aspx
To read more about memory
http://social.technet.microsoft.com/wiki/contents/articles/22316.sql-server-memory-and-troubleshooting.aspx
Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

Similar Messages

  • PL/SQL memory issues

    Hi,
    is there any way to test how many memory is consumed by PL/SQL program? I need to test memory consumption for examples listed below. Many many thanks for solution. Cejen
    declare
    type dci_test1 is table of varchar2(4000) index by binary_integer;
    type ds_test1 is record (
    i1 varchar2(4000),
    i2 varchar2(4000),
    i3 varchar2(4000)
    sTest ds_test1; -- How many memory is allocated now? 12kB + overhead or 0kB + overhead ?
    ciTest dci_test;
    begin
    sTest.i1 := 'A'; -- How many memory is allocated now? 12kB + overhead or 4kB + overhead or 1B + overhead ?
    ciTest(1) := 'A'; -- How many memory is allocated now? 4kB + overhead or 1B + overhead ?
    end;
    /

    Quick google search..
    http://oracle-online-help.blogspot.com/2006/11/precaution-while-defining-data-types-in.html

  • Result Set Causing out of memory issue

    Hi,
    I am having trouble to fix the memory issue caused by result set.I am using jdk 1.5 and sql server 2000 as the backend. When I try to execute a statement the result set returns minimum of 400,000 records and I have to go through each and every record one by one and put some business logic and update the rows and after updating around 1000 rows my application is going out of memory. Here is the original code:
    Statement stmt = con.createStatement();
    ResultSet   rs = st.executeQuery("Select * from  database tablename where field= 'done'");
                while(rs!=null && rs.next()){
                System.out.println("doing some logic here");
    rs.close();
    st.close();
    I am planning to fix the code in this way:
    Statement stmt = con.createStatement(ResultSet.TYPE_FORWARD_ONLY,
                          ResultSet.CONCUR_UPDATABLE);
    stmt.setFetchSize(50);
    ResultSet   rs = st.executeQuery("Select * from  database tablename where field= 'done'");
                while(rs!=null && rs.next()){
                System.out.println("doing some logic here");
    rs.close();
    st.close();But one of my colleague told me that setFetchSize() method does not work with sql server 2000 driver.
    So Please suggest me how to fix this issue. I am sure there will be a way to do this but I am just not aware of it.
    Thanks for your help in advance.

    Here is the full-fledged code.There is Team Connect and Top Link Api being used. The code is already been developed and its working for 2-3 hours and then it fails.I just have to fix the memory issue. Please suggest me something:
    Statement stmt = con.createStatement();
    ResultSet   rs = st.executeQuery("Select * from  database tablename where field= 'done'");
                while(rs!=null && rs.next()){
                 /where vo is the value object obtained from the rs row by row     
                if (updateInfo(vo, user)){
                               logger.info("updated : "+ rs.getString("number_string"));
                               projCount++;
    rs.close();
    st.close();
    private boolean updateInfo(CostCenter vo, YNUser tcUser) {
              boolean updated;
              UnitOfWork unitOfWork;
              updated = false;
              unitOfWork = null;
              List projList_m = null;
              try {
                   logger.info("Before vo.getId() HERE i AM" + vo.getId());
                   unitOfWork = FNClientSessionManager.acquireUnitOfWork(tcUser);
                   ExpressionBuilder expressionBuilder = new ExpressionBuilder();
                   Expression ex1 = expressionBuilder.get("application")
                             .get("projObjectDefinition").get("uniqueCode").equal(
                                       "TABLE-NAME");
                   Expression ex2 = expressionBuilder.get("primaryKey")
                             .equal(vo.getPrimaryKey());// primaryKey;
                   Expression finalExpression = ex1.and(ex2);
                   ReadAllQuery projectQuery = new ReadAllQuery(FQUtility
                             .classForEntityName("EntryTable"), finalExpression);
                   List projList = (List) unitOfWork.executeQuery(projectQuery);
                   logger.info("list value1" + projList.size());
                   TNProject project_hist = (TNProject) projList.get(0); // primary key
                   // value
                   logger.info("vo.getId1()" + vo.getId());
                   BNDetail detail = project_hist.getDetailForKey("TABLE-NAME");
                   project_hist.setNumberString(project_hist.getNumberString());
                   project_hist.setName(project_hist.getName());
                   String strNumberString = project_hist.getNumberString();
                   TNHistory history = FNHistFactory.createHistory(project_hist,
                             "Proj Update");
                   history.addDetail("HIST_TABLE-NAME");
                   history.setDefaultCategory("HIST_TABLE-NAME");
                   BNDetail histDetail = history.getDetailForKey("HIST_TABLE-NAME");
                   String strName = project_hist.getName();
                   unitOfWork.registerNewObject(histDetail);
                   setDetailCCGSHistFields(strNumberString, strName, detail,
                             histDetail);
                   logger.info("No Issue");
                   TNProject project = (TNProject) projList.get(0);
                   project.setName(vo.getName());
                   logger.info("vo.getName()" + vo.getName());
                   project.setNumberString(vo.getId());
                   BNDetail detailObj = project.getDetailForKey("TABLE-NAME"); // required
                   setDetailFields(vo, detailObj);//this method gets the value from vo and sets in the detail_up object
                   FNClientSessionManager.commit(unitOfWork);
                   updated = true;
                   unitOfWork.release();
              } catch (Exception e) {
                   logger.warn("update: caused exception, "
                             + e.getMessage());
                   unitOfWork.release();
              return updated;
         }Now I have tried to change little bit in the code. And I added the following lines:
                        updated = true;
                     FNClientSessionManager.release(unitOfWork);
                     project_hist=null;
                     detail=null;
                     history=null;
                     project=null;
                     detailObj=null;
                        unitOfWork.release();
                        unitOfWork=null;
                     expressionBuilder=null;
                     ex1=null;
                     ex2=null;
                     finalExpression=null;
    and also I added the code to request the Garbage collector after every 5th update:
    if (updateInfo(vo, user)){
                               logger.info("project update : "+ rs.getString("number_string"));
                               projCount++;
                               //call garbage collector every 5th record update
                               if(projCount%5==0){
                                    System.gc();
                                    logger.debug("Called Garbage Collectory on "+projCount+"th update");
                          }But now the code wont even update the single record. So please look into the code and suggest me something so that I can stop banging my head against the wall.

  • Spry IE Memory issue

    Hi,
    I'm having a major headache with IE 7 caching issues at the
    moment.
    I've tried adding all the no-cache headers to both the
    generated xml, and the actual spry page to no avail. I've also
    tried using random characters to the xml file and changing the
    retrieval method to post rather than get, but still no joy.
    My page basically lists a group of monitored servers, and
    their current status, retrieved as xml output from an asp script
    accessing an sql server database. The spry reload time for the data
    is set at 10 seconds. On every refresh of the xml, the IE memory
    issue jumps, and doesn't seem to release anything. Everything works
    ok in Firefox, just good old IE 7 that i'm having issues with.
    Any ideas gratefully received.

    Hi uk_jim,
    Give this a try:
    http://www.infoaccelerator.net/index.cfm?event=showEntry&entryId=F22C2021-1372-FA49-994AC8 29431456FB
    --== Kin ==--

  • SQL memory manager latch at top of the list

    select * from V$LATCH order by wait_time desc shows wait time = 451524661079 on SQL memory manager latch. What is making this so high?

    user498912 wrote:
    Thank you .. the pointer to the bug is helpful. We are having performance issues and complaints of slowness only during about a 2 or 3 hour window each weekday from about 7AM to 10AM EDT. This is definitely peak load. AWR is showing mostly disk i/o issues. Plan is to add about 20 GB more memory on the server. SGA currently at around 61GB with Automatic Shared Memory Management enabled. PGA advice showing nothing with a Cache Hit Percentage (%) 95.93
    I would look at v$sga_resize_ops (or v$memory_resize_ops) to see if there is pressure to move memory around in that time period. Your symptoms could simply be showing load on the buffer cache causing the library cache to shrink leading to agressive demands for library cache locks etc. If so, then fixing a minum shared_pool_size (or generally going to manual memory management) may be the best bet. Also worth checking if there is some inefficient SQL running at that time that results in lots of random I/O - eliminating the cause of disk reads may be the best solution to reducing demand for memory.
    Regards
    Jonathan Lewis

  • A SQL tuning issue-sql runs much slower in test than in production?

    Hi Buddies,
    I am working on a sql tuning issue. A sql runs much slower in test than in production.
    I compared the two explain plans in test and production
    seems in test, CBO refuses to use index SUBLEDGER_ENTRY_I2.
    we rebuile it and re-gether that index statistcs. run, still slow..
    I compared the init.ora parameters like hash_area_size, sort_area_size in test, they are same as production.
    I wonder if any expert friend can show some light.
    in production,
    SQL> set autotrace traceonly
    SQL> SELECT rpt_horizon_subledger_entry_vw.onst_offst_cd,
    2 rpt_horizon_subledger_entry_vw.bkng_prd,
    3 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    4 rpt_horizon_subledger_entry_vw.jrnl_id,
    5 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    6 rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    7 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    8 rpt_horizon_subledger_entry_vw.crprt_melob_cd AS corp_mlb_cd,
    rpt_horizon_subledger_entry_vw.onst_offst_cd, SUM (amt) AS amount
    9 10 FROM rpt_horizon_subledger_entry_vw
    11 WHERE rpt_horizon_subledger_entry_vw.bkng_prd = '092008'
    12 AND rpt_horizon_subledger_entry_vw.jrnl_id = 'RCS0002100'
    13 AND rpt_horizon_subledger_entry_vw.systm_afflt_cd = 'SAFF01'
    14 GROUP BY rpt_horizon_subledger_entry_vw.onst_offst_cd,
    15 rpt_horizon_subledger_entry_vw.bkng_prd,
    16 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    17 rpt_horizon_subledger_entry_vw.jrnl_id,
    18 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    19 rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    20 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    21 rpt_horizon_subledger_entry_vw.crprt_melob_cd,
    22 rpt_horizon_subledger_entry_vw.onst_offst_cd;
    491 rows selected.
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=130605 Card=218764 B
    ytes=16407300)
    1 0 SORT (GROUP BY) (Cost=130605 Card=218764 Bytes=16407300)
    2 1 VIEW OF 'RPT_HORIZON_SUBLEDGER_ENTRY_VW' (Cost=129217 Ca
    rd=218764 Bytes=16407300)
    3 2 SORT (UNIQUE) (Cost=129217 Card=218764 Bytes=35877296)
    4 3 UNION-ALL
    5 4 HASH JOIN (Cost=61901 Card=109382 Bytes=17719884)
    6 5 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_ACCNT'
    (Cost=2 Card=111 Bytes=3774)
    7 5 HASH JOIN (Cost=61897 Card=109382 Bytes=14000896
    8 7 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACC
    OUNT' (Cost=2 Card=57 Bytes=1881)
    9 7 HASH JOIN (Cost=61893 Card=109382 Bytes=103912
    90)
    10 9 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cost=
    34 Card=4282 Bytes=132742)
    11 9 HASH JOIN (Cost=61833 Card=109390 Bytes=7000
    960)
    12 11 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGE
    R_ENTRY' (Cost=42958 Card=82076 Bytes=3611344)
    13 12 INDEX (RANGE SCAN) OF 'SUBLEDGER_ENTRY_I
    2' (NON-UNIQUE) (Cost=1069 Card=328303)
    14 11 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_
    LINK' (Cost=14314 Card=9235474 Bytes=184709480)
    15 4 HASH JOIN (Cost=61907 Card=109382 Bytes=18157412)
    16 15 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_ACCNT'
    (Cost=2 Card=111 Bytes=3774)
    17 15 HASH JOIN (Cost=61903 Card=109382 Bytes=14438424
    18 17 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACC
    OUNT' (Cost=2 Card=57 Bytes=1881)
    19 17 HASH JOIN (Cost=61899 Card=109382 Bytes=108288
    18)
    20 19 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cost=
    34 Card=4282 Bytes=132742)
    21 19 HASH JOIN (Cost=61838 Card=109390 Bytes=7438
    520)
    22 21 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGE
    R_ENTRY' (Cost=42958 Card=82076 Bytes=3939648)
    23 22 INDEX (RANGE SCAN) OF 'SUBLEDGER_ENTRY_I
    2' (NON-UNIQUE) (Cost=1069 Card=328303)
    24 21 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_
    LINK' (Cost=14314 Card=9235474 Bytes=184709480)
    Statistics
    25 recursive calls
    18 db block gets
    343266 consistent gets
    370353 physical reads
    0 redo size
    15051 bytes sent via SQL*Net to client
    1007 bytes received via SQL*Net from client
    34 SQL*Net roundtrips to/from client
    1 sorts (memory)
    1 sorts (disk)
    491 rows processed
    in test
    SQL> set autotrace traceonly
    SQL> SELECT rpt_horizon_subledger_entry_vw.onst_offst_cd,
    2 rpt_horizon_subledger_entry_vw.bkng_prd,
    3 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    4 rpt_horizon_subledger_entry_vw.jrnl_id,
    5 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    6 7 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    8 rpt_horizon_subledger_entry_vw.crprt_melob_cd AS corp_mlb_cd,
    9 rpt_horizon_subledger_entry_vw.onst_offst_cd, SUM (amt) AS amount
    10 FROM rpt_horizon_subledger_entry_vw
    11 WHERE rpt_horizon_subledger_entry_vw.bkng_prd = '092008'
    12 AND rpt_horizon_subledger_entry_vw.jrnl_id = 'RCS0002100'
    AND rpt_horizon_subledger_entry_vw.systm_afflt_cd = 'SAFF01'
    13 14 GROUP BY rpt_horizon_subledger_entry_vw.onst_offst_cd,
    15 rpt_horizon_subledger_entry_vw.bkng_prd,
    16 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    17 rpt_horizon_subledger_entry_vw.jrnl_id,
    18 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    rpt_horizon_subledger_entry_vw.crprt_melob_cd,
    rpt_horizon_subledger_entry_vw.onst_offst_cd; 19 20 21 22
    no rows selected
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=92944 Card=708 Bytes
    =53100)
    1 0 SORT (GROUP BY) (Cost=92944 Card=708 Bytes=53100)
    2 1 VIEW OF 'RPT_HORIZON_SUBLEDGER_ENTRY_VW' (Cost=92937 Car
    d=708 Bytes=53100)
    3 2 SORT (UNIQUE) (Cost=92937 Card=708 Bytes=124962)
    4 3 UNION-ALL
    5 4 HASH JOIN (Cost=46456 Card=354 Bytes=60180)
    6 5 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACCOU
    NT' (Cost=2 Card=57 Bytes=1881)
    7 5 NESTED LOOPS (Cost=46453 Card=354 Bytes=48498)
    8 7 HASH JOIN (Cost=11065 Card=17694 Bytes=1362438
    9 8 HASH JOIN (Cost=27 Card=87 Bytes=5133)
    10 9 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cos
    t=24 Card=87 Bytes=2175)
    11 9 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_
    ACCNT' (Cost=2 Card=111 Bytes=3774)
    12 8 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_LI
    NK' (Cost=11037 Card=142561 Bytes=2566098)
    13 7 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGER_EN
    TRY' (Cost=2 Card=1 Bytes=60)
    14 13 INDEX (UNIQUE SCAN) OF 'SUBLEDGER_ENTRY_PK'
    (UNIQUE) (Cost=1 Card=1)
    15 4 HASH JOIN (Cost=46456 Card=354 Bytes=64782)
    16 15 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACCOU
    NT' (Cost=2 Card=57 Bytes=1881)
    17 15 NESTED LOOPS (Cost=46453 Card=354 Bytes=53100)
    18 17 HASH JOIN (Cost=11065 Card=17694 Bytes=1362438
    19 18 HASH JOIN (Cost=27 Card=87 Bytes=5133)
    20 19 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cos
    t=24 Card=87 Bytes=2175)
    21 19 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_
    ACCNT' (Cost=2 Card=111 Bytes=3774)
    22 18 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_LI
    NK' (Cost=11037 Card=142561 Bytes=2566098)
    23 17 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGER_EN
    TRY' (Cost=2 Card=1 Bytes=73)
    24 23 INDEX (UNIQUE SCAN) OF 'SUBLEDGER_ENTRY_PK'
    (UNIQUE) (Cost=1 Card=1)
    Statistics
    1134 recursive calls
    0 db block gets
    38903505 consistent gets
    598254 physical reads
    60 redo size
    901 bytes sent via SQL*Net to client
    461 bytes received via SQL*Net from client
    1 SQL*Net roundtrips to/from client
    34 sorts (memory)
    0 sorts (disk)
    0 rows processed
    Thanks a lot in advance
    Jerry

    Hi
    Basically there are two kinds of tables
    - fact
    - lookup
    The number of records in a lookup table is usually small.
    The number of records in a fact table is usually huge.
    However, in test systems the number of records in a fact table is often also small.
    This results in different execution plans.
    I notice again you don't post version and platform info, and you didn't make sure your explain is properly idented
    Please read the FAQ to make sure it is properly idented.
    Also using the word 'buddies' is as far as I am concerned nearing disrespect and rudeness.
    Sybrand Bakker
    Senior Oracle DBA

  • Association rules - memory issue

    Hi everyone,
    I'm testing oracle data mining to generate a high number of association rules, starting from a table containing around 100 columns and 1800 lines. I transformed the table in a view as shown with the market basket demo, but I'm still having a memory error each time I'm trying to launch the data mining process.
    I've tried with a table containing less columns and lines, and it worked fine.
    Is there a way to set ODM to allow it to process this big table and generate millions of association rules without having some memory errors ?
    Is it possible to find how exactly the apriori algorithm works, I mean to find if the sql process is only working with the RAM and writes in the database in the end, or if it commits into the database during the process (before exploding memory).
    Merry christmas by the way !!!!!

    Hi,
    thanks again for your reply. I tried the tweak with the shared_pool_size, and I'm still having some memory issues.
    Do you know how the model is build, and how it works technically ?
    I can see that itemsets and rules are created when you build the model. Is it stored in a procedure (which applies the apriori algorithm), computes the model and then stores the rules generated into the database during commit ?
    In this case all the process (model building) is stored in RAM, am I right ?
    And that's why the process crash when there are too much data processed. Is it possible to avoid this, I mean commit the rules even during the process to free some RAM space in order to continue the rule generation ?

  • Application is not working due to memory issue

    Hi Friesnds,
    Kindly help me regarding settings in java. Scenario is like that. I have one server having two jboss versions (jboss-4.2.3 GA and jboss- 3.2.6). Each having java memory (JVM settings) 1 GB. The total memory of that server is 3 GB.
    Problem is that every two days my application is not working due to memory issue. Once i freed the memory (through run the commands - (1). sync
    (2). echo 3 > /proc/sys/vm/drop_caches
    Application works fine.
    I heared that the above command (echo 3 > /proc/sys/vm/drop_caches) can't run frequently because of server crash.
    Kindly help me regarding this issue and provide the resolution.
    Let me know if you need more information from my side.
    Thanks
    Ashish Shukla

    All of the above.  I tested on 4 different networks and had no luck, I also had friends test my network and the other networks with their iPhones and all had no issues using FaceTime.
    Also after I wiped my phone (erased all content and settings) I was able to successfully initiate a FaceTime call with no changes to my network and was also able to initiate another call when connected to another previously tested network that didn't work before.  Once I restored from backup though, FaceTime stopped working again.
    I did find this discussion: https://discussions.apple.com/thread/5163024?start=0&tstart=0
    and tried the suggestions found there too.  That discussion describes my issue I am having as well.

  • How to correct internal memory issue

    This is related to the standard SAP program SAPRCKM_MR11 in which there is one include RCKM_MR11F01 which contains a select statement :
    SELECT *
         INTO CORRESPONDING FIELDS OF TABLE t_bhistory
         FROM v_ckmlgrir
         WHERE bukrs EQ p_bukrs
         AND   ebeln IN r_ebeln
         AND   ebelp IN r_ebelp
         AND   lifnr IN r_lifnr
         AND   ekorg IN r_ekorg
         AND   ekgrp IN r_ekgrp
         AND   werks IN r_werks
         AND   bedat IN r_bedat
         AND   bsart IN r_bsart
         AND   bstyp IN ht_bstyp
         AND   vgabe IN ('1', '2', '3')  "Wareneingang/Rechnung/Nachbel.
         AND   loekz_k NE 'L'            "Loeschkz. Bestellkopf
         AND   loekz_p NOT IN ('L', 'S') "Loeschkz. Bestellposition
         AND ( frgrl   EQ space          "Freigabe unvollständig
         OR    frgrl   IS NULL )
         AND ( memory  EQ space          "Bestellung noch nicht komplett
         OR    memory  IS NULL )         "falls Feld nicht initialisiert
         AND ( xwoff   EQ space          "Wertbildung offen
         OR    xwoff   IS NULL )         "falls Feld nicht initialisiert
    bei Dienstleistungsbestellungen werden mehrfach kontierte
    Bestell-Pos. gegen das WE/RE-Konto gebucht, deshalb dürfen
    mehrfach kontierte Pos. nicht generell überlesen werden.
          AND   vrtkz   EQ space          "keine Mehrfachkontierung
         AND   wepos   EQ 'X'            "Wareneingang wird erwartet
         AND   repos   EQ 'X'            "Rechnung wird erwartet
         AND   weunb   EQ space          "Wareneingang unbewertet
         AND   stapo   EQ space.         "Item is not statistical
    There is one job FMT612P which contains this standard program and the job is going to abend
    because of the select statement given above which is because of some memory issue with the internal table t_bhistory, please give us a suggestion how we can avoid the same.

    Hi,
    I am also facing the same issue. We had copied this report into a custom report and are trying to improve performance by breaking the view into separate queries on EKKO, EKPO and EKBE.
    Can someone advise the best way to deal with this view v_ckmlgrir or a best approach to break down the view into separate queries.
    Thanks
    Edited by: Shreyas Shrikant Kulkarni on Jun 29, 2009 4:56 PM

  • Memory issue with Podcast !!!

    Hi,
    I have several issues with the Podcast App.
    - i cannot sync podcast that i have downloaded on my laptop using itunes to the iphone,
    - when i download some podcast directly on the iphone, i cannot play them, i still have to stream them,
    - i have uge memory issues and i think that the downloaded version of the podcast might be somewhere on the iphone, but cannot access / delete them (over 16 GO, i have more than 4.5 GO used for 'other', that are not related to anything...)
    i have tried to delete / reinstall the app, but this does not change anything...
    Any help ?
    Rgds

    Quote
    I have a MSI 975X Platinum rev 2 mb with a e6300 processor overclocked to 2.24ghz.
    If your CPU operates @2.24 GHz, it means that you increased FSB frequency to about 348 MHz.
    To make sure there is no misunderstanding here, let me point out the following:
    Memory frequency is ALWAYS linked to FSB frequency by the FSB/DRAM Ratio you can select in BIOS.  If you select 1:1 for example, the BIOS will show that memory frequency is 533 MHz. This value however, only applies to the default FSB clock speed of 266 MHz:
    266 MHz x 1 = 266 MHz (or 533 MHz effective DDR2-frequency).
    If your system is overclocked, what counts is only FSB/DRAM ratio, not the memory speed displayed in BIOS.  That means, if you set FSB/DRAM ration to 1:1 and FSB frequency to 348 MHz, your RAM will operate @:
    348 MHz x 1 = 348 MHz (or 696 MHz effective DDR2-frequency).
    The main question is:
    What are you talking about when you say, you can't make your RAM operate @800 MHz?
    The BIOS does not offer a proper divider to get your RAM to 800 MHz @348 FSB clock speed to begin with.
    You have the following choices if your overclock your system:
    FSB=300 + FSB/DRAM ratio = 1.33
    300 MHz x 1.33 ~ 400 MHz (or 800 MHz effective DDR2-frequency)
    FSB=320 MHz + FSB/DRAM ratio = 1:1.25
    320 MHz x 1.25 = 400 MHz (or 800 MHz effective DDR2-frequency)
    FSB=400 MHz + FSB/DRAM ratio = 1:1
    400 MHz x 1 = 400 MHz (or 800 MHz effective DDR2-frequency)
    Use CPU-Z to monitor the DRAM frequency that is actually set if you are overclocking.

  • Memory issues

    I have a 2008 Macpro...remarkable up-to -date for its age...I plan to use it for some time yet. Lately I have been a having what seems to be a memory issue. There are 12 gig of Ram installed...6x2 gig each in pairs. Sometimes on boot the screen only shows a flashing folder. I shut down and reboot, and it seems fine...but only 8gig of ram shows installed. I opened it up, unseated the boards that hod the ram, and unseated and reset all he ram, and replaced the boards. Great, all he ram shows, 12 gig again. This morning, a few days later, it did it again, showing 8. So, when I run Crucial's memory analyzer, it shows first four slots full, last four empty. A few questions...are the first four slots the top board, the last four the bottom board? This, BTW, is what it was also showing before I reseated the ram. If so, the ram in question s on a separate board. Opinions, is the ram bad, or the board, or neither? Should I reseat again and flush with contact/tuner cleaner/degreaser? Could this hurt? (it is made to clean electronics) the machine does seem ever so slightly unstable. FCP crashed once in a 5 hour edit, which it has done before but rarely. The machine itself did not crash, just the program. Can using the machine like this do more damage, or will it just not boot? I feel it is a seating or connection issue, as I have never seen ram come and go, once it is fried it's fried. The ram is still pretty pricy, but I was going to buy 4 more gigs just to fill it out. Are the two board that hold the ram identical, top and bottom? If so, is a good diagnostic tool to swap them and see what happens? Move all the ram and swap board positions, or do I risk blowing ram ? Remember, the ram did show up when I reseated, so it seems like it is I damaged. My goal would be to see if top ram stops appearing. Does ram need to be filled in order? I know in pairs, but can I just swap the boards and leave the ram, which should look like filling slots 1,2,5,6,7,8...as opposed to 1,2,3,4,5,6.  Can the boards be swapped, or are they set somehow to be 1-4 and 5-8, or is that determined automatically by their upper or lower insertion? Any other advice is appreciated. Funds are low, so I need to keep this machine running. Other info, 3.06ghz (2 quad core) Xeon CPUs, 1tb system drive, 4 TB raid5' OSX 10.6.8 dual graphics display with 512MB (can't remember card model right now). Also, can I update graphic card in this machine? I looked at it a few years ago, but the I for I found was confusing and I never got to it. I figure I might be able to get a better card pretty cheap by now. Thanks all.

    You have typed a huge list of ideas all as one paragraph with many questions in it without a single bit of white space. My eyes are not good enough to read a question, attempt to answer it, then dig through the sea of letters to find the next question.
    Your writing needs white space, which gives it a peculiar overall shape, and allows a reader with poor eyesight to return to a particular spot in the text. So I will attempt to answer some parts of your query, but will be forced to do so from memory of what you wrote.
    The Mac Pro uses Error Correcting Code (ECC) RAM. Eight extra bits are stored with each word, and checked when read back. Single-bit errors are detected, and the check bits are used to make a correction on the fly. More complicated errors are detected, and can cause your Mac to halt on a kernel panic to prevent bad data from propagating.
    ECC RAM means your Mac can detect bad memory easily at Startup, and can declare pairs of DIMMs "absent" to avoid using them.
    Contact corrosion, which was an issue with much older DIMMs, is rarely an issue with these FBDIMMs. Once re-Inserted ONCE, corrosion will not recur for a year or more.
    The upper and lower Riser-cards are identical.
    The FBDIMMs in these machines are subject to overheating and death. If Mac OS X or any program reports DIMMs as missing, at least one of the DIMMs in that pair has died (but since they are installed in pairs, both are marked absent).

  • Question for those with memory issues on k8n neo plat

    I have been having issues with my memory on the k8n neo platinum board since i built the system last month and i just recently discovered a fact I had overlooked.
    I Have discovered that  if i test the system while it is cool (having sat turned off overnight for example), then memtest will pass the ram just fine and the system will be stable for a short period of time, generally longer if no intensive applications are used, ie. doom3 vs. websurfing).
    This system, once it heats up ( seems mostly to be the nvidia 3 nb that needs to be "hot"- that is, a probe placed in contact with the module reaches about 58+ degrees c) then the system becomes unstable and memtest, if used, fails.
    Prime95, if used on it when its cool, will usually take about an hour or so to fail, but if its used on the system when its been running a bit ( temps from nb at 58c+ and memory module hot to the touch) then it fails after about 20-30 seconds.
    I have placed an active hs on the nb in an attempt to correct this, but temps still remain high with failures occuring.
    I have spent the better part of three days testing the sytem this way and the above facts have held true in all these tests.
    I have  tested at stock ratings ( not using "auto" in bios, but using "approved ratings" of 10 x 200 for fsb and timings of 2-3-3-7 for my ram at 1:1 (which is recommended as well) since i am testing for base stability at this point;Voltages have been left at auto settings, however.
    I was wondering if anyone else with memory issues could try this out and post if this seems to be the case for you as well.
    Perhaps the memory issues on these boards are directly related to a problem with the nvidia 3 chipset running too hot?
    Does anyone have a memory issue with these boards who is using water cooling on the NB ?
     Lack of issues from people using 'problem" memory, but having no failures with water cooled nb's would support my assumption.

    Yes i did mean chipset.  
    I am using a stock hs/fan from amd, which seems to run a tad hot but, considering that I wont overclock until I can get board stable, its not an issue. (runs 35-40 idle and up to 72 intensive). I need the warranty more then the extra mhz at present.
    I removed the stock heatsink from the nvidia chipset and bought a generic chipset cooler (green, very little in the way of fins and seems to be aluminum) with a small fan.
     I used a generic silver paste from Compusa to adhere both the cpu hs and the chipset.
    I didnt lap either hs, just installed each as per instructions on artic silvers website (figured it had to go on the same way even generic).
    I also bought a kingwin thermal center which i ran probes from to the cpu, case, and nb chipset.
     I am not using it to adjust fans so those arent connected - its just a reference temp I can see at a glance, esp considering the funky bios reading this boards been known to give.
    I assume from these responces that even a nvidia 3 250 chipset shouldnt get as hot as mine?
    I asked about that from the retailer i bought this from but he seemed to think it was "ok" at 50ish celcius, but then again he didnt look too confident when he said it, lol.
    Idle right now reads on Corecenter at 28 celcius for cpu and 33 celcius for the chipset, although, I have readings of 37 c and 43 c from the kingwin thermal center, which i believe to be more accurate.
    My voltage reads at 11.67 for the 12v line at idle right now.
    The psu i got is stick from the xblade case i bought in store. (http://www.xoxide.com/xblade2.html)
    I have a dvd burner, cd drive, and two maxtor ide hd's attached to the system;The only card is the agp.
    Is there a way to check and see if the cpu ispower starved? software monitor or even an external probe perhaps?
    I havent put a probe on it, but the ram came in its own heatspreader, which gets very hot to the touch as well.
    What temps should I be looking for on these items?

  • Firefox 3.6.8 and memory issues

    After the last update to 3.6.6, I disabled plugin-container per the instructions given on the forum, and my RAM problems went away. However, after the automatic update to 3.6.8, the memory problems came back, with memory usage going up to 92%. I downgraded to 3.6.4, and turned plugin-container off again; while memory usage is currently at 84%, it is not as low as it was before the upgrade to 3.6.8. (79-80%) (There is very little difference in the websites being opened on the various tabs.)
    == This happened ==
    Every time Firefox opened
    == I upgraded to 3.6.8

    ffs guys. It's a free browser. It's cool. We love it. But can we do something about memory issues? Almost every release has huge memory leaks.
    How about halving the release points and focus on memory issues for the other half of the time. I am a long time user since way back in the day and it seems to me that instead of trying to release so many versions, we should be focused more fixing the current release.
    I say put a moratorium on new features until bloat can be fixed. These experiences are starting to give me a hardon for IE again.
    Screw new features. I can deal with plain old tabs like they are. I don't need more bells and whistles. This browser has been kicking ass and taking names for a loonnnggggg time. We don't need more bloat. Keep the security fixes coming but everything else can wait. Take a breath. Stop and smell the roses for a few weeks.

  • Regarding Memory issues((error) while scheduling the job

    Hi Friends
    I am  facing memory issue while rerunning BODS jobs in production.As we have rapid mart when jobs got failed i rerun the
    job suddenly job got failed and  in logs i found Memory issue
    please help me what i have to do in steps
    Thanks

    I think there is no one solution to buffer pool issue. Buffer pool issue happens due to many reasons like how you design the data flow, requirement of push down or even insufficient memory in the running environment.
    You can check by checking 'collect statistics for optimization' during running the job. Also in the data flow, try changing from pageable cache to memory cache by right clicking and selecting properties.  The below link give some details on caching, might be helpful.
    https://wiki.sdn.sap.com/wiki/display/BOBJ/CachinginDI
    Arun

  • Memory issues while fetching content

    HI all,
    I am using SAP KM Apis. however i am facing some memory issues while using them.
    I am using getContent() api on a resource. and then getInputStream on the Icontent object. I am using a ByteArrayOutputStream for writing the data. here is the code snippet:
    byte[] buf= new byte[4096];
    InputStream inputStream = content.getInputStream();
    ByteArrayOutputStream os = new ByteArrayOutputStream(1024)
    while ((count = inputStream.read(buf)) != -1)
        os.write(buf, 0, count);
    byte[] arr=os.toByteArray();
    THis code consuimes a lot of memory . we need to optimize this. we also tried removing BYteArrayOutputSteam and directly read the bytes in chunk :
    int n; 
    do 
         n = inputStream.read(buf, 0,(int)content.getContentLength()); 
    while (n != -1); 
    however this approach has not helped us.
    please suggest an approach where memory consumption is less and the entire content is fed.
    thanks

    Hi
    This code worked for me for reading from a KM res...
    BufferedReader reader = new BufferedReader(new InputStreamReader(is));
    StringBuilder sb = new StringBuilder();
    String line = null;
            try {
                while ((line = reader.readLine()) != null) {
                    sb.append(line + "\n");
            } catch (IOException e) {
                e.printStackTrace();
            } finally {
                try {
                    is.close();
                } catch (IOException e) {
                    e.printStackTrace();
    return sb.toString();
    Regards
    BP

Maybe you are looking for

  • Sync of Zencast items INCREDIBLY s

    I just upgraded from a Zen V Plus to the Zen. On the V Plus, I occassionally used Zencast (video casts) and aside from memory issues the process worked ok. I subscribed with Zencast, updated channels, and then converted and downloaded in Sync Manager

  • How Can I Extract a Substring in WHERE Clause of CFQUERY ?

    Is it possible to extract a substring in the WHERE clause of either <CFQUERY> or in a Query of Queries?  I am trying to use the following query to find all email addresses with the domain "comcast.net" (i.e. everything after the "@" in the email addr

  • Engage 2008: Line Graph, can't start 2nd series @ different X value

    Using Xcelsius Engage 2008 Created line chart w/ 2 series 2nd series does not start at the same X value as series 1. How do I generate the second series so the initial part of my 2nd series line does not show all 0s? Thank you Nancy Griffin

  • AT&T International Data Plans

    Is there a set plan for accessing Edge networks using a iPhone in Canada from AT&T and what is the cost?

  • Cannot play mov from final cut pro

    Hi out there, I received three mov videofiles with exactly the same length of 2.097.152 KB. This is a playout from a final cut pro editing system. When playing one of these files with quicktime 7.1 on a Win XP system there is a -2048 error message: "