[SOLVED] Systemd-fstab-generator: Out of memory...

Hello using multilib lib32-systemd 208-1 and systemd208-1...  I have a problem where I will boot and it hangs at mounting drives:
systemd-fstab-generator[135]: Out of Memory.
systemd[1]:  /usr/lib/systemd/system-generators/systemd-fstab-generator exited with exit status 1. 
I have chrooted in and upgrade stuff and modified my fstab and grub config stuff a bit since I was getting the /root read only problem..  (changed to rw) and it works... However I cannot seem to find where this problem is coming from and the only thing that I can think is that systemd cannot parse /etc/fstab correctly..  My fstab has labels like:  Label=/home, Label=/, Label=/var, Label=/boot, etc...  it was fine till a couple days ago...  Using kernel 3.11.3 x86_64...
Last edited by adramalech (2013-10-06 20:50:20)

Okay so here is my fstab:
tmpfs /dev/shm tmpfs defaults,noatime 0 0
none /tmp tmpfs defaults,noatimenodev,nosuid,mode=1777 0 0
/dev/sdd2 none swap sw 0 0
proc /proc proc nodev,noexec,nosuid 0 0
# /dev/sdc2 UUID=0cf15f7d-a6e0-4b4b-8ca0-e3fa32ca79f5
LABEL=/ / ext4 errors=remount-ro,defaults,rw,discard,noatime,data=ordered 0 1
# /dev/sdc1 UUID=22cee2cf-55fa-4721-b3da-68191c477643
LABEL=/boot /boot ext2 defaults 0 0
# /dev/sdc3 UUID=79bcca95-af9a-4f9f-99ab-78020a3960da
LABEL=/home /home ext4 errors=remount-ro,defaults,rw,noatime,discard,data=ordered 0 1
/dev/sdd3 /var ext4 errors=remount-ro,defaults,rw,relatime,data=ordered 0 2
LABEL=Storage /media/Storage ext4 errors=remount-ro,defaults,rw,relatime,data=ordered 0 2
***EDIT#1***
Storage is a WD 2TB Caviar green, same with /var and swap,  /boot, /home, / are on an M4 128GB SSD, and the rest are just stuff put in tmpfs and proc like normal...  Tried keeping the heavily read/write data off of the ssd to conservative with writes
I am wondering if I can not just update my damn Labels in fstab and through gparted and call it a day...
***EDIT#2***
Okay so I only went through and thought to update packages that needed updating since that might have fixed the problem... I saw an update for lib32-systemd 208-1 and some other non-essentials stuff so updated rebooted and still same problem     I have used this fstab for a considerable amount of years using arch and then back using gentoo for a couple years with no problems... Wondering if now it is being picky with labels...
Last edited by adramalech (2013-10-06 19:19:49)

Similar Messages

  • [SOLVED]systemd-fstab-generator and syslog error?

    here is the bootup error message:
    Nov  2 07:27:58 localhost [    9.207762] systemd-fstab-generator[151]: Failed to create unit file: File exists
    Nov  2 07:27:58 localhost [   10.119699] systemd[1]: /usr/lib/systemd/system-generators/systemd-fstab-generator exited with exit status 1.
    Nov  2 07:27:58 localhost [   10.544789] systemd[1]: Socket service syslog.service not loaded, refusing.
    Nov  2 07:27:58 localhost [   10.544974] systemd[1]: Failed to listen on Syslog Socket.
    and my fstab is here:
    # /etc/fstab: static file system information
    # <file system>    <dir>    <type>    <options>    <dump>    <pass>
    tmpfs        /tmp    tmpfs    nodev,nosuid    0    0
    /dev/sdb1 /boot ext2 defaults 0 1
    /dev/sdb5 / jfs defaults 0 1
    /dev/sdb6 /home ext4 defaults,user_xattr 0 1
    /dev/sdb7 /var reiserfs defaults 0 1
    /dev/sdb8 /tmp reiserfs defaults 0 1
    Last edited by frat (2012-11-02 03:21:07)

    systemd-fstab-generator is fixed,thank you.
    syslog errors still:
    Nov  2 10:29:44 localhost [   10.051816] systemd[1]: Socket service syslog.service not loaded, refusing.
    Nov  2 10:29:44 localhost [   10.052000] systemd[1]: Failed to listen on Syslog Socket.
    #systemctl status syslog
    syslog.service
          Loaded: error (Reason: No such file or directory)
          Active: inactive (dead)
    #systemctl status syslog.socket
    syslog.socket - Syslog Socket
          Loaded: loaded (/usr/lib/systemd/system/syslog.socket; static)
          Active: inactive (dead)
            Docs: man:systemd.special(7)
                  http://www.freedesktop.org/wiki/Software/systemd/syslog
          CGroup: name=systemd:/system/syslog.socket
    any advice?
    Last edited by frat (2012-11-02 02:56:37)

  • Weblogic generating out of memory error when commandline run doesn't

    Hello,
    I am just beginning to use weblogic. I have a jar file which runs fine if run from command line with the options
    "C:\Program Files (x86)\Java\jre6\bin\java" -jar -Xms1024m -Xmx1024m Report.jarIt connects to oracle and selects some data around (500k records) and writes it to a csv.
    When I run the same jar from within a web application (I mean obvoiusly a servlet calling the jar's main method) the webapp generates
    out of memory error after 80k records itself.
    I tried changing the configuration the server Startup arguments from the console (Server Start and then Arguments) and then restart.
    I just wrote the same thing there -Xms1024m -Xmx1024m
    I guess I am missing something. Please share your answers.
    Environment :
    Win2k8, weblogic 10gR3, jdk 1.5. The application is installed as a service.
    Thanks,
    Neetesh
    Edited by: user13312817 on 5 Dec, 2011 12:15 AM

    If you are not using NodeManager, then I don't think those server settings actually control anything. If you are just exploring WLS, then I suspect you are simply starting an AdminServer in a basic domain and don't have a cluster/nodemanager based environment. I could be wrong, please correct as needed!
    If you are simply starting your WebLogic Server instance using the startWebLogic.sh|cnd script, you can set an environment variable on the command shell that will be picked up and used when the server starts.
    set USER_MEM_ARGS=-Xmx1024 -Xms1024m
    * apply appropriate *nix changes as appropriate.
    Then use the startWebLogic.sh|cmd script to start the server and test your application.
    It may very well be the case that your particular application consumes > 1GB heap so you may need more. Remember that you now have a server environment running over your "main" class, so there is bound to be more memory used that could be just sneaking your heap use over 1GB. For example.
    -steve-

  • Getting Error Out Of Memory while importing the work repository in ODI 10g

    I exported the work repository from topology of ODI 10g of one DB and tried importing it in the another ODI 10g topology of another DB.While importing i got the error 'Out of Memory' .
    Can somebody suggest me ,how to solve the heap size out of memory issue while importing in ODI 10g.
    Thanks in Advance.

    Hi,
    you have to post your question in ODI forum
    Data Integrator
    Suresh

  • Advice needed: The way to solve out of memory problem (or the way to work with big csv files)

    Hello:)
    I'm in trouble: I have a big csv file (over 5gb of web-analytics data) and my 64 bit excel (and 6gb ram)
    I cant load file to data model because of it's size. There is an error "out of memory" in power query. 
    This is the first time when I encountered such a problem.
    What options do I have to work with such a file? To increase memory in my computer? Would it solve the problem? How much do I need to work with 6gb csv? 
    Or may be I can upload my data somewhere to azure and work with it there? 
    So the problem - is there any way to deal with big files using power query? Or I need to become a developer and learn sql or other languages? 
    Thanks in advance.
    Max

    Hi Miguel!
    Thanks for your answer. 
    I've tried to load this file on virtual pc from azure cloud with this config:
    I have increased memory limit in power query settings:
    And still, the proble is the same:
    What I do wrong? 

  • Out of Memory Error when Saving SequenceFile generated using API

    I am currently developing an application that creates a SequenceFile using .NET 4.0, C# (2010), Windows 7 Pro x64, Dual Core 2.5 GHz, 3.0Gb Memory.  Using the API EngineClass, I create a complete SequenceFile based on user/file inputs and then save it to disk.  When I'm done, I save using SequenceFile.Save() and then release using ReleaseSequenceFileEx().
    For smaller files, there is no problem.  For larger files, I sporadically get the following error:
    System.Runtime.InteropServices.COMException (0xFFFFBD98): Out of memory.
    When the file does save, it is ~2.8Mb.  Breaking the file up is not an option.  I am not looping, this is one shot.  I've tried shutting down VS2010 and restarting the computer.  It is still inconsistent on saving.
    I also had to set the Embed Interop Types to False per guidance from NI.  They said that it is unstable in 4.0.
    Questions
    1. Is there a file limit size to the Save() API function?
    2. Is there a workaround to using the Save()?
    Best Regards
    Solved!
    Go to Solution.

    I am using C#.NET.  I called them in my save method immediately prior to the actual test save sequence.
    GC.Collect();
    GC.WaitForPendingFinalizers();
    GC.Collect();
    GC.WaitForPendingFinalizers();
    try
           CurrentSequenceFile.Save(outputPath);
           catch (Exceptionerror)
           FireLogEvent(ParserLogType.Parsing_Error, string.Format("{0}", error));
    Factory.TSEngine.ReleaseSequenceFileEx(CurrentSequenceFile, ReleaseSeqFileOptions.ReleaseSeqFile_UnloadFile);
    Hope this helps

  • Generating large amounts of XML without running out of memory

    Hi there,
    I need some advice from the experienced xdb users around here. I´m trying to map large amounts of data inside the DB (Oracle 11.2.0.1.0) and by large I mean files up to several GB. I compared the "low level" mapping via PL/SQL in combination with ExtractValue/XMLQuery with the elegant XML View Mapping and the best performance gave me the View Mapping by using the XMLTABLE XQuery PATH constructs. So now I have a View that lies on several BINARY XMLTYPE Columns (where the XML files are stored) for the mapping and another view which lies above this Mapping View and constructs the nested XML result document via XMLELEMENT(),XMLAGG() etc. Example Code for better understanding:
    CREATE OR REPLACE VIEW MAPPING AS
    SELECT  type, (...)  FROM XMLTYPE_BINARY,  XMLTABLE ('/ROOT/ITEM' passing xml
         COLUMNS
          type       VARCHAR2(50)          PATH 'for $x in .
                                                                let $one := substring($x/b012,1,1)
                                                                let $two := substring($x/b012,1,2)
                                                                return
                                                                    if ($one eq "A")
                                                                      then "A"
                                                                    else if ($one eq "B" and not($two eq "BJ"))
                                                                      then "AA"
                                                                    else if (...)
    CREATE OR REPLACE VIEW RESULT AS
    select XMLELEMENT("RESULTDOC",
                     (SELECT XMLAGG(
                             XMLELEMENT("ITEM",
                                          XMLFOREST(
                                               type "ITEMTYPE",
    ) as RESULTDOC FROM MAPPING;
    ----------------------------------------------------------------------------------------------------------------------------Now all I want to do is materialize this document by inserting it into a XMLTYPE table/column.
    insert into bla select * from RESULT;
    Sounds pretty easy but can´t get it to work, the DB seems to load a full DOM representation into the RAM every time I perform a select, insert into or use the xmlgen tool. This Representation takes more than 1 GB for a 200 MB XML file and eventually I´m running out of memory with an
    ORA-19202: Error occurred in XML PROCESSING
    ORA-04030: out of process memory
    My question is how can I get the result document into the table without memory exhaustion. I thought the db would be smart enough to generate some kind of serialization/datastream to perform this task without loading everything into the RAM.
    Best regards

    The file import is performed via jdbc, clob and binary storage is possible up to several GB, the OR storage gives me the ORA-22813 when loading files with more than 100 MB. I use a plain prepared statement:
            File f = new File( path );
           PreparedStatement pstmt = CON.prepareStatement( "insert into " + table + " values ('" + id + "', XMLTYPE(?) )" );
           pstmt.setClob( 1, new FileReader(f) , (int)f.length() );
           pstmt.executeUpdate();
           pstmt.close(); DB version is 11.2.0.1.0 as mentioned in the initial post.
    But this isn´t my main problem, the above one is, I prefer using binary xmltype anyway, much easier to index. Anyone an idea how to get the large document from the view into a xmltype table?

  • Aperture runs out of memory generating previews

    Aperture 3.2.3
    65K photos in library
    OSX 10.7.3
    Model Identifier:          MacPro1,1
    Processor Name:          Dual-Core Intel Xeon
    Processor Speed:          2.66 GHz
    Total Number of Cores:          4
    L2 Cache (per Processor):          4 MB
    Memory:          10 GB
    After moving my photoes to an external drive as referenced images, and a library rebuild, Aperture is attempting to recreatea all of my previews.  After generateing a couple of hunderd previews, it runs out of memory and I can see the following errors in the system console:
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: Aperture(37843,0xb071b000) malloc: *** mmap(size=1464881152) failed (error code=12)
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** error: can't allocate region
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** set a breakpoint in malloc_error_break to debug
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: Aperture(37843,0xb071b000) malloc: *** mmap(size=1464881152) failed (error code=12)
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** error: can't allocate region
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** set a breakpoint in malloc_error_break to debug
    5/11/12 6:03:09.041 PM [0x0-0xfe0fe].com.apple.Aperture: Aperture(37843,0xb071b000) malloc: *** mmap(size=1464881152) failed (error code=12)
    5/11/12 6:03:09.041 PM [0x0-0xfe0fe].com.apple.Aperture: *** error: can't allocate region
    5/11/12 6:03:09.041 PM [0x0-0xfe0fe].com.apple.Aperture: *** set a breakpoint in malloc_error_break to debug
    Inactive memory quickly uses up all available memory and is not released.  Quitting Aperture only releases half of the inactive memory.
    Any ideas on how to get previews generated without running out of memory?  I've tried running aperture in 32-bit mode, with no difference in behavior.

    Are Previews being generated as you import, or later?  I am asking because I can identify with memory issues when editing numerous images from my D800 in succession (which are huge), but i am not seeing it while importing.
    Ernie

  • Out of Memory Error when generating JSP

    Hi,
    I have a bit of a problem. I?m trying to generate quite a large JSP file (a table with about 1500 rows). The problem is that I get an out of memory error when the JSP is being generated. And it is definitely that the JSP is too large, or rather the HTML being generated from it. Now I was under the impression that the JSP page was flush now and again, but I tried to set all the different flush options and I have also tried to manually flush the page. Nothing of this made any difference except for displaying a white page instead of an error page since the header was already sent. Now this suggests to me that the page is being cached internally by the server and then sent to the client. So is there anything I can do about this?
    Regards
    Hertz

    You are building an HTML table with 1500 rows? That is definitely a lot. And I assume you are also putting some styling on that table...
    These days programmers resource to [url en.wikipedia.org/wiki/AJAX]AJAX techniques for things like that.
    You can, for example, send the table data as a list of comma separated values, and then use Javascript on the browser to create a partial table with options to browse through the different pages of data.
    If you need to show all that data in one single page, you can still try sending just the data as CSV or XML and then building the table on the browser via a Javascript loop. Use CSS for the styling so you don't overload the HTML code with styling information.
    Marcos Broc
    Hi,
    I have a bit of a problem. I?m trying to generate
    quite a large JSP file (a table with about 1500
    rows). The problem is that I get an out of memory
    error when the JSP is being generated. And it is
    definitely that the JSP is too large, or rather the
    HTML being generated from it. Now I was under the
    impression that the JSP page was flush now and again,
    but I tried to set all the different flush options
    and I have also tried to manually flush the page.
    Nothing of this made any difference except for
    displaying a white page instead of an error page
    since the header was already sent. Now this suggests
    to me that the page is being cached internally by the
    server and then sent to the client. So is there
    anything I can do about this?
    Regards
    Hertz

  • [solved] systemd times out waiting for dev-null.device on encrypted fs

    I have a relatively new install of Arch on my laptop. The relevant sections of the drive setup are a luks encrypted root device (with associated unencrypted /boot), and a large ntfs device encrypted with truecrypt. That contains my home directory, which is shared with the Windows 8 dual-boot on the machine.
    I have all this mounted on boot, so my typical usage is to enter my password for the encrypted root, followed by the password for the truecrypt-encrypted data drive, then normal login.
    This has been working fine for a couple of weeks. This afternoon I rebooted my machine from Windows to Linux, at which point the secondary encrypted drive failed to mount. The root device mounts fine.
    On further examination, in the form of journalctl -xb, I'm getting the following errors after mounting the encrypted root device:
    Feb 12 21:29:54 kafka systemd[1]: Job dev-null.device/start timed out.
    Feb 12 21:29:54 kafka systemd[1]: Timed out waiting for device dev-null.device.
    -- Subject: Unit dev-null.device has failed
    -- Defined-By: systemd
    -- Support: http://lists.freedesktop.org/mailman/li … temd-devel
    -- Documentation: http://www.freedesktop.org/wiki/Softwar … e9d022f03d
    -- Unit dev-null.device has failed.
    -- The result is timeout.
    Feb 12 21:29:54 kafka systemd[1]: Dependency failed for Cryptography Setup for cryptdata.
    -- Subject: Unit [email protected] has failed
    -- Defined-By: systemd
    -- Support: http://lists.freedesktop.org/mailman/li … temd-devel
    -- Documentation: http://www.freedesktop.org/wiki/Softwar … e9d022f03d
    -- Unit [email protected] has failed.
    As my root home directory is on the root filesystem, I can still log in as root. If I try to mount or unmount the truecrypt device (located at /data) the command hangs. After removing "auto,x-systemd.automount" from that device from /etc/fstab and rebooting, I have a significant delay at boot, but I can then mount the /data device as normal.
    I can't find anything related to this either here or on Google. Any ideas?
    Last edited by tealeaf (2014-02-13 12:01:55)

    WonderWoofy wrote:A bit OT, but I just want to mention that having your $HOME on ntfs is probably not the best idea either.  It may work, but because it is not a POSIX compliant filesystem, there is a good chance you might run into some issues.
    Thanks for the warning. Sadly, as there are applications I need for work that only run in Windows, I need to dual boot this machine with Windows 8.1. This is the best option I can find for sharing my home directory, which is also a necessity due to the amount of data I have to share between the systems. I would much rather trust to ntfs-3g-ar and its UserMapping (and all the fiddling with ACLs that I had to do) than to the Windows ext2 drivers I can find. They all seem to be several years out of date. Linux is much better at talking NTFS than Windows is at talking EXT. (To be honest, I'm actually quite impressed with Windows 8 since the upgrade to 8.1. As a long term Arch user and a fan of tiling window managers it's interesting to see Windows moving in the right direction. )
    With 'permissions' in the /etc/fstab and the .NTFS-3G/UserMapping file in place it works almost seamlessly. (It took a lot of tinkering with ACLs in Windows and Linux, but it's working very well now. One tip that I'll write here in case it's of use is that you want the last line of the UserMapping file to be a 'generic' mapping. When I didn't have that there were very strange things going on.)
    The only two minor problems I have now are:
    1) A few applications don't like FUSE filesystems. Steam worked for a while and then broke; reinstalling it failed at every stage. When I checked, it seems that FUSE is a known problem for Steam. (I don't think it's restricted to NTFS.) My response to that was to create /home/.local/$HOME on my root (ext4) filesystem and symlink out to it for troublesome applications.
    2) There are a few characters for filenames that Windows doesn't like, making those files inaccessible in Windows. (Colons are the major culprit.) They're usually quite easily renamed. (The exception being my .maildir folder, which I have had to duplicate natively in cygwin.)
    Neither of these are anything more than minor niggles. I appreciate the warning, though.
    Having said that, if you have an alternative that lets me share a truecrypt-encrypted drive between Linux and Windows that is better than the NTFS-3G approach, I'd love to hear it for future reference.

  • Generate PSK signal without out of memory error

    Hi all,
    I want to generate a PSK modulated signal with message bits length much larger like over IM bits atleast. I have to first generate the message bits, diffrentially encode them and modulate them at the end. But when I give larger length of bits to the input of PSK modulation vi, it gives me an out of momory error.
    What I am doing to get rid of this out of memory error is... I generate 128k message bits using MT generate bits vi, Encode them and modulate them using MT PSK modulate vi and write the modulated signal data repeatedly in the binary file using binary file write vi that I placed within a for loop with loop count of 20 so that I writes modulated signal data for (20*128k= 2.56M encoded bits). But doing so introduces phase discontinuity.
    Can anyone suggest a better way of generating PSK signal with message bits atleast 1M and without the out of momory error?
    Any kind of help would be highly appretiated.
    Best regards

    Hello Sandee,
    The problem here is that when attempting to modulate a large number of bits in a single shot is that you are using the number of bits * samples per symbol * 24 bytes of memory just in creating the message and the modulated signal. In your case if you are using 16 samples per symbol this would require 384 MB of memory for a 1-shot operation of a 1M-bit message.
    Luckily this problem has been taken into consideration within the Modulation toolkit (MT). If you take the MT PSK Transmitter VI example that is installed with modulation toolkit you can make some small modifications to implement your desired functionality.
    By placing the code within a for loop and wiring the first call VI to the reset inputs of the MT VIs you can reset the VIs on the first iteration of the loop and be able to use the previous iteration states for further iterations. Also, to prevent losing samples due to filter delay you need to make sure to flush the buffers (an input to the MT Modulate PSK VI) on the last iteration.
    Let me know how this goes in your specific application.
    Regards,
    Dan King

  • [SOLVED] Cannot start Virtual Network: "out of memory"

    I am attempting to use KVM, Qemu and libvirt to run Windows VM's to test websites in IE. I have gotten everything to work fine except guest to host communication, which obviously is crucial for testing the websites.
    Currently i am using macvtap to setup networking, which works beautifully for everything except host to guest networking. The easiest way seems to be to use isolated virtual networks. However, i cannot start any virtual networks. I have tried from inside virtual manager and inside virsh command prompt. I always get "out of memory". I have googled the error and nobody seems to have this problem.
    One note, I do NOT have a swap drive. Is this needed for virtual networks?
    EDIT:
    I have now tried exactly the same on another, identical computer but with swap drive (16gb) and it did not work there either. Both computers have 8gb ram. Seems to be loads free in tops.
    Last edited by sup3rman (2015-01-23 12:29:50)

    sup3rman wrote:
    I am attempting to use KVM, Qemu and libvirt to run Windows VM's to test websites in IE. I have gotten everything to work fine except guest to host communication, which obviously is crucial for testing the websites.
    Currently i am using macvtap to setup networking, which works beautifully for everything except host to guest networking. The easiest way seems to be to use isolated virtual networks. However, i cannot start any virtual networks. I have tried from inside virtual manager and inside virsh command prompt. I always get "out of memory". I have googled the error and nobody seems to have this problem.
    One note, I do NOT have a swap drive. Is this needed for virtual networks?
    EDIT:
    I have now tried exactly the same on another, identical computer but with swap drive (16gb) and it did not work there either. Both computers have 8gb ram. Seems to be loads free in tops.
    I have the same issue on my install. I got so desperate that I actually wiped my laptop and reinstalled everything, same problem. Out of memory error when trying to create the default network. 16GB of RAM, with 96% of it free when I attempted.

  • How can I solve out of memory error on excell file in PL/SQL

    Hi,
    I'm new on PL/SQL. One of the PL/SQL code which is created excell report got out of Memory error. The first reason of this error, excell file not supported more than 65536 data. So I change the excell file separeted sheets. So that the single sheet size cannot exceed 65536 data.
    All the data are held on system cach and if many user want to take the report the they would get an out of memory error.
    So I want to change the code like that; when out of memory exception raises,
    the old excell file save to disk and new excell file is created,
    and go on to write the new file without exiting the program.
    At the end of the data all the excell file append and show only one file to the user.
    I do know how to save the file and create a new file. But I don't know how can PL/SQL program to turn back to loop again when the exception occurs.
    Is anyone help me on this issue?
    Here is my code
    Thank you
    dworkbook:=hssfworkbook.new;
    dCurrentItem := Get_Block_Property(pCurrentBlock, FIRST_ITEM);     
    while not (name_in('system.last_record')='TRUE') loop
    /* The data would be written to the excell file column order. */
    if (dRow=0) then
              /* Create a new sheet */
    elsif (dRow <= dMaxWorksheetNum) then
         /* Data of the report are written here. The data are written in column order */
    if (dRow > dMaxWorksheetNum) then
         /* give dRow and dColumn intial value */
    /* increase worksheet number */
    end if; /* End of if (dRow=1) */
    if (isWritten) and not name_in('system.last_record')='TRUE'then
         /* if not at the end of the record and the previously read record is written to the file
         , then go to next record */
         next_record;
    end if;
    /* save excell report */
    workbookwriter.save(dworkbook,global.gethome||dFileName);
    web.show_document('/users/'||dFileName,'_BLANK');
    /* when exceptions occurs */
    EXCEPTION
    WHEN ORA_JAVA.EXCEPTION_THROWN THEN
    begin
         javaException := ORA_JAVA.LAST_EXCEPTION;
         -- Print out the Exception by using the toString()
         -- Method of the exception Object
         javaException2 := Exception_.new(javaException);
         mess(27002,Exception_.getMessage(javaException2));
         -- and clean up
         ORA_JAVA.CLEAR_EXCEPTION;
    exception
         WHEN ORA_JAVA.JAVA_ERROR THEN
    -- In this case, the error would be
    -- Argument 1 can not be null
    mess(27002,ORA_JAVA.LAST_ERROR);
    --Clean up
    ORA_JAVA.CLEAR_ERROR;
    end;
    WHEN ORA_JAVA.JAVA_ERROR THEN
    -- In this case, the error would be
    -- Argument 1 can not be null
    message(ORA_JAVA.LAST_ERROR);
    --Clean up
    ORA_JAVA.CLEAR_ERROR;

    No need to double-post... most questions are answered pretty quickly...

  • How to Solve "General Error: Out of memory"

    Hi all experts,
    I have this "weird" error which I encountered for the first time.
    I converted a AVCHD file from my SD card (in mts) to .mov format using clipwrap. I played the resultant .mov file using Quick Time and it plays smoothly without any problem.
    Now, I imported this .mov into FInal Cut and it appears into my project panel. Then I click on this file expecting to see the video in the Viewer Window but it doesn't. Instead, an message saying "General Error: Out of Memory" appeared on the Viewer Window.
    Then, I tried another method by draggin rhis file from the project panel straight into the timeline. When I put my CTI onto it in the timeline, the video appears on the Canvas window but when I press the spacebar, the movement of the video is very jerky.
    Both the sequence settings on the video & sequence are matching at 1920 x 1080.
    Any idea what is going on ?
    Thanks

    MartinR wrote:
    Did you convert the AVCHD video to QuickTime using _Apple Intermediate Codec (AIC)?_
    (.mov files are QuickTime, but the video codec inside would be H.264 if all you did in Clipwrap was convert .mts to .mov without also specifying conversion to Apple Intermediate Codec. A QT/H.264 video file would be playable in QuickTime but could not be edited in FCE. Needs to be QT/AIC.)
    Suggest checking (and let us know) what your clip & sequence properties are.
    Thanks Martin.
    And you are right ! The clip converted from clipwrap is QT/H.264. But how do I instruct Clipwrap to convert it to QT/AIC ?
    Also, how come Final Cut classified this as an "Out of Memory" error ?
    Thanks

  • Error when compiling firefox...Out of memory: Kill process 6763

    I'm trying to compile the firefox version 15 as both 16.0.1 and I always get the same error, which I think leaves me with no ram even though I have 8 gigs, i try it with 8 gigs of swap but does exactly the same, here are all the facts about this problem, the only thing I have not tried is to change the compiler version, what do you think? this is a clear linkage error where the system breaks down after running out of physical memory...
    ERROR--->-using  yaourt -Sb or makepkg -s:
    /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/toolkit/library/nsUnicharUtils.cpp:275:1: warning: always_inline function might not be inlinable [-Wattributes]
    /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/toolkit/library/nsUnicharUtils.cpp:50:1: warning: always_inline function might not be inlinable [-Wattributes]
    /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/toolkit/library/nsUnicharUtils.cpp:40:1: warning: always_inline function might not be inlinable [-Wattributes]
    rm -f libxul.so
    /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/_virtualenv/bin/python /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/config/pythonpath.py -I../../config /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/config/expandlibs_exec.py --depend .deps/libxul.so.pp --target libxul.so --uselist -- c++ -pedantic -Wall -Wpointer-arith -Woverloaded-virtual -Werror=return-type -Wtype-limits -Wempty-body -Wno-ctor-dtor-privacy -Wno-overlength-strings -Wno-invalid-offsetof -Wno-variadic-macros -Wcast-align -Wno-long-long -march=native -O2 -pipe -fstack-protector --param=ssp-buffer-size=4 -D_FORTIFY_SOURCE=2 -fno-exceptions -fno-strict-aliasing -fno-rtti -ffunction-sections -fdata-sections -fno-exceptions -std=gnu++0x -pthread -pipe -DNDEBUG -DTRIMMED -g -fprofile-generate -O3 -fomit-frame-pointer -fPIC -shared -Wl,-z,defs -Wl,--gc-sections -Wl,-h,libxul.so -o libxul.so nsStaticXULComponents.i_o nsUnicharUtils.i_o nsBidiUtils.i_o nsSpecialCasingData.i_o nsUnicodeProperties.i_o nsRDFResource.i_o -lpthread -Wl,-O1,--sort-common,--as-needed,-z,relro -Wl,-rpath,/usr/lib/firefox -Wl,-z,noexecstack -fprofile-generate -Wl,-rpath-link,/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/dist/bin -Wl,-rpath-link,/usr/lib ../../toolkit/xre/libxulapp_s.a ../../staticlib/components/libnecko.a ../../staticlib/components/libuconv.a ../../staticlib/components/libi18n.a ../../staticlib/components/libchardet.a ../../staticlib/components/libjar50.a ../../staticlib/components/libstartupcache.a ../../staticlib/components/libpref.a ../../staticlib/components/libhtmlpars.a ../../staticlib/components/libidentity.a ../../staticlib/components/libimglib2.a ../../staticlib/components/libgkgfx.a ../../staticlib/components/libgklayout.a ../../staticlib/components/libdocshell.a ../../staticlib/components/libembedcomponents.a ../../staticlib/components/libwebbrwsr.a ../../staticlib/components/libnsappshell.a ../../staticlib/components/libtxmgr.a ../../staticlib/components/libcommandlines.a ../../staticlib/components/libtoolkitcomps.a ../../staticlib/components/libpipboot.a ../../staticlib/components/libpipnss.a ../../staticlib/components/libappcomps.a ../../staticlib/components/libjsreflect.a ../../staticlib/components/libcomposer.a ../../staticlib/components/libtelemetry.a ../../staticlib/components/libjsinspector.a ../../staticlib/components/libjsdebugger.a ../../staticlib/components/libstoragecomps.a ../../staticlib/components/librdf.a ../../staticlib/components/libwindowds.a ../../staticlib/components/libjsctypes.a ../../staticlib/components/libjsperf.a ../../staticlib/components/libgkplugin.a ../../staticlib/components/libunixproxy.a ../../staticlib/components/libjsd.a ../../staticlib/components/libautoconfig.a ../../staticlib/components/libauth.a ../../staticlib/components/libcookie.a ../../staticlib/components/libpermissions.a ../../staticlib/components/libuniversalchardet.a ../../staticlib/components/libfileview.a ../../staticlib/components/libplaces.a ../../staticlib/components/libtkautocomplete.a ../../staticlib/components/libsatchel.a ../../staticlib/components/libpippki.a ../../staticlib/components/libwidget_gtk2.a ../../staticlib/components/libimgicon.a ../../staticlib/components/libprofiler.a ../../staticlib/components/libaccessibility.a ../../staticlib/components/libremoteservice.a ../../staticlib/components/libspellchecker.a ../../staticlib/components/libzipwriter.a ../../staticlib/components/libservices-crypto.a ../../staticlib/libjsipc_s.a ../../staticlib/libdomipc_s.a ../../staticlib/libdomplugins_s.a ../../staticlib/libmozipc_s.a ../../staticlib/libmozipdlgen_s.a ../../staticlib/libipcshell_s.a ../../staticlib/libgfxipc_s.a ../../staticlib/libhal_s.a ../../staticlib/libdombindings_s.a ../../staticlib/libxpcom_core.a ../../staticlib/libucvutil_s.a ../../staticlib/libchromium_s.a ../../staticlib/libsnappy_s.a ../../staticlib/libgtkxtbin.a ../../staticlib/libthebes.a ../../staticlib/libgl.a ../../staticlib/libycbcr.a -L../../dist/bin -L../../dist/lib /tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/dist/lib/libjs_static.a -lffi -Wl,-rpath-link,/usr/lib -L/usr/lib -lssl3 -lsmime3 -lnss3 -lnssutil3 -lcrmf -lXrender -lfreetype -lfontconfig -lsqlite3 -ljpeg -lpng -lz -lhunspell-1.3 -L/usr/lib -levent -lpixman-1 ../../dist/lib/libgkmedias.a -lasound -lrt -L../../dist/bin -L../../dist/lib -L/usr/lib -lplds4 -lplc4 -lnspr4 -lpthread -ldl ../../dist/lib/libmozalloc.a -ldbus-glib-1 -ldbus-1 -lgobject-2.0 -lglib-2.0 -lX11 -lXext -lpangoft2-1.0 -lfreetype -lfontconfig -lpangocairo-1.0 -lpango-1.0 -lcairo -lgobject-2.0 -lglib-2.0 -lgtk-x11-2.0 -latk-1.0 -lgio-2.0 -lpangoft2-1.0 -lfreetype -lfontconfig -lgdk-x11-2.0 -lpangocairo-1.0 -lgdk_pixbuf-2.0 -lpango-1.0 -lcairo -lgobject-2.0 -lglib-2.0 -lXt -lgthread-2.0 -lfreetype -lstartup-notification-1 -lvpx -ldl -lrt -lrt
    collect2: error: ld terminated with signal 9 [Matat]
    make[6]: *** [libxul.so] Error 1
    make[6]: Leaving directory `/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu/toolkit/library'
    make[5]: *** [libs_tier_platform] Error 2
    make[5]: Leaving directory `/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu'
    make[4]: *** [tier_platform] Error 2
    make[4]: Leaving directory `/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu'
    make[3]: *** [default] Error 2
    make[3]: Leaving directory `/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release/obj-x86_64-unknown-linux-gnu'
    make[2]: *** [realbuild] Error 2
    make[2]: Leaving directory `/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release'
    make[1]: *** [profiledbuild] Error 2
    make[1]: Leaving directory `/tmp/yaourt-tmp-enric/abs-firefox/src/mozilla-release'
    make: *** [build] Error 2
    dmesg output on  ld :
    [ 1521.353469] ld invoked oom-killer: gfp_mask=0x280da, order=0, oom_adj=0, oom_score_adj=0
    [ 1521.353474] Pid: 6763, comm: ld Not tainted 3.6.0 #1
    [ 1521.353475] Call Trace:
    [ 1521.353482] [<ffffffff814cc68e>] ? dump_header.isra.11+0x5d/0x18e
    [ 1521.353486] [<ffffffff812ae3dc>] ? ___ratelimit+0xac/0x120
    [ 1521.353489] [<ffffffff810e2fb5>] ? oom_kill_process+0x275/0x3b0
    [ 1521.353492] [<ffffffff810e2b10>] ? find_lock_task_mm+0x20/0x70
    [ 1521.353494] [<ffffffff810e3455>] ? out_of_memory+0x1c5/0x290
    [ 1521.353497] [<ffffffff810e742a>] ? __alloc_pages_nodemask+0x85a/0x870
    [ 1521.353500] [<ffffffff811048c4>] ? handle_pte_fault+0x8c4/0xb10
    [ 1521.353505] [<ffffffff81075ddf>] ? select_task_rq_fair+0x4cf/0x790
    [ 1521.353509] [<ffffffff8100a21f>] ? native_sched_clock+0xf/0x70
    [ 1521.353512] [<ffffffff8102b880>] ? do_page_fault+0x130/0x460
    [ 1521.353515] [<ffffffff810e6301>] ? get_page_from_freelist+0x311/0x670
    [ 1521.353519] [<ffffffff814d2275>] ? page_fault+0x25/0x30
    [ 1521.353523] [<ffffffff810df577>] ? file_read_actor+0x67/0x1f0
    [ 1521.353526] [<ffffffff810f6915>] ? shmem_file_aio_read+0x155/0x3a0
    [ 1521.353530] [<ffffffff81128832>] ? do_sync_read+0x92/0xd0
    [ 1521.353532] [<ffffffff811290c0>] ? vfs_read+0xa0/0x160
    [ 1521.353535] [<ffffffff811291c7>] ? sys_read+0x47/0xa0
    [ 1521.353537] [<ffffffff814d2275>] ? page_fault+0x25/0x30
    [ 1521.353540] [<ffffffff814d27fd>] ? system_call_fastpath+0x1a/0x1f
    [ 1521.353541] Mem-Info:
    [ 1521.353543] DMA per-cpu:
    [ 1521.353544] CPU 0: hi: 0, btch: 1 usd: 0
    [ 1521.353545] CPU 1: hi: 0, btch: 1 usd: 0
    [ 1521.353546] CPU 2: hi: 0, btch: 1 usd: 0
    [ 1521.353547] CPU 3: hi: 0, btch: 1 usd: 0
    [ 1521.353548] DMA32 per-cpu:
    [ 1521.353550] CPU 0: hi: 186, btch: 31 usd: 0
    [ 1521.353551] CPU 1: hi: 186, btch: 31 usd: 26
    [ 1521.353552] CPU 2: hi: 186, btch: 31 usd: 59
    [ 1521.353553] CPU 3: hi: 186, btch: 31 usd: 0
    [ 1521.353554] Normal per-cpu:
    [ 1521.353555] CPU 0: hi: 186, btch: 31 usd: 30
    [ 1521.353556] CPU 1: hi: 186, btch: 31 usd: 156
    [ 1521.353557] CPU 2: hi: 186, btch: 31 usd: 169
    [ 1521.353558] CPU 3: hi: 186, btch: 31 usd: 0
    [ 1521.353562] active_anon:1599337 inactive_anon:345122 isolated_anon:0
    active_file:140 inactive_file:222 isolated_file:0
    unevictable:17 dirty:0 writeback:0 unstable:0
    free:11562 slab_reclaimable:14927 slab_unreclaimable:21150
    mapped:6681 shmem:932329 pagetables:10868 bounce:0
    [ 1521.353567] DMA free:15892kB min:20kB low:24kB high:28kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15644kB mlocked:0kB dirty:0kB writeback:0kB mapped:0kB shmem:0kB slab_reclaimable:0kB slab_unreclaimable:8kB kernel_stack:0kB pagetables:0kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes
    [ 1521.353568] lowmem_reserve[]: 0 3147 7925 7925
    [ 1521.353575] DMA32 free:23592kB min:4520kB low:5648kB high:6780kB active_anon:2754024kB inactive_anon:412764kB active_file:16kB inactive_file:44kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:3223060kB mlocked:0kB dirty:0kB writeback:0kB mapped:5892kB shmem:1294220kB slab_reclaimable:7144kB slab_unreclaimable:5108kB kernel_stack:216kB pagetables:8616kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:715 all_unreclaimable? yes
    [ 1521.353576] lowmem_reserve[]: 0 0 4778 4778
    [ 1521.353582] Normal free:6764kB min:6860kB low:8572kB high:10288kB active_anon:3643324kB inactive_anon:967724kB active_file:544kB inactive_file:844kB unevictable:68kB isolated(anon):0kB isolated(file):0kB present:4892832kB mlocked:68kB dirty:0kB writeback:0kB mapped:20832kB shmem:2435096kB slab_reclaimable:52564kB slab_unreclaimable:79484kB kernel_stack:2968kB pagetables:34856kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:2484 all_unreclaimable? yes
    [ 1521.353583] lowmem_reserve[]: 0 0 0 0
    [ 1521.353586] DMA: 1*4kB 0*8kB 1*16kB 0*32kB 2*64kB 1*128kB 1*256kB 0*512kB 1*1024kB 1*2048kB 3*4096kB = 15892kB
    [ 1521.353592] DMA32: 740*4kB 506*8kB 300*16kB 135*32kB 25*64kB 0*128kB 0*256kB 0*512kB 0*1024kB 1*2048kB 1*4096kB = 23872kB
    [ 1521.353599] Normal: 637*4kB 0*8kB 0*16kB 2*32kB 1*64kB 0*128kB 1*256kB 1*512kB 1*1024kB 1*2048kB 0*4096kB = 6516kB
    [ 1521.353605] 932908 total pagecache pages
    [ 1521.353606] 0 pages in swap cache
    [ 1521.353607] Swap cache stats: add 0, delete 0, find 0/0
    [ 1521.353608] Free swap = 0kB
    [ 1521.353609] Total swap = 0kB
    [ 1521.369740] 2094576 pages RAM
    [ 1521.369743] 80362 pages reserved
    [ 1521.369744] 36619 pages shared
    [ 1521.369745] 1993098 pages non-shared
    [ 1521.369746] [ pid ] uid tgid total_vm rss nr_ptes swapents oom_score_adj name
    [ 1521.369752] [ 2286] 0 2286 6258 269 16 0 -1000 systemd-udevd
    [ 1521.369754] [ 2293] 0 2293 153174 83 229 0 0 systemd-journal
    [ 1521.369757] [ 3585] 0 3585 4695 188 13 0 0 mount.ntfs-3g
    [ 1521.369759] [ 3615] 0 3615 3547 140 12 0 0 crond
    [ 1521.369761] [ 3616] 0 3616 6708 2523 18 0 0 preload
    [ 1521.369763] [ 3617] 0 3617 17762 297 38 0 0 cupsd
    [ 1521.369765] [ 3619] 0 3619 43562 313 52 0 0 NetworkManager
    [ 1521.369767] [ 3620] 84 3620 6988 70 20 0 0 avahi-daemon
    [ 1521.369769] [ 3621] 0 3621 6515 72 18 0 0 systemd-logind
    [ 1521.369770] [ 3622] 81 3622 4595 285 14 0 -900 dbus-daemon
    [ 1521.369772] [ 3626] 0 3626 2275 32 9 0 0 agetty
    [ 1521.369774] [ 3627] 0 3627 6693 53 19 0 0 kdm
    [ 1521.369776] [ 3628] 84 3628 6957 52 18 0 0 avahi-daemon
    [ 1521.369778] [ 3631] 0 3631 45316 12716 91 0 0 X
    [ 1521.369780] [ 4287] 0 4287 15666 128 36 0 0 kdm
    [ 1521.369782] [ 5131] 102 5131 92417 912 41 0 0 polkitd
    [ 1521.369784] [ 5132] 0 5132 53222 346 41 0 0 colord
    [ 1521.369786] [ 5163] 0 5163 129753 991 153 0 0 colord-sane
    [ 1521.369788] [11395] 0 11395 523912 258 62 0 0 console-kit-dae
    [ 1521.369790] [11468] 1000 11468 3668 111 13 0 0 startkde
    [ 1521.369792] [11480] 1000 11480 3970 37 13 0 0 dbus-launch
    [ 1521.369794] [11481] 1000 11481 4839 401 15 0 0 dbus-daemon
    [ 1521.369796] [11507] 1000 11507 4068 94 12 0 0 gpg-agent
    [ 1521.369798] [11510] 1000 11510 3760 87 11 0 0 ssh-agent
    [ 1521.369799] [11524] 1000 11524 1013 21 7 0 -300 start_kdeinit
    [ 1521.369801] [11525] 1000 11525 85821 1591 149 0 -300 kdeinit4
    [ 1521.369803] [11528] 1000 11528 188590 3072 227 0 0 kded4
    [ 1521.369805] [11534] 1000 11534 108217 2084 173 0 0 kwalletd
    [ 1521.369807] [11539] 1000 11539 107913 2348 173 0 0 kglobalaccel
    [ 1521.369809] [11542] 0 11542 55597 601 43 0 0 upowerd
    [ 1521.369811] [11555] 1000 11555 1047 18 7 0 0 kwrapper4
    [ 1521.369813] [11556] 1000 11556 127307 2185 174 0 0 ksmserver
    [ 1521.369814] [11568] 0 11568 49409 269 33 0 0 udisks-daemon
    [ 1521.369816] [11573] 0 11573 12394 89 28 0 0 udisks-daemon
    [ 1521.369818] [11577] 1000 11577 214097 12177 260 0 0 kwin
    [ 1521.369820] [11590] 1000 11590 93530 1489 131 0 0 kactivitymanage
    [ 1521.369822] [11621] 1000 11621 343123 3417 215 0 0 knotify4
    [ 1521.369824] [11631] 1000 11631 175587 4370 239 0 0 krunner
    [ 1521.369826] [11633] 1000 11633 246522 16714 342 0 0 plasma-desktop
    [ 1521.369828] [11636] 1000 11636 216148 5325 253 0 0 lancelot
    [ 1521.369830] [11639] 1000 11639 50634 268 35 0 0 mission-control
    [ 1521.369831] [11643] 1000 11643 37597 405 39 0 0 akonadi_control
    [ 1521.369833] [11645] 1000 11645 357276 722 81 0 0 akonadiserver
    [ 1521.369835] [11652] 1000 11652 377776 6206 70 0 0 mysqld
    [ 1521.369837] [11738] 1000 11738 76786 1061 110 0 0 akonadi_agent_l
    [ 1521.369839] [11739] 1000 11739 76783 1056 111 0 0 akonadi_agent_l
    [ 1521.369841] [11740] 1000 11740 75143 1033 111 0 0 akonadi_agent_l
    [ 1521.369843] [11741] 1000 11741 75144 1042 110 0 0 akonadi_agent_l
    [ 1521.369845] [11742] 1000 11742 75800 1048 112 0 0 akonadi_agent_l
    [ 1521.369847] [11743] 1000 11743 75800 1054 107 0 0 akonadi_agent_l
    [ 1521.369848] [11744] 1000 11744 76786 1064 108 0 0 akonadi_agent_l
    [ 1521.369850] [11745] 1000 11745 76780 1073 111 0 0 akonadi_agent_l
    [ 1521.369852] [11746] 1000 11746 84215 1305 155 0 0 akonadi_maildis
    [ 1521.369854] [11747] 1000 11747 92098 1211 138 0 0 akonadi_nepomuk
    [ 1521.369856] [11774] 1000 11774 59745 591 76 0 0 nepomukserver
    [ 1521.369858] [11777] 1000 11777 287924 2235 149 0 0 nepomukservices
    [ 1521.369860] [11795] 1000 11795 100823 10181 51 0 0 virtuoso-t
    [ 1521.369861] [11801] 1000 11801 93041 786 99 0 0 pulseaudio
    [ 1521.369863] [11802] 133 11802 41125 46 17 0 0 rtkit-daemon
    [ 1521.369865] [11811] 1000 11811 17253 140 36 0 0 gconf-helper
    [ 1521.369867] [11813] 1000 11813 11527 127 26 0 0 gconfd-2
    [ 1521.369869] [11816] 1000 11816 68886 1148 123 0 0 kuiserver
    [ 1521.369871] [11837] 1000 11837 58146 1047 105 0 0 nepomukservices
    [ 1521.369873] [11839] 1000 11839 53655 966 98 0 0 nepomukservices
    [ 1521.369875] [11840] 1000 11840 92263 1128 107 0 0 nepomukservices
    [ 1521.369877] [11841] 1000 11841 109750 1232 107 0 0 nepomukservices
    [ 1521.369878] [12942] 1000 12942 41975 542 78 0 0 kwrited
    [ 1521.369880] [12944] 1000 12944 246808 6774 195 0 0 ktorrent
    [ 1521.369882] [12954] 1000 12954 93430 1212 139 0 0 polkit-kde-auth
    [ 1521.369884] [12957] 1000 12957 70673 1126 130 0 0 nepomukcontroll
    [ 1521.369886] [12959] 1000 12959 92309 1654 138 0 0 kgpg
    [ 1521.369888] [12966] 1000 12966 109879 2164 176 0 0 klipper
    [ 1521.369890] [13009] 1000 13009 86477 1682 131 0 0 kio_http_cache_
    [ 1521.369892] [17302] 1000 17302 270535 59165 488 0 0 firefox
    [ 1521.369894] [17322] 1000 17322 10407 86 26 0 0 gvfsd
    [ 1521.369896] [17324] 1000 17324 50365 171 31 0 0 gvfs-fuse-daemo
    [ 1521.369897] [18462] 1000 18462 138761 5014 199 0 0 konsole
    [ 1521.369899] [18464] 1000 18464 4227 156 13 0 0 bash
    [ 1521.369901] [19535] 1000 19535 3885 324 13 0 0 yaourt
    [ 1521.369903] [19680] 1000 19680 3800 253 13 0 0 makepkg
    [ 1521.369905] [22861] 1000 22861 174765 6917 209 0 0 dolphin
    [ 1521.369907] [24940] 1000 24940 12258 2121 27 0 0 Xvfb
    [ 1521.369909] [24941] 1000 24941 2944 134 11 0 0 make
    [ 1521.369911] [25116] 1000 25116 2946 137 11 0 0 make
    [ 1521.369913] [25337] 1000 25337 3013 200 11 0 0 make
    [ 1521.369915] [ 473] 1000 473 2977 169 12 0 0 make
    [ 1521.369917] [ 3719] 1000 3719 3006 176 12 0 0 make
    [ 1521.369919] [10784] 1000 10784 3006 175 13 0 0 make
    [ 1521.369921] [12252] 1000 12252 115351 9567 175 0 0 kvirc
    [ 1521.369922] [26014] 1000 26014 4256 168 13 0 0 bash
    [ 1521.369925] [30177] 1000 30177 4255 167 14 0 0 bash
    [ 1521.369926] [30976] 1000 30976 86977 3634 138 0 0 plugin-containe
    [ 1521.369928] [ 915] 1000 915 87003 1648 132 0 0 kio_file
    [ 1521.369930] [ 916] 1000 916 109277 2635 176 0 0 kio_thumbnail
    [ 1521.369932] [ 1186] 1000 1186 4227 157 13 0 0 bash
    [ 1521.369934] [ 2105] 1000 2105 86873 1689 132 0 0 klauncher
    [ 1521.369936] [ 2473] 1000 2473 87003 1648 132 0 0 kio_file
    [ 1521.369938] [ 3468] 1000 3468 141313 5962 206 0 0 kate
    [ 1521.369940] [ 6331] 1000 6331 109512 2225 174 0 0 kate
    [ 1521.369942] [ 6733] 1000 6733 2983 150 12 0 0 make
    [ 1521.369943] [ 6760] 1000 6760 17243 1559 39 0 0 python
    [ 1521.369945] [ 6761] 1000 6761 1882 30 9 0 0 c++
    [ 1521.369947] [ 6762] 1000 6762 1816 23 9 0 0 collect2
    [ 1521.369949] [ 6763] 1000 6763 812215 809141 1588 0 0 ld
    [color=#FF40BF][ 1521.369951] Out of memory: Kill process 6763 (ld) score 402 or sacrifice child
    [ 1521.369953] Killed process 6763 (ld) total-vm:3248860kB, anon-rss:3236396kB, file-rss:168kB[/color]
    gcc -v
    Using built-in specs.
    COLLECT_GCC=gcc
    COLLECT_LTO_WRAPPER=/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2/lto-wrapper
    Target: x86_64-unknown-linux-gnu
    Configured with: /build/src/gcc-4.7.2/configure --prefix=/usr --libdir=/usr/lib --libexecdir=/usr/lib --mandir=/usr/share/man --infodir=/usr/share/info --with-bugurl=https://bugs.archlinux.org/ --enable-languages=c,c++,ada,fortran,go,lto,objc,obj-c++ --enable-shared --enable-threads=posix --with-system-zlib --enable-__cxa_atexit --disable-libunwind-exceptions --enable-clocale=gnu --disable-libstdcxx-pch --enable-libstdcxx-time --enable-gnu-unique-object --enable-linker-build-id --with-ppl --enable-cloog-backend=isl --disable-ppl-version-check --disable-cloog-version-check --enable-lto --enable-gold --enable-ld=default --enable-plugin --with-plugin-ld=ld.gold --with-linker-hash-style=gnu --disable-multilib --disable-libssp --disable-build-with-cxx --disable-build-poststage1-with-cxx --enable-checking=release
    Thread model: posix
    gcc version 4.7.2 (GCC)
    cat /proc/meminfo
    MemTotal: 8056856 kB
    MemFree: 1240336 kB
    Buffers: 278356 kB
    Cached: 4591784 kB
    SwapCached: 0 kB
    Active: 2700172 kB
    Inactive: 3860720 kB
    Active(anon): 2498432 kB
    Inactive(anon): 2441488 kB
    Active(file): 201740 kB
    Inactive(file): 1419232 kB
    Unevictable: 68 kB
    Mlocked: 68 kB
    SwapTotal: 0 kB
    SwapFree: 0 kB
    Dirty: 124 kB
    Writeback: 0 kB
    AnonPages: 1690996 kB
    Mapped: 218032 kB
    Shmem: 3249176 kB
    Slab: 172972 kB
    SReclaimable: 90332 kB
    SUnreclaim: 82640 kB
    KernelStack: 3264 kB
    PageTables: 38384 kB
    NFS_Unstable: 0 kB
    Bounce: 0 kB
    WritebackTmp: 0 kB
    CommitLimit: 4028428 kB
    Committed_AS: 6713700 kB
    VmallocTotal: 34359738367 kB
    VmallocUsed: 92972 kB
    VmallocChunk: 34359642076 kB
    HugePages_Total: 0
    HugePages_Free: 0
    HugePages_Rsvd: 0
    HugePages_Surp: 0
    Hugepagesize: 2048 kB
    DirectMap4k: 10240 kB
    DirectMap2M: 7237632 kB
    /etc/makepkg.conf
    # ARCHITECTURE, COMPILE FLAGS
    CARCH="x86_64"
    CHOST="x86_64-unknown-linux-gnu"
    CFLAGS="-march=native -O2 -pipe -fstack-protector --param=ssp-buffer-size=4 -D_FORTIFY_SOURCE=2"
    CXXFLAGS="${CFLAGS}"
    LDFLAGS="-Wl,-O1,--sort-common,--as-needed,-z,relro"
    MAKEFLAGS="-j5"
    thanks a lot
    Last edited by papu (2012-10-13 17:35:38)

    yes it's true it take 4gigs of 8 gigs i have,  but when i was using gentoo never had this problem when i was compiling any package and then my pc was 4gigs.
    df -hT
    Filesystem Type Size Used Avail Use% Mounted on
    rootfs rootfs 51G 7.7G 41G 17% /
    dev devtmpfs 3.9G 0 3.9G 0% /dev
    run tmpfs 3.9G 1.6M 3.9G 1% /run
    /dev/sdb2 ext4 51G 7.7G 41G 17% /
    tmpfs tmpfs 3.9G 76K 3.9G 1% /dev/shm
    tmpfs tmpfs 3.9G 0 3.9G 0% /sys/fs/cgroup
    tmpfs tmpfs 3.9G 56K 3.9G 1% /tmp
    /dev/sdc1 fuseblk 299G 237G 62G 80% /mnt/share
    /dev/sdb1 ext2 183M 28M 146M 16% /boot
    /dev/sdb3 ext4 18G 6.2G 11G 37% /home
    i am using vanilla kernel 3.6.0
    cat /etc/fstab
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    tmpfs /tmp tmpfs nodev,nosuid 0 0
    # /dev/sdb2
    UUID=412194cb-8953-4d02-94b4-d25d21bd7126 / ext4 rw,relatime,barrier=0 0 1
    #/dev/sdb1
    UUID=8bc28611-e3ee-4079-b311-33b9d4a0f36a /boot ext2 rw,relatime 0 2
    #/dev/sdb3
    UUID=2d6d63e0-a59f-439f-9c11-f6673055e65a /home ext4 rw,relatime,barrier=0 0 2
    #/dev/sdc1
    UUID=60F0F9D3F0F9AF80 /mnt/share ntfs-3g umask=0 0 0
    what i have to do ?
    thanks so much. friends!
    Last edited by papu (2012-10-13 19:44:17)

Maybe you are looking for