[solved] QT with QT_NO_STL?

Hello ARCH people.
I lately experienced a problem converting from std::string to QString compiling QT Applications. The following error i get:
error: conversion from `std::string' to non-scalar type `QString' requested
By searching google and reading a bit, i've found out that this problem seems to occur if QT_NO_STL is used as a compile option to QT.
Is there any specific reason why ARCH has set the option QT_NO_STL? Is it set, or am i wrong in this case?
I've never had this problem in any other distribution, so i'm asking you here for advice. Of course, i'd like if i dont have to recompile QT completely.
TIA,
STi

phrakture wrote:c++ exceptions are poor for performance.
No, you are wrong. How often do you throw exceptions? They are slower than error codes when they do occur, but faster when they don't occur (because you don't check for them with an "if"). So in average they are FASTER.
Enabling the exceptions doesn't mean it removes the use of throw and catch, it just means that Qt doesn't throw errors... you can still:
QString blah("some string");
throw std::exception(blah.latin1());
if you want to.... disabling the exception is only in the Qt library...
No, you are wrong. If you disable exceptions when compiling Qt, this means that you can't reliably throw an exception from a slot (a signal handler). If you do throw, then Qt's destructors won't be called when unwinding happens -> crash.
You will be able to use exceptions provided they don't go through Qt's code. This is restricting, actually. You can't say it's not an issue.
because you are assuming that enabling stl is 100% good with no downsides....
I can't see the downside. Enabling stl means anabling just a few additional functions in Qt, this is not bloat. Afterall, enabling stl doesn't cause any additional runtime dependency, because STL is contained in headers and you have to link to libc++ no matter what options you specify.
I am not claiming to be educated in it either, it's just that I trust the judgement calls made on these packages seeing as they work with no real hiccups....
I guess you don't develop in C++, so you should listen when somebody with greater knowledge on this subject tries to teach you something.
personally I hate c++ exceptions.... yes it's easy to use a try/catch block... but it just seems lazy to me.  It's personal preference.
Have you read Herb Sutter's "Exceptional C++?" Writing good exception-enabled code is actually a lot of additional work. But it usually leads to better, much more reliable code.
On another note, you were not "Just asking why Arch disables exceptions and stl in case of Qt?"....
vicious wrote:Seem the packager has NO idea about C++ and Qt. Qt should have another maintainer.
A much better course of action would have been "hey why is this done?" and await a response.... but the immediate calls for mutiny made me balk.
I agree here. I shouldn't talk this way. It's just the people who try to pretend they are more knowledgable than they really are who drive me crazy.
I understand it's your preference not to use exceptions. But it doesa lot of harm to disable exceptions when compiling libraries for people who want to use exceptions.
tyrade time
In my opinion, the key thing open source has going for it is this sort of "anyone can contribute" thing.  It's the exact opposite of the way us american's are taught to believe... here it's compitition that drives someone to make a good product better, it is not done for the sake of doing it.  OSS is very existensialist in a way.... now this is great, and I love the fact that someone like me could send a minor patch to linus, or kieth packard, or someone and have it accepted into code that millions of people use.  A big problem errupts when "hey here's something you did wrong, I fixed it" turns to "everything you do is wrong, I'm forking your project"... that's the attitude I get from this thread.... not an email to the maintainer or anything, but a flat out cry for a replacement....
If you want something to strive, help it, don't beat it down... you water your plants when they begin wilt, you don't chuck them in the trash and buy a cactus....
I agree completely. But I'm sure I'm not the only one responsible for the attitude you get from this thread.

Similar Messages

  • [SOLVED] Problem with "su" and "login".

    Hi, I reinstalled Arch Linux today when I updated "filesystem" and reboot appeared an error with root device, but I solved with installation CD.
    I restarted laptop after solving error with root device when I went to XFCE4 and open a terminal I tried to login as root with "su" and "su -" but showed this error:
    bash: su: command not found
    So I rebooted to start in text mode.
    When loaded SLIM I press ctrl+alt+F1 to enter text mode.
    I type root to login but did'nt appear to enter the pass, appeared again to enter login. I tried type my user but did the same.
    Someone can help me?
    Thanks!
    Last edited by surrealistic (2012-10-12 08:54:58)

    Thank you falconindy. I upgrade with --force because I search in Google the problem show "filesystem" when I tried install and I read that solution is upgrade "filesystem" with --foce.
    I tried to reinstall util-linux and now works. Thank you!

  • HT201303 last time i bought apps by using visa gift card so iam trying to use the kind of visa gift card and the system is denying the method payment ....any idea how to solve this with my itune account ?

    last time i bought apps by using visa gift card so iam trying to use the kind of visa gift card and the system is denying the method payment ....any idea how to solve this with my itune account ?

    A few reasons I can think of, but instead of shooting in the dark, lets check with experts who can look at your account.
    iTunes Store Support
    http://www.apple.com/emea/support/itunes/contact.html

  • Solve equation with one unknown parameter (x)

    I am working on a Windows Phone program that calculates taxes due. One of the calculations is a formula/equation with an unknown 'x'. I have searched a lot on solving equations with vb but this really exceeds my - admittedly small - math knob. Perhaps you
    can help me out.
    The equation to solve is (for example):
    ((25000 + x) * 0.06) + x = 16500
    The only unknown variable is x. Does someone know the proper code to solve this equation?

    Are you looking to code some super-intelligent maths equation solver that can solve any type of equation (if so - good luck!)? Or are the equations you want to solve all of the same type?
    The example you gave is of the form (A + x) * B + x = C. You can rearrange that to (1 + B) * x + A * B = C and from there to x = (C - A * B) / (1 + B).
    If all your equations are in that form, you just need to identify the value of A, B and C and use the last formula I gave to calculate x.
    I wasn't going to do that for him.
    Let's see here, he's building a program for tax calculation and he can't do 8th grade algebra??
    Still lost in code, just at a little higher level.

  • Solved problem with installation stuck at 25-30%

    I had this problem and I solved it with this:
    How to perform a "clean install" of Flash Player in Mac OS X

    @MFC_0809,
    Thanks for posting the screenshot, which confirms which installer (there are several) is being problematic.
    Please try the offline installer posted at the bottom of the Installation problems | Flash Player | Mac page, in the 'Still having problems' section.
    Haravikk the Horse,
    Do you license Flash Player?  That is not a public download URL for Flash Player, so curious as to where you obtained that URL from.
    Maria

  • When Apple solve problems with new iTunes calendar sync with windows and outlook?

    When Apple solve problems with new iTunes calendar sync with windows and outlook?

    Anything Downloaded with a Particular Apple ID is tied to that Apple ID and Cannot be Merged or Transferred to a Different Apple ID
    More Info here...
    Apple ID FAQs  >  http://support.apple.com/kb/HT5622
    If necessary...
    Contact iTunes Customer Service and request assistance
    Use this Link  >  Apple  Support  iTunes Store  Contact

  • Officejet Pro 8600 Plus: Scan to computer doesn't work with Win8.1 [solved: problem with USB3 Boost]

    Since I installed 8.1 (pro), I can't use my 8600 Pro software for scanning to the computer (via USB). It looks like there is no communication with the printer as for the scanning software, but printing works as usual.
    Tried to use the HP Printer Assistant: when I click the "Scan a Document or Photo" button, I get a "Scanner Unreachable" message. Also after a restart of the printer.
    When I try to scan from Photoshop, the system first finds the printer but then fails to connect to it.
    HP Update says that my system is up to date.
    HP Print And Scan Doctor 4.2 finds and shows the printer on the Welcome panel and seems to be able to communicate with it, to a point.
    It gives me a "Driver error" (sorry if is not called so in English--it takes over my keyboard language, which is French and says "Erreur du pilote") and states that no communication could be established between the scan driver and the printer, and that I should reboot the computer...
    So I did and reran the Doctor software, which gave me the very same error message. This time, I clicked "Ignore" and let it do its things. It ran a "Scan Test" twice and indicated that everything was fine or had been repaired. But strangely nothing happened on the printer during those "tests". And at the end, as it offered me to make a scan test myself, I clicked OK and I had the "Scanner Unreachable" error message again.
    So I made my scans on an USB key, as I had to do a few times already, and that went very well. But I'd be glad to have that "Scan to computer" function back. How?
    This question was solved.
    View Solution.

    Hi ajmch,
    I am sorry to hear that you are still having a scan issue. What antivirus program are you running on the computer? Can you try to scan through Paint and Windows Fax and Scan on the computer? When you run the HP Print and Scan Doctor what error message are you getting?
    In this document for Windows: HP Scan Software Does Not Open or Scan on a Printer Using a USB Connection are steps  for checking the WIA and for checking software.
    You can also Use the System File Checker tool to repair missing or corrupted system files that may resolve this issue.
    Let me know if this helped.
    Thank you,
    I worked on behalf of HP.

  • [Solved] Problem with mouse cursor gnome3

    Hi
    I moved back to Archlinux after a brief stay in LMDE, and test the new GNOME3, I am unpleasantly surprised that after a while the mouse is moving himself toward the top of the screen. I searched the internet, and seems to be a problem with the gnome-shell for some time, but I am not able to find the solution.
    Momentarily solved by restarting gnome with ALT + F2 + r + Enter, but after a while, same thing happens again.
    Can be well with photography programs, but of course if you do not solve this issue, GNOME, I can not use
    Can you give me a helping hand?
    Thanks and regards
    Rafa
    Last edited by Warper4 (2011-08-16 18:09:50)

    I don't know about Gnome, but my laptop had that problem.  I tracked it down the 3d accelerometer being used to simulate a joystick.  Of course, I also had (previously) set up another joystick to operate as a mouse.  In the end, the cursor "drifted" when the laptop was not level
    YMMV

  • [Not really solved] Video with Pidgin is not working

    Hi,
    I would like to use video chat with Pidgin, but it doesn't work on my computer.
    I have a webcam (:D), and it seems that it is correctly recognized because ` vlc v4l2:// ' shows me my face. I'm using XMPP (with a GMail account, but this shouldn't matter) and want to chat with someone using Ubuntu and Pidgin 2.6.6 (she has a webcam which seems to work correctly)
    In the "Voice/Video settings" plugin, there are three options in the field Video → Input → Plugins : "default", "Test Input" and "Video4Linux" (and only "default" in the field Device, for the three plugins)
    "default" and "Video4Linux" give
    (21:03:34) Error with your webcam
    (21:03:34) Conference error
    (21:03:34) The call has been terminated.
    as soon as I try to start a video call, and "Test Input" only shows me a test pattern (but I guess this is the meaning of "Test Input" )
    BTW, I wonder why there is only "Video4Linux" and not "Video4Linux2"
    The Pidgin's FAQ says this
    How do I configure my microphone/webcam?
    Currently, the command-line gstreamer-properties program is used. This is only available on GNOME-based systems. On other systems, Pidgin makes its best guess as to which device to use. A plugin is planned to support this functionality on other systems.
    But I don't use GNOME, my desktop environment is LXDE.
    What could I do to make my webcam work with Pidgin?
    Thank you!
    Edit:
    - my system is up to date
    - I have farsight2 and gstreamer0.10-{good,bad,ugly} installed
    Fractal
    Last edited by Fractal (2010-06-13 15:50:01)

    Well, the problem is not really solved actually.
    Pidgin recognizes my webcam but I still cannot open a video chat.
    When I try to start a video chat, the following errors are displayed in the debug window :
    (17:38:41) mediamanager: gst pipeline error: Could not encode stream.
    (17:38:41) mediamanager: Debug details: gstx264enc.c(737): gst_x264_enc_header_buf (): /GstPipeline:pipeline0/GstBin:conf_0x306c000/FsRtpConference:fsrtpconference1/GstBin:discover_1_97/GstX264Enc:x264enc0:
    Unexpected x264 header.
    When the other person tries to call me, I have something like that :
    (17:43:23) jingle-rtp: received codec: 32: video MPV clock:90000 channels:0
    (17:43:23) jingle-rtp: received codec: 26: video JPEG clock:90000 channels:0
    (17:43:23) jingle-rtp: received codec: 97: video H263-1998 clock:90000 channels:0
    (17:43:23) jingle-rtp: received codec: 34: video H263 clock:90000 channels:0
    (17:43:23) jingle-rtp: received codec: 98: video DV clock:90000 channels:0
    (17:43:23) jabber: setting param stun-ip for stream using auto-discovered IP: 209.85.229.126
    (17:43:23) jabber: setting param stun-port for stream using auto-discovered port: 19302
    (17:43:23) mediamanager: Couldn't read fs-element.conf: No such file or directory
    (17:43:23) backend-fs2: Couldn't read fs-codec.conf: No such file or directory
    (17:43:23) backend-fs2: Creating hash table for sessions
    (17:43:23) backend-fs2: connecting pad: success
    (17:43:23) backend-fs2: Creating hash table for participants
    (17:43:23) media: Creating hash table for sessions
    (17:43:23) gtkmedia: state: 0 sid: video-session name: (null)
    (17:43:23) jingle-rtp: state-changed: state 0 id: video-session name: (null)
    (17:43:23) backend-fs2: got new local candidate: 1
    (17:43:23) backend-fs2: got new local candidate: 1
    (17:43:23) backend-fs2: farsight-component-state-changed: component: 1 state: GATHERING
    (17:43:23) backend-fs2: farsight-component-state-changed: component: 2 state: GATHERING
    (17:43:23) jingle-rtp: jingle_rtp_codecs_changed_cb: session_id: video-session jingle_session: 0x212b6a0
    (17:43:23) jingle-rtp: jingle_rtp_candidates_prepared_cb
    (17:43:23) jingle: jingle_iceudp_finalize
    (17:43:23) jingle: jingle_transport_finalize
    (17:43:26) backend-fs2: farsight-send-codec-changed: codec: 96: video THEORA clock:90000 channels:0 delivery-method=inline
    (17:43:27) jingle-rtp: stream-info: type 1 id: video-session name: (null)
    (17:43:27) jingle-rtp: stream-info: type 1 id: (null) name: (null)
    (17:43:27) backend-fs2: farsight-component-state-changed: component: 1 state: CONNECTING
    (17:43:27) backend-fs2: farsight-component-state-changed: component: 2 state: CONNECTING
    (17:43:36) backend-fs2: farsight-component-state-changed: component: 1 state: FAILED
    (17:43:36) backend-fs2: farsight-error: 108: Could not establish connection
    (17:43:36) backend-fs2: farsight-component-state-changed: component: 2 state: FAILED
    It seems that my Pidgin want to encode the flux in H264 and the other wants Theora
    Fractal

  • [SOLVED] Problem with cron and unison

    I've got a very strange problem I cannot solve.
    I have set unison to synchronize my laptop (mercure) against my home computer (jupiter). It works really fine.
    I can run it from the terminal:
    unison -batch CERTES
    and it runs good.
    I have set ssh keys so that id doesn't ask me for a password. And that runs fine too because if I do ssh to jupiter, I log in without any password.
    That's what the unison profile looks like:
    [jordi@mercure ~]$ cat .unison/CERTES.prf
    root = /home/jordi/CERTES
    root = ssh://[email protected]//home/jordi/Macbook
    After this, everything is ok.
    Then I set a cron job so this is done once an hour everyday using crontab -e:
    0 * * * * /usr/bin/unison -batch CERTES
    And every hour I receive this mail:
    Contacting server...
    Permission denied, please try again.
    Permission denied, please try again.
    Permission denied (publickey,password).
    Fatal error: Lost connection with the server
    Ok then, this seems like cron can't log into jupiter with ssh as I thought. So I modify /etc/ssh/sshd_config to put this option:
    PermitUserEnvironment yes
    I do the same on mercure, even if I think it's useless, but...
    I always receive the same error and I can't get the 2 folders synchronzed.
    Before my hard drive crashed a week ago, I had unison running by cron and it was working nice, but now I'm unable to do it again...
    Does anyone has a clue about what's happening and how to solve it?
    Last edited by jcerdan (2008-07-04 09:52:00)

    I have tried adding source to the crontab but I got the same error.
    This is the env output from cron:
    CONSOLE=/dev/console
    SHELL=/bin/sh
    CMDLINE=root=/dev/disk/by-uuid/25d2ee78-61ac-4b43-929a-28180b156786 ro
    TERM=linux
    rootdelay=0
    USER=jordi
    INIT_VERSION=sysvinit-2.86
    ro=y
    PATH=/bin:/usr/bin:/sbin:/usr/sbin
    _=/usr/bin/env
    RUNLEVEL=3
    PWD=/home/jordi
    PREVLEVEL=N
    root=/dev/disk/by-uuid/25d2ee78-61ac-4b43-929a-28180b156786
    SHLVL=3
    HOME=/home/jordi
    kinit_params= rootfstype=ext3 rootdelay=0
    And this is the env output from the terminal:
    [jordi@mercure ~]$ env
    XDG_DATA_HOME=/home/jordi/.local/share
    SHELL=/bin/bash
    TERM=xterm
    DESKTOP_STARTUP_ID=
    HISTSIZE=1000
    QT_XFT=true
    GTK_RC_FILES=/etc/gtk/gtkrc:/home/jordi/.gtkrc-1.2-gnome2
    WINDOWID=54526543
    QTDIR=/opt/qt
    GTK_MODULES=gnomebreakpad
    USER=jordi
    GNOME_KEYRING_SOCKET=/tmp/keyring-C1rLTl/socket
    SSH_AUTH_SOCK=/tmp/keyring-C1rLTl/ssh
    KDEDIR=/opt/kde
    OOO_FORCE_DESKTOP=gnome
    SESSION_MANAGER=local/mercure:/tmp/.ICE-unix/3449
    USERNAME=jordi
    MOZ_PLUGIN_PATH=/usr/lib/mozilla/plugins:/opt/mozilla/lib/plugins
    XDG_CONFIG_DIRS=/etc/xdg:/opt/kde/etc/xdg
    PATH=/bin:/usr/bin:/sbin:/usr/sbin:/opt/java/jre/bin:/opt/kde/bin:/usr/bin/perlbin/site:/usr/bin/perlbin/vendor:/usr/bin/perlbin/core:/opt/qt/bin
    DESKTOP_SESSION=gnome
    GDM_XSERVER_LOCATION=local
    INPUTRC=/etc/inputrc
    PWD=/home/jordi
    JAVA_HOME=/opt/java/jre
    LIBXCB_ALLOW_SLOPPY_LOCK=true
    EDITOR=vi
    LANG=es_ES.utf8
    GNOME_KEYRING_PID=3448
    GDM_LANG=es_ES.utf8
    KDEDIRS=/opt/kde:/usr
    PS1=[\u@\h \W]\$
    PS3=>
    GDMSESSION=gnome
    PS2=>
    HISTCONTROL=erasedups
    SHLVL=1
    HOME=/home/jordi
    PS4=+
    XDG_CONFIG_HOME=/home/jordi/.config
    GNOME_DESKTOP_SESSION_ID=Default
    XDG_CACHE_HOME=/home/jordi/.cache
    LOGNAME=jordi
    LESS=-R
    CLASSPATH=:/opt/java/jre/lib
    DBUS_SESSION_BUS_ADDRESS=unix:abstract=/tmp/dbus-5ncgDqnWoO,guid=1e082c45cdb24be2d415016e486c820b
    XDG_DATA_DIRS=/usr/share:/usr/local/share:/usr/local/share/:/usr/share/:/usr/share/gdm/:/opt/kde/share
    PKG_CONFIG_PATH=:/opt/kde/lib/pkgconfig:/opt/qt/lib/pkgconfig
    WINDOWPATH=7
    DISPLAY=:0.0
    J2REDIR=/opt/java/jre
    G_BROKEN_FILENAMES=yes
    COLORTERM=gnome-terminal
    XAUTHORITY=/home/jordi/.Xauthority
    _=/usr/bin/env
    A lot of vars are different, but they refer to the Gnome environment. I can see that SHELL, TERM are different, but I don't know what can make the difference.
    Do you think there's a problem there?

  • [SOLVED] Problem with FGLRX and Ati HD4650

    Hi! I'm fede1587 and I'm a new user of Archlinux (but not a new user of Linux). I love it and the KISS philosophy.
    Four days ago I intalled the new version of Arch but unfortunately I have a very big problem and after 4 days of work I have not been able to solve it.
    I have a Sony Vaio VGN-FW41J with a ATI HD4650. I need Ati proprietary drivers, because with the open source ones, the laptop become too hot. After installing the catalyst legacy pakages (234k, catalyst-utils and catalyst-hook) both from AUR and from the repository (following the great guide in ArchWiki) and configuring, when I lanunch startx, sometimes Gnome starts and sometimes I got black screen with horizontal lines.
    I'm searching for solutions from 4 days and I read a lot of topic, and I try a lot of things, but nothing!
    Accourding to the various threads, I have:
    1) Load the module fglrx
    2) Blacklist the radeon module
    3) Re-run mkinitcpio
    4) Run aticonfig --initial
    5) Run aticonfig --acpi-services=off
    6) Add nomodeset to grub
    After I tried to change kernel (thinking about some incompatibilities). I installed linux-lts, and I try with this kernel, but I have the same issue. I also try whit Bridge Linux, but nothing changed.
    This is the only output found in my error.log (Xorg.X.log is empty):
    Aug 26 02:36:51 localhost kernel: [   13.326922] [fglrx:KCL_MEM_VM_MapRegion] *ERROR* remap_pfn_range failed
    Aug 26 02:36:51 localhost kernel: [   13.326932] [fglrx:__create_mapping] *ERROR* Can not get virtual address
    Aug 26 02:36:51 localhost kernel: [   13.326935] [fglrx:__mc_heap_map_virtual_space] *ERROR* Failed to map the virtual space
    Aug 26 02:36:51 localhost kernel: [   13.326938] [fglrx:mc_heap_map_virtual_space] *ERROR* Can not get virtual address
    Aug 26 02:36:51 localhost kernel: [   13.326940] [fglrx:MCIL_GetVirtualAddressInDescriptor] *ERROR* Can not get the virtual address
    Aug 26 02:36:55 localhost kernel: [   16.622882] [fglrx:KCL_MEM_VM_MapRegion] *ERROR* remap_pfn_range failed
    Aug 26 02:36:55 localhost kernel: [   16.622892] [fglrx:__create_mapping] *ERROR* Can not get virtual address
    Aug 26 02:36:55 localhost kernel: [   16.622895] [fglrx:__mc_heap_map_virtual_space] *ERROR* Failed to map the virtual space
    Aug 26 02:36:55 localhost kernel: [   16.622898] [fglrx:mc_heap_map_virtual_space] *ERROR* Can not get virtual address
    Aug 26 02:36:55 localhost kernel: [   16.622900] [fglrx:MCIL_GetVirtualAddressInDescriptor] *ERROR* Can not get the virtual address
    Aug 26 02:36:55 localhost dhcpcd[529]: wlan0: sendmsg: Cannot assign requested address
    Aug 26 02:36:56 localhost kernel: [   18.364335] [fglrx:KCL_MEM_VM_MapRegion] *ERROR* remap_pfn_range failed
    Aug 26 02:36:56 localhost kernel: [   18.364346] [fglrx:__create_mapping] *ERROR* Can not get virtual address
    AuAug 26 02:40:18 localhost /usr/sbin/crond[452]: (CRON) INFO (Syslog will be used instead of sendmail.): File o directory non esistente
    Aug 26 02:40:25 localhost dhcpcd[476]: wlan0: sendmsg: Cannot assign requested address
    I hope you can help me, because I love this distro but I can't use it with the open source driver of the ATI card (too hot) :-(
    Thanks a lot for the help and sorry for my bad english.
    Last edited by fede1587 (2012-08-30 22:15:57)

    Hi. Now I found in my Xorg.0.log the following lines:
    [    16.893] (II) fglrx(0): [uki] DRM interface version 1.0
    [    16.893] (II) fglrx(0): [uki] created "fglrx" driver at busid "PCI:1:0:0"
    [    16.893] (II) fglrx(0): [uki] added 8192 byte SAREA at 0x8000
    [    16.893] (II) fglrx(0): [uki] mapped SAREA 0x8000 to 0x7f82a04c9000
    [    16.893] (II) fglrx(0): [uki] framebuffer handle = 0x9000
    [    16.893] (II) fglrx(0): [uki] added 1 reserved context for kernel
    [    16.893] (II) fglrx(0): swlDriScreenInit done
    [    16.893] (II) fglrx(0): Kernel Module Version Information:
    [    16.893] (II) fglrx(0):     Name: fglrx
    [    16.893] (II) fglrx(0):     Version: 8.97.2
    [    16.893] (II) fglrx(0):     Date: Jul  4 2012
    [    16.893] (II) fglrx(0):     Desc: ATI FireGL DRM kernel module
    [    16.893] (II) fglrx(0): Kernel Module version matches driver.
    [    16.893] (II) fglrx(0): Kernel Module Build Time Information:
    [    16.893] (II) fglrx(0):     Build-Kernel UTS_RELEASE:        3.4.9-1-ARCH
    [    16.893] (II) fglrx(0):     Build-Kernel MODVERSIONS:        no
    [    16.893] (II) fglrx(0):     Build-Kernel __SMP__:            no
    [    16.893] (II) fglrx(0):     Build-Kernel PAGE_SIZE:          0x1000
    [    16.893] (II) fglrx(0): [uki] register handle = 0x0000a000
    [    16.916] (EE) fglrx(0): Not enough video memory to allocate primary surface (frame buffer).
    [    16.916] (II) fglrx(0): Shutdown CMMQS
    [    16.916] (II) fglrx(0): [uki] removed 1 reserved context for kernel
    [    16.916] (II) fglrx(0): [uki] unmapping 8192 bytes of SAREA 0x8000 at 0x7f82a04c9000
    [    16.916] (WW) fglrx(0): ***********************************************************
    [    16.916] (WW) fglrx(0): * DRI initialization failed                               *
    [    16.916] (WW) fglrx(0): * kernel module (fglrx.ko) may be missing or incompatible *
    [    16.916] (WW) fglrx(0): * 2D and 3D acceleration disabled                         *
    [    16.916] (WW) fglrx(0): ***********************************************************
    Do you have an idea of what it could be?
    Thanks a lot...

  • [SOLVED] Problem with Vuze and Java RE x86_64

    Hello all !
    First of all, sorry if this issue was posted before, but find no related issue similar to mine.
    Using pacman, I successfully installed JRE (version 6u16-1-x86_64) and Vuze (version 4.2.0.8-1, former Azureus), but Vuze doesn't want to start.
    Running the Vuze's executable in the gnome-terminal, I notice messages from vuze informing it can't find Java executable ("Java exec not found in PATH, starting auto-search..." AND "OOPS, unable to locate java exec in /usr/java/latest /usr/java /usr/lib/jvm/latest /usr/lib/jvm hierarchy") and therefore cannot start the Vuze UI. I also tried but failed in verify the version of Java with the command "java -version" due to command not found.
    What is the best approach to fix this problem with Java 64 bits?
    Thanks in advance
    Last edited by josephg (2009-09-28 18:55:56)

    peart wrote:Just log out and back in, most likely.  There are scripts in /etc/profile.d/ that need to be run to set up your java environment.  They get run automatically when you log in.
    Yep, perfect answer - issue solved. Thanks a lot!

  • [SOLVED] Problem with symlinking and PKGBUILD

    Hello Everyone,
    I have created the javafx-sdk package on AUR, and make the installation a lot simplier. But I am having a problem with the creationg of the symlink.
    So I have 1 binary file "javafxpackager" in the "bin" folder that I need to symlink into /usr/bin. This is the method I am using
    mkdir -p ${pkgdir}/usr/bin
    chmod +x ${pkgdir}/usr/lib/jvm/${pkgname}/bin/javafxpackager
    ln -s ${pkgdir}/usr/lib/jvm/${pkgname}/bin/javafxpackager ${pkgdir}/usr/bin/javafxpackager
    My problem is that this method doesn't work at all. The only way to me after the installation to make the symlink working is by doing this :
    ls chmod +x /usr/lib/jvm/javafx-sdk/bin/javafxpackager
    sudo ln -s /usr/lib/jvm/javafx-sdk/bin/javafxpackager /usr/bin/
    In order to make it work.
    How can I solve this problem ?
    Thanks in Advance,
    Luis Da Costa
    Last edited by aliasbody (2012-06-26 00:27:15)

    Problem solved ! Here's how to do it :
    mkdir -p ${pkgdir}/usr/bin
    chmod +x ${pkgdir}/usr/lib/jvm/${pkgname}/bin/javafxpackager
    ln -s /usr/lib/jvm/javafx-sdk/bin/javafxpackager ${pkgdir}/usr/bin/javafxpackager
    I didn't know, but you can symlink a file that doesn't exist yet, and this is what I did. I use and non existing (yet) javafxpackager binary from the non existing (yet) "/usr/lib/jvm/javafx-sdk/bin/" folder, and send it to the ${pkgdir}/usr/bin/javafxpackager, that will be copied into the original "/usr/bin/" folder.
    So this problem is solved and I'm happy
    (PS : thanks @ngoonee and @Trilby for you help ! Thanks to you I found out a little bit faster where the problem was ^^)

  • [SOLVED]problem with Bumblebee and Kernel 3.11-1-2

    Hi!
    I'm having problems with Bumblebee from Kernel 3.10 but I read that Kernel 3.11 solved the problem, I searched and tried possible solutions to my problem unsuccessfully
    The message shows me when I run "optirun":
    [ 5454.262739] [ERROR]Cannot access secondary GPU - error: [XORG] (EE) No devices detected.
    [ 5454.262776] [ERROR]Aborting because fallback start is disabled.
    I've tried with wiki's solutions, forums solutions...
    Configurations files:
    /etc/bumblebee/bumblebee.conf:
    # Configuration file for Bumblebee. Values should **not** be put between quotes
    ## Server options. Any change made in this section will need a server restart
    # to take effect.
    [bumblebeed]
    # The secondary Xorg server DISPLAY number
    VirtualDisplay=:8
    # Should the unused Xorg server be kept running? Set this to true if waiting
    # for X to be ready is too long and don't need power management at all.
    KeepUnusedXServer=false
    # The name of the Bumbleblee server group name (GID name)
    ServerGroup=bumblebee
    # Card power state at exit. Set to false if the card shoud be ON when Bumblebee
    # server exits.
    TurnCardOffAtExit=false
    # The default behavior of '-f' option on optirun. If set to "true", '-f' will
    # be ignored.
    NoEcoModeOverride=false
    # The Driver used by Bumblebee server. If this value is not set (or empty),
    # auto-detection is performed. The available drivers are nvidia and nouveau
    # (See also the driver-specific sections below)
    Driver=
    # Directory with a dummy config file to pass as a -configdir to secondary X
    XorgConfDir=/etc/bumblebee/xorg.conf.d
    ## Client options. Will take effect on the next optirun executed.
    [optirun]
    # Acceleration/ rendering bridge, possible values are auto, virtualgl and
    # primus.
    Bridge=virtualgl
    # The method used for VirtualGL to transport frames between X servers.
    # Possible values are proxy, jpeg, rgb, xv and yuv.
    VGLTransport=proxy
    # List of paths which are searched for the primus libGL.so.1 when using
    # the primus bridge
    PrimusLibraryPath=/usr/lib/primus:/usr/lib32/primus
    # Should the program run under optirun even if Bumblebee server or nvidia card
    # is not available?
    AllowFallbackToIGC=false
    # Driver-specific settings are grouped under [driver-NAME]. The sections are
    # parsed if the Driver setting in [bumblebeed] is set to NAME (or if auto-
    # detection resolves to NAME).
    # PMMethod: method to use for saving power by disabling the nvidia card, valid
    # values are: auto - automatically detect which PM method to use
    # bbswitch - new in BB 3, recommended if available
    # switcheroo - vga_switcheroo method, use at your own risk
    # none - disable PM completely
    # https://github.com/Bumblebee-Project/Bumblebee/wiki/Comparison-of-PM-methods
    ## Section with nvidia driver specific options, only parsed if Driver=nvidia
    [driver-nvidia]
    # Module name to load, defaults to Driver if empty or unset
    KernelDriver=nvidia
    PMMethod=auto
    # colon-separated path to the nvidia libraries
    LibraryPath=/usr/lib/nvidia:/usr/lib32/nvidia
    # comma-separated path of the directory containing nvidia_drv.so and the
    # default Xorg modules path
    XorgModulePath=/usr/lib/nvidia/xorg/,/usr/lib/xorg/modules
    XorgConfFile=/etc/bumblebee/xorg.conf.nvidia
    ## Section with nouveau driver specific options, only parsed if Driver=nouveau
    [driver-nouveau]
    KernelDriver=nouveau
    PMMethod=auto
    XorgConfFile=/etc/bumblebee/xorg.conf.nouveau
    /etc/bumblebee/xorg.conf.nvidia
    Section "ServerLayout"
    Identifier "Layout0"
    Option "AutoAddDevices" "false"
    Option "AutoAddGPU" "false"
    EndSection
    Section "Device"
    Identifier "DiscreteNvidia"
    Driver "nvidia"
    VendorName "NVIDIA Corporation"
    # If the X server does not automatically detect your VGA device,
    # you can manually set it here.
    # To get the BusID prop, run `lspci | egrep 'VGA|3D'` and input the data
    # as you see in the commented example.
    # This Setting may be needed in some platforms with more than one
    # nvidia card, which may confuse the proprietary driver (e.g.,
    # trying to take ownership of the wrong device). Also needed on Ubuntu 13.04.
    BusID "PCI:01:00.0"
    # Setting ProbeAllGpus to false prevents the new proprietary driver
    # instance spawned to try to control the integrated graphics card,
    # which is already being managed outside bumblebee.
    # This option doesn't hurt and it is required on platforms running
    # more than one nvidia graphics card with the proprietary driver.
    # (E.g. Macbook Pro pre-2010 with nVidia 9400M + 9600M GT).
    # If this option is not set, the new Xorg may blacken the screen and
    # render it unusable (unless you have some way to run killall Xorg).
    Option "ProbeAllGpus" "false"
    Option "NoLogo" "true"
    Option "UseEDID" "false"
    Option "UseDisplayDevice" "none"
    EndSection
    lspci |grep NVIDIA:
    01:00.0 VGA compatible controller: NVIDIA Corporation GK107M [GeForce GT 640M] (rev a1)
    I hope can help me
    Thanks!
    Last edited by surrealistic (2013-11-04 18:46:35)

    My /var/log/Xorg.8.log:
    [ 337.161]
    X.Org X Server 1.14.4
    Release Date: 2013-10-31
    [ 337.161] X Protocol Version 11, Revision 0
    [ 337.161] Build Operating System: Linux 3.11.6-1-ARCH x86_64
    [ 337.161] Current Operating System: Linux h131h-laptop 3.11.6-1-ARCH #1 SMP PREEMPT Fri Oct 18 23:22:36 CEST 2013 x86_64
    [ 337.161] Kernel command line: BOOT_IMAGE=/boot/vmlinuz-linux root=UUID=4eb69984-2228-4aba-9863-28a7fa854e51 rw quiet
    [ 337.161] Build Date: 01 November 2013 05:10:48PM
    [ 337.161]
    [ 337.161] Current version of pixman: 0.30.2
    [ 337.161] Before reporting problems, check http://wiki.x.org
    to make sure that you have the latest version.
    [ 337.161] Markers: (--) probed, (**) from config file, (==) default setting,
    (++) from command line, (!!) notice, (II) informational,
    (WW) warning, (EE) error, (NI) not implemented, (??) unknown.
    [ 337.161] (==) Log file: "/var/log/Xorg.8.log", Time: Mon Nov 4 14:20:04 2013
    [ 337.161] (++) Using config file: "/etc/bumblebee/xorg.conf.nvidia"
    [ 337.161] (++) Using config directory: "/etc/bumblebee/xorg.conf.d"
    [ 337.161] (==) ServerLayout "Layout0"
    [ 337.161] (==) No screen section available. Using defaults.
    [ 337.161] (**) |-->Screen "Default Screen Section" (0)
    [ 337.161] (**) | |-->Monitor "<default monitor>"
    [ 337.161] (==) No device specified for screen "Default Screen Section".
    Using the first device section listed.
    [ 337.161] (**) | |-->Device "DiscreteNvidia"
    [ 337.161] (==) No monitor specified for screen "Default Screen Section".
    Using a default monitor configuration.
    [ 337.161] (**) Option "AutoAddDevices" "false"
    [ 337.161] (**) Option "AutoAddGPU" "false"
    [ 337.161] (**) Not automatically adding devices
    [ 337.161] (==) Automatically enabling devices
    [ 337.161] (**) Not automatically adding GPU devices
    [ 337.161] (WW) The directory "/usr/share/fonts/OTF/" does not exist.
    [ 337.162] Entry deleted from font path.
    [ 337.162] (WW) `fonts.dir' not found (or not valid) in "/usr/share/fonts/100dpi/".
    [ 337.162] Entry deleted from font path.
    [ 337.162] (Run 'mkfontdir' on "/usr/share/fonts/100dpi/").
    [ 337.162] (WW) `fonts.dir' not found (or not valid) in "/usr/share/fonts/75dpi/".
    [ 337.162] Entry deleted from font path.
    [ 337.162] (Run 'mkfontdir' on "/usr/share/fonts/75dpi/").
    [ 337.162] (==) FontPath set to:
    /usr/share/fonts/misc/,
    /usr/share/fonts/TTF/,
    /usr/share/fonts/Type1/
    [ 337.162] (++) ModulePath set to "/usr/lib/nvidia/xorg/,/usr/lib/xorg/modules"
    [ 337.162] (==) |-->Input Device "<default pointer>"
    [ 337.162] (==) |-->Input Device "<default keyboard>"
    [ 337.162] (==) The core pointer device wasn't specified explicitly in the layout.
    Using the default mouse configuration.
    [ 337.162] (==) The core keyboard device wasn't specified explicitly in the layout.
    Using the default keyboard configuration.
    [ 337.162] (II) Loader magic: 0x7fdc20
    [ 337.162] (II) Module ABI versions:
    [ 337.162] X.Org ANSI C Emulation: 0.4
    [ 337.162] X.Org Video Driver: 14.1
    [ 337.162] X.Org XInput driver : 19.1
    [ 337.162] X.Org Server Extension : 7.0
    [ 337.162] (II) xfree86: Adding drm device (/dev/dri/card1)
    [ 337.162] (II) xfree86: Adding drm device (/dev/dri/card0)
    [ 337.162] setversion 1.4 failed
    [ 337.163] (--) PCI:*(0:1:0:0) 10de:0fd2:1025:0648 rev 161, Mem @ 0xb2000000/16777216, 0xa0000000/268435456, 0xb0000000/33554432, I/O @ 0x00002000/128
    [ 337.163] Initializing built-in extension Generic Event Extension
    [ 337.163] Initializing built-in extension SHAPE
    [ 337.163] Initializing built-in extension MIT-SHM
    [ 337.163] Initializing built-in extension XInputExtension
    [ 337.163] Initializing built-in extension XTEST
    [ 337.163] Initializing built-in extension BIG-REQUESTS
    [ 337.163] Initializing built-in extension SYNC
    [ 337.163] Initializing built-in extension XKEYBOARD
    [ 337.163] Initializing built-in extension XC-MISC
    [ 337.163] Initializing built-in extension SECURITY
    [ 337.163] Initializing built-in extension XINERAMA
    [ 337.163] Initializing built-in extension XFIXES
    [ 337.163] Initializing built-in extension RENDER
    [ 337.163] Initializing built-in extension RANDR
    [ 337.164] Initializing built-in extension COMPOSITE
    [ 337.164] Initializing built-in extension DAMAGE
    [ 337.164] Initializing built-in extension MIT-SCREEN-SAVER
    [ 337.164] Initializing built-in extension DOUBLE-BUFFER
    [ 337.164] Initializing built-in extension RECORD
    [ 337.164] Initializing built-in extension DPMS
    [ 337.164] Initializing built-in extension X-Resource
    [ 337.164] Initializing built-in extension XVideo
    [ 337.164] Initializing built-in extension XVideo-MotionCompensation
    [ 337.164] Initializing built-in extension XFree86-VidModeExtension
    [ 337.164] Initializing built-in extension XFree86-DGA
    [ 337.164] Initializing built-in extension XFree86-DRI
    [ 337.164] Initializing built-in extension DRI2
    [ 337.164] (II) LoadModule: "glx"
    [ 337.164] (II) Loading /usr/lib/nvidia/xorg/modules/extensions/libglx.so
    [ 337.172] (II) Module glx: vendor="NVIDIA Corporation"
    [ 337.172] compiled for 4.0.2, module version = 1.0.0
    [ 337.172] Module class: X.Org Server Extension
    [ 337.172] (II) NVIDIA GLX Module 325.15 Wed Jul 31 18:12:00 PDT 2013
    [ 337.172] Loading extension GLX
    [ 337.172] (II) LoadModule: "nvidia"
    [ 337.172] (II) Loading /usr/lib/xorg/modules/drivers/nvidia_drv.so
    [ 337.172] (II) Module nvidia: vendor="NVIDIA Corporation"
    [ 337.172] compiled for 4.0.2, module version = 1.0.0
    [ 337.172] Module class: X.Org Video Driver
    [ 337.172] (II) LoadModule: "mouse"
    [ 337.172] (II) Loading /usr/lib/xorg/modules/input/mouse_drv.so
    [ 337.172] (II) Module mouse: vendor="X.Org Foundation"
    [ 337.172] compiled for 1.14.0, module version = 1.9.0
    [ 337.172] Module class: X.Org XInput Driver
    [ 337.172] ABI class: X.Org XInput driver, version 19.1
    [ 337.172] (II) LoadModule: "kbd"
    [ 337.173] (WW) Warning, couldn't open module kbd
    [ 337.173] (II) UnloadModule: "kbd"
    [ 337.173] (II) Unloading kbd
    [ 337.173] (EE) Failed to load module "kbd" (module does not exist, 0)
    [ 337.173] (II) NVIDIA dlloader X Driver 325.15 Wed Jul 31 17:50:57 PDT 2013
    [ 337.173] (II) NVIDIA Unified Driver for all Supported NVIDIA GPUs
    [ 337.173] (--) using VT number 1
    [ 337.173] (II) Loading sub module "fb"
    [ 337.173] (II) LoadModule: "fb"
    [ 337.173] (II) Loading /usr/lib/xorg/modules/libfb.so
    [ 337.173] (II) Module fb: vendor="X.Org Foundation"
    [ 337.173] compiled for 1.14.4, module version = 1.0.0
    [ 337.173] ABI class: X.Org ANSI C Emulation, version 0.4
    [ 337.173] (WW) Unresolved symbol: fbGetGCPrivateKey
    [ 337.173] (II) Loading sub module "wfb"
    [ 337.173] (II) LoadModule: "wfb"
    [ 337.173] (II) Loading /usr/lib/xorg/modules/libwfb.so
    [ 337.173] (II) Module wfb: vendor="X.Org Foundation"
    [ 337.173] compiled for 1.14.4, module version = 1.0.0
    [ 337.173] ABI class: X.Org ANSI C Emulation, version 0.4
    [ 337.173] (II) Loading sub module "shadow"
    [ 337.173] (II) LoadModule: "shadow"
    [ 337.173] (II) Loading /usr/lib/xorg/modules/libshadow.so
    [ 337.173] (II) Module shadow: vendor="X.Org Foundation"
    [ 337.173] compiled for 1.14.4, module version = 1.1.0
    [ 337.173] ABI class: X.Org ANSI C Emulation, version 0.4
    [ 337.173] (II) Loading sub module "ramdac"
    [ 337.173] (II) LoadModule: "ramdac"
    [ 337.173] (II) Module "ramdac" already built-in
    [ 337.174] (II) NVIDIA(0): Creating default Display subsection in Screen section
    "Default Screen Section" for depth/fbbpp 24/32
    [ 337.174] (==) NVIDIA(0): Depth 24, (==) framebuffer bpp 32
    [ 337.174] (==) NVIDIA(0): RGB weight 888
    [ 337.174] (==) NVIDIA(0): Default visual is TrueColor
    [ 337.174] (==) NVIDIA(0): Using gamma correction (1.0, 1.0, 1.0)
    [ 337.174] (**) NVIDIA(0): Option "NoLogo" "true"
    [ 337.174] (**) NVIDIA(0): Option "ProbeAllGpus" "false"
    [ 337.174] (**) NVIDIA(0): Option "UseEDID" "false"
    [ 337.174] (**) NVIDIA(0): Option "UseDisplayDevice" "none"
    [ 337.174] (**) NVIDIA(0): Enabling 2D acceleration
    [ 337.174] (**) NVIDIA(0): Ignoring EDIDs
    [ 337.174] (**) NVIDIA(0): Option "UseDisplayDevice" set to "none"; enabling NoScanout
    [ 337.174] (**) NVIDIA(0): mode
    [ 341.309] (EE) NVIDIA(0): Failed to initialize the NVIDIA GPU at PCI:1:0:0. Please
    [ 341.309] (EE) NVIDIA(0): check your system's kernel log for additional error
    [ 341.309] (EE) NVIDIA(0): messages and refer to Chapter 8: Common Problems in the
    [ 341.309] (EE) NVIDIA(0): README for additional information.
    [ 341.309] (EE) NVIDIA(0): Failed to initialize the NVIDIA graphics device!
    [ 341.309] (EE) NVIDIA(0): Failing initialization of X screen 0
    [ 341.309] (II) UnloadModule: "nvidia"
    [ 341.309] (II) UnloadSubModule: "shadow"
    [ 341.309] (II) UnloadSubModule: "wfb"
    [ 341.309] (II) UnloadSubModule: "fb"
    [ 341.309] (EE) Screen(s) found, but none have a usable configuration.
    [ 341.309] (EE)
    Fatal server error:
    [ 341.309] (EE) no screens found(EE)
    [ 341.309] (EE)
    Please consult the The X.Org Foundation support
    at http://wiki.x.org
    for help.
    [ 341.309] (EE) Please also check the log file at "/var/log/Xorg.8.log" for additional information.
    [ 341.309] (EE)
    [ 341.309] (EE) Server terminated with error (1). Closing log file.
    dmesg |grep bbswitch:
    [ 13.570577] bbswitch: version 0.7
    [ 13.570585] bbswitch: Found integrated VGA device 0000:00:02.0: \_SB_.PCI0.GFX0
    [ 13.570591] bbswitch: Found discrete VGA device 0000:01:00.0: \_SB_.PCI0.PEG0.PEGP
    [ 13.570687] bbswitch: detected an Optimus _DSM function
    [ 13.570741] bbswitch: Succesfully loaded. Discrete card 0000:01:00.0 is on
    [ 13.572354] bbswitch: disabling discrete graphics
    [ 261.419114] bbswitch: enabling discrete graphics
    [ 278.445698] bbswitch: disabling discrete graphics
    [ 279.055257] bbswitch: enabling discrete graphics
    [ 286.292315] bbswitch: disabling discrete graphics
    [ 291.911052] bbswitch: enabling discrete graphics
    [ 310.011289] bbswitch: disabling discrete graphics
    [ 336.982794] bbswitch: enabling discrete graphics
    [ 349.526470] bbswitch: enabling discrete graphics
    I used "rcutree" option as shown in the post that you put, but I've deleted all changes made to use default configurations, because the solution of the post didn't work

  • [SOLVED] Problem with UTC and CST

    I probably made a mistake while installing Arch Linux and the clocks in my computer are all a mess.
    Right now the UTC time its 6+ hours, I changed it in the bios to the right hour but that give me -6 hours in the local time thats displayed in Arch Linux and gave me the right hour in Windows 7.
    timedatectl
    Local time: Fri 2013-12-20 12:33:12 CST
    Universal time: Fri 2013-12-20 18:33:12 UTC
             RTC time: Fri 2013-12-20 18:33:12
            Timezone: America/Mexico_City (CST, -0600)
         NTP enabled: yes
    NTP synchronized: yes
    RTC in local TZ: no
          DST active: no
    Last DST change: DST ended at
                      Sun 2013-10-27 01:59:59 CDT
                      Sun 2013-10-27 01:00:00 CST
    Next DST change: DST begins (the clock jumps one hour forward) at
                      Sun 2014-04-06 01:59:59 CST
                      Sun 2014-04-06 03:00:00 CDT
    PS: Already did everything in the wiki and tried with some [SOLVED] posts.
    Edit: The problem with this its that when i use Windows 7 (to play DotA... or using the iPod, since no player can edit the files in the ipod) the hour is 6+ hours, that causes problems with the configuration in the ipod since each time i connect it the clock (that its also my only clock) changes to 6+ hours. Its maybe irrelevant for someone but i would like to have the right hour in all my systems.
    Last edited by Senketsu (2013-12-20 19:16:54)

    I am with Scimmia, everything looks correct.   But, If you boot Windows, it will try to change the BIOS clock to local time unless you tell it otherwise.
    Edit for both of your edits.  Yes, fix windows
    Last edited by ewaller (2013-12-20 18:56:32)

Maybe you are looking for