Kernel without ALSA? Issues? I use OSS 99% yet Need easier ALSA usage

[sensible to remove alsa from kernel to allow easier loading of oss or alsa as required?]
Hi ... I use OSS for almost all audio. (99 % of the time) I did learn an awful lot about alsa on a 2 month search for sound heaven and it does have a use now and then. (some usb headset mics)
I might like an alsa-less kernel though have thought that besides the skill needed to create an alsa-less kernel as each new kernel update happens, there's the possibility that many before me have discovered what horror occurs in other areas if there's no alsa in an ARCH kernel when using pacman etc.
That's the first big question(s).
secondly, My thinking is that I would be able to change between alsa and oss easier (hahaha) by having a kernel without alsa in it. Then oss would just start or alsa would just start depending on which I had in my rc.conf...
um, yes? And the script wouldn't need to run after reinstalling OSS after uninstalling it to get alsa back.
any ideas on using alsa AND oss with ARCH.
thanks for reading!
Last edited by yvonney (2010-01-28 11:04:15)

hehehe, thanks!
um, wooo, I forgot about pacman (as you call more resolutely depmod -a)
Maybe better idea though seems like 2 kernels best.
I will have to see what OSS install/uninstall script does when there's no alsa. Maybe it will be less invasive and do nothing other than install or remove OSS.
So, I had a big think and here's why alsa sometimes.
I currently run my entire ARCH system from a 4 gig flash stick. and have a 40 gig usb drive for data than I manually mount.
When I change to one of the other machines I have gotten good at swapping video drivers and regenerating xorg.conf. I enjoy this though ama aware there's so many way to improve my ability.
So, with using OSS most of the time there has been the occasion where I though using my .asoundrc and alsa would be good because of a USB microphone I wished to use and/or just to have alsa ability refreshed.
A reboot is not a problem though I am looking to avoid having to have OSS go through it's script running to turn alsa back on etc etc. Maybe letting this happen is easier that managing 2 kernels though!!!
I appreciated the notes here about the scripts functioning and will study up on that.
still thinking though I believe it may be best to just go with one kernel and learn what OSS is doing when removed and installed. No quite sure yet what's best though feel more confident.
Thank you.
Last edited by yvonney (2010-01-29 17:45:31)

Similar Messages

  • Remove old kernels without the use of apt-get

    Greetings all, I am new to Linux. I am also trying to find a way to purge old kernels without using apt-get as /boot is entirely full and apt-get has no functionality at this point (I have tried 'apt-get -f' as well). I'll likely resort to trying one of the scripts in the bikeshed package to detect and remove unused kernels, but was wandering if there was another way to accomplish this? Thanks!
    This topic first appeared in the Spiceworks Community

    #1 Points to
    http://forums.verizon.com/t5/High-Speed-Internet-D​SL-and-Dial/Info-on-Deleting-FTP-files-from-Person​...
    #2 I would be nice, if users could remove their content without having to call Verizon.
    If you are the original poster (OP) and your issue is solved, please remember to click the "Solution?" button so that others can more easily find it. If anyone has been helpful to you, please show your appreciation by clicking the "Kudos" button.

  • Can i use another storage flash for my apple air mackbook(change from current 128Gb to 256GB without any issue)

    Can i use another storage flash for my apple air mackbook(change from current 128Gb to 256GB without any issue)

    There are some upgrade options available from do it yourself sites.
    See the MBair SSD upgrades on mid way down this page:
    http://eshop.macsales.com/shop/SSD/OWC/

  • Although I have been using Google Maps App in India for a while now (without any issues) but I still want to know if there are any plans for launching Turn by Turn Navigation under the OOTB Map App for iOS?

    Although I have been using Google Maps App in India for a while now (without any issues) but I still want to know if there are any plans for launching Turn by Turn Navigation under the OOTB Map App for iOS?
    It has been quite a while since Apple came up with Map App of their own and should be releasing improved maps with more capabilities in India.
    Cheers!
    Abhishek

    No one here can answer your question, as there is no one from Apple here, & no one here represents Apple or speaks for Apple.

  • Any issues with using LDAP on LINUX for GRC 5.2 UME?

    Our company is converting our LDAP servers from AIX to LINUX.  The DNS name used in our UME connection should not change.  Are there any issues with using LDAP on LINUX?  We are currently on GRC 5.2 SP9 (in the middle of upgrading to SP12).
    Also, I have been trying to connect our test UME system to a test LDAP box that has already been converted to LINUX but keep getting a 'connection failed' error when I try to test it. 
    Do you have to reboot the server to test changing the LDAP connections?  I've been trying it by going into UME, pulling up the LDAP tab, hitting the Modify button, entering the new userid and password for test LDAP, and hitting the Test Connection button.  I've verified that this userid and password is correct for test LDAP.
    Is there a way to get more information about why the connection failed?
    Thanks.

    I've been told by our LDAP Support group that none of the other configuration settings should have to be changed.  I should only have to change the id and password to connect to a test version of LDAP instead of our regular connection to the production LDAP.
    Can you test a connection for a different userid/password without having to reboot/restart the server?  Do I need to change these two settings, save then, reboot/restart, and then do the Test Connection button?
    Thanks.

  • Live text template render issue when using Media Encoder?

    I'm running into flash frames with Media Encoder exports of Premiere sequences using the live text templates.
    Unfortunately, AME doesn't flag it as an error, so the render completes without any issues in the render log and the only way to find it is to watch the rendered file.
    I have tried different combinations of having After Effects open or closed, with the live text template comp open or closed; I can't consistently reproduce the problem with any certainty and so am having trouble pinpointing the source.
    My only hypotheses so far are that:
    The glitches seem to be happening in the dynamic link between AME and PP, not AE and PP. The live text will pop back to the duplicated comp's text for one frame. Video playing simultaneously doesn't pop.
    In other words, the AE text template said: "name here". In PP it was edited to say "John Doe", then that comp was duplicated and edited to say "Bob Smith". Everything looks fine in PP. There are no keys on the source text in AE (I have noticed that if there are keys on the source text in AE, the text will pop back to the original AE text at that key in PP).
    When rendered out of AME, halfway through Bob Smith's title, it will read "John Doe" for a single frame, then pop back to being Bob.
    I've only had this happen so far with AME renders. The renders out of PP haven't given me any issues so far.
    It might happen more frequently on longer segments? If a title is up for 45 sec instead of 10 sec, it may be more likely to pop.
    The issue seems to occur more often in batch renders out of AME (happened to 3 of 5 files rendered in a batch, AME 2014.1) and/or with unrelated AE comps open (on a single render from AME 2014.2)
    Again, it doesn't happen every time or to the same title, or in the same spot. Files are being rendered to ProRes 422 HQ. Unchecked "Use Previews" in AME.
    Would be nice to get to the bottom of what's happening so I can recommend something to the editors when I hand off these live text templates to them.
    I realize that the live text is a new feature as of 2014.1, but haven't found anyone else talking about this. Would love if this was a very simple fix that I'm missing.
    I was able to reproduce the error on 2 different machines, with similar specs:
    Premiere CC 2014.1 and 2014.2
    Media Encoder CC 2014.1 and 2014.2
    Mac tower 2010
    2x 6-core Intel Xeon
    RAM: 64 GB
    Graphics: ATI Radeon HD 5770 and NVidia Quadro K5000 (or K4000)
    OS: 10.8.5 Mountain Lion

    I'm running into flash frames with Media Encoder exports of Premiere sequences using the live text templates.
    Unfortunately, AME doesn't flag it as an error, so the render completes without any issues in the render log and the only way to find it is to watch the rendered file.
    I have tried different combinations of having After Effects open or closed, with the live text template comp open or closed; I can't consistently reproduce the problem with any certainty and so am having trouble pinpointing the source.
    My only hypotheses so far are that:
    The glitches seem to be happening in the dynamic link between AME and PP, not AE and PP. The live text will pop back to the duplicated comp's text for one frame. Video playing simultaneously doesn't pop.
    In other words, the AE text template said: "name here". In PP it was edited to say "John Doe", then that comp was duplicated and edited to say "Bob Smith". Everything looks fine in PP. There are no keys on the source text in AE (I have noticed that if there are keys on the source text in AE, the text will pop back to the original AE text at that key in PP).
    When rendered out of AME, halfway through Bob Smith's title, it will read "John Doe" for a single frame, then pop back to being Bob.
    I've only had this happen so far with AME renders. The renders out of PP haven't given me any issues so far.
    It might happen more frequently on longer segments? If a title is up for 45 sec instead of 10 sec, it may be more likely to pop.
    The issue seems to occur more often in batch renders out of AME (happened to 3 of 5 files rendered in a batch, AME 2014.1) and/or with unrelated AE comps open (on a single render from AME 2014.2)
    Again, it doesn't happen every time or to the same title, or in the same spot. Files are being rendered to ProRes 422 HQ. Unchecked "Use Previews" in AME.
    Would be nice to get to the bottom of what's happening so I can recommend something to the editors when I hand off these live text templates to them.
    I realize that the live text is a new feature as of 2014.1, but haven't found anyone else talking about this. Would love if this was a very simple fix that I'm missing.
    I was able to reproduce the error on 2 different machines, with similar specs:
    Premiere CC 2014.1 and 2014.2
    Media Encoder CC 2014.1 and 2014.2
    Mac tower 2010
    2x 6-core Intel Xeon
    RAM: 64 GB
    Graphics: ATI Radeon HD 5770 and NVidia Quadro K5000 (or K4000)
    OS: 10.8.5 Mountain Lion

  • HR ABAP: Issue with using 'nocommit' parameter on FM HR_INFOTYPE_OPERATION

    Issue with using nocommit parameter on FM HR_INFOTYPE_OPERATION:
    My client has a requirement to create the following 4 infotypes in sequence in a LUW, i.e either all are created or none is created.
    9045   (custom infotype)
    0045
    0078
    0015
    I tried to use the nocommit parameter on FM HR_INFOTYPE_OPERATION to insert the 4 infotypes
    in a nocoomit mode and then at the end I have issued
    'Commit Work', but to my surprise only I/T 0015 is created in the database and the first three (9045, 0045 and 0078) did not make it to database.
    I searched many threads on SDN but could not find a solution.
    Please let me know if there could be any solution to implement the LUW.
    YOur inputs will be appreciated.

    Hi ,
    i think u can also try with this FM HR_MAINTAIN_MASTERDATA , see its documentations.
    no commit works like a simulation mode , what u can do is  ,
    call FM for all Infotypes and collect all error msgs if any , then finally call FM for all infotypes again without passing nocommit work ( i.e space).
    regards
    prabhu

  • HT4061 can I connect my iphone to my dad's imac and download my pictures to his machine without any issues?

    can I connect my iphone to my dad's imac and download my pictures to his machine without any issues?

    AnaMusic wrote:
    By Design you can only Sync the phone to One iTunes Library at a time...
    Syncing & importing pics/videos in the camera roll are two entirely different things & have nothing to do with each other.
    You can import iPhone camera roll photos/videos using any computer you want...this has nothing to do with iTunes or syncing.

  • I have itunes 10.5.2  Can I download and update directly to the latest version of itunes without any issues?

    I have itunes 10.5.2   Can i download and update directly to the latest version of itunes without any issues?

    Yes, but back up the library first. Note that if you're using Mac OS X 10.5.8, you won't be able to update iTunes past 10.6.3.
    (96898)

  • XML Parse issues when using Network Data Model LOD with Springframework 3

    Hello,
    I am having issues with using using NDM in conjuction with Spring 3. The problem is that there is a dependency on the ConfigManager class in that it has to use Oracle's xml parser from xmlparserv2.jar, and this parser seems to have a history of problems with parsing Spring schemas.
    My setup is as follows:
    Spring Version: 3.0.1
    Oracle: 11GR2 and corresponding spatial libraries
    Note that when using the xerces parser, there is no issue here. It only while using Oracle's specific parser which appears to be hard-coded into the ConfigManager. Spring fortunately offers a workaround, where I can force it to use a specific parser when loading the spring configuration as follows:
    -Djavax.xml.parsers.DocumentBuilderFactory=com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl But this is an extra deployment task we'd rather not have. Note that this issue has been brought up before in relation to OC4J. See the following link:
    How to change the defaut xmlparser on OC4J Standalone 10.1.3.4 for Spring 3
    My question is, is there any other way to configure LOD where it won't have the dependency on the oracle parser?
    Also, fyi, here is the exception that is occurring as well as the header for my spring file.
    org.springframework.beans.factory.xml.XmlBeanDefinitionStoreException:
    Line 11 in XML document from URL [file:/C:/projects/lrs_network_domain/service/target/classes/META-INF/spring.xml] is invalid;
    nested exception is oracle.xml.parser.schema.XSDException: Duplicated definition for: 'identifiedType'
         at org.springframework.beans.factory.xml.XmlBeanDefinitionReader.doLoadBeanDefinitions(XmlBeanDefinitionReader.java:396)
         [snip]
         ... 31 more
    Caused by: oracle.xml.parser.schema.XSDException: Duplicated definition for: 'identifiedType'
         at oracle.xml.parser.v2.XMLError.flushErrorHandler(XMLError.java:425)
         at oracle.xml.parser.v2.XMLError.flushErrors1(XMLError.java:287)
         at oracle.xml.parser.v2.NonValidatingParser.parseDocument(NonValidatingParser.java:331)
         at oracle.xml.parser.v2.XMLParser.parse(XMLParser.java:222)
         at oracle.xml.jaxp.JXDocumentBuilder.parse(JXDocumentBuilder.java:155)
         at org.springframework.beans.factory.xml.DefaultDocumentLoader.loadDocument(DefaultDocumentLoader.java:75)
         at org.springframework.beans.factory.xml.XmlBeanDefinitionReader.doLoadBeanDefinitions(XmlBeanDefinitionReader.java:388)Here is my the header for my spring configuration file:
    <?xml version="1.0" encoding="UTF-8"?>
    <beans xmlns="http://www.springframework.org/schema/beans"
           xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
           xmlns:aop="http://www.springframework.org/schema/aop"
           xmlns:tx="http://www.springframework.org/schema/tx"
           xmlns:context="http://www.springframework.org/schema/context"
           xsi:schemaLocation="http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd
           http://www.springframework.org/schema/aop http://www.springframework.org/schema/aop/spring-aop.xsd
           http://www.springframework.org/schema/tx http://www.springframework.org/schema/tx/spring-tx.xsd
           http://www.springframework.org/schema/context http://www.springframework.org/schema/context/spring-context.xsd">Thanks, Tom

    I ran into this exact issue while trying to get hibernate and spring working with an oracle XMLType column, and found a better solution than to use JVM arguments as you mentioned.
    Why is it happening?
    The xmlparserv2.jar uses the JAR Services API (Service Provider Mechanism) to change the default javax.xml classes used for the SAXParserFactory, DocumentBuilderFactory and TransformerFactory.
    How did it happen?
    The javax.xml.parsers.FactoryFinder looks for custom implementations by checking for, in this order, environment variables, %JAVA_HOME%/lib/jaxp.properties, then for config files under META-INF/services on the classpath, before using the default implementations included with the JDK (com.sun.org.*).
    Inside xmlparserv2.jar exists a META-INF/services directory, which the javax.xml.parsers.FactoryFinder class picks up and uses:
    META-INF/services/javax.xml.parsers.DocumentBuilderFactory (which defines oracle.xml.jaxp.JXDocumentBuilderFactory as the default)
    META-INF/services/javax.xml.parsers.SAXParserFactory (which defines oracle.xml.jaxp.JXSAXParserFactory as the default)
    META-INF/services/javax.xml.transform.TransformerFactory (which defines oracle.xml.jaxp.JXSAXTransformerFactory as the default)
    Solution?
    Switch all 3 back, otherwise you'll see weird errors.  javax.xml.parsers.* fix the visible errors, while the javax.xml.transform.* fixes more subtle XML parsing (in my case, with apache commons configuration reading/writing).
    QUICK SOLUTION to solve the application server startup errors:
    JVM Arguments (not great)
    To override the changes made by xmlparserv2.jar, add the following JVM properties to your application server startup arguments.  The java.xml.parsers.FactoryFinder logic will check environment variables first.
    -Djavax.xml.parsers.SAXParserFactory=com.sun.org.apache.xerces.internal.jaxp.SAXParserFactoryImpl -Djavax.xml.parsers.DocumentBuilderFactory=com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl -Djavax.xml.transform.TransformerFactory=com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl
    However, if you run test cases using @RunWith(SpringJUnit4ClassRunner.class) or similar, you will still experience the error.
    BETTER SOLUTION to the application server startup errors AND test case errors:
    Option 1: Use JVM arguments for the app server and @BeforeClass statements for your test cases.
    System.setProperty("javax.xml.parsers.DocumentBuilderFactory","com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl");
    System.setProperty("javax.xml.parsers.SAXParserFactory","com.sun.org.apache.xerces.internal.jaxp.SAXParserFactoryImpl");
    System.setProperty("javax.xml.transform.TransformerFactory","com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl");
    If you have a lot of test cases, this becomes painful.
    Option 2: Create your own Service Provider definition files in the compile/runtime classpath for your project, which will override those included in xmlparserv2.jar.
    In a maven spring project, override the xmlparserv2.jar settings by creating the following files in the %PROJECT_HOME%/src/main/resources directory:
    %PROJECT_HOME%/src/main/resources/META-INF/services/javax.xml.parsers.DocumentBuilderFactory (which defines com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl as the default)
    %PROJECT_HOME%/src/main/resources/META-INF/services/javax.xml.parsers.SAXParserFactory (which defines com.sun.org.apache.xerces.internal.jaxp.SAXParserFactoryImpl as the default)
    %PROJECT_HOME%/src/main/resources/META-INF/services/javax.xml.transform.TransformerFactory (which defines com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl as the default)
    These files are referenced by both the application server (no JVM arguments required), and solves any unit test issues without requiring any code changes.
    This is a snippet of my longer solution for how to get hibernate and spring to work with an oracle XMLType column, found on stackoverflow.

  • Licensing issue by using Applications AC and PC

    Hi,
    The SAP GRC 10.0 solution comes as bundled solution for AC, PC and RM. While downloading the SAP GRC 10.0 all the three components are downloaded but in a dormant state.
    With the release of GRC 10.0, Access Control and Process Control are offered as an integrated solution, both at the data layer and at the user interface layer. This new unified platform enables increased harmonization of key master data.
    GRC 10.0 is a harmonized platform. Although maintenance of data within AC and PC is similar in 10.0 as it was in previous releases, there are some key differences. The following master data is shared between AC and PC for the release 10.0 integration scenario.
    u2022 Organizations
    u2022 Business Processes
    u2022 Business Subprocesses
    u2022 Controls
    SAP GRC System Add-ons:
    Access Control, Process Control and Risk Management are contained in one ABAP add-on u201CGRCFND_Au201D
    SAP Backend System Plug-ins:
    A. GRCPINW: NW Function Modules hold the AC functions for ERP systems without HR (former non-HR RTA)
    B. GRCPIERP: PC relevant features are contained in the plug-in GRCPIERP, for example, for running automated controls and the HR relevant functions for AC (former HR RTA)
    There are multiple activation points for AC, PC and RM:
    u2022 SAP GRC SPRO -> activate the Application. However
    u2022 When we go for the Plug-ins (GRCPIERP) installation in backend systems with HR function module it also brings in the PC relevant features.
    u2022 Mitigating Control definition based on Organizations
    I had activated the Application AC and PC for some time assuming that PC is required to create the Mitigation Control and Organizations. However I just realized that Mitigating Control and Organization can be created even without Activating PC, which does nothing but removes the additional features of PC.
    My Query:
    u2022 Kindly let me know if I can still use PC to create and assign Mitigating Control as a shared resource for AC and PC or I have violated Licensing issue by using Applications AC and PC when I am supposed to use just Access Control.
    u2022 How is shared resources in SAP GRC 10 dealt in terms of Licensing of Applications in SAP GRC 10.
    Thanking you in advance.

    all the new pc's we've purchased for a time now, don't have a Win7 CoA product key attached, even though they come with Win7 downgrade rights, so your strategy to use the CoA pkeys for the Win7 install on the new machines might be flawed.
    We don't buy HP, and I'm in Australia, so I'm not sure if this will be the same for you.
    Also, the OEM image that you would use to re-deploy/re-image with, may need to be different for varying models (OEMs might include different drivers, or other checks or blockers, in their image and OOBE scripts), and, the embedded OEM "master" product key/certificate
    could vary (or not align) across model types/generations.
    Ours do have the SLIC table in firmware and the OA markers, and we use VL Win7 Enterprise anyway, so our 40,000 machine fleet means VL and Software Assurance as a part of our EA + Select is an (expensive) no-brainer really.
    The other thought I have for you, is the re-imaging rights that are bestowed when you have a VL agreement (and therefore, are not bestowed if you don't have one). Because we do have a VL agreement, I've never dived into how it might be worked if we didn't
    have VL.
    http://www.microsoft.com/licensing/about-licensing/briefs/reimaging.aspx
    http://www.microsoft.com/licensing/about-licensing/briefs/downgrade-rights.aspx
    If you do look at a VL agreement, I'm told they start from as few as 5 machines. And, if you purchase new pc's, you have the option to enrol the new pc's into Software
    Assurance within 90days of purchase, which I imagine would save you some cost, compared to attaching SA to them at a later time, presumably requiring you to pay full ticket price for that.
    I'd recommend you get some advice from an MS reseller, even if only to work out what the various options/costs might be.
    Don
    (Please take a moment to "Vote as Helpful" and/or "Mark as Answer", where applicable.
    This helps the community, keeps the forums tidy, and recognises useful contributions. Thanks!)

  • PROBLE WHILE RENDERING IN CC While as In CS renders without any issue the same file from same location but tried 3 CC ,machine and have the same issue.

    Please i need an Urgent Help
    I just installed for my users 5 Machines with Adobe CC and they are all crashing while rendering to QT H264.
    while in cs6 it renders the same file in the same location without any issue.
    Please help it is urgent.
    Regards

    Hi,
    I can't really read the error appearing in the screenshot you shared.  Is the application fully up to date?
    Have you read: After Effects CC 2014 - removal of H.264, MPEG-2, WMV, FLV, F4V, and SWF exporters - WHY !?
    This might also be an explanation, as i believe you are trying to use h.264 right?
    Kind Regards,
    Harrie

  • Binary kernels without Smp

    Are there binary kernels without SMP available or should I build it from source?

    Rotonen wrote:What's your main reason for needing a non-SMP kernel?
    Ok. The kernel is not the reason for my issue.
    In qemu in grafic mode Archlinux has a constant CPU load of 10% which makes the CPU run at 1500MH which makes the fan of the laptop go on frequently.
    But as soon as I go back to console mode the CPU load is far lower.

  • Is it possible to break an edit into small scenes, and then sew them together, but without losing quality by using the 'publish and share' options?

    Is it possible to break an edit into small scenes, and then sew them together, but without losing quality by using the 'publish and share' options?

    redflashred
    What is important here is that you know the properties of your source file(s), set the project preset to match the properties of those source files, and then export with settings as close as possible to the original. Often you can do this without noticing quality differences in the export.
    I am understanding and defining your issue as you wanting to maintain the quality of the original imported files in the export. If that is the case, the please supply the following, and I will give you a step by step
    a. What are the properties of your source files (including bitrate). Here I am looking for
    video compression
    audio compression
    frame size
    frame rate
    interlaced or progressive
    pixel aspect ratio
    file extension
    A quick way to get that type of information is knowing the brand/model/settings for the camera that recorded the video.
    2. Based on that information we can set manually a project preset to match those source property.
    3. When we go to export...example Publish+Share/Computer/ and your choice and its preset, we can customize the preset selected
    under the Advanced Button/Video Tab and Audio Tab of that preset in order to try for the best possible match between the properties of the
    original file and the properties of the export.
    But, please let us know if you also need help in trimming the original import at the Timeline level prior to export.
    Your mention of Publish+Share suggest that you are working with Premiere Elements 11, 12, or 13. What is the computer operating system?
    Thanks.
    ATR

  • Downgrading the kernel without breaking my system

    Hello,
    A few months ago, I was using a old (2011) version of the kernel, because of regressions in WiFi support and in sound quality. But when /lib became a symlink and the glibc update came ( http://www.archlinux.org/news/the-lib-d … a-symlink/ ), my system was broken, and I had to upgrade to a more recent version of the kernel using a LiveCD.
    So, is there a way to downgrade safely to an older version of the kernel without breaking my system?
    Thanks in advance.

    /lib is a symlink... what happens if you try to install the older kernel?
    Alternatively, re-build the package of an older kernel. (Though I doubt re-building is necessary, I offer it as a possibility.)

Maybe you are looking for