New openldap package & some questions

Hello,
Let me say first of all, I'm relatively new to the archlinux dist. That being said I've folllowed the instructions for making a abs PKGBUILD file for the latest STABLE release of openldap. I have a few questions though if I may?
1. Open ldap uses a somewhat strange versioning system .. the file is openldap-stable-<DATE>.tgz but it extracts to a openldap-2.4.16 directory.. I've worked around this  by creating a new var pgkverM  and giving it the date of the build (20090411). Per the abs package guidelines is this something that is allowed? If not how might I achieve this and stay withing.
2. Baring a fix or approval of #1, is this something I can/should share with the rest of the community? If so, whats the process for submission/approval ? I'd love nothing more than to contribute my part to the community...
The package is a very simple build all in all.. I looked at the pkgbuild from the abs repo's for the old version. which by the way is an outdated RELEASE (not stable) version. I found a workaround for solving the cpeerid error without having to patch any of the source..
Thanks

Just so you know, this actually is not quite as simple as updating one package.  There is a soname bump on libldap that requires rebuilding:
[core]
nfsidmap-0.21-2
[extra]
alpine-2.00-3
apache-2.2.11-3
apr-util-1.3.4-1
autofs-5.0.4-4
claws-mail-3.7.1-1
courier-authlib-0.62.2-2
courier-mta-0.61.1-1
cyrus-sasl-2.1.22-10
cyrus-sasl-plugins-2.1.22-10
dirmngr-1.0.2-1
dovecot-1.1.15-1
ekiga-3.2.4-1
evolution-2.26.2-1
evolution-data-server-2.26.2-1
evolution-exchange-2.26.2-1
gconf-2.26.2-1
gnash-common-0.8.5-2
gnash-gtk-0.8.5-2
gnupg-1.4.9-1
gnupg2-2.0.11-1
kdepimlibs-4.2.3-2
kdevelop-3.5.4-1
kopete-cryptography-1.3.0-2
lighttpd-1.4.22-3
mod_perl-2.0.4-3
nss_ldap-264-1
pam_ldap-184-1
pdns-2.9.22-2
php-5.2.9-3
postfix-2.5.6-2
samba-3.3.4-1
seahorse-2.26.2-1
smbclient-3.3.4-1
subversion-1.6.2-2
sylpheed-2.6.0-1
wine-1.1.21-1
[community]
balsa-2.3.28-1
cherokee-0.99.15-1
dbmail-2.2.11-2
freeradius-2.1.4-1
gq-1.2.3-1
hula-r2661-1
kdesvn-1.3.0-2
libgda3-3.1.5-2
mailutils-2.0-5
opensips-1.5.0-2
pppd-ldap-0.12b-1
pppd-ldap-simple-0.12b-1
python-ldap-2.3.6-1
qsvn-0.8.1-1
rapidsvn-0.9.8-1

Similar Messages

  • Potential new owner has some questions

    I'm possibly thinking of getting an MBA (the older revision) but I have some questions.
    I'll be trading my 07 2.4GHz model (4gig/320hd) for it. I was selling the 07 MBP because I picked up a new 2.5 MBP. I was offered a 1.8GHz 64SSD model as a straight up trade.
    Now long story short, since I've not owned one or read up on the MBA, I'm unsure if the there were any major issues with the older revision.
    I realize it will be a bit slower thanks to the FSB and GPU but were there any inherent problems, such as over-heating, or display issues that I should be aware before pulling the trigger on this puppy.
    Thanks in advance
    Mike

    Yeah, once piece of advice - don't. The old one is slow, yes there have been "some" overheating issues on some machines. I think I would sell what you have and buy a new MBA before I would trade a piece of equip for the old version. You will be sorely disappointed in the performance of the old GPU.
    Don't forget all the other limitations of this machine
    WP

  • New to macbooks, some questions

    Hi!
    I'm thinking of getting a Macbook for Uni in September because XP sickens me with its crashing and errors and general slowness! So basically I have a few questions!
    Looking at the base model and upgrading the RAM from 1GB to 2GB.
    I might want to use Bootcamp and run XP at some point. Does Bootcamp come free, and as standard with the Macbook? Also, would the base 120GB hard drive be sufficient room for running both OSX and XP if I decide to do so?
    Does the macbook come with any decent word processing/office programs as standard, or would I need to buy these? If so, is iWork good?
    I'll be using university discount, so will save £100+. I've also read that as part of the student discount there's a warranty extended to 3 years. Is this correct or just a myth?
    Also..will it take long to get used to everything? I've been using windows for 10 years+
    All I can think of for now, sure more stuff will pop into my head at some point!
    Cheers
    Paul

    Hi, Im a 1st year university student and rely on my MacBook for uni work so I can relate to your concerns (sorry for the essay like reply!)...
    Firstly to clear things up, yes!!, since 1st Oct 2007 you now get an *extended 3 years of hardware warranty as part Apple's HE (Higher Education) national contract in the UK*. I actually got my MacBook start of 2007, had a hardware issue and thought i was out of the warranty period but was thrilled to find out I had another 2 years due to the extension . All the perks of being a uni student
    http://www.apple.com/uk/education/hed/nationalcontract/
    So definitely buy your MacBook from the Apple Store's education store to take advantage of this extended warranty, and also of their discounts. Be sure to buy from the Apple Store for *Higher Education*, you'll get bigger discounts:
    http://store.apple.com/Catalog/uk/Images/routingpage.html
    Problem is you can only get onto the HE website from your University's network (are you already at university or are you starting in September?). Although on your home computer, after logging in on the UCAS website, you might be able to find a dedicated link to the Apple HE Store. I'd recommend just giving them a call and you can give them proof of applying to university if you're not there yet.
    Also i used iWork for the first term, but found i had a few formatting issues and compatibility issues when transferring and opening documents on Windows (which my University computers are). Also to save in Pages (word equivalent in iWork), u keep having to select 'Export' and there is no autosave facility to .doc format. Basically it's not very .doc friendly. I'd suggest Microsoft Office for Mac, a lot less fiddly and more convenient in this respect.
    Also forgot to mention, apparently the MacBook is due for a complete redesign this year and i suspect the next generation of MacBook to come out in June or summer time, if you're not at Uni yet, maybe it's worth the wait seeing as though you mentioned uni in Sept? Otherwise u might be gutted if you buy one now?
    Hope that helps.
    Kenny

  • New Mac Pro - some questions about setup and RAM

    After debating between iMac and Mac Pro, I decided on a refurb SINGLE 2.8 quad-core Mac Pro.
    Somewhere in these discussions (can't find it now) I thought I saw recommendations to zero out the hard drive and re-install the system when it arrives. Is this necessary?
    Also, I plan to use the single processor savings to buy RAM. When I go to the Crucial site, they have a choice of RAM for 4-core or 8-core. Mine is only 4-core, but would I choose the 8-core option to get the correct RAM for 2008 models?
    In general, when it comes to buying and installing RAM, should I follow the 8-core directions? Will having a single processor change anything to do with upgrades or expansion?
    Thanks. Hope I didn't ask too much in one post.

    Somewhere in these discussions (can't find it now) I thought I saw recommendations to zero out the hard drive and re-install the system when it arrives. Is this necessary?
    No, that really isn't necessary. If you were prepping a new bare drive that had never been used or you were installing OS X on a drive that had once been used for Windows, then zeroing the drive is quite appropriate.
    All the Mac Pros use the same type of RAM. The Late 2008 models use the faster PC2-6400 800 MHz RAM, but the other specs are the same - ECC, Fully buffered, heat sink. The RAM and installation is the same regardless of cores.
    RAM should be installed minimally in matched pairs, optimally in matched quads. The following illustrations show how it should be installed:
    Mac Pro memory arrangement photos
    Mac Pro Memory Configuration

  • New K9VGM-V; Some questions

    I installed a new MB,  AMD Athlon 64 3000+ Orleans and  CORSAIR ValueSelect 1GB 240-Pin DDR2 SDRAM DDR2.  I am using my old hard drives - an old Maxtor 40G and a new Western Digital 200G (which I loaded previously)
    After struggling with W2000K Pro issues for eight hours I got it running.  I used PC Wizard to test and verify my system. 
    After I installed my DVD Burner (Sony Optiarc SATA) when I use  PC Wizard to detect the drives PC Wizard freezes.  I have the Sata set to IDE for the DVD.  The system recognizes it but it won't play any movies and Media Player tells me "The path to the specified device is denied".   Any thoughts?  I can and have already burned new firmware for the DVD. 
    Also, is there a way to eliminate the need to push F1 to continue booting?
    Ray

    read Moan Guide 1st
    "The system recognizes it but it won't play any movies and Media Player tells me "The path to the specified device is denied".   Any thoughts?  I can and have already burned new firmware for the DVD. "
    sound like OS problem to me.
    "Also, is there a way to eliminate the need to push F1 to continue booting?"
    what support message also you get when you need to press F1 key?

  • Potential new T61 user with some questions about product/su​pport/etc.

    Hi all!
    I've been using Dell Latitude's for last 7+ years and been happy with them thus far, as I was able to stay within the C-series thru 3 systems and reuse some components such as batteries, CD-ROM, etc. but now that I'd have to go to D-series, that gives me the choice of going to new vendor...and IBM/Lenovo was the only other choice due to my OS requirement: XP Pro!
    So now I'm here asking some questions that I'm still having issues with:
    (I'm comparing the T61 to Dell D530 and/or my C610)
    1) Features that I've used on existing Dell that I don't see on T61 :
        a) Infared port --- I've used to transfer stuff from one sys to another
        b) S-Video out --- I've used to display on TV for some situations  
        c) Serial port --- can't recall using it but with my luck, I'd need it sometime. 
        d) Parallel port (C610) --- for my really old printer used on rare occasions
        e) Floppy drive --- yes, I still use it for some unusual situations
    2) On Dell, I can put a 2nd battery or floppy drive in the media bay instead of CD/DVD...what about T61?
    3) Support:
        a) Dell forums have 7+ yrs of history and are well used...this forum seems brand new...was there another one before it?
        b) My Dell's have lifetime tech support (ok, so sometimes understanding the folks from India can be an issue) but IBM seems to limit this to life of warranty.
    4) Pricing: T61 would be seem to be, at minimum, about $150 more than similar Dell even with current "sales"...I'm having trouble justifying this...
    5) Things that I see that T61 has that new Dell D530 doesn't:
       a)  built-in mike --- my current Dell has it and I use it now and then so this is a plus!
       b) keyboard light --- nice feature, but couldn't I use a USB one instead if I really needed it?
       c) fingerprint reader --- never used one...how often do people use it?
       d) 2nd card slot for Express cards --- current Dell has two standard slots and I'm using both, though one is wi-fi which C610 didn't have built-in...and other is modem to replace failed built-in...D530 has only one slot, so having 2nd of any type is good.
       e) middle mouse button --- is that what I'm seeing? Or is it something else?
       f) recovery partition --- ok, nice but how much room does it take?
    6) Things that D530 has that T61 doesn't:
       a) 4th USB --- I'm used to only 1 now so even 3 would be big step up...
       b) 15" standard screen --- T61 is only a 14.1" screen and that extra bit of size would likely help, but I'm limited on physical dimensions to 13.5" total width (unusual situation) so a 15.4" wide screen isn't possible...I'm looking at SXGA+ in either case to maximize usage.
       c) recovery CD is standard --- so I'm presuming I'd have to create my own here?
    That, I think, about does it...so...any users willing to help me decide if a switch is good for me?
    Thanks!!!
    Update: It seems that phone tech support is only available for warranty period with IBM, whereas with Dell, I can still call for hardware tech help (to diagnose problem at least) for "lifetime" or well past paid period, as I've already done in past...comments?
    Update #2: Changed the incorrect "R"s to "T"! Duh!
    Message Edited by J_Hallgren on 12-28-2007 07:50 PM
    A current Dell Latitute user who was looking to maybe switch...But didn't.

    I finally made the switch to Lenovo and the Thinkpad line of notebooks and think this was the best move I have ever made in years! I've spent far too much money maintaining a separate notebook (Dell Latitude L400) several custom built desktops.  This December, I sold my desktops, my laptops (I had 2 L400's), and my 20" Aquous LCD TV to finance my Thinkpad T61.  It was well worth it!  I will try to answer your questions in this checklist, which I found to be quite similar to my own .... frighteningly exactly like my own....:
    Hardware Requirements
    IR: This is pretty outdated but functional.  I used to use a Palm with IR as well as an IR laser printer.  I ditched both for a PDA phone with bluetooth.  The laser printer is networked now
    S-Video: You might be able to get away with a TV tuner cardbus or a expres card.  There might also be external converters that convert from your VGA/DVI (DVI via port replicator) into any variety of video outputs.
    Serial: My older Palm had a serial port for the docking station - bluetooth is a good replacement
    Parallel port: I currently have a USB to parallel adapter in my bag just in case.  These are cheap cables and work well.
    Floppy: I keep a USB one of these just in case.  I find that having fewer input devices on my laptop and leaving them external help keep the machine lighter for my mostly mobile usage (e.g., my L400 with no inputs at all except 1 USB port).
    T61 Spare Battery: Uses ultrabay battery - I rarely use the DVDRW drive and leave the ultrabay battery in there while on the go.  DVDRW drive stays in the bag until needed.  I recommend you upgrade to the 9 cell battery with the advanced ultra bay battery, which will give u something like 9 to 12 hours of operating life.  If you are not a fan of the battery pack sticking out like crazy, stick to the letter box format laptop (not widescreen), which will provide the minimum amount of battery pack stick-out with the 9 cell battery.  The 6 cell will not stick out with the letterbox laptop but will stick out with the widescreen version.
    Support Options:  I rarely call support so I have no opinion here.
    Pricing: Price was not really a concern for me.  I purchased mine for the business look mainly, which Lenovo maintained well.  I don't like the shiny/plastic/toy look of Dell and Apple computers these days.  If anything, consider the premium an investment in the build quality of these machines (or a premium just for the good looks).  I particularly like the roll cage as well as active protection system that stops the hard drive as soon as the system detects excessive motion. 
    Included Hardware Comparison:
    I like the mic, use it often for Skype and the built in 1.3MP webcam as well. 
    Keyboard light.... kinda cute but not as functional because it simply is too small.  It was a cute idea that could have been made better had Lenovo opted for a backlit keyboard instead.  That would have been more functional and added to the "cool" factor. 
    With regards to the Express Card and PCMCIA card slot, I rarely use either because everything is built in.  However, if you want that TV tuner, an express card or PCMCIA card slot is good.  If you opt for the built in media reader, you will be unable to use the express card as the media reader takes up that slot (I might be wrong, but I think the card reader is large enough that it takes up both slots, so u can't add anything anyway - but you should call pre-sales to answer that question). 
    Middle mouse button: The three buttons in the middle are used for the trackpoint/joystick mouse controller.  I think the middle button is configurable to do various things.  I personally hate the joystick and can only use the touchpad.  The lack of a touchpad was the only reason I did not stick to the smaller 12.1" X line of laptops, which was really more in tune with my size requirements (too used to using the L400 sized machine).
    Recovery Partition: This is a cheap way for manufacturers to make it easy for people to recover their systems in case of crash - without the need for any documentation or disc media.  However, it makes it hard for people like me who only use machines after a clean format and customized install of Windows.  This partition is about 10GB and can be erased with special disk format or partition tools.  Windows itself cannot read or delete this partition.  With the minimum size of the hard drives available, this should not be an excessive space killer for most people.  I personally need only 7GB total for OS and apps.  The remaining data I work with is actually streamed from a 500GB gigabit NAS.  I might switch to a 32GB solid state disk one of these days when their performance comes more in par or exceeds traditional drives.
    Recovery CD: You can create your own from the recovery partition since Lenovo does not provide any.
    4th USB: I only use 1 or 2 as well.  1 for an external micro mouse (I love my touchpad, but it's just hard to use with Photoshop or CAD apps) and a second for a USB memory key.  My other devices i.e., scanner, external mouse, keyboard, are plugged in on my desktop through a USB hub that goes to a single port on my docking station.  So yes, your docking station will probably remove the need for more than 2 ports.  Although, I do find it odd there are only 3 ports... such an odd number.
    Screen Size: I personally wish I could have gotten the X series of laptops but the 14.1 was the only other mobile solution.  For the larger screen size, docking station to my 24" Dell 2405FP.  The compact nature of the T series makes it pretty thin and compact though... it was not too much bigger than my L400 - at least, not uncomfortably bigger.
    But yeah, I think if you make the switch, you'll be getting yourself a solid machine that will keep u going for years.  The Latitude line of Dell's were really good, but overall, I feel that the nature of Dell's manufacturing process is making it too "cookie cutter" and reducing its overall quality level.  Their earlier lines were awesome, like the D lines you mentioned.  But unless you get the ATG line, you just won't get the same level of quality they used to offer (in addition, the ATG line is wickedly too expensive - I'd rather get a Durabook for military grade spec at a fraction of the price). 
    Anyway, have fun shopping! :-)
    And moderators or employees of Lenovo, don't you guys get referral bonuses for giving your employee discount code to friends and family?  *wink*
    T61_Wide | Model No. 7662 - CTO
    Core 2 Duo T7250 | 2GB OCZ DDR2-800
    82566MM Gigabit | 4965AGN Centrino Pro

  • Differences between old JavaFX and the new one: some questions

    Hi,
    I'm coming back to my JavaFX studies after 2/3 months and I see that some things are different.. Now, I have some questions:
    1) I see that javafx.ui was deprecated or rewritten in favor of javafx.ext.swing. Why there are a javafx.application.Frame and a javafx.ext.swing.SwingFrame? Which one should I use?
    2) Neither javafx.application.Frame or SwingFrame has the "*undecorated*" attribute that javafx.ui.Frame has. I got this feature only reimplementing createWindow() function like this:
    function createWindow(): java.awt.Window {
           //this is a trick, because Frame don't has a undecorated attribute.
           var f: java.awt.Window = super.createWindow();
           (f as javax.swing.JFrame).setUndecorated(true);
           return f;
    }Should I report this as a bug/RFE in [http://bugreport.sun.com/bugreport/] ? Where should I report bugs and RFEs to JavaFX?
    3) Why frame's opacity don't work in Linux? Is this a Linux restriction or only a "not yet implemented" feature?
    My java -version:
    java version "1.6.0_10"
    Java(TM) SE Runtime Environment (build 1.6.0_10-b33)
    Java HotSpot(TM) 64-Bit Server VM (build 11.0-b15, mixed mode)
    4) Where is that cool feature that permits me to easily animate in JavaFX doing only x = [1..100] dur 1000? Now, I need to use this more advanced (and more complicated) Timeline object.
    Timeline {
            keyFrames:Timeline [
                    KeyFrame{time  : 0s values: x => 0.0},
                    KeyFrame{time  : 2s values: x =< 180.0 tween Interpolator.LINEAR}
            ]//keyFrames
    }//Timeline;Sorry for the amount of questions.
    TIA,
    Bruno E. Grossi

    The new ones will be substantially faster than yours because it has 8-cores instead of four, a faster Frontside Bus, and faster RAM speed.
    You could use the same RAM, but since the new ones can accommodate a faster RAM module you would cause the machine to run slower by using the slower RAM from your old machine.
    You can use the same hard drives since both machines support SATA drives.
    On the other hand you could just be happy with what you have! It's probably more than you need.

  • [SOLVED] Some questions about a recently submited package (cgames)

    Hi there, when searching for ncurses sokoban-like games I ended up finding a great one that wasn't at the AUR yet. I submitted it (the package is named cgames) but I have some questions.
    1- The source code provided by the developer has 3 games. Should I have packaged each one individually? I decided this way because they are so small that I figure nobody would have a problem in having them all even when just wanting to play one of them (you can always change the make command in the PKGBUILD if you want a single game).
    2 - The original package had some compiling problems due to a function called getline (already existing in stdio.h) and I solved this problem with a sed command. Is this the proper way to do it? Should I have created some patch file or something?
    3- I'm 99% sure the game works in 64bit archs, but 99 isn't 100. Can anyone try it out for me, so that I change the PKGBUILD? I know I should have tested it but I don't have the time to install a 64b arch in the next few days.
    Thank you in advance
    Last edited by jlcordeiro (2010-03-29 15:33:58)

    Yeap, a bit
    Anyway, all done (i think)
    Second release:
    - Replaced sed command with a patch
    - Added x64 arch
    - Fixed a little bug with the configure command
    Last edited by jlcordeiro (2010-03-29 15:32:42)

  • I'm going to install Arch this weekend but I have some questions

    Currently I use Linux Mint on my primary PC, but I've installed Arch on my older PC at my parents house.  I like it a lot and I think I want a distro that is rolling release and also that I build myself (as opposed to installing all of the bloat on Mint).  However, I do have some questions; Linux is pretty new to me (been using it since about November) so I don't know exactly how everything works.
    1)  How do I know what packages I have to have installed for building C/C++ programs?  I know in Ubuntu install gcc/g++ by itself doesn't get the job done, you need build-essential to get all of the libraries.  Also, what packages do I need for OpenGL and SDL?  I'm a CS major and I'm in an OpenGL class so I need to be able to compile OpenGL programs; SDL is for my own purposes, but I'd still like to have it.
    2)  I use Amarok to transfer songs to my iPod.  Amarok 1.48 and libgpod 0.6.0 are in the repos and I know both of those are compatible with my 6th gen.  However, I do have a question about transferring album art.  On Mint, Amarok transfers the album art as I transfer the album, but someone told me that Amarok doesn't do that by default; there must have been a setting changed somewhere to do that.  I looked through Amarok's options and didn't see anything like that... anyone know if I can do that in Arch, and how?  There's no real information about Amarok in the wiki.
    3)  I'm going to use Gnome environment but some KDE apps (like Amarok).  Will there be any problems with that that I should know about in advance?
    I'll probably have some more questions once I actually install Arch, but that will do for now   Those are the important ones.

    1) You know by seeing its dependencies on pacman or by the developer documentation. Unlike other distros, you don't need to install dev packages. Example: You have gcc and all the basic gnu tools installed, and you want to compile a program that is not in a arch linux package yet (if theres a package, you can build it using abs / makepkg and pacman will handle the dependencies) and requires, for example, a library called "xyz", you only need to download "xyz" from pacman, and all the dev stuff will come together (like header and etc).
    I hope it helps.

  • Some questions about javacard 2.1.1 and smartcardio

    Hello i have some question about java card 2.1.1 and the smartcardio package.
    1.) I want to sign a message with the Signature.ALG_RSA_SHA_PKCS1 algorithm. I use the following code in the applet to sign the message:
    final static byte P1_CREATION_MODE = (byte) 0x01;
    final static byte INS_SIGN_MODE = (byte) 0x60;
    final static byte SmartCard_CLA = (byte) 0xB0;
    private void signMessage(APDU apdu) {
            byte[] buffer = apdu.getBuffer();
            byte byteRead = (byte) (apdu.setIncomingAndReceive());
            signature.init(privateKey, Signature.MODE_SIGN);
            short length = signature.sign(buffer, ISO7816.OFFSET_CDATA, byteRead, buffer, (short) 0);
            apdu.setOutgoingLength((short) length);
            apdu.sendBytesLong(buffer, (short) ISO7816.OFFSET_CDATA, (short) length);
            apdu.setOutgoing();
        }On the host side I use the following code to connect to the card and to send the sign apdu:
    if (TerminalFactory.getDefault().terminals().list().size() == 0) {
                LOGGER.log(Level.SEVERE, "No reader present");
                throw new NoSuchCardReader();
            /* Select the first terminal*/
            CardTerminal terminal = TerminalFactory.getDefault().terminals().list().get(0);
            /* Is a card present? */
            if (!terminal.isCardPresent()) {
                LOGGER.log(Level.SEVERE, "No Card present!");
                throw new NoSuchCard();
            /* Set the card protocol */
         Card card = terminal.connect("*");
            ATR atr = card.getATR();
            LOGGER.fine(getHexString(atr.getBytes()));
            LOGGER.fine(getHexString(atr.getHistoricalBytes()));
            CardChannel channel = card.getBasicChannel();
            CommandAPDU cmd = new CommandAPDU((byte) 0xb0, (byte) 0x60, (byte) 0x01, (byte) 0x00, new String("datadatdatadata").getBytes(), (byte) 0x40);
         ResponseAPDU response = channel.transmit(cmd);
            card.disconnect(false);But this does not work and i got the following error
    javax.smartcardio.CardException: sun.security.smartcardio.PCSCException: Unknown error 0x8010002f
            at sun.security.smartcardio.ChannelImpl.doTransmit(ChannelImpl.java:202)
            at sun.security.smartcardio.ChannelImpl.transmit(ChannelImpl.java:73)
            at de.upb.client.smartmeter.SmartMeter.initSmartCardApplet(SmartMeter.java:114)
            at de.upb.client.smartmeter.SmartMeterApplikation.main(SmartMeterApplikation.java:39)
    Caused by: sun.security.smartcardio.PCSCException: Unknown error 0x8010002f
            at sun.security.smartcardio.PCSC.SCardTransmit(Native Method)
            at sun.security.smartcardio.ChannelImpl.doTransmit(ChannelImpl.java:171)
            ... 3 more2.) 3Des encryption
    I want to use the 3Des algorithm to encrypt my data. I use
    keyDES = (DESKey) KeyBuilder.buildKey(KeyBuilder.TYPE_DES,
                        KeyBuilder.LENGTH_DES3_2KEY, false);
    cipherDES = Cipher.getInstance(Cipher.ALG_DES_CBC_ISO9797_M2, false);But i do not know what is the aquivalent on the host side??
    3.) Another problem is that i am not able to send the modulus of a public key from the host applikation to the smard card
    new CommandAPDU((byte) 0xb0, (byte) 0x20, (byte) 0x01, (byte) 0x00, modulus.toByteArray()); // create the apdu
    // the method in the applet
    private void setServerKeyMod(APDU apdu) {
            byte[] buffer = apdu.getBuffer();
            try {
                byte byteRead = (byte) (apdu.setIncomingAndReceive());
                short off = ISO7816.OFFSET_CDATA;
                // strip of any integer padding
                if (buffer[off] == 0) {
                    off++;
                    byteRead--;
                publicKeyServer.setModulus(buffer, off, byteRead);
            } catch (APDUException ex) {
                ISOException.throwIt((short) (SW_APDU_EXCEPTION + ex.getReason()));
        }The error code is 6700
    4.) My last problem ist, that i am not able to use a value bigger than 0x7F as the ne field in the apducommand, because i get the following error
    CommandAPDU((byte) 0xb0, (byte) 0x60, (byte) 0x01, (byte) 0x00, data, (byte) 0xff);
    java.lang.IllegalArgumentException: ne must not be negative
            at javax.smartcardio.CommandAPDU.<init>(CommandAPDU.java:371)
            at javax.smartcardio.CommandAPDU.<init>(CommandAPDU.java:252)I thought that it this should be possible in order to use all the bytes of the response apdu.
    If you need more code to help please let me know.
    Cheers
    Edited by: 858145 on 06.07.2011 08:23

    2) What is PKCS? what is the difference between
    PKCS#11 and PKCS#15??PKCS is the abbreviation of "Public-Key Cryptography Standards"
    PKCS #11: Cryptographic Token Interface Standard
    See http://www.rsasecurity.com/rsalabs/node.asp?id=2133
    PKCS #15: Cryptographic Token Information Format Standard
    http://www.rsasecurity.com/rsalabs/node.asp?id=2141
    If you want to use yor smartcard as secure token it doesn't have to be a JavaCard.
    BTW: I don't remember a way to access PKCS#15 tokens on a JavaCard from within an oncard JavaCard program. If you want to use keys in your oncard program, you have to transfer it onto the card or generate it oncard and export the public key by your own oncard/offcard code.
    Jan

  • CS6 Update Failure - Error U43M1D204 (new content added to question)

    Web and Design Premium install.  I get this error when running an Adobe Application Manager update in any of the products.  Per one tech note  found, I deleted the local profile folder for the logged on admin user where the updates are downloaded and then applied (AAMUpdater).  Same issue.  The logs in that folder don't shed any light on the problem that I can see.
    Addition:  This is a network command line install from a package. The first package was incorrectly configured so that updates could not be applied by an administrator logon.  Repacked with the correct setting, uninstalled on the workstation, and reinstalled via the new network package to discover this issue. 
    QUESTION:  Is their a tech note on a manual uninstall of CS6 suite in case the control panel uninstall did not remove everything and the product is in some sort of inconsistent state regarding update patches?
    Message was edited by: MichaelFaulkner

    Manish:  As I just emailed you, and in the hope of helping someone else, on the two machines I tested, I needed to do two different procedures in order to avoid the update error:
    First system:  Uninstall the incorrectly configured AAME3 package, run the Adobe cleanup tool http://www.adobe.com/support/contact/cscleanertool.html, and reinstall the new correctly configured package, run updates.
    On the second system, I did the same steps as listed above, but then also had to logon as another admin user that had never logged onto the workstation as I was receiving the same error after the reinstall with an existing admin local profile.  Then updates ran.
    NOTE:  The first system install was run manually via a batch file while an administrator was logged on.  The second one was installed via a deployment tool without any user being logged on.
    Next test will be on a workstation after hours without any previous AAME3 package installed, just using the correctly configured one with updates allowed configured, without anyone being logged on. 

  • New Mac Pro setup questions..

    Hi,
    Just pulled the trigger on a Mac Pro 3gh, 4gb, ATI 1900. I am going to that from a dual 1.42 quicksilver and had some questions.
    It seems migration assistant may not be the best way to go. Is this true, should I reinstall everything? Is it possible to drag and drop apps to the new machine?
    Is it possible to get mail info from another machine to the new one? I have an iMac at home that I have been using Mail with and would like all the email addresses on the new machine at work. My old setup, I was still using Enterouge (sp) and would like to use Mail like at home.
    I have the stock 250gb drive and I need to figure out different drive scenarios. I recently bought two external 500gb drives before even thinking about a new system. I am a commercial photographer and one drive is a backup that lives offsite from my studio for protection. Should I mirror with an external?
    Thanks for any help!
    HD

    This FAQ may be helpful:
    A Basic Guide for Migrating to Intel-Macs
    If you are migrating a PowerPC system (G3, G4, or G5) to an Intel-Mac be careful what you migrate. Keep in mind that some items that may get transferred will not work on Intel machines and may end up causing your computer's operating system to malfunction.
    Rosetta supports "software that runs on the PowerPC G3 or G4 processor that are built for Mac OS X". This excludes the items that are not universal binaries or simply will not work in Rosetta:
    Classic Environment, and subsequently any Mac OS 9 or earlier applications
    Screensavers written for the PowerPC
    System Preference add-ons
    All Unsanity Haxies
    Browser and other plug-ins
    Contextual Menu Items
    Applications which specifically require the PowerPC G5
    Kernel extensions
    Java applications with JNI (PowerPC) libraries
    See also What Can Be Translated by Rosetta.
    In addition to the above you could also have problems with migrated cache files and/or cache files containing code that is incompatible.
    If you migrate a user folder that contains any of these items, you may find that your Intel-Mac is malfunctioning. It would be wise to take care when migrating your systems from a PowerPC platform to an Intel-Mac platform to assure that you do not migrate these incompatible items.
    If you have problems with applications not working, then completely uninstall said application and reinstall it from scratch. Take great care with Java applications and Java-based Peer-to-Peer applications. Many Java apps will not work on Intel-Macs as they are currently compiled. As of this time Limewire, Cabos, and Acquisition are available as universal binaries. Do not install browser plug-ins such as Flash or Shockwave from downloaded installers unless they are universal binaries. The version of OS X installed on your Intel-Mac comes with special compatible versions of Flash and Shockwave plug-ins for use with your browser.
    The same problem will exist for any hardware drivers such as mouse software unless the drivers have been compiled as universal binaries. For third-party mice the current choices are USB Overdrive or SteerMouse. Contact the developer or manufacturer of your third-party mouse software to find out when a universal binary version will be available.
    Also be careful with some backup utilities and third-party disk repair utilities. Disk Warrior (does not work), TechTool Pro (pre-4.5.1 versions do not work), SuperDuper (newest release works), and Drive Genius (untested) may not work properly on Intel-Macs. The same caution may apply to the many "maintenance" utilities that have not yet been converted to universal binaries.
    Before migrating or installing software on your Intel-Mac check MacFixit's Rosetta Compatibility Index.
    Additional links that will be helpful to new Intel-Mac users:
    Intel In Macs
    Apple Guide to Universal Applications
    MacInTouch List of Compatible Universal Binaries
    MacInTouch List of Rosetta Compatible Applications
    MacUpdate List of Intel-Compatible Software
    Written by Kappy with additional contributions from a brody.
    Here is a list of where your important data is stored ("~" stands for "Home"):
    Your Data in Documents:
    ~/Library/Application Support/AddressBook (copy the whole folder)
    ~/Library/Application Support/iCal (copy the whole folder)
    Also in ~ / Library/ Application Support (copy whatever else you need)
    ~/Library/Keychains (copy the whole folder)
    ~/Library/Mail (copy the whole folder)
    ~/Library/Preferences/com.apple.mail.plist (This is a very important file which contains all email account settings and general mail preferences.)
    ~ / Library/iTunes (copy the whole folder)
    ~ / Library/Safari (copy the whole folder)
    ~/ Pictures/iPhoto Library
    If you want cookies:
    ~/Library/Cookies/Cookies.plist
    ~/Library/Application Support/WebFoundation/HTTPCookies.plist
    For Entourage users:
    Entourage is in Documents/Microsoft User Data
    Also in ~ /Library/ Preferences/Microsoft
    Why reward points?(Quoted from Discussions Terms of Use.)
    The reward system helps to increase community participation. When a community member gives you (or another member) a reward for providing helpful advice or a solution to their question, your accumulated points will increase your status level within the community.
    Members may reward you with 5 points if they deem that your reply is helpful and 10 points if you post a solution to their issue. Likewise, when you mark a reply as Helpful or Solved in your own created topic, you will be awarding the respondent with the same point values.

  • I have some questions regarding setting up a software RAID 0 on a Mac Pro

    I have some questions regarding setting up a software RAID 0 on a Mac pro (early 2009).
    These questions might seem stupid to many of you, but, as my last, in fact my one and only, computer before the Mac Pro was a IICX/4/80 running System 7.5, I am a complete novice regarding this particular matter.
    A few days ago I installed a WD3000HLFS VelociRaptor 300GB in bay 1, and moved the original 640GB HD to bay 2. I now have 2 bootable internal drives, and currently I am using the VR300 as my startup disk. Instead of cloning from the original drive, I have reinstalled the Mac OS, and all my applications & software onto the VR300. Everything is backed up onto a WD SE II 2TB external drive, using Time Machine. The original 640GB has an eDrive partition, which was created some time ago using TechTool Pro 5.
    The system will be used primarily for photo editing, digital imaging, and to produce colour prints up to A2 size. Some of the image files, from scanned imports of film negatives & transparencies, will be 40MB or larger. Next year I hope to buy a high resolution full frame digital SLR, which will also generate large files.
    Currently I am using Apple's bundled iPhoto, Aperture 2, Photoshop Elements 8, Silverfast Ai, ColorMunki Photo, EZcolor and other applications/software. I will also be using Photoshop CS5, when it becomes available, and I will probably change over to Lightroom 3, which is currently in Beta, because I have had problems with Aperture, which, until recent upgrades (HD, RAM & graphics card) to my system, would not even load images for print. All I had was a blank preview page, and a constant, frozen "loading" message - the symbol underneath remained static, instead of revolving!
    It is now possible to print images from within Aperture 2, but I am not happy with the colour fidelity, whereas it is possible to produce excellent, natural colour prints using its "minnow" sibling, iPhoto!
    My intention is to buy another 3 VR300s to form a 4 drive Raid 0 array for optimum performance, and to store the original 640GB drive as an emergency bootable back-up. I would have ordered the additional VR300s already, but for the fact that there appears to have been a run on them, and currently they are out of stock at all, but the more expensive, UK resellers.
    I should be most grateful to receive advice regarding the following questions:
    QUESTION 1:
    I have had a look at the RAID setting up facility in Disk Utility and it states: "To create a RAID set, drag disks or partitions into the list below".
    If I install another 3 VR300s, can I drag all 4 of them into the "list below" box, without any risk of losing everything I have already installed on the existing VR300?
    Or would I have to reinstall the OS, applications and software again?
    I mention this, because one of the applications, Personal accountz, has a label on its CD wallet stating that the Licence Key can only be used once, and I have already used it when I installed it on the existing VR300.
    QUESTION 2:
    I understand that the failure of just one drive will result in all the data in a Raid 0 array being lost.
    Does this mean that I would not be able to boot up from the 4 drive array in that scenario?
    Even so, it would be worth the risk to gain the optimum performance provide by Raid 0 over the other RAID setup options, and, in addition to the SE II, I will probably back up all my image files onto a portable drive as an additional precaution.
    QUESTION 3:
    Is it possible to create an eDrive partition, using TechTool Pro 5, on the VR300 in bay !?
    Or would this not be of any use anyway, in the event of a single drive failure?
    QUESTION 4:
    Would there be a significant increase in performance using a 4 x VR300 drive RAID 0 array, compared to only 2 or 3 drives?
    QUESTION 5:
    If I used a 3 x VR300 RAID 0 array, and installed either a cloned VR300 or the original 640GB HD in bay 4, and I left the Startup Disk in System Preferences unlocked, would the system boot up automatically from the 4th. drive in the event of a single drive failure in the 3 drive RAID 0 array which had been selected for startup?
    Apologies if these seem stupid questions, but I am trying to determine the best option without foregoing optimum performance.

    Well said.
    Steps to set up RAID
    Setting up a RAID array in Mac OS X is part of the installation process. This procedure assumes that you have already installed Mac OS 10.1 and the hard drive subsystem (two hard drives and a PCI controller card, for example) that RAID will be implemented on. Follow these steps:
    1. Open Disk Utility (/Applications/Utilities).
    2. When the disks appear in the pane on the left, select the disks you wish to be in the array and drag them to the disk panel.
    3. Choose Stripe or Mirror from the RAID Scheme pop-up menu.
    4. Name the RAID set.
    5. Choose a volume format. The size of the array will be automatically determined based on what you selected.
    6. Click Create.
    Recovering from a hard drive failure on a mirrored array
    1. Open Disk Utility in (/Applications/Utilities).
    2. Click the RAID tab. If an issue has occurred, a dialog box will appear that describes it.
    3. If an issue with the disk is indicated, click Rebuild.
    4. If Rebuild does not work, shut down the computer and replace the damaged hard disk.
    5. Repeat steps 1 and 2.
    6. Drag the icon of the new disk on top of that of the removed disk.
    7. Click Rebuild.
    http://support.apple.com/kb/HT2559
    Drive A + B = VOLUME ONE
    Drive C + D = VOLUME TWO
    What you put on those volumes is of course up to you and easy to do.
    A system really only needs to be backed up "as needed" like before you add or update or install anything.
    /Users can be backed up hourly, daily, weekly schedule
    Media files as needed.
    Things that hurt performance:
    Page outs
    Spotlight - disable this for boot drive and 'scratch'
    SCRATCH: Temporary space; erased between projects and steps.
    http://en.wikipedia.org/wiki/StandardRAIDlevels
    (normally I'd link to Wikipedia but I can't load right now)
    Disk drives are the slowest component, so tackling that has always made sense. Easy way to make a difference. More RAM only if it will be of value and used. Same with more/faster processors, or graphic card.
    To help understand and configure your 2009 Nehalem Mac Pro:
    http://arstechnica.com/apple/reviews/2009/04/266ghz-8-core-mac-pro-review.ars/1
    http://macperformanceguide.com/
    http://www.macgurus.com/guides/storageaccelguide.php
    http://www.macintouch.com/readerreports/harddrives/index.html
    http://macperformanceguide.com/OptimizingPhotoshop-Configuration.html
    http://kb2.adobe.com/cps/404/kb404440.html

  • Some questions on versioning and synchronizing metadata

    Hy all!
    I am quite new to warehousing and Oracle Warehouse Builder, and so i would have some questions regarding on some common issues. I would appriciate if you guys would who have experience in this domain to share some good practice knowledge :)
    I am using OWB 10.2
    So first of all i would like to know if you have some proposal of the way of versioning control and synchronizing projects between team memebers when working on a bigger project, team memebers that don't work on the same repository (cause i saw that OWB has an integrated multiuser support for handeling object locks and user sessions).
    I saw that a way of migrating data from one place to a nother is using the import/export options integrated in OWB. This creates mdl files wich are some kind of "dumps" of the metadata informations, but the thing with these mdl files wich i don't think is a good way to synchronize is that first of all the .mdx and .xml files contained in the .mdl (wich is kind of a zip) contains many informations in it (like creation date, some timestamps, etc) wich are always updated when exporting, and if synchronizing these files maybee using CVS, we always will get differences between the files alltough they would contain the same thing, only timestamps changed.
    Then a nother issue with this, is that we could have 2 alternatives: dump the whole project, wich is odd to have to synchronize a single file between users, especialy on a big project, then the orher way would be doing for each object from the project (each mapping, each table, etc) an separate .mdl filem then to synchronize each file of each object, wich will be unefficient on reimporting each file in part.
    So please if you can share the way you work on a big project with many implementers with OWB, i would really appriciate.
    A nother thing i would like to know is: is there a way to generate from an existing project (like one created with OWB) the OMB commands dump (maybee in a tcl script)? Cause i saw that the way the exeprienced users implement warehousing is using TCL with OMB language. I downloaded the example from oracle for warehouse project, and i saw that is entirely made from tcl scripts (so no mdl file involved). And this i think would be nice, to have the OMB commands generated from an existing projects.
    I see this OWB projects like a database wich can be built up from only OMB commands and OWB a graphical tool to do this (same as constructing a database only from DDL commands or using SQL developer to do this), this is why i am asking about a way of dumping the OMB commands for creating an OWB project.
    Please give me some advices, and correct me if i sad some dumb things :D but i really am new to warehousing and i would really appriciate if you guys with experience could share some informations.
    Thank you verry much!
    Alex21

    Depends. Having everyone working on the same project certainly simplifies things a lot regarding merging and is generally my preference. But I also recognize that some projects are complex enough that people wind up stepping on each other's toes if this is the case. In those cases, though, I try to minimize the issue of merging changes by having common structural objects (code libraries, tables, views, etc) retained in a single, strictly controlled, central project schema and having the developer's personal work areas reference them by synonym, thus being unable to alter them to the detriment of others.
    If they want to change a common object then need to drop their synonym and make a local copy which they can alter, and then there is a managed process by which these get merged back into the main project schema.
    This way any changes MUST go through a central schema, we can put processes in place to notify all of the team of any impending changes, and can also script updates across the team.
    Every hour a script runs automatically that checks for dropped synonyms and notifies the project leader. It especially checks for two developers who have built local copies of the same object and notifies each that they need to coordinate with each other as they are risking a conflict. When a structural change is submitted back to the central shared schema, it is added to a batch that is installed at end of business and a list of those impending changes is circulated to the team along with impact analysis for dependencies. The install script updates the main schema, then also drops the local copy of the object in the developer's schema who made the change and re-establishes the synonym there to get back to status quo for the change monitoring. Finally, it then updates itself in all of the developer areas via OMBPlus. So, each morning the developers return to an updated and synched environment as far as the underlying structure.
    This takes care of merging structural issues, and the management of the team should minimize other metadata merging by managing the worklist of who is to be working on a given mapping or process flow at a given time. Anyone found to be doing extraneous changes to a mapping or process flow when it is not in their job queue without getting pre-approval will be spoken to VERY firmly as this is counter to policy. And yes, OWB objects such as mappings are then also coordinated to the central project via import/export. OMBplus scripts also propogate these changes daily across the team as well.
    Yep, there is a whole lot of scripting involved to get set up.... but it saves a ton of time merging things and solvinv conflicts down the road.
    Cheers,
    Mike

  • Some questions on filter

    Hi Experts,
    I have some questions on IP. I am comfortable working with BPS but I dont know how they work in IP.
    1) Suppose we take an example
    I have put filter on some char. say "Sales Region" from values A to D. and now I associate a planning function repost values (with filter) for only A "Sales Region" (So I keep only value A in fields for condition)
    So Question is will sales region B to D will be locked for write access for some other operation ??
    And in BPS there is no concept like locking on basis of filters put on Planning Level and Planning Package right ???
    Entire transactional cube is locked for any other user if some user is running a layout in a particular Planning Area even if filter conditions in Planning Level and Planning Package do not overlap in any other layout on same Info Area ?? m i correct ??
    Ask me if I am not clear in explaining.
    Good conceptual answers would be really appriciated.
    Thanks,
    Mihir

    Hi Mayank,
    Thank you for your answer. I tried and tested ur BPS answer regarding locking at planning package. It works well.Before I used to restrict at planning level.
    But once I restrict at package and hit "save all" as well as "save modal" and log off the system those filters get removed when I relogin and see the same package. Why is this so ???
    Would like to hear from you....
    Thank again,
    Mihir

Maybe you are looking for