Solaris Live Upgrade

I didnt find a specific place to place this topic, so maybe you can replace.
I read a lot about Solaris Live Upgrade, but I'd would like to know if I can copy a BE from one computer to another over network via NFS.
Or if there is another software that can make a image of my system and replace to the others computers (they are all the same), so I will dont need to configure all of them.
Thx for any help.

Thanks for answering.
I read about JumpStart installation, but is it possible to install a customized image, but not with just the partition, packages and patches selected. I wanted to share a whole system configured. For example, the DNS, IP, User's account settings, etc.
Because, I just found a way to share the installation image with the selected packages and patches. But you will need to configure the files of the packages later in the clients.
Is it possible?

Similar Messages

  • Solaris Live Upgrade with NGZ

    Hi
    I am trying to perform a Live Upgrade on my 2 Servers, both of them have NGZ installed and those NGZ are on an diifferent Zpool not on the Rpool and also are on an external disk.
    I have installed all the latest patches required for LU to work properly but when i perform an <lucreate> i start having problems... (new_s10BE is the new BE i'm creating)
    On my 1st Server:
    I have a Global zone and 1 NGZ named mddtri.. This is the error i am getting:-
    ERROR: unable to mount zone <mddtri> in </.alt.tmp.b-VBb.mnt>.
    zoneadm: zone 'mddtri': zone root /zoneroots/mddtri/root already in use by zone mddtri
    zoneadm: zone 'mddtri': call to zoneadm failed
    ERROR: unable to mount non-global zones of ABE: cannot make bootable
    ERROR: cannot unmount </.alt.tmp.b-VBb.mnt/var/run>
    ERROR: unable to make boot environment <new_s10BE> bootable
    On my 2nd Server:
    I have a Global zone and 10 NGZ. This is the error i am getting:-
    WARNING: Directory </zoneroots/zone1> zone <global> lies on a filesystem shared netween BEs, remapping path to </zoneroots/zone1/zone1-new_s10BE>
    WARNING: Device <zone1> is shared between BEs, remmapping to <zone1-new_s10BE>
    *.This happens for all the NGZ running.*
    Duplicating ZFS datasets from PBE to ABE.
    ERROR: The dataset <zone1-new_s10BE> is on top of ZFS pool. Unable to clone. Please migrate the zone  to dedicated dataset.
    ERROR: Unable to create a duplicate of <zone1> dataset in PBE. <zone1-new_s10BE> dataset in ABE already exists.
    Reverting state of zones in PBE <old_s10BE>
    ERROR: Unable to copy file system from boot environment <old_s10BE> to BE <new_s10BE>
    ERROR: Unable to populate file systems from boot environment <new_s10BE>
    Help, I need to sort this out a.s.a.p!

    Hi,
    I have the same problem with an attached A5200 with mirrored disks (Solaris 9, Volume Manager). Whereas the "critical" partitions should be copied to a second system disk, the mirrored partitions should be shared.
    Here is a script with lucreate.
    #!/bin/sh
    Logdir=/usr/local/LUscripts/logs
    if &#91; ! -d ${Logdir} &#93;
    then
    echo ${Logdir} existiert nicht
    exit
    fi
    /usr/sbin/lucreate \
    -l ${Logdir}/$0.log \
    -o ${Logdir}/$0.error \
    -m /:/dev/dsk/c2t0d0s0:ufs \
    -m /var:/dev/dsk/c2t0d0s3:ufs \
    -m /opt:/dev/dsk/c2t0d0s4:ufs \
    -m -:/dev/dsk/c2t0d0s1:swap \
    -n disk0
    And here is the output
    root&#64;ahbgbld800x:/usr/local/LUscripts >./lucreate_disk0.sh
    Discovering physical storage devices
    Discovering logical storage devices
    Cross referencing storage devices with boot environment configurations
    Determining types of file systems supported
    Validating file system requests
    Preparing logical storage devices
    Preparing physical storage devices
    Configuring physical storage devices
    Configuring logical storage devices
    Analyzing system configuration.
    INFORMATION: Unable to determine size or capacity of slice </dev/md/RAID-INT/dsk/d0>.
    ERROR: An error occurred during creation of configuration file.
    ERROR: Cannot create the internal configuration file for the current boot environment <disk3>.
    Assertion failed: *ptrKey == (unsigned long long)_lu_malloc, file lu_mem.c, line 362<br />
    Abort - core dumped

  • DB upgrade during a Solaris upgrade using live upgrade feature

    Hello,
    Here is the scenario. We are currently upgrading our OS from Solaris 9 to Solaris 10. And there is also a major database upgrade involved too. To help mitigate downtime, the Solaris Live Upgrade feature is being used by the SA's. The DB to be upgraded is currently at 9i and the proposed upgrade end state is at 10.2.0.3.
    Does anyone know if it is possible to do at least part of the database upgrade in the alternate boot environment created during the live upgrade process? So lets say that I am able to get the database partly upgraded to 10.2.0.1. Then, I want to run the patch set for 10.2.0.3 in the alternate boot environment and do the upgrade of the instance there too so that when the alternate boot environment is booted to the primary boot environment, I can, in effect, just start up the database.
    That's sort of a high level simplified version of what I think may be possible.
    Does anyone have any recommendations? We don't have any other high availability solutions currently in place so options such as using data guard do not apply.
    Thanks in advance!

    Hi magNorth,
    I'm not a Solaris expert but I'd always recommend not to do both steps (OS upgrade and database upgrade) at the same time. I've seen to many live scenarios where either the one or the other has caused issues - and it's fairly tough for all three parties (You as the customer, Sun Support, Oracle Support) to find out the end what was/is causing the issues.
    So my recommendation would be:
    1) Do your Solaris upgrade from 9 to 10 and check for the most recent Solaris patches - once this has been finished and run for one or two weeks successfully in production then go to step 2
    2) Do your database upgrade - I would suggest not going to 10.2.0.3 but directly to 10.2.0.4 because 10.2.0.4 has ~2000 bug fixes plus the security fixes from April2008 - but that's your decision. If 10.2.0.3 is a must then at least check Note:401435.1 for known issues and alert in 10.2.0.3 and for both cases the Upgrade Companion: Note:466181.1     
    Kind regards
    Mike

  • Looking for information on best practices using Live Upgrade to patch LDOMs

    This is in Solaris 10. Relatively new to the style of patching... I have a T5240 with 4 LDOMS. A control LDOM and three clients. I have some fundamental questions I'd like help with..
    Namely:
    #1. The Client LDOMS have zones running in them. Do I need to init 0 the zone or can I just +zoneadm zone halt+ them regardless of state? I.E. if it's running a database will halting the zone essentially snapshot it or will it attempt to shut it down. Is this even a nessessary step.
    #2. What is the reccommended reboot order for the LDOMs? do I need to init 0 the client ldoms and the reboot the control ldom or can I leave the client LDOM's running and just reboot the control and then reboot the clients after the control comes up?
    #3. Oracle. it's running in several of the zones on the client LDOM's what considerations need to be made for this?
    I am sure other things will come up during the conversation but I have been looking for an hour on Oracle's site for this and the only thing I can find is old Sun Docs with broken links.
    Thanks for any help you can provide,
    pipelineadmin+*                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

    Before you use live upgrade, or any other patching technique for Solaris, please be sure to read http://docs.oracle.com/cd/E23823_01/html/E23801/index.html which includes information on upgrading systems with non-global zones. Also, go to support.oracle.com and read Oracle Solaris Live Upgrade Information Center [ID 1364140.1]. These really are MANDATORY READING.
    For the individual questions:
    #1. During the actual maintenance you don't have to do anything to the zone - just operate it as normal. That's the purpose of the "live" in "live upgrade" - you're applying patches on a live, running system under normal operations. When you are finisihed with that process you can then reboot into the new "boot environment". This will become more clear after reading the above documents. Do as you normally would do before taking a planned outage: shut the databases down using the database commands for a graceful shutdown. A zone halt will abruptly stop the zone and is not a good idea for a database. Alternatively, if you can take application outages, you could (smoothly) shutdown the applications and then their domains, detach the zones (zoneadm detach) and then do a live upgrade. Some people like that because it makes things faster. After the live upgrade you would reboot and then zoneadm attach the zones again. The fact that the Solaris instance is running within a logical domain really is mostly besides the point with respect to this process.
    As you can see, there are a LOT of options and choices here, so it's important to read the doc. I ***strongly*** recommend you practice on a test domain so you can get used to the procedure. That's one of the benefits of virtualization: you can easily set up test environments so you cn test out procedures. Do it! :-)
    #2 First, note that you can update the domains individually at separate times, just as if they were separate physical machines. So, you could update the guest domains one week (all at once or one at a time), reboot them into the new Solaris 10 software level, and then a few weeks later (or whenever) update the control domain.
    If you had set up your T5240 in a split-bus configuration with an alternate I/O domain providing virtual I/O for the guests, you would be able to upgrade the extra I/O domain and the control domain one at a time in a rolling upgrade - without ever having to reboot the guests. That's really powerful for providing continuous availability. Since you haven't done that, the answer is that at the point you reboot the control domain the guests will lose their I/O. They don't crash, and technically you could just have them continue until the control domain comes back up at which time the I/O devices reappear. For an important application like a database I wouldn't recommend that. Instead: shutdown the guests. then reboot the control domain, then bring the guest domains back up.
    3. The fact that Oracle database is running in zones inside those domains really isn't an issue. You should study the zones administration guide to understand the operational aspects of running with zones, and make sure that the patches are compatible with the version of Oracle.
    I STRONGLY recommend reading the documents mentioned at top, and setting up a test domain to practice on. It shouldn't be hard for you to find documentation. Go to www.oracle.com and hover your mouse over "Oracle Technology Network". You'll see a window with a menu of choices, one of which is "Documentation" - click on that. From there, click on System Software, and it takes you right to the links for Solaris 10 and 11.

  • Volume as install disk for Guest Domain and Live Upgrade

    Hi Folks,
    I am new to LDOMs and have some questions - any pointers, examples would be much appreciated:
    (1) With support for volumes to be used as whole disks added in LDOM release 1.0.3, can we export a whole LUN under either VERITAS DMP or mpxio control to guest domain and install Solaris on it ? Any gotchas or special config required to do this ?
    (2) Can Solaris Live Upgrade be used with Guest LDOMs ? or is this ability limited to Control Domains ?
    Thanks

    The answer to your #1 question is YES.
    Here's my mpxio enabled device.
    non-STMS device name STMS device name
    /dev/rdsk/c2t50060E8010029B33d16 /dev/rdsk/c4t4849544143484920373730313036373530303136d0
    /dev/rdsk/c3t50060E8010029B37d16 /dev/rdsk/c4t4849544143484920373730313036373530303136d0
    create the virtual disk using slice 2
    ldm add-vdsdev /dev/dsk/c4t4849544143484920373730313036373530303136d0s2 77bootdisk@primary-vds01
    add the virtual disk to the guest domain
    ldm add-vdisk apps bootdisk@primary-vds01 ldom1
    the virtual disk will be imprted as c0d0 which is the whole lun itself.
    bind, start ldom 1 and install OS (i used jumpstart) and it partitioned the boot disk c0d0 as / 15GB, swap the remaining space (10GB)
    when you run format, print command on both guest and primary domain on this disk you'll see the same slice/size information
    Part Tag Flag Cylinders Size Blocks
    0 root wm 543 - 1362 15.01GB (820/0/0) 31488000
    1 swap wu 0 - 542 9.94GB (543/0/0) 20851200
    2 backup wm 0 - 1362 24.96GB (1363/0/0) 52339200
    3 unassigned wm 0 0 (0/0/0) 0
    4 unassigned wm 0 0 (0/0/0) 0
    5 unassigned wm 0 0 (0/0/0) 0
    6 unassigned wm 0 0 (0/0/0) 0
    7 unassigned wm 0 0 (0/0/0) 0
    I havent used DMP but HDLM (Hitachi Dynamic link manager) seems not supported by ldom as i cannot make it work :(
    I have no answer on your next question unfortunately.

  • After Solaris live upgarde disks unavailable

    Hi All,
    we have two SUN Fire 6800 - cluster node. OS and Kernel Version: Solaris 9 9/05 s9s_u8wos_05 SPARC. After Solaris live upgrade the new root disks are continueously failing. During boot with new boot environment Solaris_9_905 wait and then boot disks become unavailable, cluster will crash and start with old boot environment. If the failed devices are removed then inserted, it makes the device available again.
    Could anybody help us? Any idea?
    regards
    Josef

    Hi Tim,
    We have Sun Cluster 3.1. Is it possible error? I'll seek for description of LU.
    However: it seems a simple hardware failure. it lost just the disks.
    Messages:
    root@mcl01:~#metastat -p
    d60 -m d61 d62 d63 1
    d61 1 1 c1t1d0s6
    d62 1 1 c1t5d0s6
    d63 1 1 c1t4d0s6
    d50 -m d51 d52 d53 1
    d51 1 1 c1t1d0s5
    d52 1 1 c1t5d0s5
    d53 1 1 c1t4d0s5
    d30 -m d31 d32 d33 1
    d31 1 1 c1t1d0s3
    d32 1 1 c1t5d0s3
    d33 1 1 c1t4d0s3
    d0 -m d1 d2 d3 1
    d1 1 1 c1t1d0s0
    d2 1 1 c1t5d0s0
    d3 1 1 c1t4d0s0
    0. c1t0d0
    /pci@8,600000/SUNW,qlc@2/fp@0,0/ssd@w21000000870e3de0,0
    1. c1t2d0
    /pci@8,600000/SUNW,qlc@2/fp@0,0/ssd@w21000000871421d4,0
    2. c1t3d0
    /pci@8,600000/SUNW,qlc@2/fp@0,0/ssd@w21000000871432fc,0
    3. c1t5d0
    /pci@8,600000/SUNW,qlc@2/fp@0,0/ssd@w21000000871421a1,0
    root@mcl02:~#metastat -p
    d60 -m d61 d62 d63 1
    d61 1 1 c0t1d0s6
    d62 1 1 c4t1d0s6
    d63 1 1 c4t2d0s6
    d50 -m d51 d52 d53 1
    d51 1 1 c0t1d0s5
    d52 1 1 c4t1d0s5
    d53 1 1 c4t2d0s5
    d30 -m d31 d32 d33 1
    d31 1 1 c0t1d0s3
    d32 1 1 c4t1d0s3
    d33 1 1 c4t2d0s3
    d0 -m d1 d2 d3 1
    d1 1 1 c0t1d0s0
    d2 1 1 c4t1d0s0
    d3 1 1 c4t2d0s0
    0. c0t0d0
    /ssm@0,0/pci@18,700000/pci@1/scsi@2/sd@0,0
    !!! 1. c0t1d0
    /ssm@0,0/pci@18,700000/pci@1/scsi@2/sd@1,0
    2. c0t2d0
    /ssm@0,0/pci@18,700000/pci@1/scsi@2/sd@2,0
    297. c4t0d0
    /ssm@0,0/pci@1c,700000/pci@1/scsi@2/sd@0,0
    !!! 298. c4t1d0
    /ssm@0,0/pci@1c,700000/pci@1/scsi@2/sd@1,0
    !!! 299. c4t2d0
    /ssm@0,0/pci@1c,700000/pci@1/scsi@2/sd@2,0
    And
    "format" doesn't show some of these disks or says "disk type undefined" (- something like this)
    thx
    J

  • Applying individual patches to a Live Upgrade Environment

    Hi all
    Is it possible to apply individual patches to a Live Upgrade Environment? More specifically, is it possible to apply a kernel patch to the LUE?
    I was thinking that the command would look like this:
    patchadd -R /alt_env_root /location/144500-19
    In the README I don't find anything for patching a LUE and only mention of installing it in single user mode or when the system is close to totally idle.
    Dean

    yes u can apply the individul patches anad kernel patches in the lu enviornment
    for that first u need to create ABE
    luupgrade -n mytestBE -t -s /patchesfolder 166981-17
    you can refer this below link
    http://www.oracle.com/technetwork/server-storage/solaris10/solaris-live-upgrade-wp-167900.pdf
    http://www.oracle.com/technetwork/systems/articles/lu-patch-jsp-139117.html

  • Creating Boot Environment for Live Upgrade

    Hello.
    I'd like to upgrade a Sun Fire 280R system running Solaris 8 to Solaris 10 U4. I'd like to use Live Upgrade to do this. As that's going to be my first LU of a system, I've got some questions. Before I start, I'd like to mention that I have read the �Solaris 10 8/07 Installation Guide: Solaris Live Upgrade and Upgrade Planning� ([820-0178|http://docs.sun.com/app/docs/doc/820-0178]) document. Nonetheless, I'd also appreciate pointers to a more �hands-on� documentation/howto reg. live upgrade.
    The system that I'd like to upgrade has these filesystems:
    (winds02)askwar$ df
    Filesystem 1k-blocks Used Available Use% Mounted on
    /dev/md/dsk/d30 4129290 684412 3403586 17% /
    /dev/md/dsk/d32 3096423 1467161 1567334 49% /usr
    /dev/md/dsk/d33 2053605 432258 1559739 22% /var
    swap 7205072 16 7205056 1% /var/run
    /dev/dsk/c3t1d0s6 132188872 61847107 69019877 48% /u04
    /dev/md/dsk/d34 18145961 5429315 12535187 31% /opt
    /dev/md/dsk/d35 4129290 77214 4010784 2% /export/home
    It has 2 built in harddisks, which form those metadevices. You can find the �metastat� at http://askwar.pastebin.ca/697380. I'm now planning to break the mirrors for /, /usr, /var and /opt. To do so, I'd run
    metadetach d33 d23
    metaclear d23
    d23 is/used to be c1t1d0s4. I'd do this for d30, d32 and d34 as well. Plan is, that I'd be able to use these newly freed slices on c1t1d0 for LU. I know that I'm in trouble when c1t0d0 now dies. But that's okay, as that system isn't being used anyway right now...
    Or wait, I can use lucreate to do that as well, can't I? So, instead of manually detaching the mirror, I could do:
    lucreate -n s8_2_s10 -m /:/dev/md/dsk/d30:preserve,ufs \
    -m /usr:/dev/md/dsk/d32:preserve,ufs \
    -m /var:/dev/md/dsk/d33:preserve,ufs \
    -m /opt:/dev/md/dsk/d34:preserve,ufs
    Does that sound right? I'd assume, that I'd then have a new boot environment called �s8_2_s10�, which uses the contents of the old metadevices. Or would the correct command rather be:
    lucreate -n s8_2_s10_v2 \
    -m /:/dev/md/dsk/d0:mirror,ufs \
    -m /:/dev/md/dsk/d20:detach,attach,preserve \
    -m /usr:/dev/md/dsk/d2:mirror,ufs \
    -m /usr:/dev/md/dsk/d22:detach,attach,preserve \
    -m /var:/dev/md/dsk/d3:mirror,ufs \
    -m /var:/dev/md/dsk/d23:detach,attach,preserve \
    -m /opt:/dev/md/dsk/d4:mirror,ufs \
    -m /opt:/dev/md/dsk/d24:detach,attach,preserve
    What would be the correct way to create the new boot environment? As I said, I haven't done this before, so I'd really appreciate some help.
    Thanks a lot,
    Alexander Skwar

    I replied to this thread: Re: lucreate and non-global zones as to not duplicate content, but for some reason it was locked. So I'll post here...The thread was locked because you were not replying to it.
    You were hijacking that other person's discussion from 2012 to ask your own new post.
    You have now properly asked your question and people can pay attention to you and not confuse you with that other person.

  • Sun Docs on Live Upgrade

    New from Sun:
    Upgrading With Solaris Live Upgrade:
    http://docs.sun.com/app/docs/doc/817-5505/6mkv5m1ke?a=view

    Frederick is correct that UM does not support download only of patches, however, the cli command smpatch does. Take a look at the man page for smpatch and specifically the download subcommand.
    BTW, being able to use UM to patch live upgrade images sounds like a great future feature.
    -Dave

  • Solaris 10 5/08 live upgrade only for customers with serviceplan ?

    Live upgrade fails due to missing /usr/bin/7za
    Which seems to be installed by adding patch 137322-01 on x86 according to release notes http://docs.sun.com/app/docs/doc/820-4078/installbugs-114?l=en&a=view
    But this patch (may also be the case for Sparc patch) are only available for customers with a valid serviceplan.
    Does this mean that from now on its required to purchase a serviceplan if you would run Solaris 10 and use the normal procedures for system upgrades ?
    A bit disappointing ...
    Regards
    /Flemming

    Live upgrade fails due to missing /usr/bin/7za
    Which seems to be installed by adding patch 137322-01 on x86 according to release notes http://docs.sun.com/app/docs/doc/820-4078/installbugs-114?l=en&a=view
    But this patch (may also be the case for Sparc patch) are only available for customers with a valid serviceplan.
    Does this mean that from now on its required to purchase a serviceplan if you would run Solaris 10 and use the normal procedures for system upgrades ?
    A bit disappointing ...
    Regards
    /Flemming

  • Live upgrade - solaris 8/07 (U4) , with non-global zones and SC 3.2

    Dears,
    I need to use live upgrade for SC3.2 with non-global zones, solaris 10 U4 to Solaris 10 10/09 (latest release) and update the cluster to 3.2 U3.
    i dont know where to start, i've read lots of documents, but couldn't find one complete document to cover the whole process.
    i know that upgrade for solaris 10 with non-global zones is supported since my soalris 10 release, but i am not sure if its supported with SC.
    Appreciate your help

    Hi,
    I am not sure whether this document:
    http://wikis.sun.com/display/BluePrints/Maintaining+Solaris+with+Live+Upgrade+and+Update+On+Attach
    has been on the list of docs you found already.
    If you click on the download link, it won't work. But if you use the Tools icon on the upper right hand corner and click on attachements, you'll find the document. Its content is solely based on configurations with ZFS as root and zone root, but should have valuable information for other deployments as well.
    Regards
    Hartmut

  • Solaris 10 update 9 - live upgrade issues with ZFS

    Hi
    After doing a live upgrade from Solaris 10 update 8 to Solaris 10 update 9 the alternate boot environment I created is no longer bootable.
    I have completed all the pre-upgrade steps like:
    - Installing the latest version of live upgrade from the update 9 ISO.
    - Create and test the new boot environment.
    - Create a sysidcfg file used by the live upgrade that has auto_reg=disable in it.
    There is also no errors while creating the boot environment or even when activating it.
    Here is the error I get:
    SunOS Release 5.10 Version Generic_14489-06 64-bit
    Copyright (c) 1983, 2010, Oracle and/or its affiliates. All rights reserved.
    NOTICE: zfs_parse_bootfs: error 22
    Cannot mount root on altroot/37 fstype zfs
    *panic[cpu0]/thread=fffffffffbc28040: vfs mountroot: cannot mount root*
    ffffffffffbc4a8d0 genunix:main+107 ()
    Skipping system dump - no dump device configured
    Does anyone know how I can fix this?
    Edited by: user12099270 on 02-Feb-2011 04:49

    Found the culprit... *142910-17*... breaks it
    System has findroot enabled GRUB
    Updating GRUB menu default setting
    GRUB menu default setting is unaffected
    Saving existing file </boot/grub/menu.lst> in top level dataset for BE <s10x_u8wos_08a> as <mount-point>//boot/grub/menu.lst.prev.
    File </etc/lu/GRUB_backup_menu> propagation successful
    Successfully deleted entry from GRUB menu
    Validating the contents of the media </admin/x86/Patches/10_x86_Recommended/patches>.
    The media contains 204 software patches that can be added.
    Mounting the BE <s10x_u8wos_08a_Jan2011>.
    Adding patches to the BE <s10x_u8wos_08a_Jan2011>.
    Validating patches...
    Loading patches installed on the system...
    Done!
    Loading patches requested to install.
    Done!
    The following requested patches have packages not installed on the system
    Package SUNWio-tools from directory SUNWio-tools in patch 142910-17 is not installed on the system. Changes for package SUNWio-tools will not be applied to the system.
    Package SUNWzoneu from directory SUNWzoneu in patch 142910-17 is not installed on the system. Changes for package SUNWzoneu will not be applied to the system.
    Package SUNWpsm-ipp from directory SUNWpsm-ipp in patch 142910-17 is not installed on the system. Changes for package SUNWpsm-ipp will not be applied to the system.
    Package SUNWsshdu from directory SUNWsshdu in patch 142910-17 is not installed on the system. Changes for package SUNWsshdu will not be applied to the system.
    Package SUNWsacom from directory SUNWsacom in patch 142910-17 is not installed on the system. Changes for package SUNWsacom will not be applied to the system.
    Package SUNWmdbr from directory SUNWmdbr in patch 142910-17 is not installed on the system. Changes for package SUNWmdbr will not be applied to the system.
    Package SUNWopenssl-commands from directory SUNWopenssl-commands in patch 142910-17 is not installed on the system. Changes for package SUNWopenssl-commands will not be applied to the system.
    Package SUNWsshdr from directory SUNWsshdr in patch 142910-17 is not installed on the system. Changes for package SUNWsshdr will not be applied to the system.
    Package SUNWsshcu from directory SUNWsshcu in patch 142910-17 is not installed on the system. Changes for package SUNWsshcu will not be applied to the system.
    Package SUNWsshu from directory SUNWsshu in patch 142910-17 is not installed on the system. Changes for package SUNWsshu will not be applied to the system.
    Package SUNWgrubS from directory SUNWgrubS in patch 142910-17 is not installed on the system. Changes for package SUNWgrubS will not be applied to the system.
    Package SUNWzoner from directory SUNWzoner in patch 142910-17 is not installed on the system. Changes for package SUNWzoner will not be applied to the system.
    Package SUNWmdb from directory SUNWmdb in patch 142910-17 is not installed on the system. Changes for package SUNWmdb will not be applied to the system.
    Package SUNWpool from directory SUNWpool in patch 142910-17 is not installed on the system. Changes for package SUNWpool will not be applied to the system.
    Package SUNWudfr from directory SUNWudfr in patch 142910-17 is not installed on the system. Changes for package SUNWudfr will not be applied to the system.
    Package SUNWxcu4 from directory SUNWxcu4 in patch 142910-17 is not installed on the system. Changes for package SUNWxcu4 will not be applied to the system.
    Package SUNWarc from directory SUNWarc in patch 142910-17 is not installed on the system. Changes for package SUNWarc will not be applied to the system.
    Package SUNWtftp from directory SUNWtftp in patch 142910-17 is not installed on the system. Changes for package SUNWtftp will not be applied to the system.
    Package SUNWaccu from directory SUNWaccu in patch 142910-17 is not installed on the system. Changes for package SUNWaccu will not be applied to the system.
    Package SUNWppm from directory SUNWppm in patch 142910-17 is not installed on the system. Changes for package SUNWppm will not be applied to the system.
    Package SUNWtoo from directory SUNWtoo in patch 142910-17 is not installed on the system. Changes for package SUNWtoo will not be applied to the system.
    Package SUNWcpc from directory SUNWcpc.i in patch 142910-17 is not installed on the system. Changes for package SUNWcpc will not be applied to the system.
    Package SUNWftdur from directory SUNWftdur in patch 142910-17 is not installed on the system. Changes for package SUNWftdur will not be applied to the system.
    Package SUNWypr from directory SUNWypr in patch 142910-17 is not installed on the system. Changes for package SUNWypr will not be applied to the system.
    Package SUNWlxr from directory SUNWlxr in patch 142910-17 is not installed on the system. Changes for package SUNWlxr will not be applied to the system.
    Package SUNWdcar from directory SUNWdcar in patch 142910-17 is not installed on the system. Changes for package SUNWdcar will not be applied to the system.
    Package SUNWnfssu from directory SUNWnfssu in patch 142910-17 is not installed on the system. Changes for package SUNWnfssu will not be applied to the system.
    Package SUNWpcmem from directory SUNWpcmem in patch 142910-17 is not installed on the system. Changes for package SUNWpcmem will not be applied to the system.
    Package SUNWlxu from directory SUNWlxu in patch 142910-17 is not installed on the system. Changes for package SUNWlxu will not be applied to the system.
    Package SUNWxcu6 from directory SUNWxcu6 in patch 142910-17 is not installed on the system. Changes for package SUNWxcu6 will not be applied to the system.
    Package SUNWpcmci from directory SUNWpcmci in patch 142910-17 is not installed on the system. Changes for package SUNWpcmci will not be applied to the system.
    Package SUNWarcr from directory SUNWarcr in patch 142910-17 is not installed on the system. Changes for package SUNWarcr will not be applied to the system.
    Package SUNWscpu from directory SUNWscpu in patch 142910-17 is not installed on the system. Changes for package SUNWscpu will not be applied to the system.
    Package SUNWcpcu from directory SUNWcpcu in patch 142910-17 is not installed on the system. Changes for package SUNWcpcu will not be applied to the system.
    Package SUNWopenssl-include from directory SUNWopenssl-include in patch 142910-17 is not installed on the system. Changes for package SUNWopenssl-include will not be applied to the system.
    Package SUNWdtrp from directory SUNWdtrp in patch 142910-17 is not installed on the system. Changes for package SUNWdtrp will not be applied to the system.
    Package SUNWhermon from directory SUNWhermon in patch 142910-17 is not installed on the system. Changes for package SUNWhermon will not be applied to the system.
    Package SUNWpsm-lpd from directory SUNWpsm-lpd in patch 142910-17 is not installed on the system. Changes for package SUNWpsm-lpd will not be applied to the system.
    Package SUNWdtrc from directory SUNWdtrc in patch 142910-17 is not installed on the system. Changes for package SUNWdtrc will not be applied to the system.
    Package SUNWhea from directory SUNWhea in patch 142910-17 is not installed on the system. Changes for package SUNWhea will not be applied to the system.
    Package SUNW1394 from directory SUNW1394 in patch 142910-17 is not installed on the system. Changes for package SUNW1394 will not be applied to the system.
    Package SUNWrds from directory SUNWrds in patch 142910-17 is not installed on the system. Changes for package SUNWrds will not be applied to the system.
    Package SUNWnfsskr from directory SUNWnfsskr in patch 142910-17 is not installed on the system. Changes for package SUNWnfsskr will not be applied to the system.
    Package SUNWudf from directory SUNWudf in patch 142910-17 is not installed on the system. Changes for package SUNWudf will not be applied to the system.
    Package SUNWixgb from directory SUNWixgb in patch 142910-17 is not installed on the system. Changes for package SUNWixgb will not be applied to the system.
    Checking patches that you specified for installation.
    Done!
    Approved patches will be installed in this order:
    142910-17
    Checking installed patches...
    Executing prepatch script...
    Installing patch packages...
    Patch 142910-17 has been successfully installed.
    See /a/var/sadm/patch/142910-17/log for details
    Executing postpatch script...
    Creating GRUB menu in /a
    Installing grub on /dev/rdsk/c2t0d0s0
    stage1 written to partition 0 sector 0 (abs 16065)
    stage2 written to partition 0, 273 sectors starting at 50 (abs 16115)
    Patch packages installed:
    BRCMbnx
    SUNWaac
    SUNWahci
    SUNWamd8111s
    SUNWcakr
    SUNWckr
    SUNWcry
    SUNWcryr
    SUNWcsd
    SUNWcsl
    SUNWcslr
    SUNWcsr
    SUNWcsu
    SUNWesu
    SUNWfmd
    SUNWfmdr
    SUNWgrub
    SUNWhxge
    SUNWib
    SUNWigb
    SUNWintgige
    SUNWipoib
    SUNWixgbe
    SUNWmdr
    SUNWmegasas
    SUNWmptsas
    SUNWmrsas
    SUNWmv88sx
    SUNWnfsckr
    SUNWnfscr
    SUNWnfscu
    SUNWnge
    SUNWnisu
    SUNWntxn
    SUNWnv-sata
    SUNWnxge
    SUNWopenssl-libraries
    SUNWos86r
    SUNWpapi
    SUNWpcu
    SUNWpiclu
    SUNWpsdcr
    SUNWpsdir
    SUNWpsu
    SUNWrge
    SUNWrpcib
    SUNWrsgk
    SUNWses
    SUNWsmapi
    SUNWsndmr
    SUNWsndmu
    SUNWtavor
    SUNWudapltu
    SUNWusb
    SUNWxge
    SUNWxvmpv
    SUNWzfskr
    SUNWzfsr
    SUNWzfsu
    PBE GRUB has no capability information.
    PBE GRUB has no versioning information.
    ABE GRUB is newer than PBE GRUB. Updating GRUB.
    GRUB update was successfull.
    Unmounting the BE <s10x_u8wos_08a_Jan2011>.
    The patch add to the BE <s10x_u8wos_08a_Jan2011> completed.
    Still need to know how to resolve it though...

  • Live upgrade from solaris 8 to 10

    how do i upgrade from solaris 8 to 10 using live upgrade on a sun netra t1 machine?

    THere is a good intro to LU at
    http://www.sun.com/bigadmin/collections/installation.html
    beyond that, see docs.sun.com

  • Solaris 9 live upgrade to solaris 10 segmentation fault

    pre-upgrade work:
    apply patch doc 72099
    upgrade live upgrade to solaris 10.
    lucreate-m /:/dev/dsk/c0t1d0s0:ufs -m /var:/dev/dsk/c0t1d0s1:ufs -m /opt:/dev/dsk/c0t1d0s3:ufs -n sol10
    it works well
    oot Environment Is Active Active Can Copy
    Name Complete Now On Reboot Delete Status
    sol9 yes yes yes no -
    sol10 yes no no yes -
    But luupgrade -u -n sol10 -s /sol11
    Validating the contents of the media </sol11>.
    The media is a standard Solaris media.
    The media contains an operating system upgrade image.
    The media contains <Solaris> version <11>.
    Constructing upgrade profile to use.
    Locating the operating system upgrade program.
    Checking for existence of previously scheduled Live Upgrade requests.
    Creating upgrade profile for BE <sol10>.
    Determining packages to install or upgrade for BE <sol10>.
    Performing the operating system upgrade of the BE <sol10>.
    CAUTION: Interrupting this process may leave the boot environment unstable
    or unbootable.
    /usr/sbin/luupgrade&#91;677&#93;: 6549 Segmentation Fault(coredump)
    ERROR: Installation of the packages from this media of the media failed; pfinstall returned these diagnostics:
    The Solaris upgrade of the boot environment <sol10> failed.

    I would break your mirror and do a liveupgrade on the available mirror. Then boot that disk.

  • How to upgrade from Solaris 10 u2 to Solaris 10 u3 using Live Upgrade?

    Does anyone know how I can use "live upgrade" (or any other method) for upgrading my solaris 10 u2 to Solaris 10 u3?
    I can't find instructions anywhere! ugh.
    Thank you for your help!

    Try the following documentation:
    http://docs.sun.com/app/docs/doc/819-6396
    For future reference, I found this by going to sun.com -> documentation -> Solaris Operating Systems -> Solaris 10 Operating System -> Solaris 10 11/06 Release and Installation Collection.
    -- Alan

Maybe you are looking for

  • Ipad will no longer play any videos on any site. Switch to photon and it plays. These are not flash vids!

    On safari I have a long list of blogs of news that I have viewed daily for years. About to weeks ago. These apparent you tube embeds now say this video cannot be viewed.   Since these sites haven't changed all at once I downloaded photon and they all

  • HP LaserJet 1000 not working with Airport Express

    Has anyone had any luck getting a HP LaserJet 1000 working with Airport Express on Windows XP or even OS X for that matter? I can connect to the Airport Express device using iTunes, but its pretty muched gimped for printing. uggh. MacBook Pro   Mac O

  • [SOLVED] Can't get Echinus to work

    Hello Arch users, I have installed Echinus from the AUR, but it doesn't seem to be working, or else I'm not doing something correctly.  I am using the example config from the Echinus Arch wiki.  I am using Slim as the display manager, and have config

  • ITunes language on iPhone 4s

    Hello, I'm Italian but, for work reasons, I usually live in Russian Federation, i have an iPhone 4s and even if all settings report italian as default language i have iTunes on iPhone in Russian... This is very annoying because everyday I spend a lot

  • Count in Query

    I have the following record in InCube: Vendor...PO Date.....Counter AAA......18/10/2011..1 BBB......18/10/2011..1 AAA......19/10/2011..1 AAA......20/10/2011..1 AAA......21/10/2011..1 BBB......21/10/2011..1 In my query, I want to count number unique P