Multipathing enable: migrate from netapp to hds

Our Solaris 9 boxes have been using Netapp San (Brocade switch)
We have purchased HDS SAN systems (the Virtual Storage Platform)
which also use Brocade 5100 switch
Have tested on test server by switching from the old Netapp San to the new HDS San
and see the disks okay.
But the old SAN have been using mpxio for the multipathing
So could someone outline the procedure to enable multipathing for the
new HDS San
(Presumably the mpxio won;t work and have to be disabled first)
BTW, the HBA cards are Emulex L10000
Thanks in advance

Thanks Nik
Yes, the file ssd.conf seems like what you said,
only define some netapp specifics & timeouts
(I'll list them at the bottom). I think I will simply
revert to the original ssd.conf (which I have a backup) before configuring it for HDS
Now from my understanding, the following steps
are required to enable mpxio to use the HDS San:
(Pls correct me if there're some steps not
required. I've just found this from someone's blog, but it is for solaris10)
1. Add a line to kernel/drv/fp.conf
mpxio-disable="no"
2. Edit the scsi_vhci.conf
to remove the Netapp related entries and
add those which are HDS specific (as you told
me in your first response)
3. Do the command
stmsboot -e
(Note since my solaris is 9 , so I cannot
do the better syntax stmsboot -D fp -e But my box
is a sunfire 245 whose internal disk should not
be fibre and so won't matter ??? True ?)
which will reboot for me and enable mpxio
4. Do another boot by
touch /reconfigure
init 6
I have questions about 1 & 3 which I hope
you could clarify for me
For 1, I notice that in my present working
multipathing with the Netapp san , I do
not even have the line mpxio-disable="no".
, yet it is still working
For 3, the manual says stmsboot -e
does 3 things
(a) add the line
mpxio-disable="no" to fp.conf
(b) make things right for the vfstab file
(c) enable mpxio
If so, then is step 1 really redundant ?
BTW, I know mpxio works only of my Solaris
has certain patches installed. How do I find out
which the patch pre-requisites?
P.S. Here are the added lines in my ssd.conf
(for Netapp)
throttle_max=64
not_ready_retries=30
busy_retries=30
reset_retries=30
throttle_min=8
#`ssd-config-list="NETAPP LUN", "netapp-ssd-config";
#netapp-ssd-config=1,0x9007,64,300,30,0,0,0,0,0,0,0,0,0,30,0,0,8,0,0;
#ssd-config-list=64, "not_ready_retries";
#=;
ssd-config-list="NETAPP LUN", "netapp-ssd-config";
netapp-ssd-config=1,0x9007,64,300,30,0,0,0,0,0,0,0,0,0,30,0,0,8,0,0;

Similar Messages

  • I migrated all data from old PC to new and iTunes successfully transferred to new PC. When I click on a song it says it cannot locate the original. How does one map iTunes file to iMusic to enable play from original files ?

    Subject: Migration
    I migrated all data from old PC to new and iTunes successfully transferred to new PC. When I click on a song it says it cannot locate the original. How does one map iTunes file to iMusic to enable play from original files ?

    Subject: Migration
    I migrated all data from old PC to new and iTunes successfully transferred to new PC. When I click on a song it says it cannot locate the original. How does one map iTunes file to iMusic to enable play from original files ?

  • ActiveSync stops working after migrating from Exchange 2007 to Exchange 2013

    We have started the migration from Exchange 2007 to Exchange 2013. We've followed best practices and everything is working great except ActiveSync. I've performed Exchange migrations in the past so this is nothing new for me. I've also been referring to
    a great guide which has been a big help,
    http://www.msexchange.org/articles-tutorials/exchange-server-2013/migration-deployment/planning-and-migrating-small-organization-exchange-2007-2013-part1.html.
    Once a user is migrated from Exchange 2007 to 2013, ActiveSync stops working properly. Email can be pulled to the device (Nokia Lumia 625 running Windows Phone 8) by performing a manual sync. But DirectPush is not working. The strange part is it's not affecting
    everyone who's been migrated. Anyone who is still on Exchange 2007 is not affected.
    At first I thought it was our wildcard certificate. 99% of our users are running Outlook 2013 on Windows 7 or higher but we do have a few terminal servers still running Outlook 2010. Outlook 2010 was giving us certificate errors. I realized it was the wildcard
    certificate. Rather than making changes to the OutlookProvider I simply obtained a new SAN certificate. Although that resolved the issues for Outlook 2010 users, ActiveSync was still a problem.
    Rebooting the phones and removing the email account from the user's device and re-adding it didn't resolve the issue either.
    Then I performed an iisreset on the CAS server. This didn't help either. I didn't know it at the time, but I was getting closer...
    I tried using the cmdlet Test-ActiveSyncConnectivity but it gave me the following error:
    WARNING: Test user 'extest_0d9a45b025374' isn't accessible, so this cmdlet won't be able to test Client Access server
    connectivity.
    Could not find or sign in with user DOMAIN.com\extest_0d9a45b025374. If this task is being run without
    credentials, sign in as a Domain Administrator, and then run Scripts\new-TestCasConnectivityUser.ps1 to verify that
    the user exists on Mailbox server EX02.DOMAIN.COM
    I started reviewing how Exchange 2013 proxied information from the CAS to the mailbox server and realized the issue may in fact be on the mailbox server.
    I performed an iisreset on the mailbox server and all of a sudden ActiveSync started working again. Awesome!
    I can't explain why. The only thing I can assume is when some users were migrated from 2007 to 2013 something wasn't being triggered on the Exchange 2013 side. Resetting IIS resolved the issue. I guess I'll have to do an IIS reset after I perform a batch
    of migrations. Disabling ActiveSync and re-enabling it for the affected users didn't help - only the IISRESET resolved the issue.
    If anyone has any information as to why this happens, please chime in. Also, if anyone knows why I can't run the Test-ActiveSyncConnectivity cmdlet, I'd appreciate the help.
    Thanks.

    Hi,
    In Exchange 2013, the Public Folder is changed to Public Folder mailbox instead of Public Folder in Exchange 2007 database.
    Due to the changes in how public folders are stored, legacy Exchange mailboxes are unable to access the public folder hierarchy on Exchange 2013 servers. However, user mailboxes on Exchange 2013 servers or Exchange Online can connect to legacy
    public folders. Exchange 2013 public folders and legacy public folders can’t exist in your Exchange organization simultaneously. This effectively means that
    there’s no coexistence between versions.
    For this reason, it’s recommended that prior to migrating your public folders, you should
    first migrate your all legacy mailboxes to Exchange 2013. For more information about migrating public folder from previous versions, please refer to:
    http://technet.microsoft.com/en-us/library/jj150486(v=exchg.150).aspx
    (Please note the What do you need to know before you begin part in this link)
    Regards,
    Winnie Liang
    TechNet Community Support

  • Migrating from SBS 2011 to Windows Server 2012R2 Standard with 365

    Im looking to move SBS 2011 to Server 2012R2 standard and with a 365 migration as well.
    Im thinking
    BPA for AD (health check)
    Add the Server2012R2 as a DC
    Ensure replication
    Move DHCP
    Change DNS
    Now this is where I get lost, now Im lost to use either dirsync or add essentials service, I have gone for standard and 30 cals with the option  as I have 28 users and the essentials service would make everything quite nice for remote access and backups
    etc....  however I don't know whether to just use dirsync with password replication instead ?
    Currently Im aiming towards getting the dc up and running then enabling essentials on the dc but I dont know how it will effect SBS and having essentials on the same network and sync'ing people to the cloud ? So Should I use DirSync on a different server
    instead to sync up to azure ?
    Finally with 365 I have only ever done a migration involving an export of pst's from an exchange box and then on a new domain with new pc's and reimport the pst's on the user accounts in outlook.
    However I have read when migrating to 365 was using exchange 2010 there was a application
    to move everything across however you needed an ssl cert iirc ? so I dont know how it works now since the nice new dashboard upgrade (assuming the tool has been updated like dirsync)
    along with the above how does AD now know where exahcnge if you use the migration app with an ssl cert and is it a case of still running the 365 config tool to configure each user to look at their mail in the
    cloud ? Also is it a good idea to enable essentials from the start of the 2012r2 dc for replication top 365 or to use dirsync ?
    then finally when replications is all good start transferring the FSMO roles over to the new DC and demote and remove the SBS server
    I have seen (below) but i dont think it covers this scenario and I cant be the only person in this situation =\ ?
    http://blogs.technet.com/b/infratalks/archive/2012/09/07/transition-from-small-business-server-to-standard-windows-server.aspx
    http://social.technet.microsoft.com/Forums/en-US/f552ef12-07a9-4f7a-bf5e-24500c3e1dc3/migrate-sbs-2011-to-server-2012-standard?forum=smallbusinessserver
    http://social.technet.microsoft.com/Forums/en-US/e1d4b09f-8857-4ef6-9a80-6a906e76b688/how-to-migrate-sbs-2011-to-standard-2012-server?forum=smallbusinessserver

    Hi,
    Would you please let us know current situation of this issue? If any update, please feel free to let us know.
    à
    Currently Im aiming towards getting the dc up and running then enabling essentials on the dc but I dont know how it will effect SBS and having essentials on the same network
    and sync'ing people to the cloud?
    I’m a little confused with the description. Did you mean that set up DC (Windows Server 2012 Standard) and
    then install the Windows Server Essentials Experience role? If so, when you migrate, please refer to the article:
    Migrate from Previous Versions to Windows Server 2012 R2 Essentials or Windows Server Essentials Experience
    Regarding to migrate Exchange to Office 365, please refer to following articles.
    Migrate
    All Mailboxes to the Cloud with a Cutover Exchange Migration
    Step-By-Step:
    Migrating from Exchange 2007 to Office 365
    Please Note: Since the web site is not hosted by Microsoft, the link may change without notice. Microsoft
    does not guarantee the accuracy of this information.
    Meanwhile, there is an additional article that you can refer to:
    Manage Office 365 in Windows Server Essentials. It may help us to understand Office 365 in Server Essentials clearly.
    If anything I misunderstand, please don’t hesitate to let me know.
    Hope this helps.
    Best regards,
    Justin Gu

  • IMac CPU3.06 GHz Intel Core 2 Duo, 8GB, Extremely slow and repeatedly "out of Memory" since migration from Mountain Lion to Mavericks

    Dear Apple Support Community.
    As mentioned by a number of users reporting the same problem, my 2009 iMac CPU3.06 GHz Intel Core 2 Duo, 8GB RAM, has become extremely slow and repeatedly "out of Memory" since it was migrated from Mountain Lion to Mavericks (OS 10.9 then OS 10.9.1).
    I have already run CleanMyMac, deleted 5GB of useless stuffs, which did not translate into any improvement. Following the advices collected in other Apple Support Community threads, I have applied EtreCheck the outcome of which is provided below.
    For information, I have also upgraded my 2011 MacBook CPU2.3 GHz Core i7, 8GB which became only a bit slower than before, sometimes freezing which it almost never did before. However, I am used to the fact that Mac get usually slower and more prompt fo freeze after upgrading to a more recent OS.
    Both my 2009 iMac and my 2011 MacBook have F-Secure installed. So the problem with the 2009 IMac is unlikly to be due to the presence of F-Secure. Interestingly, the few abnormalities found by Etrecheck were exacly identical before I ran CleanMyMac. There are two more findings (mentioned below) since I ran CleanMyMac on the 2009 iMac, which again didn't improve the problem.
    So please find below the details of the EtreCheck experience and some additional measurements made with iStat while running EtreCheck:
    Hardware Information:
    iMac (24-inch, Early 2009)
    iMac - model: iMac9,1
    1 3.06 GHz Intel Core 2 Duo CPU: 2 cores
    8 GB RAM
    Video Information:
    ATI Radeon HD 4850 - VRAM: 512 MB
    Audio Plug-ins:
    BluetoothAudioPlugIn: Version: 1.0 - SDK 10.9
    AirPlay: Version: 1.9 - SDK 10.9
    AppleAVBAudio: Version: 2.0.0 - SDK 10.9
    iSightAudio: Version: 7.7.3 - SDK 10.9
    System Software:
    OS X 10.9.1 (13B42) - Uptime: 0 days 2:40:57
    Disk Information:
    WDC WD1001FALS-40K1B0 disk0 : (1 TB)
    EFI (disk0s1) <not mounted>: 209.7 MB
    XooS5 HD (disk0s2) /: 999.35 GB (189.85 GB free)
    Recovery HD (disk0s3) <not mounted>: 650 MB
    OPTIARC DVD RW AD-5670S 
    USB Information:
    Apple Inc. Built-in iSight
    Canon CanoScan
    Apple Computer, Inc. IR Receiver
    Apple Inc. BRCM2046 Hub
    Apple Inc. Bluetooth USB Host Controller
    FireWire Information:
    WD My Passport 071D 800mbit - 800mbit max
    EFI (disk1s1) <not mounted>: 209.7 MB
    disk1s2 (disk1s2) <not mounted>: 2 TB
    Boot OS X (disk1s3) <not mounted>: 134.2 MB
    Thunderbolt Information:
    Kernel Extensions:
    com.f-secure.kext.fsauth          (1.0.0d1)
    Problem System Launch Daemons:
    [failed] com.apple.wdhelper.plist [REMARK: this abnormality appeared only after running CleanMyMac. It does not appear on my 2011 MacBook]
    Problem System Launch Agents:
    Launch Daemons:
    [loaded] com.adobe.fpsaud.plist
    [invalid] com.bjango.istatlocaldaemon.plist [REMARK: this abnormality pre-existed to running CleanMyMac. It appears also on my 2011 MacBook]
    [loaded] com.f-secure.fsavd.dbhelper.plist
    [loaded] com.f-secure.fsavd.plist
    [loaded] com.f-secure.fsmac.firewall.plist
    [loaded] com.f-secure.fsmac.fsupdated_guts2.plist
    [loaded] com.f-secure.fsmac.licensetool.plist
    [loaded] com.macpaw.CleanMyMac2.Agent.plist
    Launch Agents:
    [loaded] com.babylon.activation.plist
    [loaded] com.bjango.istatlocal.plist
    [loaded] com.f-secure.relauncher.plist
    [loaded] com.f-secure.trasher.plist
    User Launch Agents:
    [loaded] com.adobe.ARM.[...].plist
    [loaded] com.google.keystone.agent.plist
    [loaded] com.macpaw.CleanMyMac2Helper.diskSpaceWatcher.plist
    [loaded] com.macpaw.CleanMyMac2Helper.scheduledScan.plist
    [loaded] com.macpaw.CleanMyMac2Helper.trashWatcher.plist
    User Login Items:
    iTunesHelper
    Dropbox
    F-Secure Anti-Virus for MAC
    3rd Party Preference Panes:
    Flash Player
    Internet Plug-ins::
    FlashPlayer-10.6: Version: 11.9.900.170 - SDK 10.6
    Default Browser: Version: 537 - SDK 10.9
    AdobePDFViewerNPAPI: Version: 11.0.04 - SDK 10.6
    AdobePDFViewer: Version: 11.0.04 - SDK 10.6
    DivXBrowserPlugin: Version: 2.1
    Flash Player: Version: 11.9.900.170 - SDK 10.6
    OfficeLiveBrowserPlugin: Version: 12.2.8
    Google Earth Web Plug-in: Version: 6.1
    QuickTime Plugin: Version: 7.7.3
    iPhotoPhotocast: Version: 7.0
    User Internet Plug-ins::
    Picasa: Version: 1.0
    Bad Fonts:
    None
    Old applications:
    None
    Time Machine:
    Skip System Files: NO
    Mobile backups: OFF
    Auto backup: YES
    Volumes being backed up:
    XooS5 HD: Disk size: 930.71 GB Disk used: 753.90 GB
    Destinations:
    XooS5 TM [Local] (Last used)
    Total size: 2 
    Total number of backups: 32
    Oldest backup: 2013-09-15 01:11:39 +0000
    Last backup: 2013-12-07 12:48:45 +0000
    Size of backup disk: Too small. Backup size 2  < (Disk used 753.90 GB X 3)
    Time Machine details may not be accurate. All volumes being backed up may not be listed.
    REMARK: The above finding is unlikely to be accurate. In fact, my Time Machine is supported by a 2TB WD My Passport Studio external HD connected with a Firewire 800. As mentionned, only 754GB of the 2000GB are used.
    Top Processes by CPU:
    5%          backupd
    5%          Mail
    5%          WindowServer
    3%          Finder
    3%          Dropbox
    Top Processes by Memory:
    1.60 GB          mds_stores REMARK: this item was marked in red in the outcome of EtreCheck, so I assume there is something wrong there
    82 MB          Finder
    74 MB          Mail
    49 MB          softwareupdated
    49 MB          Dropbox
    Virtual Memory Statistics:
    97 MB          Free RAM
    1.21 GB          Active RAM
    1.19 GB          Inactive RAM
    1.21 GB          Wired RAM
    1.83 GB          Page-ins
    36 MB          Page-outs
    Please find now some additional measurements collected with iStat while running EtreCheck:
    Practically:
    The iMac can work slowly with only one software e.g. either Mail or Safari or Firefox, but still the below "out of Memory" message occurs about every 3 minutes.
    Aperture has becomre to slow to be useable
    If trying to write a mail (with Mail) while Safari or Firefox is open, then the "out of Memory" message occurs about every 20 seconds so I need to check on "Resume" several times per minute.
    Conclusions
    From the above, unless anyone in the Apple Support Community proves to be able to identify a specific problem leading to a way to find a cure, I tend to conclude that Mavericks is not suitable for computers operating with a Core 2 Duo CPU even with a 8GB of RAM.
    Unless I missed something, I did not see any warning from Apple Computer Inc. informing their customers they should refrain from upgrading Core 2 Duo Computers.
    The only solution I see so far is to downgrade from Mavericks to Mountain Lion. And the Apple Support Community would confirm that it is the only solution, I would appreciate to get your guidance on how to proceed.
    Intended action plan
    Migrate the above 2009 iMac to a 2014 Mac Mini Server operating on Mavericks to ensure that no data may be lost while downgrading to Montain Lion.
    Provided the above downgrading operation is successful, assign this 2009 iMac to partial retirement e.g. Mail, Skype, iTunes  and DVD player.
    In advance, I would like to thank the Apple Community for helping me either i) finding a cure to the problem, ii) or confirming that Mavericks cannot operate properly on Core 2 Duo computers, iii) and in such a case, guiding me on how to downgrade to Mountain Lion.
    Yours sincerely
    Core

    ... what anti-virus would you recommend using?
    Use what is already included with OS X. It already includes everything it needs to protect itself from viruses and malware. Keep it that way with software updates from Apple.
    A much better question is "how should I protect my Mac":
    Never install any product that claims to "speed up", "clean up", "optimize", or "accelerate" your Mac. Without exception, they will do the opposite.
    Never install pirated or "cracked" software, software obtained from dubious websites, or other questionable sources. Illegally obtained software is almost certain to contain malware.
    Don’t supply your password in response to a popup window requesting it, unless you know what it is and the reason your credentials are required.
    Don’t open email attachments from email addresses that you do not recognize, or click links contained in an email:
    Most of these are scams that direct you to fraudulent sites that attempt to convince you to disclose personal information.
    Such "phishing" attempts are the 21st century equivalent of a social exploit that has existed since the dawn of civilization. Don’t fall for it.
    Apple will never ask you to reveal personal information in an email. If you receive an unexpected email from Apple saying your account will be closed unless you take immediate action, just ignore it. If your iTunes or App Store account becomes disabled for valid reasons, you will know when you try to buy something or log in to this support site, and are unable to.
    Don’t install browser extensions unless you understand their purpose. Go to the Safari menu > Preferences > Extensions. If you see any extensions that you do not recognize or understand, simply click the Uninstall button and they will be gone.
    Don’t install Java unless you are certain that you need it:
    Java, a non-Apple product, is a potential vector for malware. If you are required to use Java, be mindful of that possibility.
    Disable Java in Safari > Preferences > Security.
    Despite its name JavaScript is unrelated to Java. No malware can infect your Mac through JavaScript. It’s OK to leave it enabled.
    Block browser popups: Safari menu > Preferences > Security > and check "Block popup windows":
    Popup windows are useful and required for some websites, but popups have devolved to become a common means to deliver targeted advertising that you probably do not want.
    Popups themselves cannot infect your Mac, but many contain resource-hungry code that will slow down Internet browsing.
    If you ever see a popup indicating it detected registry errors, that your Mac is infected with some ick, or that you won some prize, it is 100% fraudulent. Ignore it.
    Ignore hyperventilating popular media outlets that thrive by promoting fear and discord with entertainment products arrogantly presented as "news". Learn what real threats actually exist and how to arm yourself against them:
    The most serious threat to your data security is phishing. To date, most of these attempts have been pathetic and are easily recognized, but that is likely to change in the future as criminals become more clever.
    OS X viruses do not exist, but intentionally malicious or poorly written code, created by either nefarious or inept individuals, is nothing new.
    Never install something without first knowing what it is, what it does, how it works, and how to get rid of it when you don’t want it any more.
    If you elect to use "anti-virus" software, familiarize yourself with its limitations and potential to cause adverse effects, and apply the principle immediately preceding this one.
    Most such utilities will only slow down and destabilize your Mac while they look for viruses that do not exist, conveying no benefit whatsoever - other than to make you "feel good" about security, when you should actually be exercising sound judgment, derived from accurate knowledge, based on verifiable facts.
    Do install updates from Apple as they become available. No one knows more about Macs and how to protect them than the company that builds them.
    Summary: Use common sense and caution when you use your Mac, just like you would in any social context. There is no product, utility, or magic talisman that can protect you from all the evils of mankind.

  • JMS issues when migration from weblogic 9.2 to 10.3.5

    We are facing some issues when migration from weblogic 9.2 to 10.3.5
    In  weblogic 9.2 :_
    BMP Entity EJBs used in our project are read-only in nature using entity cache, below is the configuration details
    <!DOCTYPE weblogic-ejb-jar PUBLIC "-//BEA Systems, Inc.//DTD WebLogic 6.0.0 EJB//EN" "http://www.bea.com/servers/wls600/dtd/weblogic-ejb-jar.dtd">
    <weblogic-ejb-jar>
    <weblogic-enterprise-bean>
    <ejb-name>
    Company
    </ejb-name>
    <entity-descriptor>
    <pool>
    <max-beans-in-free-pool>300</max-beans-in-free-pool>
    <initial-beans-in-free-pool>150</initial-beans-in-free-pool>
    </pool>
    <entity-cache>
    <max-beans-in-cache>3500</max-beans-in-cache>
    <idle-timeout-seconds>100000</idle-timeout-seconds>
    <read-timeout-seconds>0</read-timeout-seconds>
    <concurrency-strategy>ReadOnly</concurrency-strategy>
    </entity-cache>
    Entity beans will get refreshed using the JMS messges. with in the MDB descriptor files(weblogic-ejb-jar.xml) we are using the provider URL directly and XA enabled connection factory is set to false.
    migration to Weblogic 10.3.5_
    With the same configurations MDB are not not getting deployed in weblogic 10 with some exception, so we removed the provider URL from weblogic-ejb-jar.xml and changed the JMS configuration to use foreign JMS and XA enable connection factory is set to true. Now when ever the JMS message is triggered Entity bean is not getting refreshed with the updated values. i.e values are stale.
    Can some one look into this and provide your inputs to resolve this issue.

    I think the Entity bean refresh problem appears to be unrelated to MDBs. The MDB is only responsible for getting the message to your application (which in turn interacts with Entity beans). You might want to try posting your question to an EJB newsgroup.
    Tom

  • How do a migrate from a Wallstreet?

    I'd love to join the ranks of those delighted with their migration to a new OSX machine. In my case, it is proving a little more complex. My old PowerBook G3 Pismo (circa 1999) died suddenly (possibly from the loss of a heat-sensitive fuse, I'm told), giving me the excuse to replace it at last with a very fast, not-quite-state-of-the-art PowerBook G4 (circa 2005).
    In order to rescue my beloved files, apps, and settings, I needed to reincarnate the 2.5" internal drive - a Travelstar 2.5" in some device that would switch on. The first thing I tried was a proprietary, USB "enclosure" from Maplins. They didn't have firewire enclosures, but a USB one claimed to be OSX compatible. Big mistake. This just sat and clicked at a fast, regular, 3 clicks per second - which bofs told me was a sure sign of "mechanical failure" (I knew they were wrong - but could not prove it, and was nearly talked into paying £75 for diagnosis plus a further £500 for "recovery" of my "data" - which sounded like extortion to me).
    Then I succeeded in plumbing it into an even older PowerBook Wallstreet (circa 1997) of my brother's. This was a straightforward operation requiring the removal of the keyboard, the heat-sink, and five screws (three phillips, two star-key - just fiddly enough to please geeks and deter richer "consumers" from attempting it at home). It powered up at once, like a dream - but only on the OS9 partition, since OSX is not retroactively compatible with the Wallstreet PowerBook. Unfortunately, this Wallstreet G3 was too ancient even to have it's own USB ports, let alone FireWire. Someone HAD installed a sort of PC card dongle that enabled both, so long as you booted up OS9 with two proprietary extensions. But there was simply not a hope in **** of getting this prehistoric notebook to emulate a firewire drive in order to perform the simple migration that most mac users rave about.
    Undeterred, I connected the two machines via Ethernet. Here I encountered another problem: my classic OS9 could log on to my new G4 - but my new G4 refused to connect to it, claiming it used "an incompatible version of the AFP protocol." Any way round this? What is AFP, anyway?
    OK. No problem. I got the old Wallstreet effectively to "push" all of its files, over the Ethernet, onto my new machine, placing them into a folder called Eric's HD. This is a slow process. As I write, there are still 8 hours to go. Nevertheless, I will have a copy, of sorts, of my complete OSX and Classic systems on my new computer.
    Now. Here's the question. How do I get my brand new OSX Tiger to "Migrate" these old files, apps and preferences, onto my new OS - since they reside neither on "another Mac" (in FireWire "Target Disk" mode), nor on "another volume on this Mac that has OSX installed." They now simply live in a little folder on my new internal hard drive.
    I know I could order a FireWire enclosure from MacAlly (@£39) and hope it behaves better than the two USB2 enclosures I have already tried. But what if, after waiting three days for delivery, I still get those clicks and nothing else? I really had hoped to complete my migation this weekend....
    Anyone know whether Migration will work from a folder on your Drive 0 - or have any other ideas? I imagine it would be be tedious and risky to copy over all my library files by hand....

    You're right, KAppy. I have been spending WAY too much time on this. Had I an enclosure which worked, I would have migrated at the drop of a hat days ago. But I AM attempting to migrate from one OSX to another (Panther to Tiger - aren't they similar enough?), and I didn't want to lose my applications and settings.
    Unfortunately, the only thing I could lay my hands on which would enclose my disk and actually drive it, was an intermediary Wallstreet (borrowed from my brother). This older computer wouldn't boot my OSX system - but would start up from the old Classic folder.
    You're also right that my mistake was trying to copy the hard drive by drag-and-drop. I thought that files invisible to OSX would still be visible to OS9, but I've now found a load of invisible files and folders were not copied. Had my G4 been able to talk to the Wallstreet, then I could have migrated, or used Carbon Copy, or whatever. But whilst the G3 could talk to the G4, the G4 annoyingly complained that the Wallstreet's 'AFP protocol' was 'incompatible' - what does that mean, anyway? Could I have chosen a different protocol on the OS9? - and refused to connect.
    Anyway, thanks for the patient advice. I'll order a FireWire enclosure, which won't arrive for a few days - but which will then, I trust, actually work with my Travelstar 40Gig (the proprietary USB2 enclosures I tried simply wouldn't).
    Meanwhile, I've now got the G3 to mount the Big Backup Drive via a FireWire device. Do you think OS9's Disk Copy could create an image of my OSX partition - and that this would then be recognised by the Migration Assistant?

  • Problem with foreign and primary keys migration from SQL Server to Oracle

    Hi folks, i'm using SQL Developer to migrate from a SQL Server database to Oracle and i'm stuck with a couple issues:
    The worst of them so far is the fact that i can't migrate any of the PKs and FKs. After successfully capturing the SQL Server DB model and converting it to Oracle, when the tool generates the scripts, all ALTER TABLE queries that add the PKs and FKs have their target columns duplicated.
    for example: when i'm trying to migrate a simple table that contains an Id (PK) and Name columns, the tool generates the following scripts:
    PROMPT Creating Table TestTable...
    CREATE TABLE TestTable (
    Id NUMBER(10,0) NOT NULL,
    Name VARCHAR2 NOT NULL
    PROMPT Creating Primary Key Constraint PK_TestTable on table TestTable ...
    ALTER TABLE TestTable
    ADD CONSTRAINT PK_TestTable PRIMARY KEY
    Id,
    Id
    ENABLE
    As for the FKs, the tool duplicates the columns as well:
    ALTER TABLE SomeTable
    ADD CONSTRAINT FK_SomeTable_SomeTable2 FOREIGN KEY
    SomeTable2Id,
    SomeTable2Id
    REFERENCES SomeTable2
    Id,
    Id
    ENABLE
    Does anyone have a clue on how to solve these issues? I'd be greatly thankful for any answers!

    Hi Fernando,
    I was unable to replicate this issue. My primary / foreign keys where defined using unique columns.
    PROMPT Creating Primary Key Constraint PK_Suppliers on table Suppliers ...
    ALTER TABLE Suppliers
    ADD CONSTRAINT PK_Suppliers PRIMARY KEY
    SupplierID
    ENABLE
    I tried a few things like
    capturing twice and renaming both models the same
    renaming the converted models
    but with no luck.
    I think this issue is occuring either at the capture or convert phase.
    1) Are you performing the capture online or offline?
    2) Can you provide a the entire DDL for one of these tables and its indexes to see if I can replicate?
    3) Did the capture or convert fail or have to be redone at any stage ?
    I all else fails I would attempt a capture and convert again using a brand new repository (create a new schema in Oracle and associate the migration repository with it).
    Regards,
    Dermot
    SQL Developer Team
    Edited by: Dermot ONeill on Oct 22, 2009 12:18 PM

  • Very difficult migration from MS Access to Oracle with SQL Developer

    Dear Developers
    I have tried to used for one Project Oracle Sql Developer Migration from Access to Oracle (Application Express). Was about an evaluation if Oracle Express (with Application Express) or Mysql (with PHP) as target database.
    I am working with Oracle for 14 years and i earn my money in projects related with Oracle. Of course i am sentimental and i push every time when an opportunity comes to have as an option Oracle.
    This time beside my sentimental approach the decision was different.
    I will tell you why and i hope that someone there will listen and make something for a better Product.
    I was trying yesterday the whole day to migrate an Microsoft Access database. A nightmare....
    1. Microsoft Access Exporter 2000 was too slow. For a 25 MB Access database, running time was 15 Minutes. And 15 minutes to obtain an undocumented error. I followed your instruction and eliminate all references etc... as you described in HELP. a lot of time necessary to read and do..... After that i obtained an Error #..16. Sorry that i don't provide you with an exact number error. i don't have it right know and personally i am not interested anymore. For this error i don't find nothing in forums about it.... Search again and again..... Every time after waiting 15 Minutes in which you don't know exactly what application does or if really does something. I decided to delete sequentially macros, modules, forms, querys from Access database and see where Exporter stops... Time and time lost for trying and searching.... Finally i obtained my XML file only by keeping my Access tables. The file was produced even if i obtained at the end Error #5:..... Because was an error i am not really sure if the file was completly consistent with my database.
    2. I followed the next steps and with some errors (enabling, disabling constraints, objects couldn't be created because of different reasons) i had my Model and Oracle Schema. I have tried to export the data from Access and import it in Oracle. data exported was in UTF8 and SQL Loader Control file doesn't have
    LOAD DATA
    character set utf8
    So i tried to add the string "character set utf8" in my dozens of CTL files - with a freeware of course.....
    Maybe you can say that if i knew that from the beginning i would have set my character set in SQL Developer. I don't know and i will not ever try with this product!
    So finally after hours of working and trying and playing with my nerves, reading a lot of forums and documentation, i had an Oracle schema, maybe incomplete because of different errors at so much migration stages.
    And now i have tried for the first time in my life MySQL Migration Toolkit. I read that for migration to MYSQL this tool is what i needed.
    I recommend you to try the product and see something about the quality required for a piece of software. Is not enough if your product runs OK with Nortwind DEMO Access database. It should run OK with all ACCESS databases.
    With the tool i have done the job in 5 MINUTES. Simple, efficiently and without errors. At the end i had the SQL Scripts and the DATA imported in the database.
    Difficult to say something now. I am sentimental but i asked mysef if you can do something to improve the quality of the product.
    What arguments can we provide for a client in Migration from Access to Oracle ???? Can you honestly answer at that?
    Please try this MYSQL Migration Toolkit.....Maybe this give you some ideas about simple tasks and about what we really need...
    Best regards
    Adrian

    Dear Barry and Kgronau
    Thank you very much for your response. My Post was after a long day of trying to import an MDB file.
    Maybe you can understand why i was so angry...
    Yes i understand that your tool try to do more than other even if actually you don't really convert too much.
    From Modules to make some procedures where the whole code is commented! .....
    For other MDB which i tried to import in APEX the forms where really not usable. Queries are full of errors requires most of the time manual corrections.
    I am sorry that i can not provide you with this MDB which cause these problems.
    Data and software doesn't belong to me and the company which own it can not provide to you.
    What i hope from your product is, even when errors come, to show a better error explanation. At what object in Access error ocures ? This is very important because i can correct my MDB or just delete the module,form,query in cause if SQL Developer doesn't like it....
    Just showing an error Error #.... doesn't help me too much. And if an error occurs i don't know if program continues to work or is dead. No information about this. Maybe you can show that your application is alive and does something (working status etc...) ! Anyway we wait long minutes in front of dead screen....
    Again, thank you very much for your attention
    Adrian

  • Can't migrate from 10.8 to 10.9 with Time Machine

    My old 2009 MacBookPro stopped working and I replaced it with the new one and hoped to be able to use the Migration Assistant to recover my backed up data I had in Time Machine. So far it jsut keeps "Looking for other sources" when I choose to migrate from Time Machine. I was told I should be able to do this going to the new system and I've done it with out any issues in the past. Is there anything I can do to get this working?

    I'm having the same issue with trying to restore a Time Machine backup of a Mavericks machine saved to my Airport Time Capsule to a newly re-installed Macbook Pro running Mavericks. Anyone have any other suggestions?
    The Migration Assistant keeps getting stuck on Step 5 of the section "Time Machine And other disk migration" in these instructions http://support.apple.com/kb/HT5872.
    The computer names have been set to the same names. My Airport Time Capsule firmware version is 7.7.2.
    Solution: I had to click on the Time Capsule icon to enable the "Continue" button. The icon doesn't look like a selectable button, unfortunately.
    Message was edited by: alfredbeta

  • Migrating from 3.1 to 3.7 - write through for a custom cache store issues

    We're migrating from 3.1 to 3.7. So far the migration and testing has been fairly uneventful, but there is one issue that came up yesterday that seems like it is going to be tricky to debug.
    We have a set of storage-enabled nodes that use a custom CacheStore to read from and write behind to a mongo database. On another node connected to that caching service, read throughs work just fine. (I can set breakpoints on the CacheStore load method and see the load calls coming through just fine) - but what's not working is when the other node does a Cache.put - the Store method on the CacheStore is never called and so far I don't see anything in the logs indicating there is a problem on either side (I'm going to make sure that the coherence logging is up to the highest level on both the nodes today when I'm doing more testing)
    I can see the cache put start to dive into the coherence jar, but I don't have source jars for coherence so it's fairly opague what might be going wrong after the Cache.put(object, object) call. I can see that it dives into various coherence methods, but
    Any ideas on where to start debugging this?
    This setup worked fine on 3.1, and as best we can tell all the API calls were converted over to their proper coherence 3.7 versions, and the coherence.xml files were migrated to use the new xsd etc.

    it seems that the issue might be related to this:
    2012-08-15 14:19:34.086 Tangosol Coherence 3.7.1.5 <Error> (thread=WriteBehindThread:CacheStoreWrapper(com.foo.cache.MongoCacheStore):Foo.com-CMS, member=13): Failed to store key="assetId=DEFAULT;assetStyle=DEFAULT;initial=c;siteId=foosite;"
    2012-08-15 14:19:34.087 Tangosol Coherence 3.7.1.5 <Error> (thread=WriteBehindThread:CacheStoreWrapper(com.foo.configrepo.cache.MongoCacheStore):Foo.com-CMS, member=13): (Wrapped) java.io.StreamCorruptedException: invalid type: 13
         at com.tangosol.util.ExternalizableHelper.fromBinary(ExternalizableHelper.java:266)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$ConverterFromBinary.convert(PartitionedCache.CDB:4)
         at com.tangosol.net.cache.BackingMapBinaryEntry.getValue(BackingMapBinaryEntry.java:124)
         at com.tangosol.net.cache.ReadWriteBackingMap$CacheStoreWrapper.storeInternal(ReadWriteBackingMap.java:5731)
         at com.tangosol.net.cache.ReadWriteBackingMap$StoreWrapper.store(ReadWriteBackingMap.java:4814)
         at com.tangosol.net.cache.ReadWriteBackingMap$WriteThread.run(ReadWriteBackingMap.java:4217)
         at com.tangosol.util.Daemon$DaemonWorker.run(Daemon.java:803)
         at java.lang.Thread.run(Thread.java:662)
    Caused by: java.io.StreamCorruptedException: invalid type: 13
         at com.tangosol.util.ExternalizableHelper.readObjectInternal(ExternalizableHelper.java:2303)
         at com.tangosol.util.ExternalizableHelper.deserializeInternal(ExternalizableHelper.java:2746)
         at com.tangosol.util.ExternalizableHelper.fromBinary(ExternalizableHelper.java:262)
         ... 7 moreLooks like it is an issue with the serialization? We're primarily using XmlBean, not POF for serialization.
    Any tips on troubleshooting this?
    Edited by: RyanGardner on Aug 15, 2012 7:37 AM
    Edited by: RyanGardner on Aug 15, 2012 7:38 AM

  • Hybrid Deployment - Migrating from Exchange Online to On-Premise Exchange

    Hi,
    I'm migrating from Exchange Online to an 2013 On-Premise Exchange. 
    Hybrid Deployment has been set up and configured successfully. 
    However, I'm encountering issues while performing a mailbox migration to the on-premise server. 
    It's has been stuck at validating and reflecting a MigrationTranscientException Error. 
    Appreciate the valuable advise and suggestions! Thanks!
    Cheers!
    Joshua
    Error message below:
    Data migrated: 
    Migration rate: 
    Error: MigrationTransientException:
    The Mailbox Replication Service could not connect to the remote server because the remote server encountered an internal error. The call to ‎'https://mail.domain.com.sg/EWS/mrsproxy.svc‎' failed. Error details: The content type text/html; charset=utf-8 of
    the response message does not match the content type of the binding ‎(application/soap+xml; charset=utf-8)‎. If using a custom encoder, be sure that the IsContentTypeSupported method is implemented properly. The first 1024 bytes of the response were: ‎'<!DOCTYPE
    html> <html> <head> <title>Runtime Error</title> <meta name="viewport" content="width=device-width" /> <style> body {font-family:"Verdana";font-weight:normal;font-size: .7em;color:black;}
    p {font-family:"Verdana";font-weight:normal;color:black;margin-top: -5px} b {font-family:"Verdana";font-weight:bold;color:black;margin-top: -5px} H1 { font-family:"Verdana";font-weight:normal;font-size:18pt;color:red } H2 { font-family:"Verdana";font-weight:normal;font-size:14pt;color:maroon
    } pre {font-family:"Consolas","Lucida Console",Monospace;font-size:11pt;margin:0;padding:0.5em;line-height:14pt} .marker {font-weight: bold; color: black;text-decoration: none;} .version {color: gray;} .error {margin-bottom: 10px;} .expandable
    { text-decoration:underline; font-weight:bold; color:navy; cursor:hand; } @media screen and ‎(max-width: 639px)‎ { pre { width: 440px; o‎'. --> The remote server returned an error: ‎(500)‎ Internal Server Error.. --> The content type text/html; charset=utf-8
    of the response message does not match the content type of the binding ‎(application/soap+xml; charset=utf-8)‎. If using a custom encoder, be sure that the IsContentTypeSupported method is implemented properly. The first 1024 bytes of the response were:
    ‎'<!DOCTYPE html> <html> <head> <title>Runtime Error</title> <meta name="viewport" content="width=device-width" /> <style> body {font-family:"Verdana";font-weight:normal;font-size: .7em;color:black;}
    p {font-family:"Verdana";font-weight:normal;color:black;margin-top: -5px} b {font-family:"Verdana";font-weight:bold;color:black;margin-top: -5px} H1 { font-family:"Verdana";font-weight:normal;font-size:18pt;color:red } H2 { font-family:"Verdana";font-weight:normal;font-size:14pt;color:maroon
    } pre {font-family:"Consolas","Lucida Console",Monospace;font-size:11pt;margin:0;padding:0.5em;line-height:14pt} .marker {font-weight: bold; color: black;text-decoration: none;} .version {color: gray;} .error {margin-bottom: 10px;} .expandable
    { text-decoration:underline; font-weight:bold; color:navy; cursor:hand; } @media screen and ‎(max-width: 639px)‎ { pre { width: 440px; o‎'. --> The remote server returned an error: ‎(500)‎ Internal Server Error. --> The Mailbox Replication Service could
    not connect to the remote server because the remote server encountered an internal error. The call to ‎'https://mail.domain.com.sg/EWS/mrsproxy.svc‎' failed. Error details: The content type text/html; charset=utf-8 of the response message does not match the
    content type of the binding ‎(application/soap+xml; charset=utf-8)‎. If using a custom encoder, be sure that the IsContentTypeSupported method is implemented properly. The first 1024 bytes of the response were: ‎'<!DOCTYPE html> <html> <head>
    <title>Runtime Error</title> <meta name="viewport" content="width=device-width" /> <style> body {font-family:"Verdana";font-weight:normal;font-size: .7em;color:black;} p {font-family:"Verdana";font-weight:normal;color:black;margin-top:
    -5px} b {font-family:"Verdana";font-weight:bold;color:black;margin-top: -5px} H1 { font-family:"Verdana";font-weight:normal;font-size:18pt;color:red } H2 { font-family:"Verdana";font-weight:normal;font-size:14pt;color:maroon }
    pre {font-family:"Consolas","Lucida Console",Monospace;font-size:11pt;margin:0;padding:0.5em;line-height:14pt} .marker {font-weight: bold; color: black;text-decoration: none;} .version {color: gray;} .error {margin-bottom: 10px;} .expandable
    { text-decoration:underline; font-weight:bold; color:navy; cursor:hand; } @media screen and ‎(max-width: 639px)‎ { pre { width: 440px; o‎'. --> The remote server returned an error: ‎(500)‎ Internal Server Error.. --> The content type text/html; charset=utf-8
    of the response message does not match the content type of the binding ‎(application/soap+xml; charset=utf-8)‎. If using a custom encoder, be sure that the IsContentTypeSupported method is implemented properly. The first 1024 bytes of the response were:
    ‎'<!DOCTYPE html> <html> <head> <title>Runtime Error</title> <meta name="viewport" content="width=device-width" /> <style> body {font-family:"Verdana";font-weight:normal;font-size: .7em;color:black;}
    p {font-family:"Verdana";font-weight:normal;color:black;margin-top: -5px} b {font-family:"Verdana";font-weight:bold;color:black;margin-top: -5px} H1 { font-family:"Verdana";font-weight:normal;font-size:18pt;color:red } H2 { font-family:"Verdana";font-weight:normal;font-size:14pt;color:maroon
    } pre {font-family:"Consolas","Lucida Console",Monospace;font-size:11pt;margin:0;padding:0.5em;line-height:14pt} .marker {font-weight: bold; color: black;text-decoration: none;} .version {color: gray;} .error {margin-bottom: 10px;} .expandable
    { text-decoration:underline; font-weight:bold; color:navy; cursor:hand; } @media screen and ‎(max-width: 639px)‎ { pre { width: 440px; o‎'. --> The remote server returned an error: ‎(500)‎ Internal Server Error. --> The call to ‎'https://mail.domain.com.sg/EWS/mrsproxy.svc‎'
    failed. Error details: The content type text/html; charset=utf-8 of the response message does not match the content type of the binding ‎(application/soap+xml; charset=utf-8)‎. If using a custom encoder, be sure that the IsContentTypeSupported method is
    implemented properly. The first 1024 bytes of the response were: ‎'<!DOCTYPE html> <html> <head> <title>Runtime Error</title> <meta name="viewport" content="width=device-width" /> <style> body {font-family:"Verdana";font-weight:normal;font-size:
    .7em;color:black;} p {font-family:"Verdana";font-weight:normal;color:black;margin-top: -5px} b {font-family:"Verdana";font-weight:bold;color:black;margin-top: -5px} H1 { font-family:"Verdana";font-weight:normal;font-size:18pt;color:red
    } H2 { font-family:"Verdana";font-weight:normal;font-size:14pt;color:maroon } pre {font-family:"Consolas","Lucida Console",Monospace;font-size:11pt;margin:0;padding:0.5em;line-height:14pt} .marker {font-weight: bold; color: black;text-decoration:
    none;} .version {color: gray;} .error {margin-bottom: 10px;} .expandable { text-decoration:underline; font-weight:bold; color:navy; cursor:hand; } @media screen and ‎(max-width: 639px)‎ { pre { width: 440px; o‎'. --> The remote server returned an error:
    ‎(500)‎ Internal Server Error.. --> The content type text/html; charset=utf-8 of the response message does not match the content type of the binding ‎(application/soap+xml; charset=utf-8)‎. If using a custom encoder, be sure that the IsContentTypeSupported
    method is implemented properly. The first 1024 bytes of the response were: ‎'<!DOCTYPE html> <html> <head> <title>Runtime Error</title> <meta name="viewport" content="width=device-width" /> <style>
    body {font-family:"Verdana";font-weight:normal;font-size: .7em;color:black;} p {font-family:"Verdana";font-weight:normal;color:black;margin-top: -5px} b {font-family:"Verdana";font-weight:bold;color:black;margin-top: -5px} H1
    { font-family:"Verdana";font-weight:normal;font-size:18pt;color:red } H2 { font-family:"Verdana";font-weight:normal;font-size:14pt;color:maroon } pre {font-family:"Consolas","Lucida Console",Monospace;font-size:11pt;margin:0;padding:0.5em;line-height:14pt}
    .marker {font-weight: bold; color: black;text-decoration: none;} .version {color: gray;} .error {margin-bottom: 10px;} .expandable { text-decoration:underline; font-weight:bold; color:navy; cursor:hand; } @media screen and ‎(max-width: 639px)‎ { pre { width:
    440px; o‎'. --> The remote server returned an error: ‎(500)‎ Internal Server Error

    Hello Joshywahcho ,<o:p></o:p>
    First Check  is the MRS proxy is enabled or not and restart the Microsoft Exchange Mailbox Replication service and
    IIS on Hybrid .<o:p></o:p>
    As now Microsoft has change the procedure to roll back the mailbox from O365, please follow the below mention Step<o:p></o:p>
    Moved the affected user's mailbox to the Hybrid Server from Office 365 <o:p></o:p>
    Connect to Exchange Online using remote PowerShell<o:p></o:p>
    Ran the command
    $Credentials = Get-Credential<o:p></o:p>
    Ran the command
    New-MoveRequest -identity <e-mail address of the affected user> -OutBound -RemoteTargetDatabase "Name of the Datbase" -RemoteHostName <Your Host Name> -RemoteCredential $Credentials -TargetDeliveryDomain <Your Target Delivery Domain><o:p></o:p>
    Ran the command
    Get-MoveRequest -Identity <e-mail address of the user> to check the status of mailbox move<o:p></o:p>
    once the mailbox move from Office 365 to Hybrid Server was completed then initiated a Mailbox move from the Hybrid Server to On-premise server <o:p></o:p>
    Mailbox was moved successfully from Hybrid Server to the On-premise server .<o:p></o:p>
    checked in the On-premise server 2007 and the affected user's mailbox was showing there<o:p></o:p>

  • Migration from PE 8 (Windows) to PE 11 - What steps?? (MAC)

    I am planning to migrate from PE 8 (Windows) to PE 11 (Mac). I will continue to use the PE catalog in PE 11 to organize my library of 16,000+ photos.  Although I had considered moving to Aperture for my Digital Asset Manager (in order to more seamlessly work in an Apple environment), I have instead decided to stay with PE since I am familiar with it and happy with its functionality. 
    I am hoping that this forum can help me: I need to know what steps to follow to perform this migration.  (I am a non-technical user.  All I know thus far is  that I need buy a boxed version or PE11 -  NOT download it from the Apple app site - in order to get the cataloging function.)  I would like the process to enable me to port both my photos AND my hierarchical keyword tagging structure.  
    (Some background on why I am asking this question proactively: I was severely burned in a previous upgrade from PE 5 to PE 8 (Windows XP to  Windows 7) where ALL of my tagging got lost in the move. This, despite hours of work with the Adobe support team who had me write the keywords into metadata and re-backup, re-port, etc, etc.  It was a nightmare of many hours spent.  All unsuccessful.  I ended up manually re-tagging each of my photos.  Needless to say, I do not want to go through that again,)
    Thanks in advance for your help.  Amy

    Elements Downloader doesn't auto run on a mac. It's set up so that you always have to launch it from organizer.
    If you know you won't use iphoto, then it's a good idea to go to the iphoto preferences>General>When a camera is connected, and choose No Application, in which case the card will mount like a drive on your desktop and you can drag the photos from it or launch PSE Downloader, or choose Image Capture, in which case apple's own OS X downloader/scanner utility launches and you can use that, then File>Get Photos>From Files and Folders to bring them into Organizer.
    For moving the catalog, the most important thing is to be sure the catalog is okay in windows first, so do go to manage catalogs>repair, and then optimize before backing up. The transfer is very easy when it works.

  • Migration from SAP BC To SAP P.I

    HI Experts ,
    I am required to do migration from SAP BC to SAP P.I. I would require the following Info/Help from you as i have not worked on such scenario before this!
    1) What kind of challenges can be faced during migration from BC to XI?
    2) What is the best practise for doing the same?
    3) Can you share some scenario from where i can have some idea how to do migration?
    4) Generally how much time it takes in migration activity if ther are not much interfaces involved in it?
    Would appreciate your valuable inputs on the same!
    I have read How to guide docs on how to do migration from BC to XI but not convinced with the content
    Regards
    Saras Jain

    Hi Saras,
    I am giving you an example senario and a how to do all the steps and why it needed. if you need more information just mail me i'll give you the concerned PDF.
    Scenario
    SAP Business Connector enables data exchange between an SAP system and
    external systems using HTTP and e-mail. SAP Exchange Infrastructure (SAP XI)
    replaces SAP Business Connector (SAP BC).
    We distinguish between standard scenarios (that do not contain customer-specific
    development) and individual scenarios (that do contain services or mappings created
    by the customer). The following sections describe how to migrate both types of
    scenarios to SAP XI.
    Scenario 1 (IDoc – HTTP, HTTP – IDoc)
    An SAP system sends an IDoc to SAP BC or to SAP XI. The IDoc is converted to
    XML. It is sent to an external business partner by using HTTP.
    An external business partner sends an XML document by using HTTP to SAP BC or
    to SAP XI. The document is converted to an IDoc and sent to a connected SAP
    system in the internal system landscape.
    Scenario 2 (IDoc – Mail, Mail – IDoc)
    An SAP system sends an IDoc to SAP BC or to SAP XI. The IDoc is converted to
    XML. It is sent to an external business partner as an e-mail.
    An external business partner sends an e-mail to SAP BC or to SAP XI. The
    document is converted to an IDoc and sent to a connected SAP system in the
    internal system landscape.
    Flow Services
    Flow services are services that are written in the proprietary graphical-based flow
    language of SAP BC.
    Java Services
    Java services are services that are written in the Java language.
    There are several platforms that support Java, including SAP XI. However, the data
    model used in the Java services in SAP BC is proprietary.
    Stylesheet Transformations
    Stylesheet transformations are often used for mapping purposes. They take an XML
    document as input, convert it according to rules defined in the stylesheet, which is
    also an XML document, and produce an XML document as output.
    The XML Stylesheet Transformation Language (XSLT) was defined by W3C (World
    Wide Web Consortium), the consortium that defines most of the common standards in modern internet communication, such as HTML and SOAP.
    SAP NetWeaver is the open integration and application platform from SAP. With
    NetWeaver ’04, there is a significant overlap between the functions of SAP
    NetWeaver and SAP BC. Many functions provided by SAP BC are offered by SAP XI
    as part of SAP NetWeaver, for example, conversion of an IDoc to an XML message,
    mapping, routing, and adapters to connect to a large number of third-party back-end
    systems. SAP BC is therefore no longer needed. New or updated SAP solutions are
    based on SAP NetWeaver. Customers looking for a strategic integration solution for
    the complete application landscape should therefore now favor SAP NetWeaver.
    The last version of SAP BC was SAP BC 4.7, which was released in June 2003. SAP
    BC has not been developed further since this release. SAP BC is not released for
    any further platforms (such as Windows 2003) and does not support any further JDK
    versions, including the current version (Version 1.4).
    For information about the supported SAP BC versions, JDKs, and operating systems,
    see SAP Note 309834. For information about the support and maintenance strategy
    for SAP BC, see SAP Note 571530. Customers who still use SAP BC need to be
    aware of the following:
    • A solution based on SAP BC cannot be a strategic, long-term solution.
    • A solution based on SAP BC is based on technology that was implemented in
    2003 and that is no longer further developed.
    • SAP BC is supported on limited platforms and for limited JDKs only. Vendor
    support for some of these platforms and JDK versions will soon be
    discontinued.
    SAP BC and SAP NetWeaver are based on different technologies: SAP BC is based
    on webMethods’ proprietary Integration Server, whereas SAP NetWeaver is based
    on open standards, such as BPEL4WS (business process execution language for
    web services). Therefore, only parts of an SAP BC implementation, such as XSLT
    mappings, can be migrated to a solution based on SAP NetWeaver. Any custom
    programming based on SAP BC, such as SAP BC flow language, is unlikely to be
    reusable in the SAP NetWeaver context.
    Purpose of this Migration Guide
    The purpose of this migration guide is to provide scenario-based descriptions of the
    migration from a company’s system landscape that uses SAP BC to exchange
    documents with business partners, to a system landscape that uses SAP XI instead.
    This migration guide outlines, for certain scenarios, the individual steps a company
    has to take to replace an existing SAP BC implementation with SAP XI.
    The basic steps in the back-end systems are:
    • Defining the logical systems
    • Maintaining the ALE distribution model
    • Maintaining partner profiles and EDI ports
    • Maintaining RFC destinations
    The basic steps in the SAP XI system are:
    • Maintaining the System Landscape Directory
    • Configuring the IDoc adapter
    • Maintaining the receiver determinations, interface determinations, and
    interface agreements
    • Maintaining the communication channels
    The type of scenario to migrate typically has an impact on the communication channels only.
    Scenario 1 IDoc/HTTP
    IDoc – HTTP Using SAP BC
    Within the system landscape of a company, an SAP system sends an IDoc
    asynchronously to SAP BC. SAP BC converts the IDoc into a standard IDoc XML
    document. The fields of the IDoc are not modified, that is, the conversion does not
    involve any mapping of source fields to target fields. After conversion, SAP BC uses
    HTTP to send the XML document synchronously to an external URL of a business
    partner.
    System Landscape
    • An SAP system and SAP BC are installed within the same system landscape
    of a company.
    • The following information is defined in the SAP system to allow data
    exchange between the SAP system and SAP BC:
    o An RFC destination and a logical system for SAP BC
    o The distribution model is maintained, if necessary
    • The following information is defined in SAP BC:
    o An SAP server with an alias is defined. For this SAP server, the
    information that the SAP BC requires to connect to the SAP system is
    specified.
    o A listener is configured to listen to requests from the SAP system.
    o A routing rule with the transport type “XML” is configured for the
    specific sender, receiver, and message type. The routing rule
    determines how a message is to be routed and where it is to be routed
    to. The scenario described does not use an ACL, (Access Control List
    to restrict execution permissions for this routing rule) or pre-processing
    or post-processing SAP BC services.
    4.2 IDoc – HTTP Using SAP XI
    Check the prerequisites described above under Prerequisites for Standard
    Scenarios.
    The IDoc adapter is used to receive an IDoc from an SAP back-end system and
    convert it to XML.
    The adapter used to post this XML to an external destination is the plain HTTP
    adapter.
    To set up the communication channel, perform the following steps:
    1. Locate the routing rule in SAP
    Business Connector.
    In the Administration View, choose
    Adapters – Routing to navigate to
    Routing Rules.
    2. Select the details of the relevant routing rule.
    To display the details, choose the green triangle (Edit).
    Transport is typically set to XML and a URL is specified. You can use this
    URL in your SAP XI settings.In some cases, Transport is set to
    B2B Service and the parameters Server Alias, Folder and Service
    exist. This is the same as when Transport is set to XML and has the
    URL http://<serverhost>:<port>/invoke/<folder>/<service>.
    You can find <serverhost> and<port> in the definition of the
    remote server under Settings –>Remote Server in the Administration
    View.
    3. Set up the communication channel in the Integration Directory using the parameters from the SAP BC routing rule.
    Use the following settings:
    • Adapter Type: HTTP Receiver
    • Transport Protocol: HTTP 1.0
    • Message Protocol: XI Payload in HTTP Body
    • Adapter Engine: Integration Server
    • Addressing Type: URL Address
    • Target Host, Service Number,and Path need to be set in accordance with the SAP BC routing rule. HTTP Proxy Host and HTTP Proxy
    Port can be set if required.All other parameters have no corresponding values in SAP BC and should be set to their default values.
    regards
    Aashish Sinha
    P : Reward points if helpful

  • Migrating from Netware to Linux - faster "sync" ?

    Hey all,
    Planning on doing a TransferID migration from netware 6.5 to linux this weekend and have done some trial data copies.. Main copy is already done, but the "sync" is still taking a large number of hours that I need to try and fit into a smaller window. Any tips on getting sync to go faster? I'm currently using the miggui but wondering if you do the commands manually, leaving things out like displaying progress bar, etc.. if it'll get to it quicker.

    Originally Posted by joebrug
    Hey all,
    Planning on doing a TransferID migration from netware 6.5 to linux this weekend and have done some trial data copies.. Main copy is already done, but the "sync" is still taking a large number of hours that I need to try and fit into a smaller window. Any tips on getting sync to go faster? I'm currently using the miggui but wondering if you do the commands manually, leaving things out like displaying progress bar, etc.. if it'll get to it quicker.
    So, we have 6-year old HP DL-385 (no G, that's old they are) that are slower than you know what.
    We're migrating to HP DL385 G6 or G7, but we're using VMware ESXi, so the target OES11 server is a Guest VM.
    Both servers are on a gigabit switch.
    We average approx. 15-20 GB/hour (it's not the AMOUNT of data, it's how many small files that affects things).
    We're able to get the sync to fit in a small window (even our large ones, maybe 2-4 hours).
    Is that not fast enough (although you may have LOTS more small files that change daily).
    Basically we cart the server down, run the sync during the day until it finishes. Obviously it skips things that were in use. Then we kick everyone off the server at 3:00 and start the sync.
    Since you're on NetWare to OES, I would NOT recommend rsync. The rsync on NetWare is very touchy/finicky and I don't think you can migrate trustees either.
    From OES2 to OES11, you could use rsync, although I'm having mixed results with the trustees at the moment.
    One thing we DO have is a "static" set of read-only data on the server, that we do NOT migrate over. We have a central office server that contains the "master" data and the clients run rsync to pull the data. So we don't even back it up either (other than the main office).
    I can't remember the EXACT numbers we got when we did NetWare to OES2, but I still think we were averaging somewhere around 12-17GB/hour depending on. The key point was to make sure you were gigabit and didn't have a mis-match because someone in your datacom unit set the **** switch to 100-full or something other than auto-auto.
    Oh, and if you have disk quotas, make sure that you either clear them or keep compression enabled. THAT will slow things down because if it tries to copy something, and there's not enough space, it'll retry like X times and then continue on and you find out like 10 hours later it was slow because you didn't have enough directory quota space.

Maybe you are looking for

  • How do i delete an icloud account off my iphone without the password

    for some reason my iphone signed onto an account that has been hacked and since no longer exists! so its almost impossible for me to find out the password, help!

  • Lsmw error - no fields with the same name

    Dear Experts, I am trying to upload master data through LSMW through Standard Batch/Direct Input but when i am executing the step READ DATA it is throwing me error as File 'C:\Documents and Settings\Guest\Desktop\DMS.TXT' and source structure 'ZEQUI_

  • Thumbnails in Web DMS

    hi dears, i am configuring web dms, now my problem is, i am unable to get thumbnails in web dms for jpg, mbp files.If i open the same files in sap dms i am able to get thumbnails, when i open them in web page i am not getting thumbnails. I already de

  • AP store states Pages 5.1 installed, but Pages indicates 4.3. How to correct?

    Ap Store indicates that Pages 5.1 has been installed on my Mac Pro,  but "About Pages" indicates the version is 4.3.  How do I install current version?

  • Speedtester test 2

    Can anyone advise me how to get to test2 please. I can run test 1 which fails during peak times,then I change my Netgear 834 router settings to the [email protected] but I just get a new window with the "server not found" message. If I use the startu