Migration trouble from 10.7 to 10.8

I want to migrate from my
old iMac using migration assistant Version 4.1.2 (447) at 10.7.5
to new iMac using Migration assistant 5 (509) at 10.8.2
connecting two iMac by ethernet cable.
But it keeps "looking other computers" even when I allowed internet sharing (ethernet) at sharing mode.
Can you help me?

kim....
You don't actually run Migration Assistant on both computers....only the new one.
I find the easiest way to move from an old Mac to a new one is via the use of a Time Machine backup on an external hard disk.  Make a Time Machine backup of your old MacBook onto the external drive.  (You might want to deauthorize this old Mac in iTunes at this point before you power it down).  Shut down the Macbook...you won't need it any more. Now connect the external drive to your new iMac.
On your new iMac, start up MIgration Assistant. You'll choose the "from another disk" option when it asks how you want to transfer everything.  Choose the Time Machine backup disk.
One tip...when you set up your account on the new iMac (once Migration Assistant's done its thing) don't use the same name as you had on your old Macbook.

Similar Messages

  • Mail migration trouble (Switching between outlook - mac to windows)

    Having troubles for migrating mails from mac outlook 2011 to windows outlook 2010.
    I thought of switching to windows becaouse of personal reasons, and email migration is one thing that has stopped me from doing so.
    I was transferring my mails from outlook 2011 to outlook 2010, and thought it would take only few minutes for migration via import/export option available in outlook 2011. But the situation seems to be more complicated than I thought. Outlook 2010 would not import tha file made while using the export mail option in outlook 2011. I am stuck with this problem and need to solve it asap. Am I doing anything wrong?? Please suggest some problem solving solutions.
    Thanks

    Hi,
    Outlook can't import .olm files directly, as a workaround, we may create an IMAP account in Outlook 2011 and upload your emails to it, then create the same IMAP account in Outlook 2007 to download the emails, then drag the emails to your own account folder.
    How to configure Outlook to receive e-mail messages from an IMAP server:
    http://support.microsoft.com/kb/286197
    Add IMAP Email Account in Microsoft Outlook 2011 for Mac OS X:
    http://www.hosting.com/support/email/add-imap-email-account-in-microsoft-outlook -2011-for-mac-os-x/
    Transfer mail using Third Party tool:
    Many mail transfer applications are available online, one i know is available on this link
    Regards.

  • Migrate wiki from MLS to MAVS

    Dear Folks,
    I want to provide a solution on how to migrate the Wiki from a Mountain Lion Server to a Mavericks Server. Both Servers are running but in different locations and they can't communicate directly with each other. Upgrading the MLS was no option - after what I had read in hundreds of threads. The MLS is a 10.8.5 (with the latest version of server.app up until the 15th of October 2013) and Mavericks a 10.9.1 (with the latest version of server.app up until New Years Eve).
    I hadn't done it without the help of Andreas from Metalab in Vienna (lovely Hackspace) - so the grace goes to him!!
    First things first - he isolated 2 issues on MLS - revolving postgresql Version AND Socket! Thank you Apple-developers you did a great job here - since this product is called „The Server for everyone“ I believe either the title is incorrect or some guys haven’t understood what the a server for everyone implies - since everyone isn’t a geek, nerd, developer etc. - think about it!
    Trying to connect to psql on MLS:
    bash-3.2# psql
    psql: could not connect to server: No such file or directory
              Is the server running locally and accepting
              connections on Unix domain socket "/var/pgsql_socket/.s.PGSQL.5432"?
    What?!
    bash-3.2# serveradmin status postgres
    postgres:state = "RUNNING"
    What?!
    So he started digging on MLS:
    bash-3.2# netstat -na|grep LISTEN
    nothing tangible here - nothing ist listening to 5432
    Dig deeper on MLS:
    bash-3.2# lsof | grep postgre
              #output omitted for clarity
    postgres_ 2546      _postgres   10u     unix 0xa319efaa0af41357       0t0         /Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432
    postgres_ 2547      _postgres   10u     unix 0xa319efaa0af3fbe7       0t0         /Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432
              #output omitted for clarity
    postgres_ 2548      _postgres   10u     unix 0xa319efaa0af3fa57       0t0         /Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432
              #output omitted for clarity
    postgres_ 2549      _postgres   10u     unix 0xa319efaa0af3f8c7       0t0         /Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432
              #output omitted for clarity
    postgres_ 3102      _postgres   10u     unix 0xa319efaa0b220357       0t0         /Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432
              #output omitted for clarity
    Hold the horses still - that got him thinking! How does that fit to the output of the psql-command just seconds earlier (and I literally mean seconds!) Can you see it?
    The truth is, the Socket is at:
    /Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432
    but is expected by psql to be in /var/psql_socket/.s.PGSQL.5432!
    Expected What?!
    Verified it again on MLS:
    bash-3.2# psql
    psql: could not connect to server: No such file or directory
              Is the server running locally and accepting
              connections on Unix domain socket "/var/pgsql_socket/.s.PGSQL.5432"?
    bash-3.2# psql --help
    Yep - when postgresql was compiled obviously somebody was sitting on his fingers - during implementation nobody had told psql the path for the socket at the "new" location?! Ergo - psql is looking for the socket in the default location - where it is not.
    Then, with the help, he managed it to get in.
    Trying to get in on MLS:
    bash-3.2# psql -p "/Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432"
    psql: invalid port number: "/Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432"
    Trying again on MLS:
    bash-3.2# psql -h "/Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432"
    psql: could not connect to server: Not a directory
              Is the server running locally and accepting
              connections on Unix domain socket "/Library/Server/PostgreSQL For Server Services/Socket/.s.PGSQL.5432/.s.PGSQL.5432"?
    And again - almost there on MLS:
    bash-3.2# psql -h "/Library/Server/PostgreSQL For Server Services/Socket/" -p 5432
    psql: FATAL:  role "root" does not exist
    Typo:
    bash-3.2# psql -h "/Library/Server/PostgreSQL For Server Services/Socket/" -p 5432 -u colab
    psql: invalid option -- u
    Try "psql --help" for more information.
    And - tata - he was in on MLS:
    bash-3.2# psql -h "/Library/Server/PostgreSQL For Server Services/Socket/" -p 5432 template1 collab
    psql (9.1.9, server 9.2.4)
    WARNING: psql version 9.1, server version 9.2.
             Some psql features might not work.
    Type "help" for help.
    template1=#
    Lets find the databases and roles on MLS:
    template1=# \l
                                       List of databases
           Name        |   Owner    | Encoding | Collate | Ctype |    Access privileges   
    -------------------+------------+----------+---------+-------+------------------ -------
    caldav            | caldav     | UTF8     | C       | C     |
    collab            | collab     | UTF8     | C       | C     |
    device_management | _devicemgr | UTF8     | C       | C     |
    postgres          | _postgres  | UTF8     | C       | C     |
    template0         | _postgres  | UTF8     | C       | C     | =c/_postgres           +
                       |            |          |         |       | _postgres=CTc/_postgres
    template1         | _postgres  | UTF8     | C       | C     | _postgres=CTc/_postgres+
                       |            |          |         |       | =c/_postgres
    webauth           | webauth    | UTF8     | C       | C     |
    (7 rows)
    And there are the roles on MLS:
    template1=# \du
                                  List of roles
    Role name  |                   Attributes                   | Member of
    ------------+------------------------------------------------+-----------
    _devicemgr | Create DB                                      | {}
    _postgres  | Superuser, Create role, Create DB, Replication | {}
    caldav     | Create DB                                      | {}
    collab     | Superuser, Create role, Create DB              | {}
    webauth    | Superuser, Create role, Create DB              | {}
    Aaaaah - so there is a role collab and a database collab! Lets connect to the "right" database and dive into it on MLS.
    template1=# \q
    bash-3.2# psql -h "/Library/Server/PostgreSQL For Server Services/Socket/" -p 5432 collab collab
    psql (9.1.9, server 9.2.4)
    WARNING: psql version 9.1, server version 9.2.
             Some psql features might not work.
    Type "help" for help.
    collab=# \d
                       List of relations
    Schema |            Name             | Type  | Owner 
    --------+-----------------------------+-------+--------
    public | blog_entity                 | table | collab
    public | document_entity             | table | collab
    public | entity                      | table | collab
    public | entity_acls                 | table | collab
    public | entity_acls_defaults        | table | collab
    public | entity_attrs                | table | collab
    public | entity_changesets           | table | collab
    public | entity_comment              | table | collab
    public | entity_lock                 | table | collab
    public | entity_preview              | table | collab
    public | entity_private_attrs        | table | collab
    public | entity_tag                  | table | collab
    public | entity_type                 | table | collab
    public | file_entity                 | table | collab
    public | filedata_entity             | table | collab
    public | filename_reservation        | table | collab
    public | global_settings             | table | collab
    public | groups                      | table | collab
    public | migration_entity            | table | collab
    public | migration_status            | table | collab
    public | migrationplaceholder_entity | table | collab
    public | notification                | table | collab
    public | page_entity                 | table | collab
    public | podcast_entity              | table | collab
    public | podcast_episode_entity      | table | collab
    public | preview_queue               | table | collab
    public | project_entity              | table | collab
    public | relationship                | table | collab
    public | savedquery_entity           | table | collab
    public | search_index                | table | collab
    public | search_stat                 | table | collab
    public | session                     | table | collab
    public | subscription                | table | collab
    public | user_activity               | table | collab
    public | user_entity                 | table | collab
    public | user_entity_favorites       | table | collab
    public | user_entity_read_status     | table | collab
    public | user_entity_unread_status   | table | collab
    public | user_entity_updates         | table | collab
    public | user_entity_watched         | table | collab
    public | user_readall_time           | table | collab
    (41 rows)
    That looks fantastic! Lets start dumping with some help through pg_dump --help and then issuing the following on MLS:
    bash-3.2# pg_dump -h "/Library/Server/PostgreSQL For Server Services/Socket/" -p 5432 -f /Volumes/USBSTICK/wikidatabase -U collab collab
    pg_dump: server version: 9.2.4; pg_dump version: 9.1.9
    pg_dump: aborting because of server version mismatch
    What?! How could that be?! The commands refer to an older version?!
    Start digging again on MLS!
    bash-3.2# which psql
    /usr/bin/psql
    bash-3.2# which pg_dump
    /usr/bin/pg_dump
    That's where I stepped in and told Andreas "Hey - ahm - I know of a path completely different…. /Applications/Server.app/.….". Thanks mate!
    Dig again on MLS!
    bash-3.2# /Applications/Server.app/Contents/ServerRoot/usr/bin/pg_dump --version
    pg_dump (PostgreSQL) 9.2.4
    Lovely - that one looks good - lets use it. All of that got us finally to the right command on MLS.
    bash-3.2# /Applications/Server.app/Contents/ServerRoot/usr/bin/pg_dump -h "/Library/Server/PostgreSQL For Server Services/Socket/" -p 5432 -f /Volumes/USBSTICK/wikidatabase -U collab collab
    There it is - the whole Database on the stick - finally - it took Andreas only 5 Minutes - approximately? I am just baffled.
    Then I copied the Folders from the /Library/Server/Wiki/FileData of the MLS to the exact same location on the MAVS and sat permission - navigating to /Library/Server/Wiki first:
    bash-3.2# chown -R _teamsserver:_teamsserver FileData/
    bash-3.2# chmod -R +a "_www allow read" FileData/
    I compared the the collab DB on the MLS with the MAVS collab DB to see the differences - BTW - the Socket has moved again on the MLS! Move up in this documentary and you’ll see on how to find it on MAVS.
    bash-3.2# lsof | grep 5432 
              #output omitted for clarity
    postgres_   382    _teamsserver    3u     unix 0x399cdd6eeea1efe1       0t0         /Library/Server/Wiki/PostgresSocket/.s.PGSQL.5432
              #output omitted for clarity
    bash-3.2# psql -h "/Library/Server/Wiki/PostgresSocket/" -p 5432 collab collab
    psql (9.2.4)
    Type "help" for help.
    collab=# \d
                       List of relations
    Schema |            Name             | Type  | Owner 
    --------+-----------------------------+-------+--------
    public | adc_device_entity           | table | collab
    public | adc_team_entity             | table | collab
    public | blog_entity                 | table | collab
    public | bot_entity                  | table | collab
    public | botgroup_entity             | table | collab
    public | botrun_entity               | table | collab
    public | build_agent_registry        | table | collab
    public | document_entity             | table | collab
    public | email_notification          | table | collab
    public | entity                      | table | collab
    public | entity_acls                 | table | collab
    public | entity_acls_defaults        | table | collab
    public | entity_attrs                | table | collab
    public | entity_auditlog             | table | collab
    public | entity_changesets           | table | collab
    public | entity_comment              | table | collab
    public | entity_lock                 | table | collab
    public | entity_preview              | table | collab
    public | entity_private_attrs        | table | collab
    public | entity_tag                  | table | collab
    public | entity_type                 | table | collab
    public | file_entity                 | table | collab
    public | filedata_entity             | table | collab
    public | filename_reservation        | table | collab
    public | global_settings             | table | collab
    public | groups                      | table | collab
    public | migration_entity            | table | collab
    public | migration_status            | table | collab
    public | migrationplaceholder_entity | table | collab
    public | page_entity                 | table | collab
    public | podcast_entity              | table | collab
    public | podcast_episode_entity      | table | collab
    public | preview_queue               | table | collab
    public | relationship                | table | collab
    public | savedquery_entity           | table | collab
    public | scm_commit_entity           | table | collab
    public | scm_server                  | table | collab
    public | scmrepogroup_entity         | table | collab
    public | search_index                | table | collab
    public | search_stat                 | table | collab
    public | session                     | table | collab
    public | subscription                | table | collab
    public | timeseries                  | table | collab
    public | timeseries_toc              | table | collab
    public | user_activity               | table | collab
    public | user_entity                 | table | collab
    public | user_entity_favorites       | table | collab
    public | user_entity_read_status     | table | collab
    public | user_entity_unread_status   | table | collab
    public | user_entity_updates         | table | collab
    public | user_entity_watched         | table | collab
    public | user_readall_time           | table | collab
    public | visible_entity_tag          | view  | collab
    public | wiki_entity                 | table | collab
    public | work_queue                  | table | collab
    public | work_schedule               | table | collab
    public | work_schedule_recurrence    | table | collab
    public | work_schedule_status        | table | collab
    (58 rows)
    Hmmm - 41 tables in the old, and 58 in the new. What am I going to do? To be able to roll back I dumped the MAVS DB before I would do anything else.
    I had to decide between trying to take the old MLS DB and just pg_restore it into the MAVS DB, or - trying to upgrade the MLS instance to MAVS. I was afraid doing the upgrade since the MLS was a mess. At the end I would decide to go for the latter - since I had dumps. It took a while but - hey - it worked. Everything in place! Wiki running - everything else, too! The database has now 58 tables - ok - lets dump it. Notice again - no need to specify the proper version of pg_dump, but the Socket has changed from MLS to MAVS - again - you could see this when I connected with psql as well!
    Please use (-F c) compression when dumping - otherwise you’ll receive a nasty error when importing it at the MAVS.
    bash-3.2# pg_dump -h "/Library/Server/Wiki/PostgresSocket/" -p 5432 -F c -f /Volumes/USBSTICK/MLSafterupgrade2MAVS_collab_db_compressed.pgdump -U collab collab
    I then copied the dump to the server - the Desktop of the administrativ account - I did it through screensharing which is accessible after establishing a VPN to the location the server resides in.
    Unfortunately the command for pg_restore has a different syntax then the pg_dump. Use pg_dump --help to see the details. I’ll explain it quickly:
    -c is to clean entries in the target tables (you must not use capital c)
    -d specifies the database into which you would like to restore
    -h specifies (again) where the DB is - the socket
    -p the port to use
    -U is the role that likes to work on the DB
    you must not use -f since it’s not needed
    bash-3.2# pg_restore -c -d collab -h "/Library/Server/Wiki/PostgresSocket/" -p 5432 /Volumes/OSXDATA/Users/macminiadmin/Desktop/MLScollabdb_compressed.pgdump -U collab
    pg_restore: [archiver (db)] Error while PROCESSING TOC:
    pg_restore: [archiver (db)] Error from TOC entry 2466; 2605 16639 CAST CAST (text[] AS public.hstore)
    pg_restore: [archiver (db)] could not execute query: ERROR:  cannot drop cast from text[] to public.hstore because extension hstore requires it
    HINT:  You can drop extension hstore instead.
        Command was: DROP CAST (text[] AS public.hstore);
    pg_restore: [archiver (db)] Error from TOC entry 1816; 2616 16636 OPERATOR CLASS hash_hstore_ops collab
    pg_restore: [archiver (db)] could not execute query: ERROR:  cannot drop operator class hash_hstore_ops for access method hash because extension hstore requires it
    HINT:  You can drop extension hstore instead.
        Command was: DROP OPERATOR CLASS public.hash_hstore_ops USING hash;
              #output omitted for clarity
    I saw quite a lot errors (some revolving functions etc.) - however, it worked. Wikis are present again on the MAVS. Not without the initial help of Andreas - thank you once again!
    I would like to recommend to everyone joining hackerspaces in their location. I do not know them in other areas - I can only speak for the metalab in Vienna. Fantastic place with fantastic, very knowledgable people. I would also recommend reading a book like PostgreSQL 9 Admin Cookbook - this helped me quit a lot as well. It’s available through various stores - virtual or real - choose what you prefer.

    Frustrated, not angry...
    Well folks. I think this is goodbye to OS X Server - the server that I thought was much easier to use.  I've followed instructions on this forum and several others over the past 7 months.  It's cost me so much time and headache.
    Migrating wikis from Mountain Lion Server to Mavericks server has yet to work
    Upgrading our Mountain Lion Server to Mavericks leaves us without wiki, calendar, and other services
    Apple support docs don't do the trick
    Phone calls with Apple support techs haven't provided any solutions.
    Our current Mountain Lion Server suddenly lost all filesharing authentication capability and Apple enterprise solutions haven't been able to figure this one out.
    I'm stuck with a server with broken mountain lion server I have never been able to upgrade and/or migrate and ultimately would be left to copy/paste wikis to a fresh Mavericks server, export/import calendars from each client station, and contacts - I don't even know how that's going to work.
    I believe in keeping IT folks employed, but...
    The support from Apple seems so marginal (just for the server) and the plethora of problems with upgrading to Mavericks server and/or the manual fight to only potentially get it done seems very un-Apple like and just downright counter to what Apple represents.
    So I'll think different...
    Our server will be relegated to a client machine with simple file sharing.  Perhaps We'll buy Daylite to assume some of the responsibilites.  Truth is, even when things were working, it was always flaky in some way.
    I really wanted to employ OS X Server in an efficient way, but it's been more trouble than I can accomodate.
    Fortunately I am not an average client user nor a novice, so working in terminal is comfortable.
    Perhaps in a few years or so, I'll reconsider, but my experience with OS X server on Mountain Lion and Mavericks has affected my trust.
    Thanks for all of the help to those who provided instructions, workarounds, directions toward solutions, etc.
    For those of you who had a successful migration - consider yourself lucky.
    ...by the way...after finishing the incredible instructions on this page (and i sincerely mean that despite them not working for me) the result was a wiki page that looked like this
    Caught exception "[<CSEntityPlaceholder 0x7fb2d15d5910> valueForUndefinedKey:]: this class is not key value coding-compliant for the key externalID." [NSUnknownKeyException] executing route /app-context/wiki/:
    0   CoreFoundation                      0x00007fff8a0fe25c __exceptionPreprocess + 172
    1   libobjc.A.dylib                     0x00007fff8da20e75 objc_exception_throw + 43
    2   CSService                           0x0000000104bc7957 -[CSLocalServiceProxy forwardInvocation:] + 1229
    3   CoreFoundation                      0x00007fff8a05c1c4 ___forwarding___ + 452
    4   CoreFoundation                      0x00007fff8a05bf78 _CF_forwarding_prep_0 + 120
    5   CSService                           0x0000000104bf9573 __27-[CSAppContextService init]_block_invoke240 + 180
    6   CSService                           0x0000000104bdf81a __53-[CSRoutingHTTPConnection httpResponseForMethod:URI:]_block_invoke + 95
    7   CSService                           0x0000000104be2d6c -[CSHTTPBackgroundResponse bounce:] + 286
    8   Foundation                          0x00007fff9073976b __NSThread__main__ + 1318
    9   libsystem_pthread.dylib             0x00007fff8bc58899 _pthread_body + 138
    10  libsystem_pthread.dylib             0x00007fff8bc5872a _pthread_struct_init + 0
    11  libsystem_pthread.dylib             0x00007fff8bc5cfc9 thread_start + 13

  • Can't migrate files from time capsule, resource temporarily unavailable

    My old iMac 20" Intel late 2006 crashed. It got slower and froze as well as lines across, so I thought it's the graphic card. At the same time the SOPHOS Anti virous detected malware and trojan on my computer which was removed manually.
    Unfortunately with the 20" late 2006 model the card cannot be replaced. Ran hardware test and did not indicate any hardware issues. Finally could not boot from the disc. When I turned on the machine, I could hear fans running and some other noice but only black screen. We back up our computers every hour onto a time capsule.
    I purchased a new iMac 21.5 " and thought it will be easy to migrate everything from the time capsule. I done it in the past when I had to replace the hard drive on the old iMac 2 years ago.
    Unfortunately when I try to access the time capsule, I received the message of data damaged and now resource temporarily unavailable. How can I retrieve my data? Any suggestions are greatly appreciated.

    You will be in deep trouble if the TM backup is corrupt..
    Run the verify disk on the sparsebundle.. Pondini has some instructions here.
    http://pondini.org/TM/Troubleshooting.html
    Look at A5
    Also there are manual methods of getting as much as possible from the backup if it is not corrupted too badly..
    Look at http://pondini.org/TM/FAQ.html
    He gives several methods to restore files.. read Q14-17.

  • Using migration assistant from HDD only to SSD+HDD?

    Hi,
    Usually, when buying a new mac and starting it for the first time, I use the migration assistant to get everything (software, settings, data...) from my time capsule backup.
    I am currently using an iMac 24'' which only has a "standard" HDD. I am however thinking about upgrading to the latest iMacs, potentially with the "SSD+HDD" option.
    The description on the "Configure" page of the online store says : "If you configure your iMac with both the solid-state drive and a Serial ATA hard drive, it will come preformatted with Mac OS X and all your applications on the solid-state drive. Then you can use the hard drive for videos, photos, and other files."
    My question is therefore the following : would the migration assistant be "smart enough" to place my OS and applications on the SSD drive, and the rest of my files ("videos, photos, and other files.") on the HDD?
    Thanks in advance!

    I had the same issue some time ago: I added a 32 gb SSD to my mac running leopard. Then wanted to use install Snow Leopard and use migration assistant to restore the sytem with Snow leopard and applications on the SSD, and the user folders on the old 250gb HD. Turns out this isn't possible. Manually accessing your time machine or other backup up to restore files by hand is a pain. Just installing Snow Leopard and pointing the user folder to the Leopard user folder on the HD creates problems too, just as pointing it to a newly created folder.
    Solution:
    -Backup your old system with time machine to an external disk
    -Note: make sure you don't mix up SSD and HD in this manual.
    -Install SSD (If this means you no longer have a cd drive, use a copy of snow leopard on an external drive).
    -BE SURE YOUR OLD SYSTEM IS BACKED UP WITH TIME MACHINE.
    -Then install OS X TWICE: once on the old HD and once the SSD. Wipe your HD with Disk Utility before install.
    -Start up OS X on the HD, use the option key at startup if necessery to select the right disk.
    -Use migration assistant to restore from the time machine backup. Now you have a fresh install of OS X, with your user files and applications intact.
    -Now startup OS X on the SSD.
    -Make the SSD  startup disk in system preferences -> Startup disk.
    -Go to System preferences -> Users. Unlock. Create an extra user, so you end up with a spare user account, and one you will actually use. Now right click on the user account thats going to be used, and select your user folder on the HD that you just restored with Migration assistant. Just leave the spare one. It is for emergencies and also disk failures: OS X is designed to have user and system files on the same disk, so you should always have one user account sticking those rules.
    -Restart OS X  on the SSD as asked.
    -Now all thats left to do is migrate applications from the HD application folder to the SSD application folder by simply copy/pasting. This is trial and error, but in my case 90% of the applications gave no trouble at all. The ones that did are library based media applications like Aperture. All I needed to do is go to their preferences and point them to their old libraries on the HD. MS Office was the only application that needed full reinstalling. The applications that ship with the OS don't need to be migrated of course since they are already there on the SSD.
    -Point your dock icons to the right folders (on the HD) and applications (on the SSD)
    -Make a new time machine backup
    -Last step, optional: clean up the HD. You don't need the system folder that is still there, but I chose to keep it for a while as a backup. I left the library in place there too, since I don't know which applications still use it, but in time I will clean it up too.
    And one more thing: Since Migration assistant doesnt' just copy files, but also seams to adapt application databases of Apple apps to work with a newer version of an application coming with a new OS, such as Mail, Safari, iCal and Address Book, this is probably going to be the best way to upgrade an SSD+HD setup to Lion too. (Lion has better SSD support). Provided of course that Apple doesn't update Migration assistant, which is what should actually be done. Upgrading this way took me 2,5 hours, most of which was spent waiting ofcourse.
    Message was edited by: jason487
    Message was edited by: jason487

  • How can I migrate everything from one account to another on same computer?

    How can I migrate everything from one account to another on same computer?

    Transferring files from one User Account to another

  • After migrating data from Time Machine, some of my photos are not showing up in iPhoto. I get a 'dashed rectangle." When I click on it I get ' ! in a Triangle" When I click on that, I actually can see the photo. I want to see my photos

    After migrating data from Time Machine, some of my photos are not showing up in iPhoto LIbrary view. I get a 'dashed rectangle." When I click on it I get ' ! in a Triangle" When I click on that, I actually can see the photo. I want to see all  my photos in 'Library' view, and I can't figure out how these photos seem to be arbitrarily hidden. Help, please.

    Try these for the first attempt:
    First be sure to have a backup copy of the library if you already don't have one.
    OT

  • Error while migrating reports from one instance to another(no torch seen)

    Hi,
    I am facing this problem when I am trying to migrate reports from UAT to Production. After migration when I try to run the report from Discoverer plus, I cannot see the torch symbol which is general if we have any parameters for that report. So, I again migrate the same .eex file and now I see the torch.
    Can any one throw some thoughts why is this happening, why I cannot see the torch symbol in the first time itself?
    Thanks,
    Srini.

    Hi Srini
    The list of values is not being associated with the item the first time around. The order in which the items are selected is important so for example if lists of values were imported first without the items upon which they depend they will be ignored. Running the import a second time will cause the LOVs to recognize their associated item and thus the association will be made.
    Generally I do two exports. I will export my folders and workbooks first and then export my item classes.
    Hope this helps explain what you are seeing.
    Best wishes
    Michael

  • Migrating menues from oracle 11i to R12

    Hi all,
    I have to migrate Menus from Oracle 11i to R12, this can be done using FND_LOAD, but we have to migrate it using API's. Please suggest me the API name and approach. Currently I am investigating on fnd_menus_pkg.load_row and fnd_menu_entries_pkg.load_row . Please let mw know if these are the correct API's to migrate Menus. Any help is appreciated.

    If the version of $FND_TOP/patch/115/import/afsload.lct and the table structure is identical on both instances, you may use FNDLOAD for your custom menus. However, I would not recommend using FNDLOAD or any API between different application releases.
    How To Download a Parent Menu and Submenus [ID 369421.1]
    Replicating Custom Menu [ID 839879.1]
    Thanks,
    Hussein

  • Migrating report from 6i to 11g

    I am migrating reports from 6i to 11g.
    I am calling report from form. All the reports works fine.
    But in this report I have to pass parameter at runtime and report should save in pdf format at a given location.
    forms & report version : 11.1.2
    Database : 11g
    Operating system : windows 7 professional.
    Below is the code i am running.
    declare
    pl_name VARCHAR2(30) := 'tempdata';
    pl_id ParamList;
    x VARCHAR2(150);
    y NUMBER;
    repid REPORT_OBJECT;
    v_rep VARCHAR2(100);
    rep_status VARCHAR2(20);
    BEGIN
    repid := FIND_REPORT_OBJECT('MALWADKARCONSUMPTIONSTMTDBR1');
    pl_id := Create_Parameter_List(pl_name);
    Add_Parameter(pl_id,'background',TEXT_PARAMETER,'Yes');
    Add_Parameter(pl_id,'PARAMFORM',TEXT_PARAMETER,'NO');
    Add_Parameter(pl_id,'DESTYPE',TEXT_PARAMETER, 'FILE') ;
    Add_Parameter(pl_id,'DESFORMAT',TEXT_PARAMETER, 'PDF') ;
    Add_Parameter(pl_id,'VCOMPANY',TEXT_PARAMETER,:block50.COMPANYNAME);
    Add_Parameter(pl_id,'VLYCOMPANY',TEXT_PARAMETER,:block50.COMPANYNAME);
    Add_Parameter(pl_id,'FROMDATE',TEXT_PARAMETER,VFROMDATE); -- VFROMDATE calculated from query
    Add_Parameter(pl_id,'TODATE',TEXT_PARAMETER,:BLOCK50.DOCDATE);
    Add_Parameter(pl_id,'VACNAME',TEXT_PARAMETER,VACNAME); -- VACNAME calculated from query
    Add_Parameter(pl_id,'DESNAME',TEXT_PARAMETER, 'D:\dbr\'||vACNAME||' '||:BLOCK50.DOCDATE||'.PDF') ;
    SET_REPORT_OBJECT_PROPERTY(repid,REPORT_COMM_MODE,SYNCHRONOUS);
    SET_REPORT_OBJECT_PROPERTY(repid,REPORT_EXECUTION_MODE,RUNTIME);
    set_report_object_property(repid,report_filename,'D:\dbr\'||vACNAME||' '||:BLOCK50.DOCDATE||'.PDF' );
    SET_REPORT_OBJECT_PROPERTY(repid,REPORT_DESTYPE,FILE);
    SET_REPORT_OBJECT_PROPERTY(repid,REPORT_DESFORMAT,'PDF');
    SET_REPORT_OBJECT_PROPERTY(repid,REPORT_SERVER,'rep_adminserver_paresh-pc_asinst_1');
    y := LENGTH ('rep_adminserver_paresh-pc_asinst_1') + 2;
    x := RUN_REPORT_OBJECT (repid, pl_id);
    WEB.SHOW_DOCUMENT('http://PARESH-PC:7001/reports/rwservlet/getjobid'||SUBSTR(x, y)||'?server=rep_adminserver_paresh-pc_asinst_1&userid=scott/TIGER@ORCL');
    Destroy_parameter_list('tempdata');
    END;
    Above code doesn't show any error but report is not generated in the provided location.
    Is modification to be done in the above code?
    Please suggest.
    Thanks in advance.
    Edited by: parapr on Oct 2, 2012 1:30 AM

    Hi,
    as stated before if the desname value is pointing to the client filesystem it wont create the output there because Reports Server is not running on the client, it is running on the server as this is the way it is.
    If you want to save output to the client you need to use the options mentioned before.
    For testing purposes only change your code as follows.
    e.g
    Add_Parameter(pl_id,'DESNAME',TEXT_PARAMETER, 'c:\'||vACNAME||' '||:BLOCK50.DOCDATE||'.PDF') ;
    As you can see I changed the original folder name to C:\ .. (I hope your reports serve is running on a Windows machine if not then changed to /tmp). Run your form and check c:\ or /tmp and see if you see the report output there. If so,
    than your code is fine the only problem is what I have already stated.
    Regards, Roberto

  • Migrating portal from 10g to 11g

    Hi,
    We are in the process of upgrading oracle mid tier from 10gAS to OFMW 11g and also migrating it from Solaris to Linux.
    With regard the portal migration from OAS 10g to OFMW 11g
    1) Is it possible to migrate portal objects from 10g (solaris) to a new install of OFMW portal 11g(linux) directly OR do we have to upgrade the existing OAS 10g to OFMW portal 11g (on Solaris) and then migrate the portal objects to OFMW Portal 11g (on Linux)
    I am new to portal migration. Any suggestions and additional details would be appreciated.
    Thanks
    RB

    Migration + upgrade in one shot is do-able, because upgrade will require a sequence of processes but they will be mostly based on repository upgrade, and hence will be transparent to the OS you are using.
    However, in practice, the standard rule-of-thumb is that you avoid dealing with multiple layers of complications in a single shot. That means, you break complications in such steps where you could recognize the sources of any later errors/problems you encounter.
    hope that helps!
    AMN

  • How to only migrate data from SQL Server 2008 to Oracle 11?

    According to our requirement, We need to only migrate data from a SQL Server database to an existed
    Oracle database user.
    1) I tried to do it with SQL Developer 3.0.04 Migration Wizard, But find an issue.
    My SQL Server database name is SCDS41P2, and my Oracle database user name is CDS41P2;
    When I used SQL Developer to do offline move data by Migration Wizard, I found all oracle user
    name in movedata files which gotten by run Migration Wizard
    is dbo_SCDS41P2. If the Oracle user name is not the same as my existed Oracle user name,
    the data can't be moved to my existed Oracle user when I run oracle_ctl.bat in command line window.
    So I had to modify the Oracle user name in all movedata files, but it's difficult to modify them because there are many tables in
    databases. So could you please tell me how to get the movedata files which the oracle user name in them is my
    expected Oracle user name?
    2) I also tried to use the 'copy to Oracle' function to copy the SQL Server database tables data
    to the existed Oracle database user. When clicked 'copy to Oracle', I selected 'Include Data' and 'Replace' option
    But I found some tables can't be copied, the error info is as below:
    Table SPSSCMOR_CONTROLTABLE Failed. Message: ORA-00955: name is already used by an existing object
    Could you please tell me how to deal with this kind of error?
    Thanks!
    Edited by: 870587 on Jul 6, 2011 2:57 AM

    Hi,
    Thanks for you replying. But the 'copy to oracle' function still can't be work well. I will give some info about the table. I also search 'SPSSCMOR_CONTROLTABLE' in the target schema, and only find one object. So why say 'name is already used by an existing object'? Could you please give me some advice? Thanks!
    What is the 'Build' version of your SQL*Developer ?
    [Answer]:
    3.0.04
    - what does describe show for the SPSSCMOR_CONTROLTABLE in SQL*Server ?
    [Answer]:
    USE [SCDS41P2]
    GO
    /****** Object: Table [dbo].[SPSSCMOR_CONTROLTABLE] Script Date: 07/18/2011 01:25:05 ******/
    SET ANSI_NULLS ON
    GO
    SET QUOTED_IDENTIFIER ON
    GO
    CREATE TABLE [dbo].[SPSSCMOR_CONTROLTABLE](
         [tablename] [nvarchar](128) NOT NULL,
    PRIMARY KEY CLUSTERED
         [tablename] ASC
    )WITH (PAD_INDEX = OFF, STATISTICS_NORECOMPUTE = OFF, IGNORE_DUP_KEY = OFF, ALLOW_ROW_LOCKS = ON, ALLOW_PAGE_LOCKS = ON) ON [PRIMARY]
    ) ON [PRIMARY]
    GO
    - what does describe show for the SPSSCMOR_CONTROLTABLE in Oracle ?
    [Answer]:
    -- File created - Monday-July-18-2011
    -- DDL for Table SPSSCMOR_CONTROLTABLE
    CREATE TABLE "CDS41P2"."SPSSCMOR_CONTROLTABLE"
    (     "TABLENAME" NVARCHAR2(128)
    ) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "USERS" ;
    -- DDL for Index SYS_C009547
    CREATE UNIQUE INDEX "CDS41P2"."SYS_C009547" ON "CDS41P2"."SPSSCMOR_CONTROLTABLE" ("TABLENAME")
    PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "USERS" ;
    -- Constraints for Table SPSSCMOR_CONTROLTABLE
    ALTER TABLE "CDS41P2"."SPSSCMOR_CONTROLTABLE" MODIFY ("TABLENAME" NOT NULL ENABLE);
    ALTER TABLE "CDS41P2"."SPSSCMOR_CONTROLTABLE" ADD PRIMARY KEY ("TABLENAME")
    USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "USERS" ENABLE;
    Edited by: 870587 on Jul 18, 2011 1:42 AM

  • My question is in regard to pc to mac migration. How do I migrate information from my pc with specific user account to an existing user account on my mac? I do not want to use multiple mac user accounts.

    My question is in regard to pc to mac migration. How do I migrate information from my pc with specific user account to an existing user account on my mac? I do not want to use multiple mac user accounts.

    https://discussions.apple.com/message/16371308#16371308

  • How can I migrate data from an old MBA to a new one without having access to the screen of the old one? The screen of the old MBA is damaged !

    How can I migrate data from an old MBA to a new one without having access to the screen of the old one? The screen of the old MBA is damaged !

    If your "older" MBA has a Thunderbolt port then it isn't that old. See Target Disk Mode about how it's used. Note that without a monitor you won't be able to tell from the screen when it has fully started, and if it succeeded in starting in TDM. If it did, then the hard drive should appear on the Desktop of your new computer. You can then access it to transfer your files. You can even use Migration Assistant or any backup utility to transfer data. Just don't try to transfer system files.

  • Migrate data from one table to another

    Hi,
    I need to move data from work space -1 to work space -2. I need to move only selected data. From work space -1 to work space -2
    Work space -1
    Table -1_
    ID NAME Description
    1 xyz one
    2 abc two
    3 mnk three
    Table -2_
    ID PID NAME Description
    1 1 x1 -
    2 1 x2 -
    3 2 a1 -
    4 2 a2 -
    5 3 m1 -
    Table -1 is parent table and Table-2 is child. Here PID is parent key of table-1 and ID's in both the table are primary keys.
    Now i need to move data from work space-1 to work space -2
    Work Space -2 table structure
    Table - 1
    ID INDUSTRY_TYPE DATE DESC
    1 TNT - -
    2 MTN - -
    Table - 2
    ID T_ID NAME DATE DESC
    1 1 T1 - -
    1 1 T2 - -
    2 2 E1 - -
    2 2 E3 - -
    I need to migrate data from work space -1 of table 1 and 2 to work space - 2 of table 1 and tabe 2
    Please suggest me a query to migrate. Please let me know if the steps mentioned above is not clear.
    Thanks
    Sudhir

    Using INSERT statement.
    insert into <target table>(<column list>)
    select <column list>
      from <source table>

Maybe you are looking for

  • Data missing in iPad

    Hi, A form is developed using livecycle and saved as static pdf. The form is populated using iText in server. When the form is rendered in iPad, the data is missing. The data is being displayed in Mac Safari browser but not in iPad browser. Kindly he

  • Not comming out of loop

    Gurus Below is the code which I have written, the problem is, when record is not found, its showing alert, but system getting hanged .. Can you suggest a solution Declare      CURSOR C_SUBCLASS IS SELECT DEPT, CLASS, SUBCLASS                 FROM SUB

  • Sometimes I like to see my calendar on paper

    I don't have MS outlook to sync with but I do use gmail, is there a way to sync my calendar with the Google calendar function?  I like my BB but there are times it would be really great to print out my calendar so I could lay things out on paper.  Do

  • Same process different sessions

    Hi, DB version is 10.2.0.1 I could see same process having four different sessions and all are inactive..why is that so ? SQL> / SPID         PROCESS      STATUS   USERNAME        MACHINE              LOGON_TIME        LAST_CALL_ET ACTION     MODULE 

  • PSE9 for MAC install Failing

    Install get about 55% then fails with error below.. Any help would be appreciated Exit Code: 7 -------------------------------------- Summary -------------------------------------- - 0 fatal error(s), 15 error(s), 2 warning(s) WARNING: File/Folder do