Azure Backup Integrates with Azure Import to ship the initial backup copy through disk

Customers can now leverage shipping of initial backup copy to nearest Azure datacenter using Azure Import service. When the initial backup copies are huge and there isn’t sufficient network bandwidth to send the backup data, this feature enables the customers
to send the backup data in disk over courier. Once the initial backup copy is stored in Azure Backup Vault, incremental backups happen seamlessly over wire.  Package @
http://support2.microsoft.com/kb/3033794
Thanks,
Giri.

Giri
Is there a blog or something that explains how to prepare the Azure Online Backup data for local storage so it can be shipped?
Or is this Azure Online Backup update feature specific to DPM workloads?
Thanks!
-Neil
neilgo

Similar Messages

  • The creation of the shadow copy on the backup destination has failed. Detailed Error: The Volume Shadow Copy operation failed with the error 0x80042316

    When using Microsoft Backup on a Windows Server 2008 SP2 (not R2 - [Microsoft version 6.0.6002]) I get the error "The creation of the shadow copy on the backup destination has failed. Detailed Error: The Volume Shadow Copy operation failed with the
    error 0x80042316"
    The odd thing is that when I start the backup manually, it works fine. When  it runs automatically the error appears.
    When doing a SFC /scannow I do get errors that can't be corrected.
    Is there anything short of rebuilding this machine (It's a DC and the only one this office has) I can do to fix this?

    Hi,
    0x80042316 means "Another shadow copy creation is already in progress". This error could occur when another VSS instance is running.
    Is there any third party backup application installed? Try to uninstall it to see the result. And please also check if there is another backup schedule which causes the conflict.
    Meanwhile please provide the result of:
    vssadmin list providers
    vssadmin list writers
    Please remember to mark the replies as answers if they help and un-mark them if they provide no help. If you have feedback for TechNet Support, contact [email protected].
    Hi,
    I do have Crashplan ProE running. It started at 1900 and the Windows Backup starts at 2000. I have modified Crashplan not to start until 2300 to see if that helps.
    Also:
    vssadmin 1.1 - Volume Shadow Copy Service administrative command-line tool
    (C) Copyright 2001-2005 Microsoft Corp.
    Provider name: 'Microsoft Software Shadow Copy provider 1.0'
       Provider type: System
       Provider Id: {b5946137-7b9f-4925-af80-51abd60b20d5}
       Version: 1.0.0.7
    AND:
    vssadmin 1.1 - Volume Shadow Copy Service administrative command-line tool
    (C) Copyright 2001-2005 Microsoft Corp.
    Writer name: 'System Writer'
       Writer Id: {e8132975-6f93-4464-a53e-1050253ae220}
       Writer Instance Id: {f0c7ee55-5885-4894-8dd3-14fffbdd4d2e}
       State: [1] Stable
       Last error: No error
    Writer name: 'SqlServerWriter'
       Writer Id: {a65faa63-5ea8-4ebc-9dbd-a0c4db26912a}
       Writer Instance Id: {017a0133-7577-4322-b513-04f72404b70a}
       State: [1] Stable
       Last error: No error
    Writer name: 'ASR Writer'
       Writer Id: {be000cbe-11fe-4426-9c58-531aa6355fc4}
       Writer Instance Id: {6b17d670-9ea8-4ff1-a59e-b474f4b893db}
       State: [1] Stable
       Last error: No error
    Writer name: 'Shadow Copy Optimization Writer'
       Writer Id: {4dc3bdd4-ab48-4d07-adb0-3bee2926fd7f}
       Writer Instance Id: {56aad1bf-9bf1-4fde-b2f4-a706577c632d}
       State: [1] Stable
       Last error: No error
    Writer name: 'COM+ REGDB Writer'
       Writer Id: {542da469-d3e1-473c-9f4f-7847f01fc64f}
       Writer Instance Id: {e942d09d-276e-450a-9bf4-b1d5abe6af8a}
       State: [1] Stable
       Last error: No error
    Writer name: 'Registry Writer'
       Writer Id: {afbab4a2-367d-4d15-a586-71dbb18f8485}
       Writer Instance Id: {3737164b-61ac-454c-b1f7-42f2620f10db}
       State: [1] Stable
       Last error: No error
    Writer name: 'DFS Replication service writer'
       Writer Id: {2707761b-2324-473d-88eb-eb007a359533}
       Writer Instance Id: {39458257-aca4-481e-a302-994f3eb1e9a3}
       State: [1] Stable
       Last error: No error
    Writer name: 'Dhcp Jet Writer'
       Writer Id: {be9ac81e-3619-421f-920f-4c6fea9e93ad}
       Writer Instance Id: {6a8ab4aa-760c-4102-a27f-58f570c9c572}
       State: [1] Stable
       Last error: No error
    Writer name: 'NTDS'
       Writer Id: {b2014c9e-8711-4c5c-a5a9-3cf384484757}
       Writer Instance Id: {53577013-8397-448e-9a25-c7cac08aa0a8}
       State: [1] Stable
       Last error: No error
    Writer name: 'IIS Metabase Writer'
       Writer Id: {59b1f0cf-90ef-465f-9609-6ca8b2938366}
       Writer Instance Id: {ee2c9173-92e7-455c-8f73-601647c0ddb2}
       State: [1] Stable
       Last error: No error
    Writer name: 'BITS Writer'
       Writer Id: {4969d978-be47-48b0-b100-f328f07ac1e0}
       Writer Instance Id: {b1f60392-9dfd-4e3a-bda6-3dc2cb622ba4}
       State: [1] Stable
       Last error: No error
    Writer name: 'IIS Config Writer'
       Writer Id: {2a40fd15-dfca-4aa8-a654-1f8c654603f6}
       Writer Instance Id: {a328aede-78f0-448e-9c36-e8379e9fe2a8}
       State: [1] Stable
       Last error: No error
    Writer name: 'WMI Writer'
       Writer Id: {a6ad56c2-b509-4e6c-bb19-49d8f43532f0}
       Writer Instance Id: {c39fa334-5c9d-4906-a3b5-33c3444948e6}
       State: [1] Stable
       Last error: No error
    Writer name: 'NPS VSS Writer'
       Writer Id: {35e81631-13e1-48db-97fc-d5bc721bb18a}
       Writer Instance Id: {3c9c5584-b5fd-48cf-8541-0c8bfa585ca0}
       State: [1] Stable
       Last error: No error

  • Priming the initial backup

    Will microsoft offer a service to prime the initial backup by accepting a couriered external disk containing the data? Specifically, my issue is that I have a number of branch sites with limited upload bandwidth, it would take weeks to complete an initial
    backup. Alternatively, is there a way I could go about taking a copy of the vhdxs containing the data that needs to be pushed to azure backup, put them on an external disk, perform the initial upload from head office (over a much faster connection), and
    then allow the branch office server to perform the subsequent incremental backups?
    Cheers

    Will microsoft offer a service to prime the initial backup by accepting a couriered external disk containing the data? Specifically, my issue is that I have a number of branch sites with limited upload bandwidth, it would take weeks to complete an initial
    backup. Alternatively, is there a way I could go about taking a copy of the vhdxs containing the data that needs to be pushed to azure backup, put them on an external disk, perform the initial upload from head office (over a much faster connection), and
    then allow the branch office server to perform the subsequent incremental backups?
    Cheers
    Seems yes on february 2015. Offline seeding – complete your initial seeding faster and off-the-network
    New features in Azure Backup – Long term retention, offline backup seeding and more
    http://azure.microsoft.com/blog/2015/02/16/new-features-in-azure-backup-long-term-retention-offline-backup-seeding-and-more/
    Offline seeding – complete your initial seeding faster and off-the-network.
    Azure Backup enables customers to complete the initial backup by integrating with the
    Azure Import/Export service. Rather than sending the initial backup copy over a congested network, customers can ship the disk containing the initial backup copy to the nearest Azure datacenter.

  • I used an ethernet cable between my iMac and Time Capsule for the initial backup.  Now that initial backup is concluded, can I simply unplug the Time Capsule from my computer?

    I used an ethernet cable between my iMac and Time Capsule for the initial backup.  Now that initial backup is concluded, can I assume I can simply unplug the Time Capsule from my computer without any glitches and Time Capsule will then run wirelessly.  Is this assumption correct? 

    can I assume I can simply unplug the Time Capsule from my computer without any glitches and Time Capsule will then run wirelessly.  Is this assumption correct?
    Yes, if the computer will be able to receive a strong wireless signal from the Time Capsule.
    I did not disable wireless.  So I assume the backups are occurring both with the ethernet and wirelessly.
    It is inot possible to have both an "active" Ethernet and "active" wireless signal simultaneously. If an Ethernet connection is established, your computer will default to the Ethernet connection. If the Ethernet cable is unplugged, the computer will revert to the wireless connection.
    I also show 2 wireless networks available through Airport.  Shouldn't there only be one now?
    You have a simultaneous dual band Time Capsule. That means that it will always be providing both a 2.4 GHz and a 5 GHz network.
    I named my new network the same name as the old network.  This enabled all of our devices to continue with their connection without entering the new network and/or passwords; however, I am afraid this may pose problems in the future somehow.  Any input on this issue as well?
    Things should be fine. Devices will connect to the best signal.....either 2.4 GHz or 5 GHz....based on their capabilties and location in relation to the Time Capsule.

  • Is Time Machine under Snow Leopard faster only for the initial backup?

    A few days ago, I upgraded to Snow Leopard -- via clean install, manual reinstallation of apps, etc. So understandably, the first Time Machine backup to a preexisting store on an original 500 GB Time Capsule is going to be huge and take a long time. Apple claims that TM is 40% faster on the initial backup, but are there speed improvements as well? I was surprised by the slowness; I started it before going to bed, and it wasn't nearly done when I woke up:
    http://support.apple.com/kb/HT1770
    Jan 6 01:29:52 Musa com.apple.backupd[438]: Copied 8.3 GB of 59.3 GB, 9299 of 250057 items
    Jan 6 02:29:52 Musa com.apple.backupd[438]: Copied 9.6 GB of 59.3 GB, 35209 of 250057 items
    Jan 6 03:29:52 Musa com.apple.backupd[438]: Copied 11.5 GB of 59.3 GB, 48441 of 250057 items
    Jan 6 04:29:52 Musa com.apple.backupd[438]: Copied 12.9 GB of 59.3 GB, 109846 of 250057 items
    Jan 6 05:29:53 Musa com.apple.backupd[438]: Copied 17.4 GB of 59.3 GB, 140388 of 250057 items
    Jan 6 06:29:54 Musa com.apple.backupd[438]: Copied 26.1 GB of 59.3 GB, 151723 of 250057 items
    Jan 6 07:29:54 Musa com.apple.backupd[438]: Copied 36.9 GB of 59.3 GB, 167431 of 250057 items
    Jan 6 08:20:04 Musa com.apple.backupd[438]: Copied 219433 files (38.0 GB) from volume Gigas.
    The throughput rate is quite variable, but at its best, it's only about 2.7 MB/sec. Is that to be expected, or would I have to start a new backup to see speed improvements?

    Kappy wrote:
    i'd say you are concerned over nothing.
    Not sure I agree. Look at this system.log output:
    Jan 6 19:54:22 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 19:54:22 Musa com.apple.backupd[7571]: Attempting to mount network destination using URL: afp://odysseus@Time%20Capsule.afpovertcp.tcp.local/odysseus
    Jan 6 19:54:30 Musa com.apple.backupd[7571]: Mounted network destination using URL: afp://odysseus@Time%20Capsule.afpovertcp.tcp.local/odysseus
    Jan 6 19:54:33 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 19:54:33 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 19:54:42 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 20:00:50 Musa com.apple.backupd[7571]: Compacting storage: 72.07 GB requested (including padding), 58.36 GB available before compacting
    Jan 6 20:00:50 Musa com.apple.backupd[7571]: Stopping backup.
    Jan 6 20:00:50 Musa com.apple.backupd[7571]: Backup canceled.
    Jan 6 20:00:53 Musa com.apple.backupd[7571]: Ejected Time Machine disk image.
    Jan 6 20:00:53 Musa com.apple.backupd[7571]: Compacting backup disk image to recover free space
    Jan 6 20:01:12 Musa com.apple.backupd[7571]: Completed backup disk image compaction
    Jan 6 20:01:12 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 20:01:12 Musa com.apple.backupd[7571]: Network destination already mounted at: /Volumes/odysseus
    Jan 6 20:01:14 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 20:01:14 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 20:01:20 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 20:07:26 Musa com.apple.backupd[7571]: Starting pre-backup thinning: 72.07 GB requested (including padding), 58.36 GB available
    Jan 6 20:10:46 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2010-01-03-131536: 58.42 GB now available
    Jan 6 20:12:10 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2010-01-04-094152: 58.52 GB now available
    Jan 6 20:13:02 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2010-01-04-073331: 58.57 GB now available
    Jan 6 20:13:02 Musa com.apple.backupd[7571]: Stopping backup.
    Jan 6 20:13:02 Musa com.apple.backupd[7571]: Backup canceled.
    Jan 6 20:13:08 Musa com.apple.backupd[7571]: Ejected Time Machine disk image.
    Jan 6 20:13:08 Musa com.apple.backupd[7571]: Compacting backup disk image to recover free space
    Jan 6 20:13:30 Musa com.apple.backupd[7571]: Completed backup disk image compaction
    Jan 6 20:13:30 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 20:13:30 Musa com.apple.backupd[7571]: Network destination already mounted at: /Volumes/odysseus
    Jan 6 20:13:33 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 20:13:33 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 20:13:41 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 20:19:36 Musa com.apple.backupd[7571]: Starting pre-backup thinning: 72.07 GB requested (including padding), 58.57 GB available
    Jan 6 20:20:38 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2010-01-04-071252: 58.58 GB now available
    Jan 6 20:20:38 Musa com.apple.backupd[7571]: Removed all 1 expired backups, more space is needed - deleting oldest backups to make room
    Jan 6 20:25:17 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2009-07-31-231155: 60.81 GB now available
    Jan 6 20:25:17 Musa com.apple.backupd[7571]: Deleted 2 backups: oldest backup is now Aug 15, 2009
    Jan 6 20:25:17 Musa com.apple.backupd[7571]: Stopping backup.
    Jan 6 20:25:20 Musa com.apple.backupd[7571]: Backup canceled.
    Jan 6 20:25:24 Musa com.apple.backupd[7571]: Ejected Time Machine disk image.
    Jan 6 20:25:24 Musa com.apple.backupd[7571]: Compacting backup disk image to recover free space
    Jan 6 20:26:39 Musa com.apple.backupd[7571]: Completed backup disk image compaction
    Jan 6 20:26:39 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 20:26:39 Musa com.apple.backupd[7571]: Network destination already mounted at: /Volumes/odysseus
    Jan 6 20:26:41 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 20:26:41 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 20:26:49 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 20:32:46 Musa com.apple.backupd[7571]: Starting pre-backup thinning: 72.07 GB requested (including padding), 60.81 GB available
    Jan 6 20:32:46 Musa com.apple.backupd[7571]: No expired backups exist - deleting oldest backups to make room
    Jan 6 20:47:38 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2009-08-15-182619: 65.12 GB now available
    Jan 6 20:47:38 Musa com.apple.backupd[7571]: Deleted 1 backups: oldest backup is now Aug 25, 2009
    Jan 6 20:47:38 Musa com.apple.backupd[7571]: Stopping backup.
    Jan 6 20:47:41 Musa com.apple.backupd[7571]: Backup canceled.
    Jan 6 20:47:47 Musa com.apple.backupd[7571]: Ejected Time Machine disk image.
    Jan 6 20:47:47 Musa com.apple.backupd[7571]: Compacting backup disk image to recover free space
    Jan 6 20:49:36 Musa com.apple.backupd[7571]: Completed backup disk image compaction
    Jan 6 20:49:36 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 20:49:36 Musa com.apple.backupd[7571]: Network destination already mounted at: /Volumes/odysseus
    Jan 6 20:49:38 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 20:49:38 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 20:49:46 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 20:55:48 Musa com.apple.backupd[7571]: Starting pre-backup thinning: 72.06 GB requested (including padding), 65.12 GB available
    Jan 6 20:55:48 Musa com.apple.backupd[7571]: No expired backups exist - deleting oldest backups to make room
    Jan 6 21:05:11 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2009-08-25-185228: 67.32 GB now available
    Jan 6 21:05:11 Musa com.apple.backupd[7571]: Deleted 1 backups: oldest backup is now Sep 1, 2009
    Jan 6 21:05:11 Musa com.apple.backupd[7571]: Stopping backup.
    Jan 6 21:05:13 Musa com.apple.backupd[7571]: Backup canceled.
    Jan 6 21:05:22 Musa com.apple.backupd[7571]: Ejected Time Machine disk image.
    Jan 6 21:05:22 Musa com.apple.backupd[7571]: Compacting backup disk image to recover free space
    Jan 6 21:06:45 Musa com.apple.backupd[7571]: Completed backup disk image compaction
    Jan 6 21:06:45 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 21:06:45 Musa com.apple.backupd[7571]: Network destination already mounted at: /Volumes/odysseus
    Jan 6 21:06:48 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 21:06:48 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 21:06:57 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 21:12:57 Musa com.apple.backupd[7571]: Starting pre-backup thinning: 72.06 GB requested (including padding), 67.32 GB available
    Jan 6 21:12:57 Musa com.apple.backupd[7571]: No expired backups exist - deleting oldest backups to make room
    Jan 6 21:25:51 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2009-09-01-060212: 69.89 GB now available
    Jan 6 21:25:51 Musa com.apple.backupd[7571]: Deleted 1 backups: oldest backup is now Sep 8, 2009
    Jan 6 21:25:51 Musa com.apple.backupd[7571]: Stopping backup.
    Jan 6 21:25:53 Musa com.apple.backupd[7571]: Backup canceled.
    Jan 6 21:26:00 Musa com.apple.backupd[7571]: Ejected Time Machine disk image.
    Jan 6 21:26:00 Musa com.apple.backupd[7571]: Compacting backup disk image to recover free space
    Jan 6 21:27:30 Musa com.apple.backupd[7571]: Completed backup disk image compaction
    Jan 6 21:27:30 Musa com.apple.backupd[7571]: Starting standard backup
    Jan 6 21:27:30 Musa com.apple.backupd[7571]: Network destination already mounted at: /Volumes/odysseus
    Jan 6 21:27:33 Musa com.apple.backupd[7571]: Disk image /Volumes/odysseus/Musa_002332d5c37e.sparsebundle mounted at: /Volumes/Backup of Musa
    Jan 6 21:27:33 Musa com.apple.backupd[7571]: Backing up to: /Volumes/Backup of Musa/Backups.backupdb
    Jan 6 21:27:41 Musa com.apple.backupd[7571]: Node requires deep traversal:/ reason:must scan subdirs|
    Jan 6 21:33:37 Musa com.apple.backupd[7571]: Starting pre-backup thinning: 72.06 GB requested (including padding), 69.89 GB available
    Jan 6 21:33:37 Musa com.apple.backupd[7571]: No expired backups exist - deleting oldest backups to make room
    Jan 6 21:45:33 Musa com.apple.backupd[7571]: Deleted backup /Volumes/Backup of Musa/Backups.backupdb/Musa/2009-09-08-081225: 72.34 GB now available
    Jan 6 21:45:33 Musa com.apple.backupd[7571]: Pre-backup thinning completed successfully: 1 backups were deleted
    Jan 6 21:45:33 Musa com.apple.backupd[7571]: Backup date range was shortened: oldest backup is now Sep 15, 2009

  • HT201250 The initial backup is taking 2 days.  Can I stop and restart without any issues?

    Just got an iomega external hard disk backup and the initial backup is said to take two days.  Can I stop and restart without any issues?

    On a big backup, the initial estimate is often wrong and gets better after a little while. But, yeah, there is a "Stop backing up" command in the Time Machine menu, right?  That is safe as far as I know.
    charlie

  • Azure remote App with Azure SQL Backend

    Hi, If I had an application which utilised a SQL backend could I host the front end in Azure remote app and the backend on an Azure SQL database or would I need to use IAAS to provide the backend?
    Many thanks
    James 

    Hi James,
    If the application is designed to be compatible with Azure SQL Database then yes, you can use Azure RemoteApp for the frontend and Azure SQL Database for the backend.  If the application is not compatible with Azure SQL Database then you would
    need to use SQL Server running in an IaaS VM, or SQL Server running in an on-premises VM.  The IaaS VM or SQL Database should be in the same Azure region as the RemoteApp collection in order to minimize latency.
    For example, say you have an internally-developed application that uses SQL Server 2012 as its backend database.  Before you would be able to use Azure SQL Database you would first need to migrate the database to Azure SQL Database and make any changes
    necessary to the application and/or database for compatibility.  The SQL Database Migration Wizard is helpful for this process.
    -TP

  • Clone from hot backup failing with ORA-19870, 19505 unable to find backup

    Ok, the steps I am taking worked previously, the only thing that I know changed was the underlying file system and I am perplexed but then again I am not the most experienced dba so please forgive if I miss something obvious.
    My steps are: create backup using:
    backup database tag= tagname format '/archive/stage/DBNAME/%U';
    backup archivelog all format '/archive/stage/DBNAME/%U';
    backup current controlfile '/archive/stage/DBNAME/%U';
    Then on the new server, I copy the files over to the new server and:
    connect target username@prod
    connect auxillary
    set until time
    duplicate target database to "PRODSBX"
    etc...
    The problem is the duplicate is not looking at my '/archive/stage/DBNAME' directory for the backup. It's looking in the normal backup directory on that prod box that it can't connect to... can anyone give me clues as to what I should be looking at to figure out why? I looked at file permissions in case it couldn't access this directory, I looked at id, group etc... and I am not seeing anything. These exact same scripts worked before... the only thing I have edited is the set until time so I just can't figure out why it's not going to the /archive/stage.
    Any light shed in my general direction would be greatly appreciated.
    Thanks!
    Va.

    I have no way to copy the output of the files as this is on a closed network. If there is any particular information you are looking for I'd be happy to post it.
    The restore is looking for the default backup location as specified in my rman configuration but in my backup script, I specify a different location that I do the hot backup to. The first time it took this location and successfully cloned to the new server and now it seems to be ignoring it. I am not sure why, any suggestions in what I should look at, if this is a possible configuration issue on my clone server or on my primary?
    Thanks,
    Va.

  • How do I restore an old Iphone 3gs backup if it does not appear in the "restore backup" drop-down list?

    Hi
    Last night I tried to install ios 6.1.2. The install failed. I tried again and this time it recommended restoring the phone - I agreed and it worked, but after the restore was completed, I could not get out of the set-up assistant screens on the phone. It kept saying either connect to Itunes, or connected to Itunes - there was no conclusion of the set-up process.
    So again I chose to restore the phone - this time I got through the process and asked it to restore my last back up. It did so, but it turns out the back-up is empty. Looking at the time of the back-up, I think what happened is the phone  synced while it was stuck in the set-up assistant, and the old backup with all my data was overwritten with a new empty back up.
    If I use windows explorer I can see a back-up set from earlier in the evening that should contain all my data (6pm)  but when I look in iTunes that backup set is not available for selection in the "restore back-up" drop down list. Perhaps its only a partially completed backup, but at 1.55gb it sure contains a lot of valuable data. How can I make iTunes see or use the 6pm backup set?

    yes - sadly I think that's what happened. As the phone was connected, it was syncing and backing up automatically while I tried to complete the set-up process. So an empty back-up overwrote the full one. But the full back up is still there in the application data files - it must have done a back-up automatically before it restored the phone. So why can't I see it and use it in iTunes? Very frustrating!

  • I want to create an external video widget with iAd. I followed the steps and copied the Url and paste it in the Video pop-up menu in the Properties section of the inspector. When i select 'preview' the video doesn't play and the Url that i pasted is gone.

    i want to create an external video widget with iAd to use in an iBook. I followed the steps and copied the Url and paste it in the Video pop-up menu in the Properties section of the inspector. When i select 'preview' the video doesn't play and the Url that i pasted is gone.
    hope someone can help!

    May have something to do with whether or not the URL can be reached while building your widget.  I've got a video on the web. When I enter the URL into the field for the video object and click to another field,  the URL is replaced with some text indicating that it is accessing the video then updates itself to displaying just the name of the video file.
    It shows just fine in preview and I can see the full path to the video when I export the widget so it should work for you. 
    Is the video accessible from your Mac while you're laying out your widget?
    -Mark

  • In azure cloud service with Location West Europe, why the IP address shows the server is locate in United States?

    We have 9 projects in azure cloud service, and all are use West Europe location, but with different subscription, and we found that one of those cloud service which its IP address is United
    States(Why?) and others are all in Amsterdam(this is correct)
    Can someone explain why? the server which it's ip located in United
    States is very slow, BTW i'm in Amsterdam.

    Hi LH,
    I have saw the same problem on Brazil. There has some comments about this issue, you can refer to it:
    Microsoft owns large ranges of IP addresses which are typically registered in Redmond, so usually Azure IP addresses around the world show up as being
    physically located in Redmond when using these types of tools.
    It's more or less an issue with the way our IP are registered. They all "belong" to Microsoft in the US and Brasil
    IP locator tools like whatismyip are sometimes incorrect. Some will give the real location, some will give the location of the ISP, etc …
    IP geo lookup tools typically rely on a static database of IP address range registrations.
    It's a Microsoft issue in the sense that we may publish the correct location for our datacenters
    (eg. Amsterdam should locate in Europe and not in Redmond…) – but it's also a 3rd party tools issue.
    If you want to be sure – use
    http://msdn.microsoft.com/en-us/library/windowsazure/dn175718.aspx 
     And you could also  see this blog about this issue :
    http://azure.microsoft.com/blog/2014/06/11/windows-azures-use-of-non-us-ipv4-address-space-in-us-regions/
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • What do I do if imported AAC files wont copy onto my iPhone? It has worked in the past with other imported files of the same type.

    I have tried multiple times to sync a bunch of RHCP songs that i imported, but they wont copy onto my iPhone. I have tried isolating them in a playlist and syncing just that playlist and still nothing. please help. I have done this with other songs before and it worked fine but for some reason these ones wont copy.

    I suggest taking it back to Best Buy. Once it has been opened, Apple will not handle the phone. Best Buy or another 3rd party are your only choices. Best Buy probably made an error when replacing the screen. Understand they are not an authorized Apple Service Center.

  • Only import constraint from the full backup dump using datapump !

    Dear Firends ,
    I am using Oracle 10g in my production server . I firstly export the full database backup using expdp . Now I want to import just the constraints of my production server to a test server . Is it possible to import constraints only to a test server from a full dump backup ?
    plz inform ... ...

    Yet another 'I refuse to read the documentation, unless some volunteer spoon feeds me' question.
    What happened to the DBA community. Is there some special virus spreading all over the globe?
    Or is the virus called 'OTN' and is OTN making DBAs permanently lazy, trying constantly how to do a little as possible for their money.
    You can easily construct the answer yourself from this overview.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_import.htm#sthref389
    And please re-read the Forums Etiquette post again, or read it for the first time, stating you need to peruse the documentation prior to asking a question.
    Your assistance in diminishing the boot load of RTFM questions is appreciated.
    Sybrand Bakker
    Senior Oracle DBA

  • Report with reportData tag stil user the initial data

    I have created a report based upon a file, this works fine.
    when I want to schedule it using the <reportData>...</reportData> tag it still uses my initial data.
    The data that is provided is encode.
    This is my soap envelope:
    <soapenv:Envelope xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/">
    <soapenv:Body>
    <pub:scheduleReport xmlns:pub="http://xmlns.oracle.com/oxp/service/PublicReportService">
    <scheduleRequest>
    <deliveryRequest>
    <printOption>
    <printerName>Ontwik</printerName>
    <printNumberOfCopy>1</printNumberOfCopy>
    <printTray>Tray 2</printTray>
    <printSide>Double Sided Long Edge (Duplex)</printSide>
    </printOption>
    </deliveryRequest>
    <reportRequest>
    <attributeFormat>pdf</attributeFormat>
    <attributeTemplate>DefaultTemplate</attributeTemplate>
    <reportAbsolutePath>/~administrator/basedonFile/basedonFile.xdo</reportAbsolutePath> <reportData>50453576626C4E6F615842745A573530506A784462476C6C626E512B50474669596E4A6C646D6C6864476C76626A34354F544D3350433968596D4A795A585A700D0A595852706232342B504535686257552B51695A686258413755794242546C525852564A515043394F5957316C506A7776513278705A573530506A784A5A4756750D0A64476C6D61574E6864476C76626A3551515535454D446B7A4E5459794D444D384C306C6B5A57353061575A705932463061573975506A786862573931626E512B0D0A4D6A51334E4449344D7A77765957317664573530506A784F645731695A584A505A6B6C305A57317A506A457A4D6A7776546E5674596D567954325A4A644756740D0A637A3438546C4E514C55465556464A4A516C565552564D2B50464E5A553052424C564E356333526C5A57316B59585231625434784E79314B565577744D6A41770D0A4F54777655316C545245457455336C7A6447566C6257526864485674506A7856546B6C554C57566C626D686C6157512B517A77765655354A5643316C5A57356F0D0A5A576C6B506A785454314A555279317A623239796443426E6232566B5A584A6C626A354852447776553039535645637463323976636E51675A32396C5A4756790D0A5A57342B50464242546B52484C584268626D52325A584A7A64484A6C6132746C636A34354F544D335043395151553545527931775957356B646D5679633352790D0A5A5774725A58492B5045684A553152504C55687063335276636D6C6C617A34384C30684A553152504C55687063335276636D6C6C617A343856466C51543049740D0A64486C775A5342685A6D64705A6E526C596D56336157707A506C41384C31525A554539434C5852356347556759575A6E61575A305A574A6C64326C71637A34380D0A51306843525667745132686C59327469623367675A5868775A584A3061584E6C506C6B384C304E49516B56594C554E6F5A574E72596D393449475634634756790D0A64476C7A5A54343851306843533038745132686C59327469623367676132397A64475675506C6B384C304E49516B74504C554E6F5A574E72596D3934494774760D0A6333526C626A34384C303554554331425646525353554A5656455654506A7776546D3975553268706347316C626E512B</reportData><sizeOfDataChunkDownload>-1</sizeOfDataChunkDownload>
    </reportRequest>
    <userJobName>PrintJobs</userJobName></scheduleRequest><userID>Administrator</userID><password>XXXXXXXXXXX</password>
    </pub:scheduleReport>
    </soapenv:Body>
    </soapenv:Envelope>
    any one has a cleu ?

    same here; if i run runReport it works okay..when is use ScheduleReport it shows the intial data(preview)
    Do you already know of a solution?

  • Backup Deletion with DB13?

    Hey there,
    I have planned some AllOnlineRedolog Backups to a disk. But there is only enough space for ONE Backup. So, after the Backup on the disk has been saved to a tape, it should be deleted. Not directly after the Tape-Backup but  relatively before the new AllOnlineRedolog Backup.
    Example:
    01.01.2008 - 3:00am - AllOnline + RedoLog Backup (to Disk)
    01.01.2008 - 8:00pm - Backup to Tape
    08.01.2008 - 1:00am - Deletion of old Backup
    08.01.2008 - 3:00am - AllOnline.... and so on.
    Is this possible just with DB13? How could I realize this?
    Thanks in advance!

    Hi there,
    If you are using the DB13 you probably are familiar with the profile file init<SID>.sap in that file you can configure the disk path if you didn't then the backups goes to the default path that is %SAP_DATA%\sapbackup.  In the same file you can configure the retention period of that files and all the logs created during backup activities.
    There is an activity pattern in the DB13 called Clean up logs, that pattern could be configured to delete old backup files.
    We implement exactly the way you want, and in the init<SID>.sap located in %ORACLE_HOME%\database\ we set the parameters as follow:
    retention period in days for archive log files saved on disk
    default: 30
    cleanup_disk_archive = 14
    This means that offline readologs will be deleted if they are 14 days old
    retention period in days for database files backed up on disk
    default: 30
    cleanup_disk_backup = 1
    This means that all datafiles only will be available 1 day and at the next day when you run "Clean up logs" activity the previous backup will be deleted.
    You can configure that for example the Clean up runs every day one our before the backup and with that you first delete the old backup and then you make a new one.
    You can use the scheduler task if you want but if you use the DB13 you dont need to enter to you server every day to review the logs. We configure the scheduler task for our Portal because has no ABAP instance .
    the instructions in the .bat file coul be "brconnect -u / -c force -p init<SID>.sap -f cleanup " and that is actually the pattern in DB13.
    You can review the documentation of brconnect for more details.
    http://help.sap.com/saphelp_sm40/helpdata/EN/50/7dd41742210144aee3fdee21c553eb/content.htm
    Regards.
    Gustavo Balboa

Maybe you are looking for