Importing Virtualbox VM into pool fails with any errors VDI 3.0

I'm at the last stage of importing my VM's into mypool which I created and the importing fails with Job failed. I do not have large space /var/tmp and have setup a link to the VM's
FINER: THROW
com.sun.vda.service.api.ServiceException: File copy failed for server 192.168.0.25.
at com.sun.vda.service.vbox.Storage.copyFileToVolume(Storage.java:396)
at com.sun.vda.service.vbox.VBDesktopProvider.importImage(VBDesktopProvider.java:1300)
at com.sun.vda.service.vbox.VBDesktopProvider.newDesktopInstance(VBDesktopProvider.java:225)
at com.sun.vda.service.core.jobs.ImportJob.runJob(ImportJob.java:49)
at com.sun.vda.service.core.jobs.Job.run(Job.java:122)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:417)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:269)
at java.util.concurrent.FutureTask.run(FutureTask.java:123)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:650)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:675)
at java.lang.Thread.run(Thread.java:595)
Caused by: java.io.FileNotFoundException: (No such file or directory)
at java.io.RandomAccessFile.open(Native Method)
at java.io.RandomAccessFile.<init>(RandomAccessFile.java:212)
at com.sun.vda.service.vbox.Storage.copy(Storage.java:773)
at com.sun.vda.service.vbox.Storage.copyFileToVolume(Storage.java:386)
... 10 more
Apr 4, 2009 1:40:15 PM com.sun.vda.service.persistence.PersistentJob commit

Thomas,
Thanks for your reply. I have finished the VDI 3.0 Install and everything works fine now. For some reason even though the Storage OS was opensolaris 2008.11 but running on a VM it did not like that and once I dedicated a seperate box, then things worked fine. I just redid the install of the VM to try if it goes through and this time it works well. So my setup as of now has everything running on a ASUS MBD whitebox with Open Solaris running as a VM in Virtual Box and acting as a ZFS Storage (Virtualbox is running inside the GA release. However I would like to run everything in one OS which would be the GA. I tried setting up the iSCSI interface in GA and it did not work. Any ideas on how I can make this work on just one platform ?

Similar Messages

  • Import CTS+ transport into Portal failed with Returncode:'12'

    Hi, Gurus!
    I have encountered the problem while deploying CTS+ transport to the Portal (system XO3): Returncode:'12' class java.io.IOException:
    The file (
    \solman.mycompany.com\sapmnt\trans\data\BIGK900007\ctsplus.par) does not exist on the filesystem.
    Full log text:
    | ######################################
    | Deployment
    | Transport request : BIGK900007
    | System : XO3
    | tp path : tp | Version and release: 372.04.71 700 
    | Deploy Service called with following parameters:
    | CommunicationData-Type:0
    | CommunicationData-Address:http://portalXO3.mycompany.com:50218
    | CommunicationData-Host:portalXO3.mycompany.com
    | CommunicationData-Port:50218
    | CommunicationData-User:'filled'
    | CommunicationData-Password:'filled'
    | Deployable(0) | Deployable-Type:0
    | Deployable-Id:
    solman.mycompany.com\sapmnt\trans\data/BIGK900007/ctsplus.par
    | Returncode:'12' | class java.io.IOException:The file (
    \solman.mycompany.com\sapmnt\trans\data\BIGK900007\ctsplus.p
    | >>>ar) does not exist on the filesystem.
    | Deployment
    | End date and time : 20100106154329
    | Ended with return code: ===> 12 <===
    | ######################################
    The file ctsplus.par does exists at the right place and available from the Portal machine.
    Especially strange looks the Deployable-Id: *
    solman.mycompany.com\sapmnt\trans\data/BIGK900007/ctsplus.par*
    Maybe the different slashes "\" and "/" are the cause of the porblem? Both systems (Portal and Solution manager) run on Windows 2003 Servers.
    I have checked the "Note 1232674 - class java.io.IO exception during import" and set up the full control permissions to access "\sapmnt\" for the local (solman) user SAPServiceXO3 but it does not help.
    I've created a local "SAPServiceXO3" user in the server of Solman and have given to it highest permissions to the file share.  Has anybody any idea how to resolve the problem?
    Thanks in advance.
    Edited by: Mikhail Palstsianau on Jan 28, 2010 8:47 AM

    Hi,
    Due you have common trans directory for all the system ???
    Thanks
    Anil

  • WebLogic 11g data source connection pooling failed with IO error:socket read timed out.

    Hi all,
    We encountered IO Error: Socket read timed out( <Received exception while creating connection for pool "DS_1": IO Error: Socket read timed out> ) during the creation of data sources in WebLogic 11g. Manual data source testing seems to indicate intermittent connection and the server seems to take a long time to start up with multiple IO errors. We increased the timeout at the database side but it does not seems to help. The database is 11g (11.2.0.3). The database services and listener are up which does not indicate that the database instance is down.

    This particular error means your database is down and socket is timedout. Or the query takes so long that the reader timedout -- unlikely.
    Also, the general weblogic socket write errors, which you might see sometimes means that the client got disconnected before the server wrote results back. This is harmless

  • IMP-00017: following statement failed with ORACLE error 25153

    Sirs,
    Any one can guide me on below errors getting while importing the data
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    Export file created by EXPORT:V09.02.00 via conventional path
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P1 character set (possible charset conversion)
    . importing SWTHIST's objects into SWTHIST
    . . importing partition "AUTHHIST_2006":"AUTH_09_06" 2857788 rows imported
    IMP-00017: following statement failed with ORACLE error 25153:
    "CREATE INDEX "AUTH2006_1" ON "AUTHHIST_2006" ("BR_ID" , "CARD_NO" , "SYSTEM"
    "_REQ_DTTIME" ) PCTFREE 10 INITRANS 2 MAXTRANS 255 STORAGE(INITIAL 65536 FR"
    "EELISTS 1 FREELIST GROUPS 1) TABLESPACE "SWTIDX" LOGGING"
    IMP-00003: ORACLE error 25153 encountered
    ORA-25153: Temporary Tablespace is Empty
    IMP-00057: Warning: Dump file may not contain data of all partitions of this table
    Import terminated successfully with warnings.

    Hi,
    I have seen this error when you have a temporary tablespace without any datafile associated with it. I don't know how that happens, but that is what I have seen.
    ORA-25153: Temporary Tablespace is Empty
    Try altering the temp tablespace and add a tempfile.
    Dean

  • IMP-00017: following statement failed with ORACLE error 1031

    Hello,
    I export a schema from prod (database A) and I try to imported to onther test database (databse B).
    database A: single database - Non ASM - 10gR1
    database B: RAC database - ASM -10gR2
    To import database I use this:
    imp osiris/b@osiris file=prod_osiris070601.dmp full=y ignore=y log=osiris.log
    All the tables are imported withot problem but when he start to imprt views I get this error:
    IMP-00017: following statement failed with ORACLE error 1031:
    Can you please help?
    Thanck you in advance.

    I thing that the problem is that the user can not create views!!!
    I create the user normaly and I give him this rols: "CREATE SESSION", "CONNECT" and "RESOURCE"
    it's ok?
    . . importing table "ZIP_CODE" 70768 rows imported
    IMP-00017: following statement failed with ORACLE error 1031:
    "CREATE FORCE VIEW "OSIRIS"."DETAILED_FRANKING_SYSTEM_VIEW" "
    " ("CONFIGURATION_SYSTEM_ID","SYSTEM_ID","DEVICE_SERIAL_NO","OPCO_DE"
    "VICE_SERIAL_NO","STATUS","BO_CUST_ID","CUSTOMER_NAME","CUSTOMER_ADDRESS_NAM"
    "E","CUSTOMER_STREET1","CUSTOMER_STREET2","CUSTOMER_STREET3","CUSTOMER_STREE"
    "T4","CUSTOMER_CITY","CUSTOMER_STATE","CUSTOMER_ZIP4","CUSTOMER_ZIP","CUSTOM"
    "ER_COUNTRY","CUSTOMER_NAME_CAP","CUSTOMER_ADDRESS_NAME_CAP","CUSTOMER_STREE"
    "T1_CAP","CUSTOMER_STREET2_CAP","CUSTOMER_STREET3_CAP","CUSTOMER_STREET4_CAP"
    "","CUSTOMER_ADDRESS_STREET_CAP","CUSTOMER_CITY_CAP","CUSTOMER_STATE_CAP","C"
    "USTOMER_COUNTRY_CAP","INSTALL_ADDRESS_NAME","INSTALL_STREET1","INSTALL_STRE"
    "ET2","INSTALL_STREET3","INSTALL_STREET4","INSTALL_CITY","INSTALL_STATE","IN"
    "STALL_ZIP4","INSTALL_ZIP","INSTALL_COUNTRY","INSTALL_ADDRESS_NAME_CAP","INS"
    "TALL_STREET1_CAP","INSTALL_STREET2_CAP","INSTALL_STREET3_CAP","INSTALL_STRE"
    "ET4_CAP","INSTALL_ADDRESS_STREET_CAP","INSTALL_CITY_CAP","INSTALL_STATE_CAP"
    "","INSTALL_COUNTRY_CAP","DESCRIPTION","CONFIG_MODEL_ID","TYPE","PART_NUMBER"
    "AND DEVICE_SYSTEM.DEVICE_TYPE='METER'"
    "AND US_SERIAL_NO_CONVERTER.DEVICE_SERIAL_NO(+) = DEVICE_SYSTEM.DEVICE_SERIA"
    "L_NO"
    IMP-00003: ORACLE error 1031 encountered
    ORA-01031: insufficient privileges
    IMP-00017: following statement failed with ORACLE error 942:
    "GRANT DELETE ON "DETAILED_FRANKING_SYSTEM_VIEW" TO "APP""
    IMP-00003: ORACLE error 942 encountered
    Message was edited by:
    ASkika

  • Can not log into server computer with any accounts - "You are unable to login to the user account "abcdefg" at this time. Logging in to the account failed because an error occurred."

    I have a Mac mini running the latest version of OS X and Server. Been running fine and flawlessly. However, I had a strange problem with the iCloud preferences panel crashing when I tried to access it, so I rebooted. Now I can not log into the system with any accounts. My master admin account (along with all the others) gives me the error:
    You are unable to login to the user account "abcdefg" at this time. Logging in to the account failed because an error occurred."
    I am able to see the server from other macs and I can log into it using the same account, but it only shows me a few of the shared folders I have access to but NOT to my main directories.
    Rebooting into Command-R and doing a disk utility, I try and repair permission on that drive and get a bunch of errors like:
    ACL found but not expected on Users
    Repaired "Users"
    ACL found but not expected on Users/.localized
    Repaired "Users/.localized"
    ACL found but not expected on Users/Shared
    Repaired "Users/shared"
    ACL found but not expected on Users/Shared/.localized
    Repaired "Users/Shared/.localized"
    Permissions repair complete.
    But rebooting is no joy...same problem. Any idea what is going on or how to repair it? Should I do a time machine restore? Complete new OS X install? Any idea what is causing this or how to salvage it?

    Got everything to re-install and it worked fine...for a few hours. Then I came in to find ALL of my network users deleted. Just GONE. Then found out the Open Directory was trashed and was unable to open, recover or restore from a backup. Looks like I may have a bad drive here.
    I installed a new drive in the system, re-installed and so far (for a couple of hours anyway) the system seems to be working and stable.

  • Import error imp-00017: following statement failed with oracle error 1659:

    Hi all,
    I am trying to import a file (dmp) which is of 1.63GB size,
    when I type
    imp
    dbname/pwd@db
    c:\filepath
    30720
    no
    no
    yes
    yes
    yes
    I am getting the error imp-00017: following statement failed with oracle error 1659:
    -- end it says Ora-01659 unable to allocate minextents beyond 1 in tablespace USERS
    --unable to create INITIAL extent for segment in tablespace USERS
    --import terminated successfully with warnings
    when I looked into
    select maxbytes,file_name from dba_data_files where tablespace_name='USERS';3.4360+10 c:\oracle\product\10.2.0\oradata\orcl\users01.dbf
    kindly help me in solving this is.
    Regards,
    aak
    Edited by: AAK 460425 on Apr 12, 2009 6:33 PM

    Thanks,
    select bytes from user_free_space where tablespace_name='USERS'
    BYTES
    458752
    720896
    458752
    655360
    43384832
    is the size less?
    My apology I have updated this in the below thread, which is same as this one.
    Re: Import error

  • Import failing with no errors

    Import failing with no errors
    All it says in logs for each table is
    . . skipping table "<table_name>"
    It's a full import from a 10.2.03 db into a 10.2.0.3 db

    Hi,
    "INDEXFILE" , this is parameter is specified, index-creation statements for the requested mode are extracted and written to the specified file, rather than used to create indexes in the database. No database objects are imported.
    Due to that that Objects are skipping.
    The INDEXFILE parameter can be used only with the FULL=y, FROMUSER, TOUSER, or TABLES parameters.
    Else you can perform a two step process as stated by Amardeep.
    - Pavan Kumar N

  • DbSl Trace: CONNECT failed with sql error '12518'  at IMPORT ABAP

    Hi
    We are trying to install ECC 6.0 on windows 2008 SP2 and oracle 10.2.0.4. We get an error at IMPORT ABAP phase.
    content of import_monitor.java :
    java version "1.4.2_26"
    Java(TM) Platform, Standard Edition for Business (build 1.4.2_26-b03)
    Java HotSpot(TM) 64-Bit Server VM (build 1.4.2_26-b03, mixed mode)
    Import Monitor jobs: running 1, waiting 1, completed 26, failed 0, total 28.
    Loading of 'SAPSLOAD' import package: ERROR
    Import Monitor jobs: running 0, waiting 1, completed 26, failed 1, total 28.
    content of SAPSLOAD.log:
    E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe: START OF LOG: 20100705173108
    E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe: sccsid @(#) $Id: //bas/701_REL/src/R3ld/R3load/R3ldmain.c#5 $ SAP
    E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe: version R7.01/V1.4 [UNICODE]
    Compiled Feb 24 2009 23:41:07
    E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe -i SAPSLOAD.cmd -dbcodepage 4103 -l SAPSLOAD.log -stop_on_error
    DbSl Trace: OCI-call 'OCIServerAttach' failed with rc=12518
    DbSl Trace: CONNECT failed with sql error '12518'
    DbSl Trace: OCI-call 'OCIServerAttach' failed with rc=12518
    DbSl Trace: CONNECT failed with sql error '12518'
    (DB) ERROR: db_connect rc = 256
    DbSl Trace: OCI-call 'OCIServerAttach' failed with rc=12518
    DbSl Trace: CONNECT failed with sql error '12518'
    DbSl Trace: OCI-call 'OCIServerAttach' failed with rc=12518
    DbSl Trace: CONNECT failed with sql error '12518'
    (DB) ERROR: DbSlErrorMsg rc = 99
    E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe: job finished with 1 error(s)
    E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe: END OF LOG: 20100705173108
    content of sqlnet.log file:
    Fatal NI connect error 12560, connecting to:
    (DESCRIPTION=(ADDRESS=(PROTOCOL=BEQ)(PROGRAM=oracle)(ARGV0=oracleUED)(ARGS='(DESCRIPTION=(LOCAL=YES)(ADDRESS=(PROTOCOL=beq)))'))(CONNECT_DATA=(SID=UED)(CID=(PROGRAM=E:\oracle\UED\102\bin\sqlplus.exe)(HOST=U34SSAPUED)(USER=uedadm))))
      VERSION INFORMATION:
         TNS for 64-bit Windows: Version 10.2.0.4.0 - Production
         Oracle Bequeath NT Protocol Adapter for 64-bit Windows: Version 10.2.0.4.0 - Production
      Time: 05-JUL-2010 16:19:39
      Tracing not turned on.
      Tns error struct:
        ns main err code: 12560
        TNS-12560: TNS:protocol adapter error
        ns secondary err code: 0
        nt main err code: 530
        TNS-00530: Protocol adapter error
        nt secondary err code: 2
        nt OS err code: 0
    Fatal NI connect error 12518, connecting to:
    (DESCRIPTION=(ADDRESS=(COMMUNITY=SAP.WORLD)(PROTOCOL=TCP)(HOST=u34ssapued)(PORT=1527))(CONNECT_DATA=(SID=UED)(GLOBAL_NAME=UED.WORLD)(CID=(PROGRAM=E:\usr\sap\UED\SYS\exe\uc\NTAMD64\R3load.exe)(HOST=U34SSAPUED)(USER=uedadm))))
      VERSION INFORMATION:
         TNS for 64-bit Windows: Version 10.2.0.2.0 - Production
         Windows NT TCP/IP NT Protocol Adapter for 64-bit Windows: Version 10.2.0.2.0 - Production
      Time: 05-JUL-2010 17:28:38
      Tracing not turned on.
      Tns error struct:
        ns main err code: 12564
        TNS-12564: TNS:connection refused
        ns secondary err code: 0
        nt main err code: 0
        nt secondary err code: 0
        nt OS err code: 0
    Best Regards...,

    Hi,
    TNS-12560: TNS:protocol adapter error
    Probably your listener or database is down...
    If both are running, just stop the installation and reboot the server.
    Then again start listener and database and continue SAP installation with "Continue old installation" option..
    Also make sure that below environment variables are set properly...
    ORACLE_HOME
    ORACLE_SID
    DBSID
    Regards.
    Rajesh Narkhede

  • Hello, i need help i'm doing experiment on ilab i have managed to put my experiment on the platform but when i try to launch it fails with an error 404 that the page not found. any one with idea of what i should do to solve this problem. thank you

    hello, i need help i'm doing experiment on ilab i have managed to put my experiment on the platform but when i try to launch it fails with an error 404 that the page not found. any one with idea of what i should do to solve this problem. thank you

    hey Sally,
    Can you elaborate on the ilab and what you are trying to do? I have found this page to explain error 404, but I am not sure if it directly applies to  you
    http://digital.ni.com/public.nsf/allkb/D0998D3734926FA08625782300016E40?OpenDocument
    Regards,
    A. Zaatari
    National Instruments
    Applications Engineer

  • Disco command line import fails with parser error 7

    Hi, we have discoverer version 10.1.2.48.18 installed to multiple environments.
    Developers have created eex-file and we try to import it to all the test environments, it goes through fine with some but with some it fails with parser error 7 :
    Here is the full error log:
    11/13/2008 7:30:22 AM
    D:\disco1012\BIToolsHome_1\bin\DIS51ADM.EXE /connect /import oudata_disco.eex /identifier /refresh /preserve_workbook_owner /show_progress /log import.log
    oudata_disco.eex:Could not locate the Folder with identifier 'INSTRUCTOR_ASSIGNMENTS1' in the target End User Layer
    oudata_disco.eex:An imported Folder had display name 'Dc transactional sales q2' renamed to 'Dc transactional sales q2 1'
    oudata_disco.eex:An imported Item Class had display name 'Eval_Delivery_Region' renamed to 'Eval_Delivery_Region 1'
    oudata_disco.eex:An imported Item Class had display name 'Quarter' renamed to 'Quarter 1'
    oudata_disco.eex:An imported Item Class had display name 'Track List with ALL' renamed to 'Track List with ALL 1'
    The Item Hierarchy with identifier 'INSTRUCTOR_ASSIGNMENTS_LAST_REFRESH_DATE_DEFAULT_DATE_HIERARCHY' has not been imported because: There are no items in this hierarchy node
    Import completed, but with warnings. Please check the result.
    File(s) imported partially :
    oudata_disco.eex
    11/13/2008 7:36:38 AM
    11/13/2008 7:37:29 AM
    D:\disco1012\BIToolsHome_1\bin\DIS51ADM.EXE /connect /import oudata_disco_workbooks_84.eex /identifier /refresh /preserve_workbook_owner /show_progress /log import.log
    oudata_disco_workbooks_84.eex:A parsing failure has occurred in file 'oudata_disco_workbooks_84.eex'.
    Parser error: '7'
    The import has failed (your End User Layer has not been modified) - oudata_disco_workbooks_84.eex:A parsing failure has occurred in file 'oudata_disco_workbooks_84.eex'.
    Parser error: '7'
    If anyone knows what can be teh cause of this please let me know.
    thanks,
    Nina

    Hi,
    The most likely cause is that the export file has become corrupted. Are there very big export files?
    Another possibility is that Discoverer Administrator has run out of memory while parsing the file.
    Rod West

  • External Drive Disk Erase Failed with the Error Input/Output Error

    I have 2 hard drives in an external FW800 enclosure that I am unable to format. When I go to initialize the drives in Disk Utility, I get the following error message: "Disk Erase failed with the error: Input/output error."
    The drives show up in Disk Utility, but I can't repair them (that option is grayed out). Disk Utility correctly ID's the manufacturer of the drives (Maxtor), their size (200gb each), so it's obviously seeing that the drives are there. But it won't let me format them.
    The drives are new, by the way; they don't have any data/files on them. I have Disk Warrior, but the drives don't show up there to be repaired -- probably because they aren't formatted yet.
    After looking at other posts, I tried switching the jumper settings around on the drives -- from Master/Slave to cable select and back again, but it didn't help. I also tried doing a zero erase (even though the drives are new), zapping the PRAM -- again, no help.
    One question I had is whether this could be a bad FW800 cable? The cable is new -- it came with the enclosure, which is an OWC Dual FW 800 enclosure. Other than that, does anyone have any other thoughts about what's causing this? Any help would be greatly appreciated.
    Matthew

    SOLUTION!!!!
    I had the exact same problem. I have the original 20 GB hard drive that came in my Powerbook G4 550MHz and a couple of years ago I traded up for a 60 GB drive and bought a FW/USB enclosure for my original drive to use it to backup my important files. I hadn't backed up in over a year (shame on me!) and I decided maybe I should erase the drive and start from scratch. It was connected via USB.
    At that point DiskUtility gave me the exact same Input/Output error. I tried partitioning the drive into 1 or more partitions but came up with the same error. I couldn't figure out what was wrong so I decided to startup in OS 9.2.2, I did that and let it start up, then plugged in the hard drive and it gave me the standard "This disk is unrecognizable, do you want to eject or erase?" so I clicked Initalize. It worked!
    Just make sure you choose the MacOS Extended option when initializing out of OS 9 (instead of the MacOS Standard option) so it can be read and viewed in OS X.
    If your computer is too new to be able to boot from an OS 9 folder on your drive or an OS 9 CD, then see if a friend or a local library has older computers that are running OS 9 or can boot from it. If not let me know and you can send me your drive and I'll reformat it.
    Kind of crazy...I haven't used the OS 9 partition on my HD in YEARS...was even thinking about erasing it since I don't use any Classic applications anymore...good thing I didn't!
    Nick
    Powerbook G4 550Mhz   Mac OS X (10.4.6)  

  • Transfer ID failed with LUM errors

    I have 4 netware servers to migrate to OES2. One was completed successfully using Transfer id a few weeks ago. This weekend I attempted another.
    The last part of the transfer id failed with LUM errors that it couldnt access lum objects admingroup, www etc. As we have had one successful migration these objects do exist in the tree and are lum enabled. The /etc/nam.conf was modified prior to starting the transfer id to point to the ldap server with the master replica. I did create too many projects which I think caused me the lum errors which I couldn't get past. With the source server turned off I followed a TID dated 2009 for this LUM error and modified the unix object which had been created to be members of the groups it required, (luckily I had the previously migrated server unix object to compare to) quit the migration software and rebooted the server and then ran ndsrepair. No errors. I did not migrate any services.
    Further investigation revealed errors relating to obtaining the SSLCert but it has not expired and despite the cant get it messages the files .der and .pem appeared in the correct folders on the target server. I did copy the .der file to the required location naming it with the ipaddress as per manual instructions. The target server now has a case of dual identity. The nss volumes are mounted ok and the users can map drives using the source server name but the Suse Linux server still thinks that its the pre-migration server when I run a terminal session. Ifconfig shows the correct source server ip. Consoleone is another story. The volume objects are the old ones and the new ones with pools are not showing. From the users viewpoint it works.
    Before I do more damage can someone please point me in the right direction. The target name did change to the source before the restart so therefore I should just need to do the hostname rename manually. Should I start at step 7 in the manual instructions in an attempt to fix this server or should I abandon it and start again?

    $(UKevin
    This defect will make into the next OES2SP2 Patch (post September). We didn't have enough time for QA to regress the fix for September Patch.
    -Ramesh
    >>>
    From:
    kjhurni<[email protected]>
    To:
    novell.support.open-enterprise-server.migration
    Date:
    10/05/2010 09:06 PM
    Subject:
    Re: Transfer ID failed with LUM errors
    Great, thanks!
    As I mentioned previously, we DID find a bug about 3 weeks ago with LUM
    repairfailing during the "group" (admingroup). We do have a field
    patch for it, but obviously it's not made it into the channel yet.
    So if you DO still have that problem, you should probably ask for the
    patch (it's one file).
    The defect #635574
    Should help them reference the bug and get the patch for you.
    I'm not sure when it will be released into the patch channel.
    Not to say this fixes any and all LUM problems but since we have that
    patch, we've not had any issues with the remaining OES2 SP2 64-bit
    servers we've done with ID Transfer.
    kjhurni
    kjhurni's Profile: http://forums.novell.com/member.php?userid=734
    View this thread: http://forums.novell.com/showthread.php?t=422454

  • IMP-00017: following statement failed with ORACLE error 1276

    i have exported the full database without rows from live database and tried to import it into training database.
    both use ASM file system.i have precreated tablespace in training database.
    but im getting the error IMP-00017: following statement failed with ORACLE error 1276.
    it tries to create tablesapce and datafile on live ASM location rather than creating it on training database ASM location.
    live database asm location is WEBD and training database is at TRAIND.
    it tries to import on WEBD RATHER THAN TRAIND and throws the error.
    any help is appreciated.i have googled it and cant find anything helpful.
    Im bit new to DBA stuff.
    is there anyway to import and export data in ASM file system?
    Thanks

    When datapump exports the tablespace information, all it knows about is the currently location of your datafiles. Let's take a simple example. Let's say you have tablespace foo with a file called foo.dbf and it lives in a windows directory c: if you are now trying to import that into a linux box, Data Pump does not know that. All it knows is that it needs to create the tablespace data file in c:. If you are changing the locations of the datafiles, you need to use remap_datafile=old_datafile:new_datafile ... in your impdp command. Or, precreate them and use exclude=tablespace. The old_datafile and new_datafile need to be the complete path.
    Hope this helps.
    Dean

  • Payables Open Interface Pgm fails without any error(Very urgent)

    All,
    I'm into an integration project where in I'm pushing Invoice doc from a third party appln to Oracle apps (11.5.10.2) thru payables open interface. The integration is smooth till today. We pushed a record today, but the concurrent neither picks up the data but failed with 'completed error' status. When I delete the same record from the interface, the concurrent runs fine. So we are sure there is some issue with the data in my record. Since concurrent is failed without any reason, from where I can get the log to know what is the actual problem with that record? Pls help me with your inputs.
    Thanks,
    Sen

    Sen,
    To enable trace, please refer to this document (skip the part about generating TKPROF, just check the concurrent request output file if more details about the error are logged after enabling trace and setting the profile option).
    Note: 453527.1 - How To Trace a Concurrent Request And Generate TKPROF File
    https://metalink2.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=453527.1
    The database log file can be usually found under $ORACLE_HOME/admin/<SID>_<hostname>/bdump directory (or from SQL*Plus, issue "show parameter background_dump_dest").
    Regards,
    Hussein

Maybe you are looking for