Expdp fails to create .dmp files in NFS mount point in solaris 10,Oracle10g

Dear folks,
I am facing a wierd issue while doing expdp with NFS mount point. Kindly help me on this.
===============
expdp system/manager directory=exp_dumps dumpfile=u2dw.dmp schemas=u2dw
Export: Release 10.2.0.4.0 - 64bit Production on Wednesday, 31 October, 2012 17:06:04
Copyright (c) 2003, 2007, Oracle. All rights reserved.
Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
ORA-39001: invalid argument value
ORA-39000: bad dump file specification
ORA-31641: unable to create dump file "/backup_db/dumps/u2dw.dmp"
ORA-27040: file create error, unable to create file
SVR4 Error: 122: Operation not supported on transport endpoint
I have mounted like this:
mount -o hard,rw,noac,rsize=32768,wsize=32768,suid,proto=tcp,vers=3 -F nfs 172.20.2.204:/exthdd /backup_db
NFS=172.20.2.204:/exthdd
given read,write grants to public as well as specific user

782011 wrote:
Hi sb92075,
Thanks for ur reply. pls find the below. I am able to touch the files while exporting log files also creating having the error msg as i showed in previous post.
# su - oracle
Sun Microsystems Inc. SunOS 5.10 Generic January 2005
You have new mail.
oracle 201> touch /backup_db/dumps/u2dw.dmp.test
oracle 202>I contend that Oracle is too dumb to lie & does not mis-report reality
27040, 00000, "file create error, unable to create file"
// *Cause:  create system call returned an error, unable to create file
// *Action: verify filename, and permissions                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

Similar Messages

  • Unable to do expdp on NFS mount point in solaris Oracle db 10g

    Dear folks,
    I am facing a wierd issue while doing expdp with NFS mount point. Kindly help me on this.
    ===============
    expdp system/manager directory=exp_dumps dumpfile=u2dw.dmp schemas=u2dwExport: Release 10.2.0.4.0 - 64bit Production on Wednesday, 31 October, 2012 17:06:04
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-39001: invalid argument value
    ORA-39000: bad dump file specification
    ORA-31641: unable to create dump file "/backup_db/dumps/u2dw.dmp"
    ORA-27040: file create error, unable to create file
    SVR4 Error: 122: Operation not supported on transport endpoint
    I have mounted like this:
    mount -o hard,rw,noac,rsize=32768,wsize=32768,suid,proto=tcp,vers=3 -F nfs 172.20.2.204:/exthdd /backup_db
    NFS=172.20.2.204:/exthdd

    Hi Peter,
    Thanks for ur reply.. pls find the below. I am able to touch the files while exporting log files also creating having the error msg as i showed in previous post.
    # su - oracle
    Sun Microsystems Inc. SunOS 5.10 Generic January 2005
    You have new mail.
    oracle 201> touch /backup_db/dumps/u2dw.dmp.test
    oracle 202>

  • The distribution agent failed to create temporary files in 'C:\Program Files\Microsoft SQL Server\110\COM' directory. System returned errorcode 5.

    One of my replication job failed with below error, which means job failed because of permission issue on the folder. To resolve that, I have granted the permission and it started working fine. Now the question is as my job was running fine since long time
    without any issue then how suddenly the permission of the folder got omitted automatically? I am sure that no one deleted the permission from the folder and as the server is part of SQL cluster no failover occurs.
    Error:-The distribution agent failed to create temporary files in 'C:\Program Files\Microsoft SQL Server\110\COM' directory. System returned errorcode 5.
    Rahul

    Hi Rahul,
    Error code 5 indicates that the error is "access is denied."  Please make sure that the COM folder is excluded from any antivirus scan that occurs on the system.
    If the issue still persists, I recommend you use
    Process Monitor to find out a bit more about the accounts and access requests involved. Then grant write permission to the exact COM folder for the account that is running the SQL Server Agent service. For more details, please review the following similar
    thread.
    http://www.sqlservercentral.com/Forums/Topic1364832-1550-1.aspx
    Thanks,
    Lydia Zhang
    If you have any feedback on our support, please click
    here.

  • Distribution agent failed to create temporary files in

    Hello,
    we are configured transnational replication in sql server 2005 location, all replication jobs are running fine,
    while we checking sql server errorlog showing below error :
    Distribution agent failed to create temporary files in  'C:\Program Files\Microsoft SQL Server\100\COM' directory. System returned errorcode 5.
    NOte: normally we are configured replication 2005 [90] , but while see the above error it shows [100] comparability level . 

    Hi,
    If the account that is running SQL Server Agent does not have write access to the COM folder, the Distribution Agent will fail
    when it is running as a job. If you run the Distribution Agent from a command line by using an account that does not have write access to the COM folder, the same failure will occur.
    To work around this
    issue, grant write permissions to the COM folder for the account that is running the SQL Server Agent service. If you run the Distribution Agent from a command line, grant write permissions to the COM folder for the account that you use to run the Distribution
    Agent.
    For more information, please go through the below URL:
    http://support.microsoft.com/kb/956032/en-us  

  • Expdp fail and create table SYS_EXPORT_SCHEMA_20

    Hi Gurus
    I am using Oracle 10.2.0.3 in AIX env
    My database size is around 1600 GB. Sometime my expdp fail and create table like SYS_EXPORT_SCHEMA_20, SYS_EXPORT_SCHEMA_05. As I run expdp from system user , I notice that it create this type of table into system tablespace. It time it consume around 5gb space. Now my system tablespace size is 68 GB.
    Can I drop those table? If I drop these table then it create any problem? This is my production database.
    Regards
    Rabi

    user13134974 wrote:
    Hi Gurus
    I am using Oracle 10.2.0.3 in AIX env
    Regards
    RabiThose tables you were mentioning, SYS_EXPORT_SCHEMA_nn , are the data pump master tables used for data pump jobs;their purpose is to
    hold the info about the job details.
    Once the job has finished table should be droped, but in case of a job failure table remains so every new dp job must create new SYS_EXPORT_SCHEMA_nn table
    with the +1 nn iteration depending on the name of the last master table that was left due to the dp job failure.
    Cleaning those tables can be done with the dbms_datapump STOP_JOB Procedure, check the docs about the details :
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10825/dp_export.htm
    You can also go visit youroracle support to see examples and instructions for cleaning your db from those dismised master tables,
    How To Cleanup Orphaned DataPump Jobs In DBA_DATAPUMP_JOBS ? [ID 336014.1]

  • ORA-19506:failed to create sequential file.

    RMAN offline full backup failed.
    We had setup of Oracle RMAN backup with TSM and it was working perfect but few days backup, our team did testing with RMAN-NetBackup on database server and after that our existing backup process started giving error.
    Our RMAN backup script is point to TSM but while executing, it shows to point NetBackup and not sure what dependencies are forcing them to point NetBackup.
    Oracle- 9i
    OS- AIX
    Error Message:-
    Starting backup at 04-SEP-12 09:21:51
    channel t1: starting full datafile backupset
    channel t1: specifying datafile(s) in backupset
    input datafile fno=00004 name=/u03/oradata/Test/datap_index01.dbf
    input datafile fno=00003 name=/u03/oradata/Test/datap_data01.dbf
    channel t1: starting piece 1 at 04-SEP-12 09:21:52
    released channel: t1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of backup command on t1 channel at 09/04/2012 09:37:07
    ORA-19506: failed to create sequential file, name="Test_T_FULL_OFFLINE_set2057_piece1_20120904_09nkbf1g_1_1", parms=""
    ORA-27028: skgfqcre: sbtbackup returned error
    ORA-19511: Error received from media manager layer, error text:
    VxBSAValidateFeatureId: Failed with error:
    Server Status: cannot connect on socket

    Thanks for respond but issue is slightly different, we have RMAN backup setup with TSM and it was working fine earlier but last week one of our team installed and configured NetBackup with RMAN for testing oracle backup with RMAN-NetBackup and after testing they revert back the changes.
    After this testing mine RMAN-TSM backup started giving error and in error it seems to point to NetBackup instead of TSM, however I did check the RMAN script but there is no any changes that point to NetBackup.
    I am not sure what changes are causing to point backup to NetBackup instead of TSM.
    Please see the complete error message below-
    allocated channel: t1
    channel t1: sid=11 devtype=SBT_TAPE
    channel t1: Veritas NetBackup for Oracle - Release 7.5 (2012050906)*
    old RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    new RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    new RMAN configuration parameters are successfully stored
    starting full resync of recovery catalog
    full resync complete
    old RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE 'SBT_TAPE' TO 'bk_ctl_%d_%F';
    new RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE 'SBT_TAPE' TO 'bk_ctl_%d_%F';
    new RMAN configuration parameters are successfully stored
    starting full resync of recovery catalog
    full resync complete
    Starting backup at 04-SEP-12 09:21:51
    channel t1: starting full datafile backupset
    channel t1: specifying datafile(s) in backupset
    input datafile fno=00004 name=/u03/oradata/TEST_T/datap_index01.dbf
    input datafile fno=00003 name=/u03/oradata/TEST_T/datap_data01.dbf
    channel t1: starting piece 1 at 04-SEP-12 09:21:52
    released channel: t1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of backup command on t1 channel at 09/04/2012 09:37:07
    ORA-19506: failed to create sequential file, name="TEST_T_FULL_OFFLINE_set2057_piece1_20120904_09nkbf1g_1_1", parms=""
    ORA-27028: skgfqcre: sbtbackup returned error
    ORA-19511: Error received from media manager layer, error text:
    VxBSAValidateFeatureId: Failed with error:
    Server Status:  cannot connect on socket
    RMAN>

  • Failed to create TRS file?

    Hello all,
    We have an automated build for our Help system, in which Final Builder checks the files out of source control and then calls up RH 8 to build our projects. This is all done locally (FB, source control, RH) to the build machine.
    (I also manually build our .CHMs when needed. This is done locally on my own machine.)
    I'm encountering the following error message on one of our child projects when it's run by command line on our build machine:
    Error: Failed to create TRS file. Please use RoboHelp to recover the project.
    Unexpected Error: Failed to prepare single source data for RHCL. Please try to compile it in RoboHTML.
    This project works perfectly on my system - I can manually open it and generate the CHM. This error only seems to occur when it's run via the command line in Final Builder. This is also the only project this occurs on.
    I've tried googling a bit on .trs files (is this even correct?). There's none in the problem project...but then there's also none in any other project (which all compile fine, both manually and by command line), although it looks like it may be a temp file, which may explain why there's none.
    I'm at a loss here as to what's missing or wrong with this project - I've compared my local system to the source control's, and there's no differences between the files (i.e. there's no folder or file that's on my system but not in source control). Nor are there any obvious (to me) differences between this project and the others, and I don't know enough about this file to know what 'parent' files or folders to look at to see why it's not generating.
    Anyone have an idea what exactly the .trs file is, and what might cause a project to be unable to build it?
    Thanks,

    This post suggests the solution is to remove a read only attribute as part of the build process:
    Robohelp 7 HTML command line error
    This one provides some explanation of a bunch of file types (temporary build files):
    Strange new RoboHelp file types
    Hope this helps,
    Amebr
    P.S. Google seems to be doing a better job searching the forums than the Adobe search at the moment. The below seems to work well at keeping it focused on the Robohelp forums.
    site:forums.adobe.com robohelp <enter your search terms here>
    Message was edited by: Amebr

  • Nfs mount point does not allow file creations via java.io.File

    Folks,
    I have mounted an nfs drive to iFS on a Solaris server:
    mount -F nfs nfs://server:port/ifsfolder /unixfolder
    I can mkdir and touch files no problem. They appear in iFS as I'd expect. However if I write to the nfs mount via a JVM using java.io.File encounter the following problems:
    Only directories are created ? unless I include the user that started the JVM in the oinstall unix group with the oracle user because it's the oracle user that writes to iFS not the user that creating the files!
    I'm trying to create several files in a single directory via java.io.File BUT only the first file is created. I've tried putting waits in the code to see if it a timing issue but this doesn't appear to be. Writing via java.io.File to either a native directory of a native nfs mountpoint works OK. ie. Junit test against native file system works but not against an iFS mount point. Curiously the same unit tests running on PC with a windows driving mapping to iFS work OK !! so why not via a unix NFS mapping ?
    many thanks in advance.
    C

    Hi Diep,
    have done as requested via Oracle TAR #3308936.995. As it happens the problem is resolved. The resolution has been not to create the file via java.io.File.createNewFile(); before adding content via an outputStream. if the File creation is left until the content is added as shown below the problem is resolved.
    Another quick question is link creation via 'ln -fs' and 'ln -f' supported against and nfs mount point to iFS ? (at Operating System level, rather than adding a folder path relationship via the Java API).
    many thanks in advance.
    public void createFile(String p_absolutePath, InputStream p_inputStream) throws Exception
    File file = null;
    file = new File(p_absolutePath);
    // Oracle TAR Number: 3308936.995
    // Uncomment line below to cause failure java.io.IOException: Operation not supported on transport endpoint
    // at java.io.UnixFileSystem.createFileExclusively(Native Method)
    // at java.io.File.createNewFile(File.java:828)
    // at com.unisys.ors.filesystemdata.OracleTARTest.createFile(OracleTARTest.java:43)
    // at com.unisys.ors.filesystemdata.OracleTARTest.main(OracleTARTest.java:79)
    //file.createNewFile();
    FileOutputStream fos = new FileOutputStream(file);
    byte[] buffer = new byte[1024];
    int noOfBytesRead = 0;
    while ((noOfBytesRead = p_inputStream.read(buffer, 0, buffer.length)) != -1)
    fos.write(buffer, 0, noOfBytesRead);
    p_inputStream.close();
    fos.flush();
    fos.close();
    }

  • NFS4: Problem mounting NFS mount onto a Solaris 10 Client

    Hi,
    I am having problems mounting NFS mount point from a Linux-Server onto a Solaris 10 Client.
    In the following
    =My server IP ..*.120
    =Client IP ..*.100
    Commands run on Client:
    ==================
    # mount -o vers=3 -F nfs 172.25.30.120:/scratch/pvfs2 /scratch/pvfs2
    nfs mount: 172.25.30.120: : RPC: Rpcbind failure - RPC: Unable to receive
    nfs mount: retrying: /scratch/pvfs2
    nfs mount: 172.25.30.120: : RPC: Rpcbind failure - RPC: Unable to receive
    nfs mount: 172.25.30.120: : RPC: Rpcbind failure - RPC: Unable to receive
    # mount -o vers=4 -F nfs 172.25.30.120:/scratch/pvfs2 /scratch/pvfs2
    nfs mount: 172.25.30.120:/scratch/pvfs2: No such file or directory
    # rpcinfo -p
    program vers proto port service
    100000 4 tcp 111 rpcbind
    100000 3 tcp 111 rpcbind
    100000 2 tcp 111 rpcbind
    100000 4 udp 111 rpcbind
    100000 3 udp 111 rpcbind
    100000 2 udp 111 rpcbind
    1073741824 1 tcp 36084
    100024 1 udp 42835 status
    100024 1 tcp 36086 status
    100133 1 udp 42835
    100133 1 tcp 36086
    100001 2 udp 42836 rstatd
    100001 3 udp 42836 rstatd
    100001 4 udp 42836 rstatd
    100002 2 tcp 36087 rusersd
    100002 3 tcp 36087 rusersd
    100002 2 udp 42838 rusersd
    100002 3 udp 42838 rusersd
    100011 1 udp 42840 rquotad
    100021 1 udp 4045 nlockmgr
    100021 2 udp 4045 nlockmgr
    100021 3 udp 4045 nlockmgr
    100021 4 udp 4045 nlockmgr
    100021 1 tcp 4045 nlockmgr
    100021 2 tcp 4045 nlockmgr
    100021 3 tcp 4045 nlockmgr
    100021 4 tcp 4045 nlockmgr
    # showmount -e 172.25.30.120 (Server)
    showmount: 172.25.30.120: RPC: Rpcbind failure - RPC: Unable to receive
    Commands OnServer:
    ================
    program vers proto port
    100000 2 tcp 111 portmapper
    100000 2 udp 111 portmapper
    100021 1 tcp 49927 nlockmgr
    100021 3 tcp 49927 nlockmgr
    100021 4 tcp 49927 nlockmgr
    100021 1 udp 32772 nlockmgr
    100021 3 udp 32772 nlockmgr
    100021 4 udp 32772 nlockmgr
    100011 1 udp 796 rquotad
    100011 2 udp 796 rquotad
    100011 1 tcp 799 rquotad
    100011 2 tcp 799 rquotad
    100003 2 udp 2049 nfs
    100003 3 udp 2049 nfs
    100003 4 udp 2049 nfs
    100003 2 tcp 2049 nfs
    100003 3 tcp 2049 nfs
    100003 4 tcp 2049 nfs
    100005 1 udp 809 mountd
    100005 1 tcp 812 mountd
    100005 2 udp 809 mountd
    100005 2 tcp 812 mountd
    100005 3 udp 809 mountd
    100005 3 tcp 812 mountd
    100024 1 udp 854 status
    100024 1 tcp 857 status
    # showmount -e 172.25.30.120
    Export list for 172.25.30.120:
    /scratch/nfs 172.25.30.100,172.25.24.0/4
    /scratch/pvfs2 172.25.30.100,172.25.24.0/4
    Thank you, ~al

    I also tried to run Snoop on the client and wireshark on Server and following is what I see:
    One Server: Upon issuing mount command on client:
    # tshark -i eth1
    Running as user "root" and group "root". This could be dangerous.
    Capturing on eth1
    0.000000 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    0.205570 172.25.30.100 -> 172.25.30.120 Portmap V2 GETPORT Call MOUNT(100005) V:3 UDP
    0.205586 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    0.207863 172.25.30.100 -> 172.25.30.120 Portmap V2 GETPORT Call MOUNT(100005) V:3 UDP
    0.207869 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    2.005314 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    4.011005 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    5.206109 Dell_70:ad:29 -> SunMicro_70:ff:17 ARP Who has 172.25.30.100? Tell 172.25.30.120
    5.206277 SunMicro_70:ff:17 -> Dell_70:ad:29 ARP 172.25.30.100 is at 00:14:4f:70:ff:17
    5.216157 172.25.30.100 -> 172.25.30.120 Portmap V2 GETPORT Call MOUNT(100005) V:3 UDP
    5.216170 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    On Clinet Upon issuing mount command on client:
    # snoop -d bge1
    Using device /dev/bge1 (promiscuous mode)
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    atlas-pvfs2 -> pvfs2-io-0-3 PORTMAP C GETPORT prog=100005 (MOUNT) vers=3 proto=UDP
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (UDP port 111 unreachable)
    atlas-pvfs2 -> pvfs2-io-0-3 PORTMAP C GETPORT prog=100005 (MOUNT) vers=3 proto=UDP
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (UDP port 111 unreachable)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    pvfs2-io-0-3 -> * ARP C Who is 172.25.30.100, atlas-pvfs2 ?
    atlas-pvfs2 -> pvfs2-io-0-3 ARP R 172.25.30.100, atlas-pvfs2 is 0:14:4f:70:ff:17
    atlas-pvfs2 -> pvfs2-io-0-3 PORTMAP C GETPORT prog=100005 (MOUNT) vers=3 proto=UDP
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (UDP port 111 unreachable)
    Also I see the following on Client:
    # rpcinfo -p pvfs2-io-0-3
    rpcinfo: can't contact portmapper: RPC: Rpcbind failure - RPC: Failed (unspecified error)
    When I try the above rpcinfo command on Client and Server Snoop And wireshark(ethereal) outputs are as follows:
    Client # snoop -d bge1
    Using device /dev/bge1 (promiscuous mode)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    atlas-pvfs2 -> pvfs2-io-0-3 TCP D=111 S=872 Syn Seq=2065245538 Len=0 Win=49640 Options=<mss 1460,nop,wscale 0,nop,nop,sackOK>
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (TCP port 111 unreachable)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=2004 (Unknown), size = 48 bytes
    ? -> (multicast) ETHER Type=0003 (LLC/802.3), size = 90 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    pvfs2-io-0-3 -> * ARP C Who is 172.25.30.100, atlas-pvfs2 ?
    atlas-pvfs2 -> pvfs2-io-0-3 ARP R 172.25.30.100, atlas-pvfs2 is 0:14:4f:70:ff:17
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    atlas-pvfs2 -> pvfs2-io-0-3 TCP D=111 S=874 Syn Seq=2068043912 Len=0 Win=49640 Options=<mss 1460,nop,wscale 0,nop,nop,sackOK>
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (TCP port 111 unreachable)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    Server # tshark -i eth1
    Running as user "root" and group "root". This could be dangerous.
    Capturing on eth1
    0.000000 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    0.313739 Cisco_3d:68:10 -> CDP/VTP/DTP/PAgP/UDLD CDP Device ID: MILEVA Port ID: GigabitEthernet1/0/16
    2.006422 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    3.483733 172.25.30.100 -> 172.25.30.120 TCP 865 > sunrpc [SYN] Seq=0 Win=49640 Len=0 MSS=1460 WS=0
    3.483752 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    4.009741 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    6.014524 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    6.551356 Cisco_3d:68:10 -> Cisco_3d:68:10 LOOP Reply
    8.019386 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    8.484344 Dell_70:ad:29 -> SunMicro_70:ff:17 ARP Who has 172.25.30.100? Tell 172.25.30.120
    8.484569 SunMicro_70:ff:17 -> Dell_70:ad:29 ARP 172.25.30.100 is at 00:14:4f:70:ff:17
    10.024411 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    12.030956 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    12.901333 Cisco_3d:68:10 -> CDP/VTP/DTP/PAgP/UDLD DTP Dynamic Trunking Protocol
    12.901421 Cisco_3d:68:10 -> CDP/VTP/DTP/PAgP/UDLD DTP Dynamic Trunking Protocol
    ^[[A 14.034193 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00  Cost = 0  Port = 0x8010
    15.691119 172.25.30.100 -> 172.25.30.120 TCP 866 > sunrpc [SYN] Seq=0 Win=49640 Len=0 MSS=1460 WS=0
    15.691138 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    16.038944 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    16.550760 Cisco_3d:68:10 -> Cisco_3d:68:10 LOOP Reply
    18.043886 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    20.050243 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    21.487689 172.25.30.100 -> 172.25.30.120 TCP 867 > sunrpc [SYN] Seq=0 Win=49640 Len=0 MSS=1460 WS=0
    21.487700 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    22.053784 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    24.058680 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    26.063406 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    26.558307 Cisco_3d:68:10 -> Cisco_3d:68:10 LOOP Reply
    ~thank you for any help you can provide!!!

  • Problems creating dmp file from script

    Hi,
    I am running portal 9.0.4.1 on solaris and am trying to create the export dmp file from the downloaded unix script. I use the following command line:
    gc.sh -mode export -s portal -p &lt;portal schema password) -c &lt;sid&gt;
    I then accept the default dmp file name. However, the following error is then returned:
    ld.so.1: exp: fatal: libclntsh.so.9.0: open failed: No such file or directory
    Killed
    This part of the export/ import operation is now complete.
    I am then returned to the command prompt, but with no dmp file created and the .out file containing just the word "error".
    Any ideas where I'm going wrong????
    Thanks for your help.

    Your LD_LIBRARY_PATH is most likely not set properly. Try to execute sqlplus from command line and make sure that works. Once you're able to connect to the database via sqlplus from command line, retry the export/import again.

  • Series Selection Creating .DMP File for Customised Form

    Hello ,
    I have created a Customised form with series selection from SAP .
    While registration bydefault "Primary" named series is created for that customised form.
    Then i added one more Series named " 10-11" with indicator as '2010-11' and set it as default series.
    Then i opened the customised form and entered the data it gets saved without any problem.
    After that i again went to the Document Settings and change the series from '10-11' to 'Primary' and
    select the indicator as 'Default' and saved the settings.
    When i opened the customised form the series shown is the changed one as 'Primary' but while entering the data and saving
    it system shows ".dmp file created" and sap shuts down.
    Then i have to unregister the UDO for that form and to delete all data entered for that form from table to make it work agian.
    This is the Problem which i am facing while using the Series in the customised form and its been a trouble some
    for me .
    Suggest any path to sort out this problem arising.
    Thanks & Regards
    Amit

    Hello Manish,
    Thanks for your quick reply.
    I am using SAP 2007 B Patch - 13
    But my question is if this problem occurs even after Patch upgradation then what will be the step for resolving that.
    In future if this problem arises what steps i have to take to remove this problem  without affecting my data.
    Suggest.
    Thanks & Regards,
    Amit

  • Dreamweaver fails to create cache file

    I have reinstalled Dreamweaver 8 and my main site fails to
    create a site cache file. I am running windows xp fully updated. My
    other sites can create a cache file with no difficulty. I can
    create and destroy files within the sitecache directory. The file
    appears when the error mesage is displayed but disappears as soon
    as the error dialog is acknowledged.
    Any thoughts?

    quote:
    Originally posted by:
    Newsgroup User
    Do you have the cache enabled in your Site definition?
    HTH,
    Randy
    > I have reinstalled Dreamweaver 8 and my main site fails
    to create a site cache
    > file. I am running windows xp fully updated. My other
    sites can create a
    > cache file with no difficulty. I can create and destroy
    files within the
    > sitecache directory. The file appears when the error
    mesage is displayed but
    > disappears as soon as the error dialog is acknowledged.

  • Fail to create temp file in Oracle 10g on CentOS

    My disk is full so I delete six temp files and create only one, but when I execute this SQL give me this error. I know i have more than 4GB free space but i don´t know what problem is happen.
    ALTER TABLESPACE temp ADD TEMPFILE '/oracle/oradata/ral/temp01.dbf' SIZE 512m AUTOEXTEND ON NEXT 250m MAXSIZE 2048m;
    ERROR
    ALTER TABLESPACE temp ADD TEMPFILE '/oracle/oradata/ral/temp01.dbf' SIZE 512m AUTOEXTEND ON NEXT 250m MAXSIZE 2048m
    Informe de error:
    Error SQL: ORA-01119: error create database file '/oracle/oradata/ral/temp01.dbf'
    ORA-27044: no se ha podido escribir el bloque de cabecera del archivo
    Linux-x86_64 Error: 28: No space left on device
    Additional information: 3
    +01119. 00000 - "error in creating database file '%s'"+
    *Cause:    Usually due to not having enough space on the device.+
    *Action:+
    Could you help me plz?
    Thanx in advance

    993296 wrote:
    My disk is full so I delete six temp files and create only one, but when I execute this SQL give me this error. I know i have more than 4GB free space but i don´t know what problem is happen.
    ALTER TABLESPACE temp ADD TEMPFILE '/oracle/oradata/ral/temp01.dbf' SIZE 512m AUTOEXTEND ON NEXT 250m MAXSIZE 2048m;
    ERROR
    ALTER TABLESPACE temp ADD TEMPFILE '/oracle/oradata/ral/temp01.dbf' SIZE 512m AUTOEXTEND ON NEXT 250m MAXSIZE 2048m
    Informe de error:
    Error SQL: ORA-01119: error create database file '/oracle/oradata/ral/temp01.dbf'
    ORA-27044: no se ha podido escribir el bloque de cabecera del archivo
    Linux-x86_64 Error: 28: No space left on device
    Additional information: 3
    +01119. 00000 - "error in creating database file '%s'"+
    *Cause:    Usually due to not having enough space on the device.+
    *Action:+
    Could you help me plz?
    Thanx in advanceHi,
    please check space under /oracle/oradata/ral filesystem
    Linux-x86_64 Error: 28: No space left on device
    Regards

  • Failing to create temp file in Perl cgi script

    This used to work, perhaps before the Apache 2.4 upgrade?
    From the httpd error log:
    AH01215: Error in tempfile() using template /tmp/XXXXXXXXXX: Could not create temp file /tmp/4VmuZsw3Wt: No such file or directory at...
    Here's the offending code:
    #!/usr/bin/perl
    use strict;
    use warnings;
    use File::Temp;
    use CGI;
    my $q = new CGI();
    print $q->header();
    print "Hi mom\n";
    my( $foo, $bar ) = File::Temp::tempfile();
    print $foo "Test content\n";
    close( $foo );
    print "Hi dad\n";
    1;
    If invoked from the command-line, it works as intended. Tried as users me, root and http (the Apache user).
    If run as CGI script, I get this:
    Hi mom
    then it aborts with an HTTP 500, and emits the above message in the error log.
    Before anybody asks:
    drwxrwxrwt 7 root root 240 May 22 20:51 /tmp
    The Apache configuration has this, which /some/path/cgi contains my above script.
    ScriptAlias "/cgi" /some/path/cgi
    Baffled ...

    Restarting Apache fixed the issue. Ran out of file descriptors perhaps?
    The question is: will it happen again?
    Last edited by jernst (2014-05-22 21:20:12)

  • Creating DMP file

    Hi All,
    We are taking dmp file. Is there anyway to edit DMP file?.
    If any Please Suggest.
    Thanks,
    Rajiv Birari.

    Hi,
    >>We are taking dmp file. Is there anyway to edit DMP file?.
    It depends. Although it is not advisable by Oracle and millions DBAs around the world, in fact, using a hexadecimal editor like Editpad or others you are able to edit it. The question is: What are you trying to change in the DMP file?
    Cheers
    Legatti

Maybe you are looking for