Mounting /stage directory on other server?

Hi to all,
regarding multi-node installation on oracle application 11i,
can anyone knows on how to mount /stage(rapidwiz installer) directory from server1 to server2.?
information of server:
Server1 = 192.168.0.1 / 255.255.255.0 host:oracle.mabuhay.com
Server2 = 192.168.0.2 / 255.255.255.0 host:applmgr.mabuhay.com
thanks in advance,
BJCDR

Hi,
If you're running Linux or one of the Unixes, you should be able to use NFS to export the directory on server1 for mounting on server2. There are a lot of docs that describe how to do this, just google "Setting up NFS share." Here's one example that looks fairly straightforward: http://tldp.org/HOWTO/NFS-HOWTO/server.html
I assume you'll be running rapidwiz as root, so you will probably need to use the no_root_squash option on the NFS mount. The docs caution that using no_root_squash can be a security risk, but in this case (temporary share, just for reading the install media), it should be safe enough.
You could also try using Samba/CIFS to share the directory, but my experiences attempting that route have been somewhat unpleasant, so I wouldn't really recommend it.
If this is a Windows platform, I can't be of much help, but there are lots of other people on the forum who are very experienced with installing EBS on Windows. I have no idea if standard Windows folder sharing is adequate in this case. :-)
Regards,
John P.

Similar Messages

  • Receiver file adapter NFS - Access directory of other server

    HI,
    I need access from an file adapter type NFS to a shared folder on the SAP ECC server. This is possible?. Users who are connecting to assign permissions on the folder. The 2 servers are on the network and the server is not enabled SAP ECC FTP. All run on Windows 2003 server environments and the SAP PI Server is 7.1 EHP 1.
    Regards,
    Johnny

    To access a Unix NFS from Windows you can install "Microsoft Services for Unix". Those are free of charge provided by Microsoft.
    See
    http://technet.microsoft.com/en-us/library/bb496506.aspx
    Markus

  • Steps to mount a directory

    Hi all,
    I have two solaris server. One is file server and another is oracle server.
    I want to mount a directory of file server in oracle server.
    Please provide me the steps to perform it. And let me know what privilege should I have.
    Thanks.

    on server, check if directory to mount is exported,
    see file e.g. /etc/dfs/sharetab (run man sharetab)
    run exportfs to export the filesystem to the outside
    verify if nfsd daemon is started (run man nfsd)
    on 'client',
    mount the remote dir:
    mount <servername>:/<remotedir> <mountpoint> <options>
    see man mount
    if you want the mount to happen automatically at access attempt,
    consider the automount
    run man automount
    Guido
    ps this is a Linux tech forum

  • How to Stage directory for Oracle EBS R12.1.1 on Window Server 2008 ?

    Hi Every Body,
    1- How to Stage directory for Oracle EBS R12.1.1 downloaded from e-delivery.com on Window Server 2008.?
    -- I have download all files from e-delivery and check the md5 and all files are OK.
    I extracted the entire files with winrar program to the same directory and during the unzipping I was notified to replace these files (dvd.label and component.label) and I agreed each time. So >>
    2- Was that a problem ?
    -- I have setup vc++ from VS6 , the sp6, cygwin , ... and set up the active directory and the domain.. and all other requirements.
    BUT when i started the ebs set up it raise the "*Os User and Group Check*" error.
    even if I am the only user and sure I am the Administrator. and also I run this Command to be sure
    C:\>net localgroup administrators and it said that the administrator is one of the administrators group.
    3- Somebody said that I have to change the temp and tmp folder to be on c:\ drive >> might this solve the problem?
    thanks in advance.

    Hi,
    Most of the instructions you have mentioned , already was configured.
    except for the tmp and temp directories.
    And even I changed them from System Properties >> Advanced >>  Environment Variable under User variable for Administrator and System variables , the problem still exist Os User and Group Check .
    I attach the conf_VIS.txt file because I thought it may help.
    http://www.4shared.com/document/1bS6aio5/conf_VIS.html
    Actually I am not sure about the rapid install patch version but I have download the EBS R12 from the following link and I think it contains the recent installer version
    Link >> Oracle E-Business Suite Release 12.1.1 Media Pack for Microsoft Windows (32-bit)
    http://edelivery.oracle.com/EPD/Download/get_form?egroup_aru_number=11250138
    Note If you still can not open the link just suggest a suitable file hosting and I will upload to you. here I used 4shared and hoped it suits you.

  • Move large database to other server using RMAN in less downtime

    Hi,
    We have large database around 20TB. We want to migrate (move) the database from one server to other server. We do not want to use standby option.
    1)     How can we move database using RMAN in less downtime
    2)     Other than RMAN is there any option is available to move the database to new server
    For option 1 (restore using RMAN),
    Whether below options are valid?
    If this option is valid, how to implement this?
    1)     How can we move database using RMAN in less downtime
    a)     Take the full backup from source (source db is up)
    b)     Restore the full backup in target (source db is up)
    c)     Take the incremental backup from source (source db is up)
    d)     Restore incremental backup in target (source db is up)
    e)     Do steps c and d, before taking downtime (source db is up)
    f)     Shutdown and mount the source db, and take the incremental backup (source db is down)
    g)     Restore last incremental backup and start the target database (target is up and application is accessing this new db
    database version: 10.2.0.4
    OS: SUN solaris 10
    Edited by: Rajak on Jan 18, 2012 4:56 AM

    Simple:
    I do this all the time to relocate file system files... But the principle is the same. You can do this in iterations so you do not need to do it all at once:
    Starting 8AM move less-used files and more active files in the afternoon using the following backup method.
    SCRIPT-1
    RMAN> BACKUP AS COPY
    DATAFILE 4 ####"/some/orcl/datafile/usersdbf"
    FORMAT "+USERDATA";
    Do as many files as you think you can handle during your downtime window.
    During your downtime window: stop all applications so there is no contention in the database
    SCRIPT-2
    ALTER DATABASE DATAFILE 4 offline;
    SWITCH DATAFILE 4 TO COPY;
    RECOVER DATAFILE 4;
    ALTER DATABASE DATAFILE 4 online;
    I then execute the delete of the original file at somepoint later - after we make sure everything has recovered and successfully brought back online.
    SCRIPT-3
    DELETE DATAFILECOPY "/some/orcl/datafile/usersdbf"
    For datafiles/tablespaces that are really busy, I typically copy them later in the afternoon as there are fewer archivelogs that it has to go through in order to make them consistent. The ones in the morning have more to go through, but less likelihood of there being anything to do.
    Using this method, we have moved upwards 600G at a time and the actual downtime to do the switchover is < 2hrs. YMMV. As I said, this can be done is stages to minimize overall downtime.
    If you need some documentation support see:
    http://docs.oracle.com/cd/E11882_01/server.112/e18951/asm_rman.htm#CHDBDJJG
    And before you do ANYTHING... TEST TEST TEST TEST TEST. Create a dummy tablespace on QFS and use this procedure to move it to ASM to ensure you understand how it works.
    Good luck! (hint: scripts to generate these scripts can be your friend.)

  • [SOLVED] mount.nfs4: access denied by server

    Hi folks. I seem to be having a bit of a problem getting nfs4 to work. I am trying to mount a share from alpha (my fileserver) onto charlie (my workstation). Both of these are new Arch systems and I haven't had any nfs working yet, although I have with other distros on the same hardware.
    Fileserver (alpha) config:
    # /etc/exports
    /files 192.164.1.0/24(rw,sync,fsid=0,no_subtree_check)
    # /etc/hosts.allow
    sshd: 192.168.1.0/255.255.255.0
    nfsd: 192.168.1.0/255.255.255.0
    rpcbind: 192.168.1.0/255.255.255.0
    mountd: 192.168.1.0/255.255.255.0
    idmapd: 192.168.1.0/255.255.255.0
    statd: 192.168.1.0/255.255.255.0
    [General]
    Verbosity = 3
    Pipefs-Directory = /var/lib/nfs/rpc_pipefs
    Domain = localdomain
    [Mapping]
    Nobody-User = nobody
    Nobody-Group = nobody
    [Translation]
    Method = nsswitch
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    none /dev/pts devpts defaults 0 0
    none /dev/shm tmpfs defaults 0 0
    #/dev/cdrom /media/cd auto ro,user,noauto,unhide 0 0
    #/dev/dvd /media/dvd auto ro,user,noauto,unhide 0 0
    #/dev/fd0 /media/fl auto user,noauto 0 0
    /dev/sda1 /boot ext3 defaults 0 1
    /dev/sda2 swap swap defaults 0 0
    /dev/sda5 / ext3 defaults 0 1
    /dev/sda6 /var ext3 defaults 0 1
    /dev/sda7 /home ext3 defaults 0 1
    /dev/sda8 /files ext3 defaults 0 1
    rpc_pipefs /var/lib/nfs/rpc_pipefs rpc_pipefs defaults 0 0
    nfsd /proc/fs/nfsd nfsd rw,nodev,noexec,nosuid 0 0
    DAEMONS=(syslog-ng network netfs rpcbind nfs-common nfs-server hal @alsa @crond @openntpd @sshd)
    [root@alpha ~]# df
    Filesystem 1K-blocks Used Available Use% Mounted on
    /dev/sda5 19228276 879492 17372036 5% /
    none 507792 140 507652 1% /dev
    none 507792 0 507792 0% /dev/shm
    /dev/sda1 93307 15887 72603 18% /boot
    /dev/sda6 19228276 372632 17878896 3% /var
    /dev/sda7 19228276 176224 18075304 1% /home
    /dev/sda8 902688436 204872 856629640 1% /files
    [root@alpha ~]#
    Workstation (charlie) config:
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    none /dev/pts devpts defaults 0 0
    none /dev/shm tmpfs defaults 0 0
    #/dev/cdrom /media/cd auto ro,user,noauto,unhide 0 0
    #/dev/dvd /media/dvd auto ro,user,noauto,unhide 0 0
    #/dev/fd0 /media/fl auto user,noauto 0 0
    UUID=437982b2-5c84-4f53-954d-cf43f8b4e707 / ext3 defaults 0 1
    UUID=97d79d76-357a-4f4e-8513-f181bff6af62 /boot ext3 defaults 0 1
    UUID=d8525095-9b97-4439-932f-8f4e0236cce1 /home ext3 defaults 0 1
    UUID=ffba933b-af93-407c-b1b8-69d1cc5be146 swap swap defaults 0 0
    rpc_pipefs /var/lib/nfs/rpc_pipefs rpc_pipefs defaults 0 0
    alpha:/ /files nfs4 defaults 0 0
    [General]
    Verbosity = 3
    Pipefs-Directory = /var/lib/nfs/rpc_pipefs
    Domain = localdomain
    [Mapping]
    Nobody-User = nobody
    Nobody-Group = nobody
    [Translation]
    Method = nsswitch
    DAEMONS=(syslog-ng network crond alsa hal fam rpcbind nfs-common netfs)
    [root@charlie ~]# mount -a
    mount.nfs4: access denied by server while mounting alpha:/
    [root@charlie ~]#
    This happens even after both systems are rebooted. Can anyone spot what I am missing?
    Thanks for looking.
    Last edited by dgregory46 (2009-10-21 01:04:09)

    Now I really feel stupid. A little proofreading would have saved me a big headache. In /etc/exports I was exporting to 192.164.1.0/24 while my network is the more standard 192.168.1.0/24.
    It works fine now, although I did take phaul's suggestion and added my main share "inside" the nfs4 root.

  • Why war.DDInit's result put to screen, not to a stage directory

              Hi, I am new in weblogic. I run "java weblogic.ant.taskdefs.war.DDInit stage" to automatically
              create web.xml & weblogic.xml, but it only create one xml file and print it to screen,
              not to the stage/WEB-INF directory, I have create stage/WEB-INF direcoty and put
              jsp file under stage directory and set enviroment just as the doc(Developing Weblogic
              Server Applications) describe, I use Weblogic 7.0, How to solve this problem?
              

              Hi, I use the absolute directory "c:\stage" instead of relative directory "stage",
              but it still not work.
              "ludovic le goff" <[email protected]> wrote:
              >Hello Rafe,
              >
              >In my opinion, there is maybe a mistake in your java command line.
              >As described at this link :
              >http://e-docs.bea.com/wls/docs70/programming/packaging.html#1068003, the
              >staging-dir parameter is a directory.
              >I would suggest to type the following java command instead of yours (
              >assuming that you put the stage directory in the C drive ) :
              >- java weblogic.ant.taskdefs.war.DDInit c:\stage
              >
              >I hope this helps.
              >Best regards,
              >Ludovic.
              >Developer Relations Engineer
              >BEA Customer support
              >"rafe lin" <[email protected]> a écrit dans le message news:
              >3d0d5f77$[email protected]..
              >>
              >> Hi, I am new in weblogic. I run "java weblogic.ant.taskdefs.war.DDInit
              >stage" to automatically
              >> create web.xml & weblogic.xml, but it only create one xml file and print
              >it to screen,
              >> not to the stage/WEB-INF directory, I have create stage/WEB-INF direcoty
              >and put
              >> jsp file under stage directory and set enviroment just as the
              >doc(Developing Weblogic
              >> Server Applications) describe, I use Weblogic 7.0, How to solve this
              >problem?
              >
              >
              

  • OEL 6.3 - mount.nfs: access denied by server

    Hi,
    I am trying to mount an NFS directory on a server running OEL 6.3, pointing to another OEL 6.3 server. I get the following error:
    [oracle@csdowmsdb503 etc]$ mount 192.x.x.x:/home/oracle/m501/m501_f /home/oracle/m501_f/
    mount.nfs: access denied by server while mounting 192.x.x.x:/home/oracle/m501/m501_f
    The UIDs and GIDs on each server match. If we use "nfsvers=3" it work, so it seems to be something specific to vers 4. We would like to not use the vers 3 workaround.
    Thanks!

    Hi,
    When I added "- v" I realized it is actually erroring out on vers=4 and defaulting to vers=3 instead. The error it shows is now "No such file or directory."
    [oracle@xxx503 m501_g]$ sudo mount -v -t nfs xxx.xxx.xxx.35:/home/oracle/m501/m501_g /home/oracle/m501_g/
    mount.nfs: timeout set for Fri Nov 9 10:46:27 2012
    mount.nfs: trying text-based options 'vers=4,addr=xxx.xxx.xxx.35,clientaddr=xxx.xxx.xxx.37'
    mount.nfs: mount(2): No such file or directory
    mount.nfs: trying text-based options 'addr=xxx.xxx.xxx.35'
    mount.nfs: prog 100003, trying vers=3, prot=6
    mount.nfs: trying xxx.xxx.xxx.35 prog 100003 vers 3 prot TCP port 2049
    mount.nfs: prog 100005, trying vers=3, prot=17
    mount.nfs: trying xxx.xxx.xxx.35 prog 100005 vers 3 prot UDP port 37692
    xxx.xxx.xxx.35:/home/oracle/m501/m501_g on /home/oracle/m501_g type nfs (rw)

  • Why ejb20.DDInit's result put to screen, not to a stage directory

    Hi, I am new in weblogic. I run "java weblogic.ant.taskdefs.ejb20.DDInit stage" to
    automatically create ejb-jar.xml, weblogic-ejb-jar.xml, but it only create one xml
    file and print it to screen, not to the stage/META-INF directory, I have create stage/META-INF
    direcoty and put correct class file under stage directory and set enviroment just
    as the doc(Developing Weblogic Server Applications) describe, I use Weblogic 7.0,
    How to solve this problem?
    And if I use weblogic.marathon.ddinit.EJBInit, it produce no xml file, and display:
    [ModuleInit]: Searching for class files
    [ModuleInit]: 4 classes found
    [ModuleInit]: Discovered module type for E:\java\usingj2ee\pricing\stage
    [ModuleInit]: Found EJB components. Initializing descriptors
    [ModuleInit]: Found EJBHome: usingj2ee.pricing.PricingHome
    [ModuleInit]: Creating relations
    [ModuleInit]:
    Found no EJBs in the specified directory.
    [ModuleInit]:
    Be sure the EJB classes are in the proper package hierarchy under the specified directory
    'E:\java\usingj2ee\pricing\stage.'
    [ModuleInit]: Writing descriptors
    [ModuleInit]: Building module with newly created descriptors
    [ModuleInit]: Finished building module
    [EJBJarCMBean] Writing descriptors
    why?

    You found a bug. I file a bug request: CR080702.
    We will fix it very soon.
    Thanks.
    Cedric
    "Rafe Lin" <[email protected]> wrote in news:[email protected]:
    >
    Hi:
    Now I attach the stage directory and source java file, Thank you.
    Cedric Beust <[email protected]> wrote:
    Please post the updated jar file (preferably with the source files in
    it). Are you sure the XML descritprs are not created inside the jar
    file? The trace seems to say the contrary...
    Cedric
    "Rafe Lin" <[email protected]> wrote in news:3d128fb6
    @newsgroups2.bea.com:
    >>
    >>
    >>>
    Hi, now I rename PringImpl.java to PricingEJB.java, and there is no
    error when DDInit run, but the two xml still not appear, why? when I
    run DDInit, screen display:
    [ModuleInit]: Searching for class files
    [ModuleInit]: 4 classes found
    [ModuleInit]: Discovered module type for
    e:\java\usingj2ee\pricing\stage [ModuleInit]: Found EJB components.
    Initializing descriptors [ModuleInit]: Found EJBHome:
    usingj2ee.pricing.PricingHome [ModuleInit]: Bean class:
    usingj2ee.pricing.PricingEJB [ModuleInit]: Adding Session bean
    'PricingEJB' [ModuleInit]:
    [ModuleInit]: Creating relations
    [ModuleInit]: Writing descriptors
    [ModuleInit]: Building module with newly created descriptors
    [ModuleInit]: Finished building module
    [EJBJarCMBean] Writing descriptors
    Thanks!
    Cedric Beust <[email protected]> wrote:
    "Rafe Lin" <[email protected]> wrote in
    news:[email protected]:
    Hi, now I attach the jar file, of course, I run DDInit before the
    achieve file is created, it just present the stage dir struct, I
    did
    not run DDInit on the jar file. Thanks!Your bean has unusual naming conventions: your bean is called
    PricingImpl.
    Rename it to PricingEJB and DDInit should recognize it.
    Admittedly, EJBInit should do a better job at guessing, or at least at
    producing a better error message.
    CedricAttachment decoded: untitled-2.txt
    ---=_news3d128fb5
    Attachment decoded: Pricing.jar
    ---=_news3d128fb5--Attachment decoded: untitled-2.txt
    ---=_news3d145663
    Attachment decoded: for_chk.jar
    ---=_news3d145663--

  • Can ssh into client iMac but cannot mount home directory ...

    I've got a server running 10.4.11 and a bunch of client iMacs (10.5.n). If I ssh into a client iMac it lets me log in but cannot mount my home directory (which lives on the server). Obviously ssh is getting my credentials from the server but isn't getting (or cannot get) the information to mount my disk. If I log in through the gui on the client it all works fine. I suspect this is less of an error and more of a design. Is there a way to get the clients to mount my home when logging in with ssh? If not then Apple presumably has a reason for disallowing this behaviour. Is there a way around this? Perhaps with mount_afp?
    We want to be able to log into many client iMacs at once to run computations on them. Eventually we want to do this on a suite of XServes and XGrid. This is all very easy with linux and nfs and I was expecting the same sort of ability through OS X. I know that I can export everything from the server with nfs (though that may not solve my problem) but I only have the one system and multiple users and I can't really experiment while they are connected.
    Suggestions? Comments? If there are obvious examples of this elsewhere please point me toward them.

    Well it's never worked that way for me. Do you actually use this method to log in?
    I've gone through all of the server logs, watching what gets added as I log on to a client. I don't see anything there that suggests it is even attempting to mount the home directory. I also watched on the client and didn't see any errors. When I ssh into a client I see the following message on the terminal, "Could not chdir to home directory /Network/Servers/server.some.place/Volumes/R1/UsersR1/username: Unknown error: 118" This suggests that the client knows where my home should be but doesn't know how to mount it.
    I have enabled the basic setup on my server. User accounts have Home URLs like "afp://server.some.place/UsersR1/username" and Full Paths like "/Network/Servers/server.some.place/Volumes/R1/UsersR1" (R1 is a RAID disk). UsersR1 has been made into a Share Point and Server Admin tells me that the disk "will be automatically mounted using the AFP protocol as /Network/Servers/R1 on client machines". This part doesn't seem to be true as when I log into a client through the GUI the actual path that is mounted as my home is /Network/Servers/server.some.place/R1" (pwd in my home directory returns /Network/Servers/server.some.place/Volumes/R1/UsersR1/username which is a actually rather inconvenient).
    Under the Advanced tab in the Workgroup Manager, Accounts pane, I've got "Allow simultaneous login on managed computers". This seems to work but also seems only to apply to the GUI, that is I can log onto several clients at once.
    How do I ask a client machine what home directory information it is getting from the server for any particular user? Is there some setup on the client machines that will enable automatic mounting of homes through ssh? On the clients, the Directory Utility, which I have used to connect the client to a Directory Server, has a Mounts pane which allows me to "Edit automatic NFS mounts for this computer". I am not exporting any NFS from the server, but perhaps this is the way to make it work?
    Perhaps what I want to requires Kerberos? I don't have that enabled.
    When I have done this on a cluster of linux machines I exported the home directory from the server with NFS (exportfs) and then, using NIS and automount on the clients it "just worked". It was very simple. If it is supposed to work with Macs and OS X, it may be simple but it's not obvious.

  • Why Database should be in mount stage for enable Flashback

    Oracle 11.1.0.6.0
    Windows
    Why database should be in mount stage for enabling flashback feature but we can disable this feature even in Open mode?
    Kindly share your inputs .
    Thanks,
    Alok

    >
    As per doc this is prerequisites that database should be in mount mode to enable flashback but we can disable flashback even in open mode.
    I would like to know the concept behind this.
    >
    Only Oracle knows why and why, for 11g, you can now have the database open. Any 'guesses' would only be guesses.
    You can get some possible clues from the DBA guide section for 10g 'Starting Up and Shutting Down' the database.
    http://docs.oracle.com/cd/B19306_01/server.102/b14231/start.htm
    Traditionally MOUNT mode (database closed) is used for administrative activities by the DBA that need to be done either when there is no user activity that might conflict with them.
    >
    Starting Up an Instance
    •Start the instance and mount the database, but leave it closed. This state allows for certain DBA activities, but does not allow general access to the database.
    >
    Or this from the same doc a little later on
    >
    Starting an Instance and Mounting a Database
    You can start an instance and mount a database without opening it, allowing you to perform specific maintenance operations. For example, the database must be mounted but not open during the following tasks:
    •Enabling and disabling redo log archiving options. For more information, please refer to Chapter 7, "Managing Archived Redo Logs".
    •Performing full database recovery. For more information, please refer to Oracle Database Backup and Recovery Basics
    The following command starts an instance and mounts the database, but leaves the database closed:
    STARTUP MOUNT
    >
    Perhaps part of the flashback enable process required exclusive access to ALL tablespaces or datafiles; meaning you can't have users modifying them.
    Perhaps the enable process for 10g needed to perform simultaneous updates to all datafile headers, the controlfiles and undo or cache or data dictionary entries; meaning you can't have users getting in the way.
    Dont' know.

  • Stage directory not created after clean restart of WLS

    I deployed the application in stage mode.
    After clean restart of the weblogic server, the stage folder is not getting created.
    But application is up and running.
    Workaround Tried:
    -Checked the Java options whether any of "-nostage,-external-stage" options are used, but none are used.
    -Checked the deployment tab under server->config->deployment, it has the stage path specified
    Can anyone point what could be preventing the stage folder creation??

    If you are deploying through the Admin Console then while deploying try using the below option:
    Recommended selection.
    Copy this application onto every target for me     
    Though the default option for each server is stage mode, but you can try to use the above option and check whether the stage directory is created or not.

  • How to store Live stream on the other Server

    Hello Everyone,
    This is very important for me to store all live streams in the archive server at F4V sormat.
    to do that I record live stream in the F4V:
    mystreams.publish('mp4:' + netstream.name + 'f4v');
    mystream.recod();
    This cods is work currently in the rootinstall directory's but I wan to store on the archive directory in other linux server.
    1- I share the archive folder in the archive server.
    2- Create folder on the AMS linux server in the applications/livepkgr/streams/_definst_/archive
    3- mount archive folder to the archive folder in other server:
    mount -t cifs -o administrator=root,password=myams //192.168.1.100/Archive /opt/adobe/ams/applications/livepkgr/streams/_definst_/Archive
    So Archive Server is: 192.168.1.100/Archive.
    Now I have problem to store live streams in the Archive server...!!!
    When I start publish live events on the Flash Media Server, AMS is make video file with f4v format in the Archive server (ams have access to write in other server) But this file has 0kb. there is no capacity.
    Please advise me to solve this problem.
    Many Thanks.

    This seems like a good guide:
    http://www.marcoach.nl/index.php/nl/tech-tips/55-os-x-mountain-lion-webmail.html
    Lion Server (10.7) does still have webmail on board, although this is not the forum, this is.

  • Reconfigure Open Directory in Yosemite Server

    Is it possible to delete and reconfigure Open Directory in Yosemite server?
    The host name and configuration were modified after Open Directory was activated and I get the message "Unable to load replica list" in the Settings Tab of Open Directory on the Server App (Server 4.0.3 (Build 14S350)). I think the best way would be to start over the automatic configuration.

    Many Open Directory problems can be resolved by taking the following steps. Test after each one, and back up all data before making any changes.
    1. The OD master must have a static IP address on the local network, not a dynamic address. It must not be connected to the same network with more than one interface; e.g., Ethernet and Wi-Fi.
    2. You must have a working DNS service, and the server's hostname must match its fully-qualified domain name. To confirm, select the server by name in the sidebar of the Server application window, then select the Overview tab. Click the Edit button on the Host Name line. On the Accessing your Server sheet, Domain Name should be selected. Change the Host Name, if necessary. The server must have at least a three-level name (e.g. "server.yourdomain.com"), and the name must not be in the ".local" top-level domain, which is reserved for Bonjour.
    3. The primary DNS server used by the server must be itself, unless you're using another server for internal DNS. The only DNS server set on the clients should be the internal one, which they should get from DHCP if applicable.
    4. Only if you're still running Mavericks server, follow these instructions to rebuild the Kerberos configuration on the server.
    5. If you use authenticated binding, check the validity of the master's certificate. The common name must match the hostname and domain name. Deselecting and then reselecting the certificate in Server.app has been reported to have an effect in some cases. Otherwise delete all certificates and create new ones.
    6. Unbind and then rebind the clients in the Users & Groups preference pane. Use the fully-qualified domain name of the master.
    7. Reboot the master and the clients.
    8. Don't log in to the server with a network user's account.
    9. Disable any internal firewalls in use, including third-party "security" software.
    10. If you've created any replica servers, delete them.
    11. If OD has only recently stopped working when it was working before, you may be able to restore it from the automatic backup in /var/db/backups, or from a Time Machine snapshot of that backup.
    12. As a last resort, export all OD users. In the Open Directory pane of Server, delete the OD server. Then recreate it and import the users. Ensure that the UID's are in the 1001+ range.
    If you get this far without solving the problem, then you'll need to examine the logs in the Open Directory section of the log list in the Server app, and also the system log on the clients.

  • How can I create a directory on my server using FTP in applescript?

    Hi
    I have created a script which successfully uploads files to my server using FTP and a curl command.
    The one thing I haven't worked out how to do is how to create a directory on my server using applescript. At the moment I have to manually create the correct directory for the script to work, but would like to be able to automate everything.
    Would this be possible?
    Thanks
    Nick

    Thanks Bernard.
    I have tried that and still get the error message:
    "curl: Can't open '--ftp-create-dirs'!
    curl: try 'curl --help' or 'curl --manual' for more information
    % Total % Received % Xferd Average Speed Time Time Time Current
    Dload Upload Total Spent Left Speed
    0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0
    curl: (9) Server denied you to change to the given directory"
    The other relevant parts of the script are:
    set ftp_url to "ftp.****.org.uk/artists/" & artistName2 & "/assets/"
    set ftp_username to "*****@*****.org.uk"
    set ftp_password to "****"
    set ftp_entire to ftp_username & ":" & ftp_password & " ftp://" & ftp_url
    set outPath to outputFolder & fileCounter & ".jpg"
    and then...
    do shell script ("curl -T --ftp-create-dirs " & outPath & " -u " & ftp_entire)
    Like I said, all works fine when the directories are already in place. I have permissions for the artists folder set to "777", i.e. writable. Because the script is trying to create two directories at once, i.e. "artistName2" and then "assets" under it, I wondered if this was significant, but then I tried with just one new directory, and still got the error.

Maybe you are looking for

  • BAPI for Price Change using Transaction VK12

    Hi All.. Can any one help in locating a BAPI for making price changes using transaction VK12. Rgds

  • Standard Network Report

    Hello, Is there any report available, where the std Network is the input and output should be the operative WBS elemnets .i,e a standard network is used in some 100 projects,if I the std network as input the report should display all WBS elements in

  • Netflix problems with macbook pro

    does anyone have a problem accessing netflix? I had a geek squad agent install silverlight which he said is the operating system netflix uses, and it worked for awhile but now im again getting the error message that the netflix site is temporarily un

  • First aim..if combobox onchange send data to listbox that I create before

    ow to send a data from combobox to list box in java when I click datas in combobox

  • Question about Level Meter Plug in

    I know how to change the channel meter between pre/post fader metering is there a way to do this with Logic's Level Meter plug in and with Inspector?