Generate File in two node apps architecture.

Hi All,
We have two node apps environment. Node 1: Database Tier and Node 2 Application Tier. APPL_TOP is not shared.
We want to generate a file from database (says payment batch) and transfer this file to another server for sending to bank
Currently we are using a database procedure called from concurrent request to generate the file and is generated in the utl_file directory which is on database server. The same concurrent request will do the file transfer, but sincne the file is generated in the database server the scripts in the apps tier will not be able to see it. We are not intended to use SFTP or NFS. So how can we do this using standard oracle apps method?
Thanks

thanks for the update.
One option in my mind is to call anonymous pl/sql script on apps tier from concurrent program and spool the output.
What you think?

Similar Messages

  • Best way to transfer file between two web apps

    I'm looking for some architectural suggestions. Seems like a simple problem, but I'm really struggling.
    -- Problem:
    We have two web servers, both running apps in JBoss. I need to pass a file of roughly 20MB from one server to the other, and get the other server to process it as soon as it receives it.
    -- Details:
    Seems simple enough... My plan was to write a web service to pass it. Our company is all onboard the SOA train. I wrote a service that encoded the file as a byte[] and passed it. Works amazing for small files, but throws an Axis OutOfMemory error when trying to base64 encode the file. I then looked at using attachments for web services, but dismissed that. It creates malformed XML since it treats the service as a multipart MIME message wraps the webservice inside. It quickly became messy and I couldn't get it working. Seemed to defeat the purpose of an XML service too.
    I've considered just doing a HTTP POST of the file to the other server, but that's poorly documented and a bit of a hack way to do it. I also considered just dropping the file in a web accessible folder, then using a web service to pass a URL and get the other server to retrieve it via HTTP. Again, not a very nice solution.
    Anyone have any suggestions? Thanks.

    Kimos2 wrote:
    I don't have access to run/configure services on these machines, so FTP is out. It has to be JVM to JVM. Even if I did, I'd have to worry about triggering JBoss to process the file right away after it had been dropped there. I know i could poll the directory or send a message to trigger the process, there'd be synchronization issues etc to work through and seems to introduce unneeded complication.Yes, I agree with that. Having been through the polling directories and grabbing partially uploaded files scenario ad nauseam. It's much worse with large files too.
    Barring any suggestions I haven't thought of, I'll probably end up settling on the HTTP POST solution. My problem is that it's providing a service without a specification and is not self-describing. No WSDL file to pass to clients that plan on using it. I would do HTTP POST as well. I didn't understand the part about "without a specification". You do have a specification, don't you? Wasn't this it: "I need to pass a file of roughly 20MB from one server to the other, and get the other server to process it as soon as it receives it." Informal, yes, but I don't have a problem with that. Post it to a URL on the receiving server and set up the receiving server to do whatever it's supposed to do with the file.

  • Accessing separate databases for two Wb Apps.

    I have two entries in my weblogic.properties file for two web apps:
    # Netstride portal deployment
    weblogic.httpd.webApp.webapp=D:/dev/portal/public_html/
    # Defines the ISPS portal WAR deployment.
    weblogic.httpd.webApp.isps=D:/dev/ISPS_app/public_html/
    and my connection Pool entry in weblogic.properties file is:
    weblogic.jdbc.connectionPool.commercePool=\
    url=jdbc:weblogic:oracle,\
    driver=weblogic.jdbc.oci.Driver,\
    loginDelaySecs=0,\
    initialCapacity=3,\
    maxCapacity=50,\
    capacityIncrement=1,\
    allowShrinking=true,\
    shrinkPeriodMins=2,\
    testConnsOnReserve=true,\
    testTable=dual,\
    refreshMinutes=1,\
    props=user=inet2;password=in3t2;server=ispsdevl;weblogic.t3.waitForConnection=true;weblogic.t3.waitSecondsForConnection=999999999999,\
    weblogic.jts.waitSecondsForConnectionSecs=999999999999,\
    verbose=false
    The webLogic JDriver entry in weblogiccommerce.properties file is:
    #------WebLogic jDriver for Oracle 8.1.5 jdbc20---------#
    commerce.usermgmt.RDBMSRealm.driver=weblogic.jdbc.oci.Driver
    commerce.usermgmt.RDBMSRealm.dbUrl=jdbc:weblogic:oracle
    commerce.usermgmt.RDBMSRealm.dbServer=ispsdevl
    #commerce.usermgmt.RDBMSRealm.dbUser=inet
    #commerce.usermgmt.RDBMSRealm.dbPassword=in3t
    # FOR GOING AGAINST INET2 SCHEMA
    commerce.usermgmt.RDBMSRealm.dbUser=inet2
    commerce.usermgmt.RDBMSRealm.dbPassword=in3t2
    If I want to point the 1st app to database schema inet and the 2nd app to inet2
    schema, what changes in weblogiccommerce.properties.
    I think in weblogic.properties file i will have to have two different connection
    pools for the two apps. am I right?
    I zipped the weblogic.properties and weblogiccommerce.properties files into webLogic.zip
    and attaching the zip file.
    Thanks.
    [WebLogic.zip]

    i think you need to use link server to create the bridge.
    http://msdn.microsoft.com/en-us/library/ms188279.aspx

  • 11gR2- webutil upload file to AS on two nodes RAC?

    Hellow experts plz help with the following issue,
    we are using 11gR2 forms on two node rac,webutil configured on both nodes. upload/download files to AS folder(UP_FILES) using our forms. node 2 is actually replica of node 1(forms,reports,UP_FILES). we are now facing a problem that the form which upload file to AS, only upload that to the node from where it is running say if it is running from node1 then it upload files to folder UP_FILES on that node( as the entry in webutil.cfg file) but we want that UP_FILES folder should be synced with each other on both nodes and a form running whether from node1 or node2 upload the file to both nodes at time.
    How this will be accomplished?
    --------webutil.cfg entry
    transfer.appsrv.read.3=D:\UP_FILES
    transfer.appsrv.write.3=D:\UP_FILES
    -----FORM UPLOAD CODE
          IF :CONTROL.FILE_LOC IS NOT NULL THEN
            acyr3 := :CONTROL.TXTVOUCHERNO||'-'||acyr2 ;
           FILE_RESULT := WEBUTIL_FILE_TRANSFER.CLIENT_TO_AS_WITH_PROGRESS(CLIENTFILE=>:CONTROL.FILE_LOC,
           SERVERFILE =>'D:\UP_FILES\'||acyr3||'.PDF',
          PROGRESSTITLE=>'UPLOAD TO DATABASE IN PROGRESS',
          PROGRESSSUBTITLE=>'PLEASE WAIT' );
          END IF;
    --FORM DOWNLOAD CODE
            FILE_RESULT := WEBUTIL_FILE_TRANSFER.AS_TO_CLIENT_WITH_PROGRESS
                 CLIENTFILE=>'D:\UP_FILES\'||acyr3||'.PDF',
            SERVERFILE =>'D:\UP_FILES\'||acyr3||'.PDF',
          PROGRESSTITLE=>'DOWNLOAD FROM DATABASE IN PROGRESS',
          PROGRESSSUBTITLE=>'PLEASE WAIT'
                 CLIENT_HOST('rundll32.exe url.dll,FileProtocolHandler D:\UP_FILES\'||acyr3||'.PDF');
                 IF FILE_RESULT THEN
      message('File downloaded successfully from the Application Server');
      END IF;

    Well, you're uploading the file to one node, so this is no surprise. What you can do is to store your file on a shared folder (which might be a bit tricky on windows) or synchronize a folder between your nodes. Unfortunately Forms can't access ASM directly (assuming you are using ASM), so you can't store your files directly in ASM where it would be accessible from both nodes.
    cheers

  • Oracle Apps RDF Report generates file with ^M (Carriage return) characters

    We are upgrading from 11i to r12 (also moving from HP unix to AIX on hardware side) and this rdf report generates a text file which is then eventually sent to a third party system for printing. Report works just fine in 11i as it generates a file without any ^M chars but in R12 environment, the same report generates file with ^M chars.
    Also want to understand how the Oracle Apps system hands over the output files to O/S which is saved.
    Pls advice
    Thanks
    Ram M.

    What type of concurrent program are we talk about here?
    This is a Report type of Concurrent program
    How does the text file get generated?
    Report is generating a Text file on :desname, a custom folder. 
    Was the binary mode used to copy the files from the old server to the new one?
    I believe RDF Reports were copied over in Binary format but none of the text files generated from old server are copied over.

  • How to generate xml file with multiple nodes using sqlserver as database in SSIS..

    Hi ,
    I have to generate the xml file using multiple nodes by using ssis and database is sqlserver.
    Can some one guide me on to perform this task using script task?
    sudha

    Why not use T-SQL for generating XML? You can use FOR XML for that
    http://visakhm.blogspot.in/2014/05/t-sql-tips-fun-with-for-xml-path.html
    http://visakhm.blogspot.in/2013/12/generating-nested-xml-structures-with.html
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My Wiki User Page
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • Single JVM, two nodes, separate override files configuration?

    This question might have surfaced before in the forums but even after going through several posts I couldn't get a decent understanding on how this works. Here is my question...
    I wanna bring up two coherence nodes in a single JVM, both independent and unrelated to each other. These two nodes have different cache-config xml files, and also override files. My question is how do I configure these nodes such that they work with two separate override files. I don't wanna use -Dtangosol.coherence.override because my doing this, this system property applies to both the nodes so I can't keep the override xml files separate.
    So, what is best way to accomplish this such a way that the nodes don't collide with each other by using a common override files?
    I hope I am clear.
    Thank you.

    Hi,
    Both the solution I posted and Jon's solution rely on a ClassLoader to isolate the Coherence nodes within the JVM. We do not use this in Production but we do use it as part of our build process and run over 18,000 test cases with a single JVM cluster, so in that respect it works well. I think if you used it in Production you would need to have done a lot of testing to make sure that there are no very subtle issues that might come back to bite you later.
    The reason doing what you want is difficult is because Coherence uses a number of singletons at runtime - for example the Cluster instance, and without using a class loader it is difficult to have more than one of these. Coherence also uses a lot of System properties to override configuration values so these may need to be isolated too.
    Finally using a class loader would make if difficult for your application code to interact with the two isolated nodes inside the single JVM.
    It might be better if you told us why you think you need to run two nodes in a single JVM - what requirement are you trying to meet? There might be a much more sensible solution to your requirements. As NJ has suggested using Extend might be one solution but without knowing more about what you want to do I couldn't say for sure.
    If you don't know what Extend is then read the documentation - The Client Guide covers Extend http://docs.oracle.com/cd/E24290_01/coh.371/e22839/toc.htm and the rest of the documentation is here http://docs.oracle.com/cd/E24290_01/index.htm
    JK

  • Control file not updated till date in RAC two node database

    Hi,
    Good day.
    To my surprise i found that control files, in our database RAC with two nodes, are not updated till date except one from node 1 and another from node 2.
    But we got all the controlfile names in the init parameter file, in both the nodes.
    Node 1
    bash-3.00$ ls -lrt
    total 922128
    -rw------- 1 oracle dba 157286400 Nov 3 2007 cntrl03.dbf
    -rw------- 1 oracle dba 157286400 Nov 3 2007 cntrl02.dbf
    -rw------- 1 oracle dba 157286400 Apr 12 17:21 cntrl01.dbf
    Node 2
    bash-3.00$ ls -lrt
    total 922128
    -rw------- 1 oracle dba 157286400 Mar 30 09:17 cntrl03.dbf
    -rw------- 1 oracle dba 157286400 Mar 30 09:17 cntrl02.dbf
    -rw------- 1 oracle dba 157286400 Apr 12 17:20 cntrl01.dbf
    As you could see cntrl01.dbf is the only file updated till date in both the nodes. The database is running fine.
    what is actually happening?
    Any suggestions?
    Thanks,
    Balu.

    As you're using RAC it is very likely that you are using an SPFILE.
    Connect as a DBA account (or an account with SELECT ANY DICTIONARY privileges) in both the instances and do "show parameter spfile" and "show parameter control" (or query V$parameter)
    It is likely that the spfile is referencing only 1 controlfile.

  • How to add a second database along with existing two node RAC environment

    Hi,
    I was wondering if anyone can help me with this.
    My Environment:
    1. Two node RAC Cluster database (11.2.0.2) with ASM running perfectly (Oracle Sid = test-1)
    2. I have installed a second single instance db on node 2 (Oracle Sid = test-2) with NTFS file system for datafiles
    3. Database is up and running, but I am not able to connect it from any client.
    4. I am getting ORA-12514:  TNS:listener does not currently know of service requested in connect descriptor
    5. Database (test2) is registered with grid LISTENER
    5. TNSPING from client machine response is ok
    Am I missing something here, happy to provide more info if requested.
    Thanks,
    PS

    C:\Users\root.test_prod>lsnrctl
    LSNRCTL for 64-bit Windows: Version 11.2.0.2.0 - Production on 23-FEB-2012 11:58:05
    Copyright (c) 1991, 2010, Oracle. All rights reserved.
    Welcome to LSNRCTL, type "help" for information.
    LSNRCTL> status
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))
    STATUS of the LISTENER
    Alias LISTENER
    Version TNSLSNR for 64-bit Windows: Version 11.2.0.2.0 - Production
    Start Date 18-FEB-2012 19:51:47
    Uptime 4 days 16 hr. 6 min. 24 sec
    Trace Level off
    Security ON: Local OS Authentication
    SNMP OFF
    Listener Parameter File C:\app\11.2.0\grid\network\admin\listener.ora
    Listener Log File C:\app\11.2.0\grid\log\diag\tnslsnr\IRIS11G-DB-2\listener\alert\log.xml
    Listening Endpoints Summary...
    (DESCRIPTION=(ADDRESS=(PROTOCOL=ipc)(PIPENAME=\\.\pipe\LISTENERipc)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=10.30.0.202)(PORT=1520)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=10.30.0.215)(PORT=1520)))
    Services Summary...
    Service "+ASM" has 1 instance(s).
    Instance "+asm2", status READY, has 1 handler(s) for this service...
    Service "test-11gRAC.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Service "test-11gRXDB.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Service "irisapps.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Service "test-2" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    Service "test-2XDB" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    The command completed successfully
    LSNRCTL> service
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))
    Services Summary...
    Service "+ASM" has 1 instance(s).
    Instance "+asm2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:1322 refused:0 state:ready
    LOCAL SERVER
    Service "test-11gRAC.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:1829 refused:0 state:ready
    LOCAL SERVER
    Service "test-11gRXDB.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "D000" established:0 refused:0 current:0 max:1022 state:ready
    DISPATCHER <machine: IRIS11G-DB-2, pid: 4496>
    (ADDRESS=(PROTOCOL=tcp)(HOST=IRIS11G-DB-2.test_prod.internal)(PORT=57695))
    Service "irisapps.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:1829 refused:0 state:ready
    LOCAL SERVER
    Service "test-2" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:0 refused:0 state:ready
    LOCAL SERVER
    Service "test-2XDB" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "D000" established:0 refused:0 current:0 max:1022 state:ready
    DISPATCHER <machine: IRIS11G-DB-2, pid: 9340>
    (ADDRESS=(PROTOCOL=tcp)(HOST=IRIS11G-DB-2.test_prod.internal)(PORT=58653))
    The command completed successfully
    LSNRCTL>

  • Two node installation with 2 concurrent managers..

    Need your help. I have 12.1.1 Two node installation on Linux 5.9 with 64 bit.
    Linux2- Database server + Concurrent manager
    Linux1- Apps Node
    Port - 8002 (port pool -2)
    on my app node - linux1, I have 2 directories <SID>_linux1 and <SID>_linux2. First one is for primary apps node and second one is for 2nd concurrent manager. It is a shared directory structure. I have 2 questions:
    1) If I run the autoconfig on Linux1, I can run all the services and connect to apache. However, If I run the autconfig on linux2, I will not be able to connect to web server (for a short time request will be sent to Linux1 and then it gets forwarded to Linux2 - even though Linux2 is not app server).
    2) How do I run the second concurrent manager on LInux2? After I run the 2 concurrent manager, how do I test both concurrent managers are working?
    Your help is much appreciated.

    Hi Hussein,
    Thank you for the info. I was wondering whether the Linux2 concurrent manager was installed correctly or not. I went to xml file and found these entries and it made me think, it didn't install correctly because only concurrent manager should have been yes and the rest should be no. Isn't it? If it is wrong, how do I correct this?
    <oa_system_config>
             <TIER_DB oa_var="s_isDB">NO</TIER_DB>
             <TIER_ADMIN oa_var="s_isAdmin">YES</TIER_ADMIN>
             <TIER_WEB oa_var="s_isWeb">YES</TIER_WEB>
             <TIER_FORMS oa_var="s_isForms">YES</TIER_FORMS>
             <TIER_NODE oa_var="s_isConc">YES</TIER_NODE>
             <TIER_FORMSDEV oa_var="s_isFormsDev">YES</TIER_FORMSDEV>
             <TIER_NODEDEV oa_var="s_isConcDev">YES</TIER_NODEDEV>
             <TIER_WEBDEV oa_var="s_isWebDev">YES</TIER_WEBDEV>
             <config_option type="techstack" oa_var="s_techstack">as1013</config_option>
             <config_option type="techstack" oa_var="s_tnsmode">generateTNS</config_option>
             <config_option type="adx" oa_var="s_apps_version">12.1.1</config_option>
             <config_option type="techstack" oa_var="s_apache_mode">NORMAL</config_option>
             <ias_version oa_var="s_ias_version">10.1.3.4.0</ias_version>
          </oa_system_config>

  • R12 two node installation

    Hi All,
    I have to do two node R12 installation, please guide me how to procede because I could not see any option like "Sinlge node " and "Mult node" as it was in 11i, I want to have my database on one server and my application on second server
    Regds

    Hi
    I am new to Oracle E-business suite. I am trying to install R12 on single node/sincgle user istall on Linux 32bit. I followed the Oracle Doc ID:4012310.1 and follwoing the instructions from http://download-east.oracle.com/docs/cd/B40089_09/current/acrobat/120oaig.pdf.
    I created stage area by extracting the files downloaded from Oracle edelivery site.
    Created user "Oracle" and gave permissions to access stageR12 directory.
    Logged in as Oracle and try to run ./rapidwiz throws the follwoing error:
    Rapid Install wizard is validating your filesystem.....
    4 DVD labels found
    Rapid install wizard will now launch the java interface .....
    [oracle@localhost rapidwiz]$ Exceprion in the thread "main" java.lang.noclassdeffounderror:
    Orcle/apps/ad/util/logfile
    at oracle.apps.as.rapidwiz.RIWizard.main(RIWIZard.java.1419)
    Please help
    Thanks
    Raj

  • [b]Migrating the DB-Tier (DB and CM) to Two node non RAC cluster[/b]

    Hi,
    The current set-up of our E-business suite is a two node install:-
    The DB Tier (Database and Concurrent Manager) on one node
    The Apps Tier (Forms /Web Server) on another node.
    For the HA solution (NON ORACLE RAC) we are planning to:-
    Move the DB Tier (Database and Concurrent Manager) to a three node hardware sun cluster managed by veritas cluster manager (NOT ORACLE RAC). We need to know will the Database Tier (Database and Concurrent Manager) work on Hardware cluster node and will it support COLD FAILOVER from one Node to another. We know the database on its own would be fine with a cold failover because we have tested the database cold failover on the three node cluster for non E-business suite database. But here we have the added thing of the Concurrent manger sitting on the node along with the Database on the DB Tier.
    The Apps Tier (Forms / Web server) will be put on a separate set of server using Load balances etc.
    Has anybody implemented similar HA set-up and will this planned set-up work or are there any issues with this.
    Any help / info would be appreciated.
    Thanks

    Hi,
    Yes, you can do the cold failover the database and all the 11i services also.
    1. In the concurrent manager service
    == when you failover the concurrent manager do the following things before failover.
    create the listener/tnsnames files which includes the new hostname and keep it with the veritas failover service. i mean, when you failover, these files should replace the existing files, before the existing files should backedup. and create a script to change the hostname,logfile_hostname,outfile_hostname in fnd_concurent_processes table.
    add the nodes in the install->nodes navigation
    do the failover manually and check the listener files are properly pinging using tnsping. and start the concurrent manager.
    so tatally, you have to prepare two sql scripts
    one for change the hostname from node b to node a
    one for node a to node b.
    and 2 listener + 2 tnsnames files which contains the seperate hostname accordingly.
    use the adcmctl.sh top stop and start
    finally, create a shell script to kill the sysmgr for the current managers when the manager takes long time to shutdown. before run the kill script, wait for 5 mts atleast.
    I done the same scenario many times, with veritas failover service with 2 sun v880 servers.
    regards,
    Pandian

  • Simple two node Cluster Install - Hung after reboot of first node

    Hello,
    Over the past couple of days I have tried to install a simple two node cluster using two identical SunFire X4200s, firstly following the recipe in: http://www.sun.com/software/solaris/howtoguides/twonodecluster.jsp
    and when that failed referring to http://docs.sun.com/app/docs/doc/819-0912 and http://docs.sun.com/app/docs/doc/819-2970.
    I am trying to keep the install process as simple as possible, no switch, just back to back connections for the internal networking (node1 e1000g0 <--> node2 e1000g0, node1 e1000g1 <--> node2 e1000g1)
    I ran the installer on both X4200s with default answers. This went through smoothly without problems.
    I ran scinstall on node1, first time through, choosing "typical" as suggested in the how to guide. Everything goes OK (no errors) node2 reboots, but node1 just sits there waiting for node2, no errors, nothing....
    I also tried rerunning scinstall choosing "Custom", and then selecting the no switch option. Same thing happened.
    I must be doing something stupid, it's such a simple setup! Any ideas??
    Here's the final screen from node1 (dcmds0) in both cases:
    Cluster Creation
    Log file - /var/cluster/logs/install/scinstall.log.940
    Checking installation status ... done
    The Sun Cluster software is installed on "dcmds0".
    The Sun Cluster software is installed on "dcmds1".
    Started sccheck on "dcmds0".
    Started sccheck on "dcmds1".
    sccheck completed with no errors or warnings for "dcmds0".
    sccheck completed with no errors or warnings for "dcmds1".
    Configuring "dcmds1" ... done
    Rebooting "dcmds1" ...
    Output from scconf on node2 (dcmds1):
    bash-3.00# scconf -p
    Cluster name: dcmdscluster
    Cluster ID: 0x47538959
    Cluster install mode: enabled
    Cluster private net: 172.16.0.0
    Cluster private netmask: 255.255.248.0
    Cluster maximum nodes: 64
    Cluster maximum private networks: 10
    Cluster new node authentication: unix
    Cluster authorized-node list: dcmds0 dcmds1
    Cluster transport heart beat timeout: 10000
    Cluster transport heart beat quantum: 1000
    Round Robin Load Balancing UDP session timeout: 480
    Cluster nodes: dcmds1
    Cluster node name: dcmds1
    Node ID: 1
    Node enabled: yes
    Node private hostname: clusternode1-priv
    Node quorum vote count: 1
    Node reservation key: 0x4753895900000001
    Node zones: <NULL>
    CPU shares for global zone: 1
    Minimum CPU requested for global zone: 1
    Node transport adapters: e1000g0 e1000g1
    Node transport adapter: e1000g0
    Adapter enabled: no
    Adapter transport type: dlpi
    Adapter property: device_name=e1000g
    Adapter property: device_instance=0
    Adapter property: lazy_free=1
    Adapter property: dlpi_heartbeat_timeout=10000
    Adapter property: dlpi_heartbeat_quantum=1000
    Adapter property: nw_bandwidth=80
    Adapter property: bandwidth=70
    Adapter port names: <NULL>
    Node transport adapter: e1000g1
    Adapter enabled: no
    Adapter transport type: dlpi
    Adapter property: device_name=e1000g
    Adapter property: device_instance=1
    Adapter property: lazy_free=1
    Adapter property: dlpi_heartbeat_timeout=10000
    Adapter property: dlpi_heartbeat_quantum=1000
    Adapter property: nw_bandwidth=80
    Adapter property: bandwidth=70
    Adapter port names: <NULL>
    Cluster transport switches: <NULL>
    Cluster transport cables
    Endpoint Endpoint State
    Quorum devices: <NULL>
    Rob.

    I have found out why the install hung - this needs to be added into the install guide(s) at once!! - It's VERY frustrating when an install guide is incomplete!
    The solution is posted in the HA-Cluster OpenSolaris forums at:
    http://opensolaris.org/os/community/ha-clusters/ohac/Documentation/SCXdocs/relnotes/#bugs
    In particular, my problem was that I selected to make my Solaris install secure (A good idea, I thought!). Unfortunately, this stops Sun Cluster from working. To fix the problem you need to perform the following steps on each secured node:
    Problem Summary: During Solaris installation, the setting of a restricted network profile disables external access to network services that Sun Cluster functionality uses, ie: The RPC communication service, which is required for cluster communication
    Workaround: Restore external access to RPC communication.
    Perform the following commands to restore external access to RPC communication.
    # svccfg
    svc:> select network/rpc/bind
    svc:/network/rpc/bind> setprop config/local_only=false
    svc:/network/rpc/bind> quit
    # svcadm refresh network/rpc/bind:default
    # svcprop network/rpc/bind:default | grep local_only
    Once I applied these commands, the install process continued ... AT LAST!!!
    Rob.

  • Add a file upload to spreadsheet app

    Created a wizard-generated out of the box app based on spreadsheet.
    Now I want the ability to store a file attachment for each row. So I added the file_id column to the table as a FK to my custom table which stores the uploaded files.
    What is a good UI to handle updates to this?
    On the Create page, I added a File Browse box to upload the file, works fine.
    Added a Download link on the report page, also works fine.
    On the Update page, how should this file_id field be handled? There need to be 2 options (that I can see)
    a. delete the file (delete it from the custom table and null out the file_id column)
    b. replace it with a new file (delete the old file from the custom table, upload the new file and update the file_id pointer)
    How can this be done on the wizard-generated Update Form page?
    Thanks

    OK this is what I ended up doing
    On the Update form, I added a Display as Text item and created an OnLoad computation to populate it with a Download link
    select '< a href="foo.download_file?p_file_id='||file_id||
           '">< img src="/i/themes/theme_10/showhide_show.gif" / >< /a >'
    from mytab
    where id=:P3_IDAdded a File Browse item (P3_FILENAME).
    Added a After Submit process that fires after the wizard-generated Automatic DML process that does
    DECLARE
      l_file_id  NUMBER;
    BEGIN
      IF (:P3_FILENAME IS NULL ) THEN
        RETURN;
      END IF;
      DELETE FROM myfiles
      WHERE       file_id = (SELECT file_id
                             FROM   mytab
                             WHERE  id = :p3_id);
      UPDATE mytab
      SET    file_id = NULL
      WHERE  id = :p3_id;
      INSERT INTO myfiles (file_id,file_name,mime_type,file_size,created_by,created_at,blob_content )
      SELECT id,
             filename,
             mime_type,
             doc_size,
             updated_by,
             updated_on,
             blob_content
      FROM   htmldb_application_files
      WHERE  NAME = :p3_filename;
      UPDATE mytab
      SET    file_id = (SELECT id
                        FROM   htmldb_application_files
                        WHERE  NAME = :p3_filename)
      WHERE  id = :p3_id;
      DELETE FROM htmldb_application_files
      WHERE       NAME = :p3_filename;
    END;In plain English:
    1. Allow the user to view/download the file
    2. If they select/enter a file in the File Browse box, delete the existing file and replace it with the new file and store a pointer to it in the table.
    It is working like a charm.
    Hope this helps someone.

  • How Do RoboHelp 9 WebHelp Generated Files Handle Map IDs and Aliases?

    The text below was written by our team's developer/architect. I am the help author who uses RoboHelp to write content and generate the help files, but I am clueless how it all gets generated and is deployed. Please help. We use RoboHelp 9. I use it in Windows XP and our app and help run on IE 7, 9, and Firefox (multiple versions).
    "Our application uses the numeric identifiers associated with the Map ID. For example, to get to the <appname>_home_page.htm file, we use the number 1053. <appname> = pecs, in this example.
    All of this is used in a call to a RoboHelp method defined in the RoboHelp_CSH.js file. The mehtod we are calling is the RH_ShowHelp() JavaScript method and the code to perform the call, when you click on Page Help, is this:
    RH_ShowHelp(0, ''/pecsHelp/index.htm>pecsHelp',HH_HELP_CONTEXT,topic);
    Topic is translated to the Map ID number for the page help. HH_HELP_CONTEXT is defined in the RoboHelp_CSH.js file. This method translates into a URL and from what I have seen, the URL that gets generated is this:
    http://{server}[:port]/pecsHelp/index.htm/{server}[:port]/pecsHelp/index.htm#<id=1053>>pecsHelp
    Server and port get replaced with the appropriate values. I have no clue how id=1053 is supposed to get translated to mean "pecs_home_page.htm". If you check the PECS_help.h file, you will see the following entry:
    #define PECS_Home_Page1 1053
    Then in the RoboHelp alias file (PECS 3.0.ali), the following line is in the file:
    <alias name="PECS_Home_Page1" link="pecs_home_page.htm"> </alias>
    But both of these files are used during the WebHelp generation process and I don't know how the WebHelp generated files handle the Map ID and aliases."

    You need to assign the numbers you find in the pecs_help.h file to topics in your help. You do this in Context Sensitive Help > Map Files > All Map IDs. (From RH7, but I assume the location is similar in RH9.) This creates the entries in the .ali file.
    Peter Grainge suggests a couple of sites to read for a greater understanding here:
    http://www.grainge.org/pages/authoring/calling_webhelp/using_map_ids.htm
    (Although the second  site is based on RH X5, the basic concepts and procedures should be very similar. )
    HTH,
    Amber

Maybe you are looking for

  • Windows 7 bootcamp show black screen with a small white bar

    I just upgraded my iMac mid 2011 21.5" with a crucial m4 SSD. The SSD is plugged in the disk 1 DATA port and on the SSD power on the motherboard. I fully installed Windows 7 with all the driver and updates. My mac is upgraded with the lastest update.

  • Running out of harddrive space - here's a fix.

    Twice now I've run out of harddrive space on my t500 with a 160GB harddrive installed. The first time I called the IBM support people who were as useful as a chocolate teapot. This time I turned to the Microsoft forums and finally got it figured out.

  • Unable to save imported contacts file or sync

    I imported a contacts .csv file into palm desktop for a Z22.  It opens well in palm desktop, but does not seem to save or sync. I tried just entering a contact into desktop and even this would not sync with the Z22. I have my drive partitioned with W

  • Add or remove elements from a list attached to session duplicated the size

    Hi All, I have a jsp page that in there I'd like to show three columns in a table:      <c:forEach var="thesaurus" items="${sessionScope.collection.thesaurusList}">        <tr>                     <td class="check"><input type="checkbox" name="word"

  • Can't create netinstall image

    Hi, I'm currently trying to create a Net Install image from the system volume of another computer in Target Disk Mode. After selecting the source and entering the info, I am told that the source does is missing a required "archive file." I'm thinking