OSB and server-less SAN based backup.

Hi All,
This is my first question in this forum. Is SAN based backup possible in OSB. If so how can I configure this. Thank you very much.
Regards
Aminul

Thanks Donna. Since we are planning to implement OSB in our environment, hopefully we will get more support from this forum. could please put some lights on"OSB supports SAN environments dynamically sharing tape drives between SAN attached media servers".Thanks once again.
Regards
Aminul
Edited by: user12203512 on Oct 19, 2011 8:55 PM

Similar Messages

  • OSB and SOA Server (Mediator) Design question

    As OSB is the strategic ESB and is used for external services does it make sense that all communication must go out via the OSB.
    For example if I have a request that comes in gets routed through the OSB and then calls a backend composite which contains a BPEL that sits on the Oracle SOA server that calls multiple external services (CRM, SAP etc..) Would all the calls out from the BPEL then go back to the OSB and out to the external services (via proxy/business services.) This seems to add a lot of network hops to the whole architecture.
    I can't really seem to find a diagram to explain this but an external service consumer could call the Oracle SOA stack that will come in via the OSB (for security gateway as OWSM 11g doesn't support gateway yet.) Then get transformed into the CDM and then passed on to the Oracle SOA server that has a composite service with BPEL that orchestrates multiple calls to internal and external services. The external ones requiring a transformation back into the service providers format.
    How does this flow work?

    If you are talking to external B2B systems you should use the Oracle B2B Server, have a look at this link,
    http://download.oracle.com/docs/cd/E14571_01/integration.1111/e10229/b2b_intro.htm#CEGGAGJA
    Section 1.6 in there gives a good example of doing just that. The OSB mediator is responsible for talking between CDM's in this case so it does some internal work, within one CDM set of services and then when ready the mediator translates from CDM to the format needed for the B2B. The B2B Server then does all the clever stuff needed to talk to the B2B partner.
    In the internal case where the call is within the Enterprise the Oracle documentation states that there may be cases when doing this translation from CDM to legacy format may take place within the SOA server but recommends that it is normally done using the mediator in the same way as with partner calls. The 11g developer documentation states "The stated direction by Oracle is for the Oracle Service Bus to be the preferred ESB for interactions outside the SOA Suite. Interactions within the SOA Suite may sometimes be better dealt with by the Oracle Mediator component in the SOA Suite, but we believe that for most cases, the Oracle Service Bus will provide a better solution and so that is what we have focused on within this book."
    Just what those exceptions are is an interesting question. If you have a CDM and all of your legacy world has translators you will spend all of your time hoping up to the mediator and back down maybe that is the exception. If everything is using the mediator then the SOA orchestration power is lost. If on the other hand most of your system lives within a canonical schema and there is one exception perhaps you make it pay the price for non-conformance.

  • Cannot login to network account (leopard client and server)

    Up until now, I have used local accounts on my leopard server. I want to start experimenting with OD prior to implementing. I created a new user account in the /LDAPv3/127.0.0.1 domain, and have bound my leopard client to the server using directory utility. On the login screen, "Network Accounts Available" has a green button to the left of it. When I try and login to the server account, the login window just shakes. At first, I could enter the password and then it would prompt me for a new password. Trying to enter a new password would not allow me to login. I went back to server admin and disabled the "require new password" setting, (as well as the other good security policies)...
    I have also reset the password in WGM, and made sure to disable all the security stuff there too...
    Lastly, I have deleted the server in directory utility, rebooted, then added it back in, and rebooted again...
    I still cannot login to the server account, the login screen just shakes
    Does anyone have an idea of what settings and or logs I can check to try and narrow down what is going on?
    Thanks in advance....

    to close out the thread, I have working dns on my network, but I did not have dns enabled on my server. I enabled the dns service and entered just the info for my server, then assigned my server and client to use the server's ip addy as the primary dns server. Next, I created the home directory.
    Once both steps were done, I was able to log in from my client to my server based account...
    FYI-I found a document on afp548.com called "leopard server: advanced setup, rsync backup and automated reporting" that walks you right thru the process...Here is the link, it's a very useful doc....
    http://www.afp548.com/filemgmt_data/files/Leopard%20Server%20Quickstart%20Guide. pdf
    thanks again boomboom_uk and woVi, your suggestions were spot on....

  • [OSB and OWSM] - External Web service stacks and frameworks

    Hi everyone ! I'm starting to read about OSB and OWSM and I'm having some doubts. I've some developments of Web services with external Stacks like CXF, JBossWS, Metro and I'd like to ask some questions:
    1- Will I be able to productively leverage all features of OSB and OWSM like creating a proxy service to add WS-* standards policies and features (WS-Security, for instance) even with these web services implemented in different stacks other than Weblogic's ?
    2- If it is possible, do these web services need to be deployed at the Weblogic server to enable the OWSM and OSB to work effectively ?
    3- Even if it is possible to use the OSB and OWSM nicely with webservices developed at external stacks and deployed at other application servers is there any reason to quit using these external web service stacks in favor of Weblogic's (like features only enabled on OSB/OWSM when the services use the Weblogic stack) ?
    I suppose that if I ignore the JAX-WS stack from Weblogic and use an external framework (like CXF) I'll lose most of the application server administration capabilities since the Weblogic server won't be able to recognize the CXF stack as it does with its own.But, the main doubt I'm in is, since OSB and OWSM might be used with external providers I probably won't need to develop my web services using the Weblogic implementation (which my team does not yet know) since there will be no features of the OSB and OWSM which can only be used with the Weblogic's stack. I would like, please, to know your oppinions about these considerations. Sorry about the long post and possible errors (I just started learning).
    Thank you!

    Hi Lupan,
    I can speak mostly to OWSM as my experience with OSB is small thus far.
    +1- Will I be able to productively leverage all features of OSB and OWSM like creating a proxy service to add WS-* standards policies and features (WS-Security, for instance) even with these web services implemented in different stacks other than Weblogic's ?+
    OWSM (10gR3) has two types of policy enforcement point (PEP) -- Gateway and Agent. The Gateway acts as a remote proxy and is neutral to the service implementation technology as long as it adheres to SOAP 1.1. In this regard you can use OWSM freely with CXF, MS implementations, etc. Agents run in-process with the service and thus have far greater restrictions on what service implementation frameworks and containers that are supported. There is some certification for AXIS 1.x running in OAS and Tomcat; but practically speaking, my recommendation for Agents is to only use within OAS 10gR3 where it is built in (and using OC4J Web Services through JAX-RPC).
    OWSM 11gR1 initially supports only agent-style (in process) PEP and is built-in to Fusion Middleware and WLS. It is only for Fusion and WLS Web Service implementations.
    OWSM is quite full-featured for WS-*.
    +2- If it is possible, do these web services need to be deployed at the Weblogic server to enable the OWSM and OSB to work effectively ?+
    "No" if using OWSM Gateway PEP. A qualified "Yes" if using OWSM Agent PEP.
    +3- Even if it is possible to use the OSB and OWSM nicely with webservices developed at external stacks and deployed at other application servers is there any reason to quit using these external web service stacks in favor of Weblogic's (like features only enabled on OSB/OWSM when the services use the Weblogic stack) ?+
    There is the manageability that you mention, but also there is the identity propagation scenario and tight security integration. Both OAS and WLS hosted services in the native stacks (JAX-RPC and JAX-WS) allow sophisticated and secure passing of identity in the request -- for instance, via SAML Assertions in the WS-Sec header -- and built-in capabilities to map the passed identity into the running service's Subject (i.e. enabling JAAS security etc.).
    In my experience this type of identity propagation functionality has either been absent or less complete in other typical implementation frameworks not closely aligned with the container security mechanisms.
    Hope this helps,
    Todd

  • Running OSB and SOA Suite as Windows Services?  Node Manager

    Hi,
    Based on our current hardware configuration, we need to deploy our SOA Managed Server and Admin Server on the same machine as our OSB. Currently for development we just start the servers using the start scripts and keep a window open for them (3 of them: 1 - admin, 1 - soa_server1, and 1 - OSB.) We want to be able to run these as services and allow the admins to log out of the box. It seems like the Node Manager would allow us to do that but can we configure our Windows server to have a Node Manager running for the OSB and another for the SOA stuff on the same box? Do we even need to do that? Can we have a single Node Manager manage all three on port 5556?
    I hope this makes sense. Thanks in advance for any quick responses.
    Jeremy

    Definitely you can (should) have a single node manager. You may refer below links -
    http://blogs.oracle.com/jamesbayer/2010/01/weblogic_nodemanager_quick_sta.html
    WLS 10.3.2 Adminserver startup failed when trying to run as Windows Service
    Regards,
    Anuj

  • RMAN, RAC, NFS, and server lock ups

    Good day. My environment is:
    --a 2-node RAC
    --Enterprise Edition 11.2.0.3
    --RHEL 5.1
    The goal is to use RMAN to push backups to a shared NFS mount (on a different server). Both nodes will have access to this location (in the event one node goes down, the other can still run backups). Easy, right?
    Wrong.
    I've tried every NFS mount option in the book. Most work just fine, some don't. When I use the recommended NFS mount options:
    rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp, vers=3,timeo=600, actimeo=0
    or
    rw,bg,hard,nointr,rsize=32768,wsize=32768,proto=tcp,noac,forcedirectio, vers=3,suid
    The mount works normally. I can "ls" and "mkdir" and "touch" and "vi" and "cp" files back and forth from the NFS backup location to the RAC node all day long. No problems. However, when I try to do almost anything in RMAN which requires writing to the NFS backup location such as the command "backup archive all delete input;" (or even things as simple as a Crosscheck or RMAN configuration change which writes any changes back to the autobackup ControlFile) the node locks up. There are no errors (or if there are, I don't know where to find them), even when I use RMAN log.
    Just to recap: I run a Crosscheck (or any RMAN process that writes to the NFS backup location), the node will lock up, and I can let it sit for a day, inaccessible, with CRSCTL on the other node saying it's offline, and the node will never come out of a "frozen" state. It cannot be pinged or connected to.
    I think I can safely rule out NFS mount options at this point.
    I understand (after extensive reading of MOS docs and testing) that RAC RMAN can and does suffer from inefficient I/O when writing to an NFS mount. I don't think that's the culprit either. The autobackup ControlFile is not that big and I cannot see how running a simple Crosscheck would lock an entire node.
    I am hoping someone has encountered this in the past and hopefully it's just a simple misconfiguration somewhere.

    My NFS line in /etc/fstab is (these options are for supporting 11.2.0.3, 11.1.0.7, and 10.2.0.4/5 simultaneously): server.domain:/NFS_Export /backup nfs rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp,actimeo=0,vers=3,timeo=600 0 0
    Before you installed GI, did you by chance do a yum update? I've encountered a similar issue which ended up being due to mkinitrd creating a corrupted kernel; mkinitrd is invoked during the GI installation when the ADVM drivers are added and in my case mkinitrd created a new kernel prior to the new kernel being installed. Second to that, make sure you have the matching kernel headers to your kernel version. If they are different then you could probably get away with just creating a new kernel with mkinitrd and relinking GI/RDBMS homes, but be prepared to wipe GI and reinstall.

  • Connection Problem while client is behind proxy and server out side proxy

    hello
    i implemented ChatApplication in JAVA, for that i used socket connection when client and server both are in same network then it's working fine.
    but when my server is on internate and client is behind proxy and try to connect with server
    it not able to connect i get exception.
    i serch most of forum i got same answer and i try it but i was not success.
    any kind of help is appriciated.
    i attached my code(which i implement for testing ) pls reply me
    thanks in advance.
    Server.java
    import java.lang.*;
    import java.io.*;
    import java.net.*;
    class Server {
       public static void main(String args[]) {
          String data = "you are successfully connected with server.";
          try {
             ServerSocket srvs = new ServerSocket(1234);
             Socket skt = srvs.accept();
             System.out.print("Server has connected!\n");
             PrintWriter out = new PrintWriter(skt.getOutputStream(), true);
             System.out.print("Sending string: '" + data + "\n");
             out.print(data);
             out.close();
             skt.close();
             srvs.close();
          catch(Exception e) {
             System.out.print("Whoops! It didn't work!\n");
    ProxyClient.java
    import java.io.*;
    import java.net.*;
    import java.util.*;
    class ProxyClient{
       public static void main(String args[]) {
         String host="61.17.212.29";
         int port=1234;
               String line;
         Properties properties = System.getProperties();
         /*properties.put("firewallSet", "true");
         properties.put("firewallHost", "192.168.0.1");
         properties.put("firewallPort", "808");*/
         properties.put("socksProxySet","true");
         properties.put("socksProxyHost", "192.168.0.1");
         properties.put("socksProxyPort", "1080");
         System.setProperties (properties);
         try {
         /*SocketAddress addr = new InetSocketAddress("192.168.0.1", 1080);
         Proxy proxy = new Proxy(Proxy.Type.SOCKS, addr);
         Socket skt = new Socket(proxy);
         InetSocketAddress dest = new InetSocketAddress("61.17.212.29",1234);
         skt.connect(dest);*/
             System.out.println("before socket taken");
             Socket skt = new Socket(host,port);
             System.out.println("after socket taken");
             BufferedReader networkBin = new BufferedReader(new InputStreamReader(skt.getInputStream()));
             System.out.println("Received string: '");
             line = networkBin.readLine();     // Read one line
          while(true)
                 System.out.println(line);     // Output the line
                 line = networkBin.readLine(); // Read the next line, if any
          catch(Exception e) {
             System.out.print("Whoops! It didn't work!\n");
         e.printStackTrace();
    }

    Now look here. I could not care less about this
    code. I don't know anything about it, I don't
    want to know, I have already recommended you don't
    use it, and I have also given you a simpler and
    better solution. If you don't want to take my advice
    that is your privilege and your problem.ya i has understand system propertis i have setted and u can see it in the code i have tried by both system properties and also J2SE 5.0 proxy class but i got a same problem malformed Exception server refuse to connection.
    is there any problem at sever side?
    can u tell me in which way u r teling to set the propery i m looking forward for ur reply.
    ya i m sure u will give me.................reply "ejp".
    Thnx in advance.

  • Database and Server Report

    Hi,
    I want to assist my friend (he is same like me, newbie in oracle) to give a report on health of oracle database and server; which is running in his small software development and data feeding company. There are around
    1. 50 users,
    2. windows XP professional,
    3. Oracle9i Enterprise Edition Release 9.2.0.1.0 – Production,
    4. one database, No Archive Mode, oracle home is : c:\oracle\ora92 and database is stored at D:\ORACLE\oradata\orcl, 3 control files, 3 redo groups and 11 dbfs
    5. 2.27 GB is free space on the drive of oracle home,
    6. more than 1 GB RAM is in the system,
    7. <what else more; I should describe, please let me know>
    He has to submit a report on current status, health, performance etc. of oracle database and server. How should we design and get the data, please guide me.
    Thanks & Regards

    You Mean Helath Check Report of the DB
    Tablespace size?
    Datafile Size?
    Default tablespace?
    UNDO management?
    SGA?
    buffer cache hit ratio?
    library cache hit ratio?
    There are so many other parameters..Performance cannot be predicted from remote site
    But you can provide certain information
    Here are some views which could be helpful
    DBA_TABLESPACE
    DBA_FREE_SPACE
    DBA_DATA_FILES
    V$CONTROLFILE
    V$LOG--Meaning less since you run in archived mode
    Number of Oracle Process allocated
    UNDO RETENTION
    SQL>show parameter process
    SQL>show parameter UNDO
    Do you have METALINK acces
    1019546.6
    this has everything then you can pass whatever they need
    and many more

  • How to make site root-relative links work in DW and Server both?

    See details on buggy DW image link behavior, below. My question is:
    1) how to make site root-relative links work in DW and Server both? Or…
    2) how to reliably automate the change of several hundred legacy root-relative links of the form
    /images/image.jpg  to document-relative?
    That is, to
    ../images/image.jpg or
    ../../images/image.jpg or
    ../../../images/image.jpg etc…depending on where the directory is.
    The old format (/images/image.jpg ) used to work fine in my previous DW 8 configuration but appear grey in DW after “upgrading” to DW cs5.5 mac. (they look fine on the server, but it’s hard to edit image-heavy pages locally when they are all grey).
    I tried changing the files to how DW creates root relative links now:
    /public_html/images/image.jpg, which is a very easy, attractive root flow since there’s a one-to-one mapping. These look great in DW but are broken on the server!
    I looked at the “advanced” site setup, and it looked like it might be possible to nuke the /public_html/ part of my server info…but it also looked like there was the potential for doing damage changing these settings, which are automatically generated from our server connection settings, which seem to work.
    The “links relative to document/ site root” toggle…does that change how DW interprets existing links, or just change the default when you are adding a link?  I have made 80% of the file links document relative…before wondering if root-relative isn’t better?
    It sure seems less ambiguous for all those images if theres a way to make root relative work for DW design view, DW link check, and server.
    Summary of buggy behavior: (see test with images here)
    "old style" site root link
          /images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: broken (grey w/ broken icon)
          Link check in DW: "external link" (i.e., uncheckable, + file could appear orphaned)
          Browser: good
          Ease of switching: n/a (existing format)
    "new style" site root relative link
          /public_html/images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: good
          Link check in DW: good
          Browser: broken
          Ease of switching: easy
    Document relative link
          ../../images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: good
          Link check in DW: good
          Browser: good
          Ease of switching: hard (how to automate?)
    Absolute link
          http://www.oasisdesign.net/images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: broken (grey w/ broken icon)
          Link check in DW: external (i.e., uncheckable, + file could appear orphaned)
          Browser: good
          Ease of switching: n/a...not a real option
    Thanks!
    Similar discussion on "/"

    Hello again Jon!
    Thanks for jumping on this.
    All clear and understood about where publc_html is etc.
    No contemplation of nuking the actual public_html directory on the server, just the "/public_html" text string at the start of the DW-generated links.
    "/public_html" is automatically added to the front of the link in DW if I create the link with any of the GUI tools, if I have "site root relative" selected. And ""/public_html" ends up in the code, and gets uploaded that way to the server, where it (obviously) doesn't work.
    Doesn't sound like it is supposed to work this way. Also, what seems to be the usual root relative format (/images/image.jpg) shows as a broken link in the GUI and an external link in the DW link check. All this togther makes me thinkI have some obscure setting incorrect?
    The setting that caught my eye is manage sites/ site setup/ advance settings/ local info/ web url,  which is automatically set to http://www.oasisdesign.net/public_html/
    it gives an option to change it but it makes every effort to make this NOT look like something users should mess with:
    Having gone through the more careful thought process during this post, I'm ready to do the experiment of changing the remote server web URL (why is it wrong by default?)...think I'll eat dinner first so there's 45 min to avert disaster if anyone knows this to be a bad idea!
    Art
    PS--don't  have a local testing server...don't think this will solve the GUI broken link/ link shows as external problems.
    Is there an easy, automated way to change links sitewide from document to root relative?

  • Windows Deployment Services PXE Installation of Windows 10 Preview and Server vNext Technical Preview Failed

    The automated answer file associated with Windows 10 Enterprise Technical Preview x64 (9481) and the Windows Server Technical Preview x64 (9841) seem to be incompatible. No problem adding both wim files to WDS, this completes without issue.
    When trying to install both the Server vNext Technical Preview and Windows 10 Enterprise Technical Preview Via Network Installation (PXE) both encounter errors. The Server PXE Installation wont even commence after choosing the vNext Operating system installation,
    and choosing the Windows 10 operating system encounters a "answer configuration error" towards the end (maybe the generalization pass?). 
    Anyone have a work around for configuring the answer file? or maybe a more positive experience?

    We're playing around with Windows 10 TP and Server vNext also. For Windows 10 TP i got PXE based deployment running with unattend.xml from Windows 8.1. But i still did not get Server vNext deployment working. 
    I would also be happy if there's somebody out who has an idea on that topic. 

  • Mavericks (10.9.5) time machine and server 3.2.1 user permissions conflict

    We are using a Mac mini Server (mid 2010) 2.66 GHz Intel Core 2 Duo, 4GB 1067MHz DDR3, NVIDIA GeForce 320M 256MB machine. There are 2 500GB hard disks which have each been partitioned into 2 separate logical disks of 250GB each (in total 4 logical 250GB disks).
    We having been using this server to file share on the local network for years using Snow Leopard, Lion and Mountain Lion (although there were teething problems with Mountain Lion using the new Server app).
    After experiencing some issues with an iCloud password change and App Store failing to load, we decided to try updating to Mavericks. Upon update we were prompted to update the Server app to Mavericks version (Server app 3.2.1), this seemed to work to begin with, preserving file sharing and network user records and permissions, however we soon started to experience issues with permissions and Time Machine. We noticed that the Open Directory set up to store our network users was no longer available in the Server app and so decided to erase one of the alternate partitions and perform a clean install of Mavericks and Server app 3.2.1. This proved to produce the same issues with Time Machine permissions and also difficulties with the shared folder permissions.
    Any help with this would be greatly appreciated.
    Thanks

    Hi Linc,
    Thanks for the reply.
    The Time Machine folders shown are from the Server app Storage tab which shows all connected hard drives. I haven't set TM drives up to be shared and hopefully the following screenshots will illustrate the current setup. This setup was created from a clean install of all involved software i.e. drives were erased and setup from scratch and the TM drives were simply connected and TM took control from there.
    As you can see from these screenshots, when viewing the permissions through the Server app, there appears to be an extra user when compared to the OS Get Info window.
    After leaving TM to its own devices over the weekend, I was greeted with a 'preparing backup' message this morning(Mon 13th October), with a last backup completed on Sat 11th October at 15:19:16. Stopping this backup and starting a manual backup has now resulted in a new, never before experienced error message stating that the backup disks can not be found:
    Console reports:
    I must re-iterate, the disk was erased and simply connected and Time Machine did its thing. Now as 'admin' it is clear that the folders created by TM do not have permissions so that I can simply view them in Finder, but does this directly affect TMs ability to access folders it has created itself?
    One useful bit of information I seem to have neglected is that the erased TM drive is encrypted, however the password was entered when the drive was connected to the machine and there has been no prompt thereafter.
    Please help. Only other comment I can make is that before the update to Mavericks and Server app 3.2.1, none of this was a problem.

  • 1 Admin console to manage OSB and SOA installed on different servers

    Dear All,
    We are planning to have OSB installed on Host 1 and SOA server installed on Host 2. But there will be one Admin console to manage these 2 for the ease of maintenance. When i was trying to install this configuration I had few issues. Hope SOA experts can help me with the configuration.
    1)
    As SOA and OSB are installed on 2 different host both the servers must have weblogic installed as pre requisite. So I will have
    on host 1:
    http://host1:7001/console
    http://host1:7001/sbconsole
    On host 2:
    http://host2:7001/console (this is redundant)
    http://host2:8001/soa-infra
    2) Is it possible to inactivate http://host2:7001/console ? and add SOA server running on host2 to http://host1:7001/console. In the sense, I would like to control the SOA server from host 1.
    Issue No :2
    When I create a domain on Host 1 I just get the options for OSB and not SOA. It is obvious as SOA is not installed on Host1. but if that is the case, Is it possible to add SOA server running on host2 to host1. Is it going to be a cluster?
    Is this configuration seems realistic? Has anyone implemented. Any pointers please.
    Reagrds,
    Aani

    Hi Aani,
    Please go through the blog series -
    http://blogs.oracle.com/reynolds/2010/09/installing_an_11g_soa_cluster.html
    You may also refer EDG -
    http://download.oracle.com/docs/cd/E17904_01/core.1111/e12036/toc.htm
    Regards,
    Anuj
    Edited by: Anuj Dwivedi on Feb 23, 2011 9:14 PM

  • Synchronizing local database and server database

    Hi,
    I am working on an application where users store information on their laptops in a local database. Once a day, synchronizing has to take place with the server database. The clients will send their mutated data to the server and the server will send its mutated data, e.g. changed item file.
    The choice of database is Mysql for the laptop, and server resides on AS400, so DB2 is the platform.
    I am wondering what techniques could be used to accomplish this. Should it be done by sql or via text files? Through a socket or something else?
    Does anyone have an idea what the best approach is for this matter?
    And if possible, some sample code would be very much appreciated.
    Thanks in advance.

    Hi,
    There are numerous solutions to this. I have worked on a similar project a couple of years ago. The plan is as follows:
    1. A java client is run on the laptop (user) which connects to the remote server on calls a procedure passing the localdata as the parameter.
    2. the server parses the local data and updates itself and returns the updated data to the client.
    The main considerations:
    1. Connectivity: Connecting via internet to the server tunnelling through firewall or VPN ( I have no adequate knowledge on this).
    2. The amount of data transfer. I was working on relatively less transactional data and hence the connection thru a dial up was not a problem.
    3. The frequency of this operation. TO me it was once a day process, and hence I scheduled it during vee hours when the network was relatively free.
    This is all my thoughts about the scenario. Any comments and suggestions are welcome.
    Cheers,
    Sekar

  • How to create client and server app using node.js in firefox os.

    I want to create a client and server app using node.js for firefox os. Please suggest me how to create a apps.
    I want to send a request from client to server and based on client request, service reply to client.
    If any links available, Please share it.
    Thanks

    Hi sb00349044,
    The SUMO forums focuses on end-user support.
    For questions about developing for Firefox OS, you will find better guidance in MDN, StackOverflow, and mailing lists:
    * [https://developer.mozilla.org/en-US/Firefox_OS MDN - Firefox OS]
    * [http://stackoverflow.com/questions/tagged/firefox-os StackOverflow - Firefox OS]
    * [https://lists.mozilla.org/listinfo Mailing Lists at Mozilla]
    Thanks,
    - Ralph

  • Generating a PDF file from the date entered on line and serving it to user

    All,
    I need to create a page where the data will be entered by the user outside the company.
    The requirement is to create a pdf file which dynamic images and a bar code and server ir once the user clicks a button.
    In short, they do not want the user to go to the report manager or process monitor to see the image.
    Any thought on how this can be accomplished ?

    Is there a special reason to create PDF file? I have generated dynamic images and HTML content including Barcode based on user data entry on the fly.If you need more information, please contact me at [email protected]

Maybe you are looking for