Problem synchronizing windows clients with solaris 10 NTP server

Hello everyone,
I realyy need help on the following:
I tried to synchronize windows 2003 as client with NTP server running in Solaris 10
My Time Zone in the NTP server is GMT+3, and in my windows is also set to GMT+3
When I start NTP server and clients , I got 6 hours delay between the clients and the server
I tried to synchronize the Solaris server with another solaris system, the synchronization is working fine
Your help
Regards
Hakim Fourar

True (but only for later releases). But since he mentioned getting a 6 hour offset after starting them, that makes me think that the systems are communicating, and that the time zones are incorrect.
This is easily verified by using 'ntpq -p' and 'date ; date -u'. The first will show if the client thinks it is synchronized to a server (and what the UTC offset is in miliseconds). The second will show the date in the default timezone and in UTC.
If you're not using POSIX, you'd think that GMT+XX meant east of UTC, but for POSIX, it means west of UTC.
Darren

Similar Messages

  • ASA5500 using Windows 7 computer as NTP server

    I have an ASA5510 connected to a computer running Windows 7 (the NTP Server) on its "inside" interface.
    Using the ASDM, I have configured the ASA5510 to use the Windows 7 as its NTP server (my architecture forces me to use a local machine as an NTP server):
         -IP address: 192.0.99.1 (the ASA5510 has an IP address of 192.0.99.40)
         -Interface: inside
         -Key number: None
         -Enable NTP authentication: no.
    I have other Windows computers on the "inside" interface using the NTP Server, so NTP traffic is relayed without any problem. But somehow, the ASA5510 is not able to synchronize with the NTP Server.
    I see the following log entry:
         -Source IP: 192.0.99.1
         -Source Port: 123
         -Destination IP: 192.0.99.40
         -Destination port: 65535
         -Description: Teardown UDP connection 3905 for inside: 192.0.99.1/123 to identity: 192.0.99.40/65535 duration 0:02:01 bytes 96
    so it seems like the ASA5510 sends a request to the NTP Server, but I am not sure whether the reply doesn't get processed correctly, or the connection stays open too long (my UDP connection timeout is the default, 2 minutes).
    I had trouble getting SonicWALL NSA2400s to use Windows 7 devices as NTP servers. I had to get a firmware version where there was no MD5 authentication (which I think is OK in this case), and change a setting in the Windows registry (HKEY_LOCAL_MACHINE/SYSTEM/CurrentControlSet/services/W32Time/Config/AnnounceFlags: from 0xa to 0x5)
    any insight is welcome.
    Pedro

    I have turned on all ntp debugging, and run "show ntp status" and "show ntp assoc":
    CCG-SHIP-FWL(config)# show debug
    debug ntp adjust enabled at level 1
    debug ntp authentication enabled at level 1
    debug ntp events enabled at level 1
    debug ntp packets enabled at level 1
    debug ntp params enabled at level 1
    debug ntp select enabled at level 1
    debug ntp sync enabled at level 1
    debug ntp validity enabled at level 1
    CCG-SHIP-FWL(config)# show ntp status
    Clock is unsynchronized, stratum 16, no reference clock
    nominal freq is 99.9984 Hz, actual freq is 99.9984 Hz, precision is 2**6
    reference time is 00000000.00000000 (06:28:16.000 GMT Thu Feb 7 2036)
    clock offset is 0.0000 msec, root delay is 0.00 msec
    root dispersion is 0.00 msec, peer dispersion is 0.00 msec
    CCG-SHIP-FWL(config)# show ntp assoc
          address         ref clock     st  when  poll reach  delay  offset    disp
    ~192.16.99.1      .LOCL.            1   380  1024  337    73.0  255671    71.0
    * master (synced), # master (unsynced), + selected, - candidate, ~ configured
    CCG-SHIP-FWL(config)# NTP: xmit packet to 192.16.99.1:
    leap 3, mode 3, version 3, stratum 0, ppoll 1024
    rtdel 0000 (0.000), rtdsp 10400 (1015.625), refid 00000000 (0.0.0.0)
    ref 00000000.00000000 (06:28:16.000 GMT Thu Feb 7 2036)
    org d34ac42f.4609d99d (20:21:03.273 GMT Tue May 1 2012)
    rec d34a6050.3598360c (13:14:56.209 GMT Tue May 1 2012)
    xmt d34a6250.22e73ba2 (13:23:28.136 GMT Tue May 1 2012)
    NTP: rcv packet from 192.16.99.1 to OWN_FWL_LAN_PORT on inside:
    leap 0, mode 4, version 3, stratum 1, ppoll 1024
    rtdel 0000 (0.000), rtdsp a0bf4 (10046.692), refid 4c4f434c (76.79.67.76)
    ref d34abbb0.bb426e39 (19:44:48.731 GMT Tue May 1 2012)
    org d34a6250.22e73ba2 (13:23:28.136 GMT Tue May 1 2012)
    rec d34ac62f.3977adb6 (20:29:35.224 GMT Tue May 1 2012)
    xmt d34ac62f.3977adb6 (20:29:35.224 GMT Tue May 1 2012)
    inp d34a6250.233258a0 (13:23:28.137 GMT Tue May 1 2012)
    NTP: nlist 0, allow 0, found 0, low 0.000000, high 0.000000
    NTP: no select intersection
    NTP: synchronization lost

  • Probelm client auth from jsse client with open ssl server

    I tried to connect jsse client with a openssl server.. with clientAuth
    This is what i did ..
    Using openssl req comand i created a X509 certificate for server and imported the same to java keystore..
    The communication works fine without client authentication.
    To enable client auth i create client private/public key pair using keytool and exported the public key to a file client.public. and used it in open ssl server .
    This is how i invoke the client ..
    java
    -Djavax.net.debug=all
    -Djavax.net.ssl.trustStore=cacerts
    -Djavax.net.ssl.trustStorePassword=changeit
    -Djavax.net.private -Djavax.net.ssl.keyStorePassword=password EchoClient
    After which i get following error in server
    SSL3 alert write:fatal:handshake failure
    SSL_accept:error in SSLv3 read client certificate B
    SSL_accept:error in SSLv3 read client certificate B
    ERROR
    17246:error:140890C7:SSL routines:SSL3_GET_CLIENT_CERTIFICATE:peer did not return a certificate:s3_srvr.c:1666:
    shutting down SSL
    CONNECTION CLOSED
    The client debug says it is recieving a certificate request.. what could be the problem.. can anybody help...

    i also have that problem. I was trying to configure SSL in apache in Win XP machine, but this error occurs. Is there anyone, who can help on it?

  • Problem using CORBA clients with RMI/EJB servers..!!!???

    Hi,
    I have a question on using EJB / or RMI servers with CORBA clients using
    RMI-IIOP transport, which in theory should work, but in practice has few
    glitches.
    Basically, I have implemented a very simple server, StockTreader, which
    looks up for a symbol and returns a 'Stock' object. In the first example, I
    simplified the 'Stock' object to be a mere java.lang.String, so that lookup
    would simply return the 'synbol'.
    Then I have implemented the above, as an RMI-IIOP server (case 1) and a
    CORBA server (case 2) with respective clients, and the pair of
    client-servers work fine as long as they are CORBA-to-CORBA and RMI-to-RMI.
    But the problem arises when I tried using the RMI server (via IIOP) with the
    CORBA client, when the client tries to narrow the object ref obtained from
    the naming service into the CORBA idl defined type (StockTrader) it ends up
    with a class cast exception.
    This is what I did to achieve the above results:
    [1] Define an RMI interface StockTrader.java (extending java.rmi.Remote)
    with the method,
    public String lookup( String symbol) throws RMIException;
    [2] Implement the StorckTrader interface (on a PortableRemoteObject derived
    class, to make it IIOP compliant), and then the server to register the stock
    trader with COS Naming service as follows:
    String homeName =....
    StockTraderImpl trader =new StockTraderImpl();
    System.out.println("binding obj <" homeName ">...");
    java.util.Hashtable ht =new java.util.Hashtable();
    ht.put("java.naming.factory.initial", args[2]);
    ht.put("java.naming.provider.url", args[3]);
    Context ctx =new InitialContext(ht);
    ctx.rebind(homeName, trader);
    [3] Generate the RMI-IIOP skeletons for the Implementation class,
    rmic -iiop stock.StockTraderImpl
    [4] generate the IDL for the RMI interface,
    rmic -idl stock.StockTraderImpl
    [5] Generate IDL stubs for the CORBA client,
    idlj -v -fclient -emitAll StockTraderImpl.idl
    [6] Write the client to use the IDL-defined stock trader,
    String serverName =args[0];
    String symList =args[1];
    StockClient client =new StockClient();
    System.out.println("init orb...");
    ORB orb =ORB.init(args, null);
    System.out.println("resolve init name service...");
    org.omg.CORBA.Object objRef
    =orb.resolve_initial_references("NameService");
    NamingContext naming =NamingContextHelper.narrow(objRef);
    ... define a naming component etc...
    org.omg.CORBA.Object obj =naming.resolve(...);
    System.out.println("narrow objRef: " obj.getClass() ": " +obj);
    StockTrader trader =StockTraderHelper.narrow(obj);
    [7] Compile all the classes using Java 1.2.2
    [8] start tnameserv (naming service), then the server to register the RMI
    server obj
    [9] Run the CORBA client, passing it the COSNaming service ref name (with
    which the server obj is registered)
    The CORBA client successfully finds the server obj ref in the naming
    service, the operation StockTraderHelper.narrow() fails in the segment
    below, with a class cast exception:
    org.omg.CORBA.Object obj =naming.resolve(...);
    StockTrader trader =StockTraderHelper.narrow(obj);
    The <obj> returned by naming service turns out to be of the type;
    class com.sun.rmi.iiop.CDRInputStream$1
    This is of the same type when stock trader object is registered in a CORBA
    server (as opposed to an RMI server), but works correctly with no casting
    excpetions..
    Any ideas / hints very welcome.
    thanks in advance,
    -hari

    On the contrary... all that is being said is that we needed to provide clearer examples/documentation in the 5.1.0 release. There will be no difference between the product as found in the service pack and the product found in the 5.1.1. That is, the only substantive will be that 5.1.1 will also
    include the examples.
    "<=one way=>" wrote:
    With reference to your and other messages, it appears that one should not
    expect that WLS RMI-IIOP will work in a complex real-life system, at least
    not now. In other words, support for real-life CORBA clients is not an
    option in the current release of WLS.
    TIA
    "Eduardo Ceballos" <[email protected]> wrote in message
    news:[email protected]...
    We currently publish an IDL example, even though the IDL programmingmodel in Java is completely non-functional, in anticipation of the support
    needs for uses who need to use IDL to talk to the Weblogic server,
    generically. This example illustrates the simplest connectivity; it does not
    address how
    to integrate CORBA and EJB, a broad topic, fraught with peril, imo. I'llnote in passing that, to my knowledge, none of the other vendors attempt
    this topic either, a point which is telling if all the less happy to hear.
    For the record then, what is missing from our distribution wrt RMI-IIOPare a RMI-IIOP example, an EJB-IIOP example, an EJB-C++. In this you are
    correct; better examples are forth coming.
    Still, I would not call our RMI-IIOP implementation fragile. I would saythat customers have an understandably hard time accepting that the IDL
    programming model is busted; busted in the sense that there are no C++
    libraries to support the EJB model, and busted in the sense that there is
    simply no
    support in Java for an IDL interface to an EJB. Weblogic has nothing to doit being busted, although we are trying to help our customers deal with it
    in productive ways.
    For the moment, what there is is a RMI (over IIOP) programming model, aninherently Java to Java programming model, and true to that, we accept and
    dispatch IIOP request into RMI server objects. The way I look at it is this:
    it's just a protocol, like HTTP, or JRMP; it's not IDL and it has
    practically nothing to do with CORBA.
    ST wrote:
    Eduardo,
    Can you give us more details about the comment below:
    I fear that as soon as the call to narrow succeeds, the remainingapplication will fail to work correctly because it is too difficult ot
    use an idl client in java to work.It seems to me that Weblogic's RMI-IIOP is a very fragile
    implementation. We
    don't need a "HelloWorld" example, we need a concrete serious example(fully
    tested and seriously documented) that works so that we can get a betteridea
    on how to integrate CORBA and EJB.
    Thanks,
    Said
    "Eduardo Ceballos" <[email protected]> wrote in message
    news:[email protected]...
    Please post request to the news group...
    As I said, you must separate the idl related classes (class files and
    java
    files) from the rmi classes... in the rmic step, you must set a newtarget
    (as you did), emit the java files into that directory (it's not clearyou
    did this), then remove all the rmi class files from the class path... ifyou
    need to compile more classes at that point, copy the java files to theidl
    directly is you must, but you can not share the types in any way.
    I fear that as soon as the call to narrow succeeds, the remainingapplication will fail to work correctly because it is too difficult otuse
    an idl client in java to work.
    Harindra Rajapakshe wrote:
    Hi Eduardo,
    Thanks for the help. That is the way I compiled my CORBA client, by
    separating the IDL-generated stubs from the RMI ones, but still I
    get a
    CORBA.BAD_PARAM upon narrowing the client proxy to the interfacetype.
    Here's what I did;
    + Define the RMI interfaces, in this case a StockTrader interface.
    + Implement RMI interface by extendingjavax.rmi.PortableRemoteObject
    making
    it IIOP compliant
    + Implemnnt an RMI server, and compile using JDK1.2.2
    + use the RMI implementation to generate CORBA idl, using RMI-IIOPplugin
    utility rmic;
    rmic -idl -noValueMethods -always -d idl stock.StockTraderImpl
    + generate Java mappings to the IDL generated above, using RMI-IIOPplugin
    util,
    idlj -v -fclient -emitAll -tf src stocks\StockTrader.idl
    This creates source for the package stock and also
    org.omg.CORBA.*
    package, presumably IIOP type marshalling
    + compile all classes generated above using JDK1.2.2
    + Implement client (CORBA) using the classes generated above, NOTthe
    RMI
    proxies.
    + start RMI server, with stockTrader server obj
    + start tnameserv
    + start CORBA client
    Then the client errors when trying to narrow the obj ref from the
    naming
    service, into the CORBA IDL defined interface using,
    org.omg.CORBA.Object obj =naming.resolve(nn);
    StockTrader trader =StockTraderHelper.narrow(obj); // THIS
    ERRORS..!!!
    throwing a CORBA.BAD_PARAM exception.
    any ideas..?
    Thanks in advance,
    -hari
    ----- Original Message -----
    From: Eduardo Ceballos <[email protected]>
    Newsgroups: weblogic.developer.interest.rmi-iiop
    To: Hari Rajapakshe <[email protected]>
    Sent: Wednesday, July 26, 2000 4:38 AM
    Subject: Re: problem using CORBA clients with RMI/EJBservers..!!!???
    Please see the post on june 26, re Errors compiling... somewherein
    there,
    I suspect, you are referring to the rmi class file when you are
    obliged
    to
    completely segregate these from the idl class files.
    Hari Rajapakshe wrote:
    Hi,
    I have a question on using EJB / or RMI servers with CORBA
    clients
    using
    RMI-IIOP transport, which in theory should work, but in practice
    has
    few
    glitches.
    Basically, I have implemented a very simple server,
    StockTreader,
    which
    looks up for a symbol and returns a 'Stock' object. In the firstexample, I
    simplified the 'Stock' object to be a mere java.lang.String, so
    that
    lookup
    would simply return the 'synbol'.
    Then I have implemented the above, as an RMI-IIOP server (case
    1)
    and a
    CORBA server (case 2) with respective clients, and the pair of
    client-servers work fine as long as they are CORBA-to-CORBA andRMI-to-RMI.
    But the problem arises when I tried using the RMI server (via
    IIOP)
    with
    the
    CORBA client, when the client tries to narrow the object ref
    obtained
    from
    the naming service into the CORBA idl defined type (StockTrader)
    it
    ends
    up
    with a class cast exception.
    This is what I did to achieve the above results:
    [1] Define an RMI interface StockTrader.java (extending
    java.rmi.Remote)
    with the method,
    public String lookup( String symbol) throws RMIException;
    [2] Implement the StorckTrader interface (on a
    PortableRemoteObject
    derived
    class, to make it IIOP compliant), and then the server to
    register
    the
    stock
    trader with COS Naming service as follows:
    String homeName =....
    StockTraderImpl trader =new StockTraderImpl();
    System.out.println("binding obj <" homeName ">...");
    java.util.Hashtable ht =new java.util.Hashtable();
    ht.put("java.naming.factory.initial", args[2]);
    ht.put("java.naming.provider.url", args[3]);
    Context ctx =new InitialContext(ht);
    ctx.rebind(homeName, trader);
    [3] Generate the RMI-IIOP skeletons for the Implementation
    class,
    rmic -iiop stock.StockTraderImpl
    [4] generate the IDL for the RMI interface,
    rmic -idl stock.StockTraderImpl
    [5] Generate IDL stubs for the CORBA client,
    idlj -v -fclient -emitAll StockTraderImpl.idl
    [6] Write the client to use the IDL-defined stock trader,
    String serverName =args[0];
    String symList =args[1];
    StockClient client =new StockClient();
    System.out.println("init orb...");
    ORB orb =ORB.init(args, null);
    System.out.println("resolve init name service...");
    org.omg.CORBA.Object objRef
    =orb.resolve_initial_references("NameService");
    NamingContext naming=NamingContextHelper.narrow(objRef);
    ... define a naming component etc...
    org.omg.CORBA.Object obj =naming.resolve(...);
    System.out.println("narrow objRef: " obj.getClass() ":"
    +obj);
    StockTrader trader =StockTraderHelper.narrow(obj);
    [7] Compile all the classes using Java 1.2.2
    [8] start tnameserv (naming service), then the server to
    register
    the
    RMI
    server obj
    [9] Run the CORBA client, passing it the COSNaming service ref
    name
    (with
    which the server obj is registered)
    The CORBA client successfully finds the server obj ref in the
    naming
    service, the operation StockTraderHelper.narrow() fails in thesegment
    below, with a class cast exception:
    org.omg.CORBA.Object obj =naming.resolve(...);
    StockTrader trader =StockTraderHelper.narrow(obj);
    The <obj> returned by naming service turns out to be of the
    type;
    class com.sun.rmi.iiop.CDRInputStream$1
    This is of the same type when stock trader object is registeredin a
    CORBA
    server (as opposed to an RMI server), but works correctly with
    no
    casting
    excpetions..
    Any ideas / hints very welcome.
    thanks in advance,
    -hari

  • Advantages and Disadvantages of Using Oracle 7 Client with Oracle 9 Server

    Hello,
    I was wondering what are the advanatges of using Oracle 7 client with Oracle 9 Server versus using Oracle 9 Client with Oracle 9 server.

    Some years ago I used 7.3 Client with 9i DB, and I don't remember of particular problems, but I'm speaking of Sql*Plus only. Are there specific problems ?
    Of course the best thing is to install one 9i Client (did you already do it ?) : if there are same problems, then that's not compatibility issue.
    And what are advantages of using 9 Client with 9 server other than 7 is no longer supported.You're sure you won't have compatibility problems.
    If you have access to Metalink have a look at [url http://metalink.oracle.com/metalink/plsql/ml2_documents.showFrameDocument?p_database_id=NOTE&p_id=207303.1]Note 207303.1

  • Synchronize with  External NTP server.

    Dear All Good morning,
    Environment:
    SunOS CSF-2 5.10 Generic_138888-03 sun4u sparc SUNW, Sun-Fire-V245 system.
    Sun Cluster 3-2 Two node.
    Question:
    How to Synchronize Cluster timing with external NTP server/device? If external NTP device is down will have any impact in the cluster setup?

    epmuneer wrote:
    Question:
    How to Synchronize Cluster timing with external NTP server/device?The configuration for NTP on Solaris Cluster is explained here:
    [http://docs.sun.com/app/docs/doc/820-4676/cacbdgeg?l=en&a=view|http://docs.sun.com/app/docs/doc/820-4676/cacbdgeg?l=en&a=view]
    and
    [http://docs.sun.com/app/docs/doc/820-4677/cbhijhbh?l=en&a=view|http://docs.sun.com/app/docs/doc/820-4677/cbhijhbh?l=en&a=view]
    If external NTP device is down will have any impact in the cluster setup?You should configure the cluster nodes as peers as well as getting the time from the external NTP server.
    If the external NTP server fails, then time will drift, but at least the cluster nodes keep themselves in sync to have a consistent view.
    Regards
    Thorsten

  • Unable to connect the crm client with the IPC server

    Hi,
    I am unable to connect with one CRM client with the IPC server.
    my Dispatcher and all the IPC servers are working all.
    Plz helpme out.
    With thanks and Regards,
    Yusuf.

    Hi Muhammed,
    Check in the file dispatcher.properties if the RFC names are correctly maintained.
    Regards,
    Patricia
    PS Your post was in the wrong forum. I moved it to the correct one where the chance of an answer is much higher.
    Hope you understand and good luck.
    Regards,
    Patricia
    SDN Moderator

  • Is it possible to run a Windows domain with a Mac Server?

    Looking for some advice on getting rid of our Windows domain controller (DHCP server, AD, etc.) in favour of a Mac Server (which we own).  Is it possible to run a Windows domain with a Mac Server?

    Snow Leopard Server (10.6) and earlier included Samba which supported acting as an NT Domain Controller (this pre-dates Active Directory). Lion (10.7) and later as standard can no longer act as an NT Domain Controller nor can they act as an Active Directory server.
    So the short answer is no.
    Theoretically you could port Samba4 to a Mac and Samba4 can act as an Active Directory server. In reality this is not going to be easy but is theoretically possible.
    An easier option might be to run a copy of Linux (e.g. Ubuntu) in a virtual machine e.g. using VirtualBox and then run Samba4 in that and configure it to act as and Active Directory server. Then you could connect the Mac side to the Samba4 Active Directory server. This approach would use all Mac hardware, and not have any extra costs of licenses to buy and as far as all the rest of the network is concerned would be a standard Active Directory server environment.

  • Managing 10.4 clients with 10.5 Server

    Before I jump in....any feedback on managing 10.4 clients with 10.5 server (will post the vice-versa in the 10.4 server forum) from those who are doing it currently?
    Thanks,
    k

    I checked out the link you mentioned. It's really dealing w/ managing the client. I'm have weird issues w/ clients logging in. Today I had a grade level try to log in. Ten of the computers would let that grade level in, two would not. Of the Ten computers approx 5 wouldn't let them log in the first time, but once they were restarted they were able to log in. All of the computers are on a wired network as well as a wireless backup. Of the two that would not let them login I had just restored one and set the settings up the same as the others.
    I'll be damned, this is the weirdest I have had with Apple client management.

  • NFS problem between RedHat Client and Solaris Server

    Hi all, we are experiencing a problem between a RedHat client and a Solaris 10 server. For the purposes of this post, I'll call the Redhat client server A and the Solaris 10 server B.
    Server B is exporting a filesystem that server A is trying to mount. Server A can successfully mount the exported file system, however, strange things are happening. If I change to the exported mount point on server A and create a file, the file is owned by nobody:nobody, not the user that created the file.
    A look at the file on server B shows the file has the correct UID and GID (ie the UID & GID of server A).
    The fstab file on server A looks like this:
    serverB:/data /data nfs4 rsize=32768,wsize=32768,hard,nointr,rw,bg,actimeo=0,timeo=300,suid 0 0
    Does anyone have a explanation for this?
    NB: There is a firewall between server A and server B. A firewall rule is in place to allow traffic between the two servers on port 2049
    Stewart

    Hi
    If I change to the exported mount point on server A and create a file, the file is owned by nobody:nobody, not the user that created the file.On a NFS share, for security reasons, you normally dont have root provileges.
    A file createt as root user will be mapped to nobody:nobody.
    The behaviour you see is correct.
    If you want the file to be createt as root, you have to export the filesystem with -o ro,anon=0
    NFSv3 will be blocked by your firewall.
    Franco

  • Getting problems importing universes - Desginer with the BOE server

    Hey Guys,
    Im getting a problem, i have 2 clients Windows XP with Designer.
    One client is connecting normaly, I can connect in the server, import universes and work on them.
    The other one, its connecting as well, but when i try to import, i get the folow error:
    Failed to initialize import/export dialog.
    Previous Error: Failed  to retrieve root universes folder. You may not have permission to view this folder.
    Some consideration:
    1) Both computers are using the same user, Adminitrator.
    2) The second client, connect in the server.
    Have you guys got this problem?
    Thanks

    Hey Guys,
    I still trying to solve the problem.
    Do you know if i need to have the oracle client installed so the Designer could work?
    the computer that works has the oracle client installed. Thats the only difference.
    Thanks

  • ORADC problem on Windows 2000 with Service pack 1 and service pack 2

    Hi,
    I have a VB application running on NT4.0 workstations. It uses
    ORADC control to talk to ORACLE 7.3 When I load it on Windows
    2000 with service pack 1 or service pack 2, I encounter some
    problems.
    The ORADC control is used in most of the VB forms and the
    Refresh method of the ORADC conrol is used to refresh the APEX
    TRUE DBGRID PRO control. When this action is executed, I receive
    a run time error
    "Method Refresh of object '_DORADC' failed Err#: -2147417848
    It works fine when no service pack is applied on Win2K.
    Can anyone tell me what's going on?
    Thanks

    satheesh_p wrote:
    Thanks for your reply.
    When I checked the system requirements. I couldn't find information on windows 2003 service pack 2
    link http://www.adobe.com/products/coldfusion/systemreqs/.
    Whether your OS is windows 2003 with service pack 2?
    Are you sure coldfusion 8 will work on windows 2003 with service pack 2?
    Regards,
    Satheesh.
    Well it was there, if you understand the naming properly:
    ColdFusion 8 system requirements
            Windows
    Intel® Pentium® II or AMD Athlon processor
    Microsoft®
    Windows® 2000 Professional with Service Pack 3; Windows 2000 Server,
    Advanced Server, or Datacenter Server with Service Pack 3; Windows
    Server® 2003 Web, Standard, or Enterprise Edition with Service Pack 1
    and R2; Windows Server® 2008; Windows XP Professional or Home Edition;
    or Windows Vista®
    R2 is the "offical" Microsoft name for service pack 2.

  • Interoperability: WSDL-based Java client with MS SOAP server

    Does anybody have the experience with a Java client communicating with Microsoft
    SOAP server based on the WSDL protocol? My initial experiece shows a lot of problems.
    Any information in this regard is appreciated.
    Thanks

    I'm experiencing the same sort of problem too.
    I downloaded Bea Web Service Broker in order to test some MS Soap-based web services
    before developing some Weblogic EJBs which are to use them, since both of them
    (WS Broker and Weblogic EJBs) are said to use the same underlying package of Java
    classes to access any web service.
    Most of the MS web services won't work with this tool. I have problems with remote
    methods that have one or more parameters (the remote methods seem not to get them)
    and also with some web services that raise FileNotFoundExceptions from URLs (rather
    odd to me) on the Broker, despite the fact those URLs are valid.
    Any hint or suggestion will be very appreciated.

  • Windows-Client with readonly disk

    Hello,
    some of our Windows-Clients do not allow writing to local disk (e.g.
    download of TTAWinPrint.exe) or executing regedit. I already removed the
    print- and CDM-applet from our webtop as suggested in another article and
    those downloads and regedit-calls are now gone.
    But there is still one download, it is "TTAwinClientInfo.exe"
    What is this for and how can I prevent Tarantella from downloading it?
    We are using Enterprise 3.4 on SuSE 9.0 and WinClients with IE.

    Barbara wrote:
    Hello,
    some of our Windows-Clients do not allow writing to local disk [...] there
    is still one download, it is "TTAwinClientInfo.exe"
    What is this for and how can I prevent Tarantella from downloading it?
    We are using Enterprise 3.4 on SuSE 9.0 and WinClients with IE.
    Barbara,
    TTAWinClientInfo : provides information that we can't get via the java
    interface as well as an interface to the windows RDP license system.
    You can not remove it from being downloaded.
    Regards,Nevertheless Tarantella seems to work fine even without that information:
    On startup there are some error popups of the kind "download not allowed",
    but after clicking them away the application (X11-Session on
    Linux-Appserver) starts as normal.
    So my main problem is, that the users complain about those messages, they
    don't want to have to quit several popups before they can start to work ...

  • Using a Mac on windows network with a proxy server

    Hi I’m very new to Macs to apologies if this is a really silly question. I’m
    trying to setup a Mac on our network and I’m having an issue getting it to work
    properly with our proxy server. I have connected the Mac to the network and
    selected automatic proxy configuration in the network settings using the URL of
    our Pac file. I am asked for a proxy username and password when I initially try and
    access a website but once I have entered my credentials they are saved in
    keychain and anyone coming after me can browse using my account. I work in a
    hospital so there are many different people accessing different devices. We
    monitor and trace all users internet browsing so my question is can I configure
    the Mac to ask for proxy credentials for different users? On our windows devices
    a login box will appear if the browser has been closed and reopened. Can I do
    this with the Mac?????
    Any help you can offer will be very much appreciated.
    Thanks

    The short answer is that the IT department is a strict Microsoft shop and also incompetent. Almost everything in that message is patently false. Unfortunately, when accessing Microsoft Server services like this, you will need the support of the IT staff to tell you various settings to use. I think it is pretty clear that they have no intention of doing that.
    There is nothing "special" about bootcamp. It is just a boot loader. A Mac running Windows via bootcamp is a Windows PC, not a Mac. As such, it will work perfectly fine in the network. You could also try running Parallels. I believe Parallels can be configured to use a bootcamp partition so you could run Windows via either bootcamp or Parallels.
    One of the few things that is correct is that Apple does have to provide driver support for bootcamp and there is the possibility of random incompatibilities - but no more so than any other PC manufacturer. Parallels/VMWare may work around that with a different set of incompatibilities. Don't be too worried about this part. These "incompatibilities" are minor and almost always with funky hardware devices. What you want to do should work perfectly with either Parallels/VMWare or Bootcamp.

Maybe you are looking for