Zenworks Database - High Availability

Hello,
we just use Zenworks 10 with an Oracle Database. We have two primary Zenworks Servers at two different Locations (Other Town - Linked via VPN).
Is it possibile to configure the Database High Availability - so that the external primary Zenworks Server is still available/manageable after a VPN Connection breakdown.
Best regerds,
Alex Sommer

This would need to be done with Oracle.
ZCM talks to a single Database.
If you can configure Oracle so that ZCM is unaware that there are
multiple back-end databases and Oracle can somehow figure out how to
resolve changes when the different DBs are not talking, I presume this
would work. This would all need to be handled by Oracle, however.
Normally, All Primaries would be in the Data Center with the Database.
Satellite Servers would be in the remote offices.
If the VPN connection was down, users would authenticate with Cached
Credentials and have their cached bundles/policies.
On 5/4/2011 11:06 AM, alexsommer wrote:
>
> Hello,
>
> we just use Zenworks 10 with an Oracle Database. We have two primary
> Zenworks Servers at two different Locations (Other Town - Linked via
> VPN).
>
> Is it possibile to configure the Database High Availability - so that
> the external primary Zenworks Server is still available/manageable after
> a VPN Connection breakdown.
>
>
> Best regerds,
>
> Alex Sommer
>
>
Craig Wilson - MCNE, MCSE, CCNA
Novell Knowledge Partner
Novell does not officially monitor these forums.
Suggestions/Opinions/Statements made by me are solely my own.
These thoughts may not be shared by either Novell or any rational human.

Similar Messages

  • Database high availability (Best solution in this sceneio)

    Dear All,
    We have 2 database servers, in 2 different continents. Operating system is Linux and Oracle 11gR1 is the database. Connectivity among the servers is over the internet.
    There are many possible ways to provide high availability of data to applications that are connected to both database. Obviously people near to the server will be accessing data from their nearest server to improve performance of the applications as applications connecting database will also be on server nearest to their geographical location.
    However inserts, updates and deletes functions should be online and changes must reflect to the users accessing any database server.
    There are several ways to achieve this objectives, like oracle streams, oracle data guard and also there are several options available within these main categories.
    I want to find out the best possible way to perform this functionality. Please guide and help me in this.
    You can ask me any question you want that can help to understand the scenario. The main concern is the performance of database servers.
    Thanks, Imran

    As i read the whole thread there is always require you need absolute current data on both end, non of them can be far away or behind to each other, the solution you are proposing yourself to replicate the data itself require more bandwidth between these two node, not only require more bandwidth but require constant traffic on both end, if you go with data guard then there is one way traffic , you cannot go tro and far , as far as concern about multi master replication then ask yourself how much the data change rate will imply to identify network overhead between them, what would you do if network goes down between both node? it would be like pulling yours hair if any complexity occurs between two node like constant updating the same data at both node.
    As you said
    First: All the entries are stored in one database i.e. the primary database, and using some method these records are also transfered to the other database. However the queries and reporting should be done from local database as per user's/application's geographical location.+
    Second: Entries are records locally and immediately replicated to other database, as the employee table we have has employee_id the primary key and this has to be always unique. There are many other tables like this. reporting and query will obviously from the local database.+
    From above two point there is confusion with yours first point . You cannot ignore current data availability at both end due to application integrity like (primary key , foreign key etc) otherwise alter yours application with unique site key "PK","UK" within yours all application primary key and foreign key.I am not much aware about replication but i am sure you cannot handle real time data availability at both end, there is always some delay at both end.
    If you ask my opinion then i would go one centralized location for yours scenario irrespective the geographical disbursement cause there are applications that have already seen over the net which access the centralized location from too far from the centralized server i.e yahoo server, google server and many more.There are various dedicated submarine optical fiber cable link exist there , you need to contact yours regional vendors like wateen,PTCL etc for such fibre link.
    Khurram

  • Making connect against SID high available

    Hi Forum,
    is there any way to make applications that use host/port/SID to connect to the database high available? The background of my question is that we have a legacy application that does not support to connect via servicename to our RAC database. The configuration dialog has no options to use a servicename, you can just enter host/port/sid.
    Now is there anyway to trick this application so that the sid/host can be failed over? The host could be tricked via virtual IP that could be provided by the grid infrastructure, but what about the SID? Is there some kind of "virtual SID" out there?
    Thanks in advance
    Thomas

    Hi,
    If you are lucky, you may be able to trick the application in giving a whole (DESCRIPTION........) tnsnames entry as the host and not using the "port" and "sid" entry in that masks.
    As a normal SQLNET/JDBC client will understand
    @(DESCRIPTION = (ADDRESS = ...))
    equally as
    @host:port:sid
    this could work. If it does then you got no problems, because you can specify the services there with all failover stuff.
    If this does not work, maybe 11gR2 server pools can help you. It will not provide you a load balancing though, just HA.
    With 11gR2 server pools/policy managed, the database SID is not fixed on a server, but will failover to other nodes.
    So lets assume you have 3 servers, and your policy managed database is running on node 1 and 2.
    Furthermore assume, that the policy for the database is minimum 2 servers and maximum 2 servers.
    In this constellation instance on node 1 will be DBNAME_1 and on node 2 DBNAME_2.
    If first server fails, the DBNAME_1 instance will automatically be started on the free node (server 3), and your clients would still be able to connect to the SID.
    The tricky part now is to manage that the additional VIP and listener you created are failing over to server 3 as well and not to server 2.
    But with a correct settings in the clusterware for the newly created VIP that should be addressable (look at the clusterware administration guide Chapter 5).
    Hope that helps
    Sebastian

  • ASM Instance high availability.

    what woululd be the best practice to keep the ASM instance high available,
    Right now i have 3 dbs running on one ASM instance, i am wondering if this ASM instance has problem all the 3 dbs will be down....

    Have you looked at the Oracle® Database High Availability Best Practices Guide?
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b25159/toc.htm
    Other areas to look would be clustering, either using Oracle clustering or clustering from another vendor such as Veritas or PollyServ.

  • Many VLFs in high availability database

    Hi,
    I have a database in high availability mode with a Log file of 16GB. Running DBCC SQLPERF (LOGSPACE) reveals that only 0.03% of the file is used. So I'd like to shrink the file. I performed full and transaction backups and tried to shrink, but nothing happens.
    I executed DBCC OPENTRAN but no transaction is open on the DB. Executing SELECT name,log_reuse_wait_desc FROM sys.databases; returns "NOTHING". But If I run DBCC LOGINFO I see 320 Virtual log files, with about 200 being marked with STATUS 2 (not
    reusabale).
    Looking at the Always-on availability dashboard shows that replication is fine.
    Does somebody know why these VLFs are marked as such?
    Thanks

    with about 200 being marked with STATUS 2 (not reusabale).
    No need to worry about this, you can read below link
    http://blog.moserit.com/virtual-log-file-monitoring-with-dbcc-loginfo-in-alwayson
    Though we mark the log records as available for cleanup, the actual process of cleaning up is deferred.  Since the new available space is tracked but the VLFs themselves are not
    yet marked as inactive it is not reported as such by DBCC LOGINFO directly.  The other commands such as DBCC SQLPERF(‘LOGSPACE’) accurately report the free space since they include the VLFs marked for deferred cleanup when accounting for space. 
    There is, unfortunately, no equivalent of the DBCC LOGINFO currently that can track VLFs marked for deferred cleanup.

  • High Availability for EBS Database 11gR2

    Hello Gurus,
    We have the below environment:
    Oracle Applications - 12.1.3
    Oracle Database - 11.2.0.3
    OS - Oracle Enterprise Linux 5.7
    I would like to know how to achieve high availability for the database. We don't want to go with RAC at this stage. What are the other alternatives to achieve high availability at the database side.
    Thanks in advance!
    -Khan

    I have come across the note id
    How To Install An Oracle Database In An Active/Passive Configuration Without RAC? [ID 734361.1]
    Can you please confirm if this can be used to configure Database for HA?
    Also please confirm if I need to get the two servers configured with any cluster software at Linux level?

  • High Availability - Active Passive setup for EBS Database

    Hello Gurus,
    We have the below environment:
    Oracle Applications - 12.1.3
    Oracle Database - 11.2.0.3
    OS - Oracle Enterprise Linux 5.7
    I would like to know how to achieve high availability (*Active/Passive*) for the database. We don't want to go with RAC at this stage. Can you please guide on a specific note id which explains how to setup an active passive database environment for HA.
    Thanks.

    Hi;
    We have the below environment:
    Oracle Applications - 12.1.3
    Oracle Database - 11.2.0.3
    OS - Oracle Enterprise Linux 5.7
    I would like to know how to achieve high availability (*Active/Passive*) for the database. We don't want to go with RAC at this stage. Can you please guide on a specific note id which explains how to setup an active passive database environment for HA.You can go with Dataguard. Please review:
    Business Continuity for Oracle E-Business Release 12 Using Oracle 11g Physical Standby Database [ID 1070033.1]
    Also see:
    MAA Roadmap for the E-Business Suite [ID 403347.1]
    Regard
    Helios

  • Choosing VM as high availability for BizTalk's SQL Server databases

    Hi,
    I'm lloking to choose the architecture of our BizTalk 2013 solution.
    The application server will be build on virtual machines, but I still have interrogation about the SQL Server.
    Is a SQL Server on a VM supported for BizTalk 2013 ?
    Because in case of physical failure, the VM moved to another server can we considered the VM as a solution of High-Availability ?
    thanks for your reply.

    When the SQL VM fails over what attributes of the server will change? If everything including the server name, the mapped SAN locations (for Databases) and/or the versions of SQL will remain the same then it will behave like a temporary network outage between
    the Front-end and the SQL.
    If however, the failed over VM has a different set of mapped location for DB then you would need to establish a BizTalk Log Shipping between SQL VM 1 and SQL VM 2. In this scenario, the recovery will require time and cannot be automatic.
    A word of caution though. If you have SLA's pertaining to transactions/messages per second with the customer then you might want to evaluate having dedicated boxes of SQL. The SQL Licensing is PER CORE and in a VIRTUAL ENVIRONMENT all the
    CORES have to be licensed since the Virtualization does not permit you to BIND the VM to a specific set of CORES. The same would apply for your BizTalk Servers. This just might work out costlier as opposed to the dedicated server environment.
    Regards.

  • Possible to use Azure SQL as database for RD Connection broker high availability?

    I am looking in to setting up a highly available Remote Desktop services service on Azure and note that the high availability for the Connection Broker service needs a full SQL database (rather than the LocalDB used for one instance). Is it possible
    to use an Azure SQL database for these purposes (i.e. opposed to VMs running SQL Server)?
    Has anyone had any experience with this? This would be nice as I am not looking at thousands of users and the Azure SQL would provide the high availability for the database without creating multiple (expensive) VMs running SQL server.

    So far there's no such a feature in azure sql, you can vote for it
    http://feedback.azure.com/forums/217321-sql-database/suggestions/423275-enable-sql-service-broker-in-sql-azure

  • Grid Control 12c and high availability databases

    Hello,
    In the context of high availability databses, my configuration is as follow:
    2 AIX servers on 2 sites with shared storage -> Site1 and site2.
    During normal operations databases are on site1. In case of disaster, databases can fallover on site2.
    On my Grid Control, 2 agents are installed: One for AIX on site1 and one for AIX on site2.
    Databases have been discovered on AIX:site1
    Listener is a virtual address and can also fallover.
    What are the possibility on the grid side ?
    Do I have to discover also the databases on site2 ?
    Actually, Listener is seen down on the site who has not the databases.
    Is there a way to only configure the Listener virtual address on the Grid ?
    Regards

    Hi,
    Basically you have to relocate the targets to the new agent (the one installed on site2). I haven't done that, but I expect to work fairly easy and smooth. Refer to the following section of Cloud Control 12c manual:
    Configuring Targets for Failover in Active/Passive Environments
    Regards,
    Sve

  • How to perform up high availability upgrade with database changes?

    Hi,
    Can someone point me to pertinent documentation for performing live upgrades? Specifically I'm looking for best practice for setting up and attaining High Availability especially during application upgrades which have db changes.
    Thanks in advance,
    Rich

    Hi,
    For internal clients, yes, your requirement can be met by setting the autodiscover site affinity by using the “AutoDiscoverSiteScope” parameter in Set-ClientAccessServer command:
    https://technet.microsoft.com/zh-cn/library/bb125157(v=exchg.150).aspx
    For external clients, I am afraid “No” because the “autodiscover.domain.com” only points to one CAS server.
    Thanks,
    Simon Wu
    TechNet Community Support

  • Advice Requested - High Availability WITHOUT Failover Clustering

    We're creating an entirely new Hyper-V virtualized environment on Server 2012 R2.  My question is:  Can we accomplish high availability WITHOUT using failover clustering?
    So, I don't really have anything AGAINST failover clustering, and we will happily use it if it's the right solution for us, but to be honest, we really don't want ANYTHING to happen automatically when it comes to failover.  Here's what I mean:
    In this new environment, we have architected 2 identical, very capable Hyper-V physical hosts, each of which will run several VMs comprising the equivalent of a scaled-back version of our entire environment.  In other words, there is at least a domain
    controller, multiple web servers, and a (mirrored/HA/AlwaysOn) SQL Server 2012 VM running on each host, along with a few other miscellaneous one-off worker-bee VMs doing things like system monitoring.  The SQL Server VM on each host has about 75% of the
    physical memory resources dedicated to it (for performance reasons).  We need pretty much the full horsepower of both machines up and going at all times under normal conditions.
    So now, to high availability.  The standard approach is to use failover clustering, but I am concerned that if these hosts are clustered, we'll have the equivalent of just 50% hardware capacity going at all times, with full failover in place of course
    (we are using an iSCSI SAN for storage).
    BUT, if these hosts are NOT clustered, and one of them is suddenly switched off, experiences some kind of catastrophic failure, or simply needs to be rebooted while applying WSUS patches, the SQL Server HA will fail over (so all databases will remain up
    and going on the surviving VM), and the environment would continue functioning at somewhat reduced capacity until the failed host is restarted.  With this approach, it seems to me that we would be running at 100% for the most part, and running at 50%
    or so only in the event of a major failure, rather than running at 50% ALL the time.
    Of course, in the event of a catastrophic failure, I'm also thinking that the one-off worker-bee VMs could be replicated to the alternate host so they could be started on the surviving host if needed during a long-term outage.
    So basically, I am very interested in the thoughts of others with experience regarding taking this approach to Hyper-V architecture, as it seems as if failover clustering is almost a given when it comes to best practices and high availability.  I guess
    I'm looking for validation on my thinking.
    So what do you think?  What am I missing or forgetting?  What will we LOSE if we go with a NON-clustered high-availability environment as I've described it?
    Thanks in advance for your thoughts!

    Udo -
    Yes your responses are very helpful.
    Can we use the built-in Server 2012 iSCSI Target Server role to convert the local RAID disks into an iSCSI LUN that the VMs could access?  Or can that not run on the same physical box as the Hyper-V host?  I guess if the physical box goes down
    the LUN would go down anyway, huh?  Or can I cluster that role (iSCSI target) as well?  If not, do you have any other specific product suggestions I can research, or do I just end up wasting this 12TB of local disk storage?
    - Morgan
    That's a bad idea. First of all Microsoft iSCSI target is slow (it's non-cached @ server side). So if you really decided to use dedicated hardware for storage (maybe you do have a reason I don't know...) and if you're fine with your storage being a single
    point of failure (OK, maybe your RTOs and RPOs are fair enough) then at least use SMB share. SMB at least does cache I/O on both client and server sides and also you can use Storage Spaces as a back end of it (non-clustered) so read "write back flash cache
    for cheap". See:
    What's new in iSCSI target with Windows Server 2012 R2
    http://technet.microsoft.com/en-us/library/dn305893.aspx
    Improved optimization to allow disk-level caching
    Updated
    iSCSI Target Server now sets the disk cache bypass flag on a hosting disk I/O, through Force Unit Access (FUA), only when the issuing initiator explicitly requests it. This change can potentially improve performance.
    Previously, iSCSI Target Server would always set the disk cache bypass flag on all I/O’s. System cache bypass functionality remains unchanged in iSCSI Target Server; for instance, the file system cache on the target server is always bypassed.
    Yes you can cluster iSCSI target from Microsoft but a) it would be SLOW as there would be only active-passive I/O model (no real use from MPIO between multiple hosts) and b) that would require a shared storage for Windows Cluster. What for? Scenario was
    usable with a) there was no virtual FC so guest VM cluster could not use FC LUs and b) there was no shared VHDX so SAS could not be used for guest VM cluster as well. Now both are present so scenario is useless: just export your existing shared storage without
    any Microsoft iSCSI target and you'll be happy. For references see:
    MSFT iSCSI Target in HA mode
    http://technet.microsoft.com/en-us/library/gg232621(v=ws.10).aspx
    Cluster MSFT iSCSI Target with SAS back end
    http://techontip.wordpress.com/2011/05/03/microsoft-iscsi-target-cluster-building-walkthrough/
    Guest
    VM Cluster Storage Options
    http://technet.microsoft.com/en-us/library/dn440540.aspx
    Storage options
    The following tables lists the storage types that you can use to provide shared storage for a guest cluster.
    Storage Type
    Description
    Shared virtual hard disk
    New in Windows Server 2012 R2, you can configure multiple virtual machines to connect to and use a single virtual hard disk (.vhdx) file. Each virtual machine can access the virtual hard disk just like servers
    would connect to the same LUN in a storage area network (SAN). For more information, see Deploy a Guest Cluster Using a Shared Virtual Hard Disk.
    Virtual Fibre Channel
    Introduced in Windows Server 2012, virtual Fibre Channel enables you to connect virtual machines to LUNs on a Fibre Channel SAN. For more information, see Hyper-V
    Virtual Fibre Channel Overview.
    iSCSI
    The iSCSI initiator inside a virtual machine enables you to connect over the network to an iSCSI target. For more information, see iSCSI
    Target Block Storage Overviewand the blog post Introduction of iSCSI Target in Windows
    Server 2012.
    Storage requirements depend on the clustered roles that run on the cluster. Most clustered roles use clustered storage, where the storage is available on any cluster node that runs a clustered
    role. Examples of clustered storage include Physical Disk resources and Cluster Shared Volumes (CSV). Some roles do not require storage that is managed by the cluster. For example, you can configure Microsoft SQL Server to use availability groups that replicate
    the data between nodes. Other clustered roles may use Server Message Block (SMB) shares or Network File System (NFS) shares as data stores that any cluster node can access.
    Sure you can use third-party software to replicate 12TB of your storage between just a pair of nodes to create a fully fault-tolerant cluster. See (there's also a free offering):
    StarWind VSAN [Virtual SAN] for Hyper-V
    http://www.starwindsoftware.com/native-san-for-hyper-v-free-edition
    Product is similar to what VMware had just released for ESXi except it's selling for ~2 years so is mature :)
    There are other guys doing this say DataCore (more playing for Windows-based FC) and SteelEye (more about geo-cluster & replication). But you may want to give them a try.
    Hope this helped a bit :) 
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • 11.1.2 High Availability for HSS LCM

    Hello All,
    Has anyone out there successfully exported an LCM artifact (Native Users) through Shared Services console to the OS file system (<oracle_home>\user_projects\epmsystem1\import_export) when Shared Services is High Available through a load balancer?
    My current configuration is two load balanced Shared Services servers using the same HSS repository. Everything works perfectly everywhere except when I try to export anything through LCM from the HSS web console. I have shared out the import_export folder on server 1. And I have edited the oracle_home>\user_projects\epmsystem1\config\FoundationServices1\migration.properties file on the second server to point to the share on the first server. Here is a cut and paste of the migration.properties file from server2.
    grouping.size=100
    grouping.size_unknown_artifact_count=10000
    grouping.group_by_type=Y
    report.enabled=Y
    report.folder_path=<epm_oracle_instance>/diagnostics/logs/migration/reports
    fileSystem.friendlyNames=false
    msr.queue.size=200
    msr.queue.waittime=60
    group.count=10000
    double-encoding=true
    export.group.count = 30
    import.group.count = 10000
    filesystem.artifact.path=\\server1\import_export
    I perform an export of just the native users to the file system, the export fails stating and I find errors in the log stating the following.
    [2011-03-29T20:20:45.405-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11001] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: /server1/import_export/msr/PKG_34.xml] Executing package file - /server1/import_export/msr/PKG_34.xml
    [2011-03-29T20:20:45.530-07:00] [EPMLCM] [ERROR] [EPMLCM-12097] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [SRC_CLASS: com.hyperion.lcm.clu.CLUProcessor] [SRC_METHOD: execute:?] Cannot find the migration definition file in the specified file path - /server1/import_export/msr/PKG_34.xml.
    [2011-03-29T20:20:45.546-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11000] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: Completed with failures.] Migration Status - Completed with failures.
    I go and look for the path that it is searching for "/server1/import_export/msr/PKG_34.xml" and find that this file does exist, it was infact created by the export process so I know that it is able to find the correct location, but it then says that it can't find it after the fact. I have tried creating mapped drives and editing the migration.properties file to reference the mapped drive but it gives me the same errors but with the new path. I also tried the below article that I found in support with no result. Please advise, thank you.
    Bug 11683769: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE
    Bug Attributes
    Type B - Defect Fixed in Product Version 11.1.2.0.000PSE
    Severity 3 - Minimal Loss of Service Product Version 11.1.2.0.00
    Status 90 - Closed, Verified by Filer Platform 233 - Microsoft Windows x64 (64-bit)
    Created 25-Jan-2011 Platform Version 2003 R2
    Updated 24-Feb-2011 Base Bug 11696634
    Database Version 2005
    Affects Platforms Generic
    Product Source Oracle
    Related Products
    Line - Family -
    Area - Product 4482 - Hyperion Lifecycle Management
    Hdr: 11683769 2005 SHRDSRVC 11.1.2.0.00 PRODID-4482 PORTID-233 11696634Abstract: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE*** 01/25/11 05:15 pm ***Hyperion Foundation Services is set up for high availability. Lifecycle management migrations fail when running Hyperion Foundation Services - Managed Server as a Windows Service. We applied the following configuration changes:Set up shared disk then we modify the migration.properties.filesystem.artifact.path=\\<servername>\LCM\import_export We are able to get the migrations to work if we set filesystem.artifact.path=V:\import_export (a maped drive to the shared disk) and we run Hyperion Foundation Services in console mode.*** 01/25/11 05:24 pm *** *** 01/28/11 03:32 pm *** (CHG: Sta->11)*** 01/28/11 03:32 pm *** (CHG: Comp. Ver-> 11.1.2 -> 11.1.2.0.00)*** 01/28/11 03:34 pm *** (CHG: Base Bug-> NULL -> 11696634)*** 02/02/11 01:53 pm *** (CHG: Sta->41)*** 02/02/11 01:53 pm ****** 02/02/11 05:18 pm ****** 02/03/11 10:30 am ****** 02/04/11 03:53 pm *** *** 02/04/11 03:53 pm *** (CHG: Sta->80 Asg->NEW OWNER)*** 02/10/11 08:49 am ****** 02/15/11 11:59 am ****** 02/15/11 04:56 pm ****** 02/16/11 09:58 am *** (CHG: Sta->90)*** 02/16/11 09:58 am ****** 02/16/11 02:12 pm *** (CHG: Sta->93)*** 02/24/11 06:14 pm *** (CHG: Sta->90)Back to top

    it is not possible to implement a kind of HA between two different appliances 3315 and 3395. 
    A node in HA can have the 3 persona. 
    Suppose Node A (Admin(primary), monitoring(Primary) and PSN). 
    Node B(Admin(Secondary), Monitoring(Secondary) and PSN). 
    If the Node A is unavailable, you will have to promote manually the Admin role to Primary. 
    Although the best way is to have
    Node A Admin(Primary), Monitoring(Secondary) and PSN
    Node B Admin(Secondary), Monitoring (Primary) and PSN. 
    Rate if helpful and Marked As correct if it is correct for the experts. 
    Regards,

  • Oracle Berkeley DB Java Edition High Availability (White Paper)

    Hi all,
    I've just read Oracle Berkeley DB Java Edition High Availability White Paper
    http://www.oracle.com/technetwork/database/berkeleydb/berkeleydb-je-ha-whitepaper-132079.pdf
    In section "Time Consistency Policy" (Page 18) it is written:
    "Setting a lag period that is too small, given the load and available hardware resources, could result in
    frequent timeout exceptions and reduce a replica's availability for read operations. It could also increase
    the latency associated with read requests, as the replica makes the read transaction wait so that it can
    catch up in the replication stream."
    Can you tell me why those read operations will not be taken by the master ?
    Why will we have frequent timeout ?
    Why should read transaction wait instead of being redirect to the master ?
    Why should it reduce replica's availability for read operations ?
    Thanks

    Please post this question on the Berkeley DB Java Edition (BDB JE) forum Berkeley DB Java Edition. This is the Berkeley DB Core (BDB) forum.
    Thanks,
    Andrei

  • Error while installing NW 7.4 SR2 in High availability

    Hello Guys,
    We are installing NW 7.4 SR2 Enterprise Portal with sybase database on windows 2012 R2 server under high availability i.e Microsoft Cluster.
    JEP is the system ID of the server, SAPJEP is the virtual host and X.X.X.36 is the virtual I.P for Central Instance node.
    Physical I.P of First node is X.X.X.21
    We have started the installation as first cluster node and get stuck at phase 3.
    The error in sapinst.log is below:-
    FCO-00011 The step assignNetworknameToClusterGroup with step key
    |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_GetSidFromCluster|ind|ind|ind|ind|getSid|0|assignNetworknameToClusterGroup
    was executed with status ERROR ( Last error reported by the step: Could
    not set properties on ClusterResource 'SAP JEP IP'.).
    The error log of sapinst_dev.log is is below:-
    TRACE      2015-03-18 11:30:29.193
    NWInstance.getStartProfilePath(true)
    TRACE      2015-03-18 11:30:29.194
    NWInstance.getStartProfileName()
    TRACE      2015-03-18 11:30:29.194
    NWInstance.getStartProfileName() done: JEP_ERS10_SAPJEP
    TRACE      2015-03-18 11:30:29.194
    NWERSInstance.getDirProfile()
    TRACE      2015-03-18 11:30:29.194
    NW.getDirProfile(false)
    TRACE      2015-03-18 11:30:29.195
    NW.getDirProfile() done: \\SAPJEP/sapmnt/JEP/SYS/profile
    TRACE      2015-03-18 11:30:29.204 [synxcpath.cpp:925]
              CSyPath::getOSNodeType(bool) lib=syslib module=syslib
    Path \\SAPJEP/sapmnt/JEP/SYS/profile is on a non-existing share.
    TRACE      2015-03-18 11:30:29.204
    NWERSInstance.getDirProfile() done: \\SAPJEP/sapmnt/JEP/SYS/profile
    TRACE      2015-03-18 11:30:29.205
    NWInstance.getStartProfilePath() done: \\SAPJEP/sapmnt/JEP/SYS/profile/JEP_ERS10_SAPJEP
    TRACE      2015-03-18 11:30:29.206
    NWInstance.getProfilePath() done: \\SAPJEP/sapmnt/JEP/SYS/profile/JEP_ERS10_SAPJEP
    TRACE      2015-03-18 11:30:29.206
    NWProfiles(JEP.ERS10.SAPJEP).getPersistentDefaultProfile()
    TRACE      2015-03-18 11:30:29.207
    NW.getPersistentDefaultProfile()
    TRACE      2015-03-18 11:30:29.207
    NW.isUnicode(true)
    TRACE      2015-03-18 11:30:29.208
    NW.isUnicode() done: true
    TRACE      2015-03-18 11:30:29.208
    NW.getPersistentDefaultProfile() done
    TRACE      2015-03-18 11:30:29.208
    NWProfiles(JEP.ERS10.SAPJEP).getPersistentDefaultProfile() done
    TRACE      2015-03-18 11:30:29.208
    NWProfiles(JEP.ERS10.SAPJEP)._getPersistentProfile(INSTANCE) done
    TRACE      2015-03-18 11:30:29.209
    NWProfiles(JEP.ERS10.SAPJEP).getPersistentProfile() done
    TRACE      2015-03-18 11:30:29.209
    PersistentR3Profile.commit()
    TRACE      2015-03-18 11:30:29.209
    PersistentR3Profile.commit() done
    TRACE      2015-03-18 11:30:29.209
    NWProfiles(JEP.ERS10.SAPJEP).getPersistentDefaultProfile()
    TRACE      2015-03-18 11:30:29.210
    NW.getPersistentDefaultProfile()
    TRACE      2015-03-18 11:30:29.210
    NW.isUnicode(true)
    TRACE      2015-03-18 11:30:29.211
    NW.isUnicode() done: true
    TRACE      2015-03-18 11:30:29.211
    NW.getPersistentDefaultProfile() done
    TRACE      2015-03-18 11:30:29.211
    NWProfiles(JEP.ERS10.SAPJEP).getPersistentDefaultProfile() done
    TRACE      2015-03-18 11:30:29.211
    PersistentR3Profile.commit()
    TRACE      2015-03-18 11:30:29.211
    PersistentR3Profile.commit() done
    TRACE      2015-03-18 11:30:29.212
    NWProfiles(JEP.ERS10.SAPJEP).commit() done
    TRACE      2015-03-18 11:30:29.212
    NWInstanceInstall.addSAPCryptoParametersToProfile() done
    TRACE      2015-03-18 11:30:29.260
    The step setProfilesForSAPCrypto with key |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_Unpack2|ind|ind|ind|ind|unpack|0|NW_SAPCrypto|ind|ind|ind|ind|crypto|0 has been executed successfully.
    INFO      2015-03-18 11:30:29.262 [synxccuren.cpp:887]
              CSyCurrentProcessEnvironmentImpl::setWorkingDirectory(const CSyPath & C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1)
              lib=syslib module=syslib
    Working directory changed to C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1.
    TRACE      2015-03-18 11:30:29.265 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:29.267 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:29.270 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:29.280 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:29.281 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:29.284 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:29.293 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:29.295 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:29.315 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    INFO      2015-03-18 11:30:29.330 [csistepexecute.cpp:1024]
    Execute step askUnpack of component |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_Unpack2|ind|ind|ind|ind|unpack|0
    TRACE      2015-03-18 11:30:29.339 [csistepexecute.cpp:1079]
    Execution of preprocess block of |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_Unpack2|ind|ind|ind|ind|unpack|0|askUnpack returns TRUE
    TRACE      2015-03-18 11:30:29.364
      Call block: NW_Unpack2_ind_ind_ind_ind
        function: NW_Unpack2_ind_ind_ind_ind_DialogPhase_askUnpack
    is validator: false
    TRACE      2015-03-18 11:30:29.364
    NWInstall.getSystem(JEP)
    TRACE      2015-03-18 11:30:29.366
    NWOption(localExeDir).value()
    TRACE      2015-03-18 11:30:29.366
    NWOption(localExeDir).value() done: true
    TRACE      2015-03-18 11:30:29.367
    NWOption(collected).value()
    TRACE      2015-03-18 11:30:29.367
    NWOption(collected).value() done: true
    TRACE      2015-03-18 11:30:29.369
    NWInstall({
      sid:JEP
      sapDrive:E:
      dbtype:ind
      sapmnt:/sapmnt
      hasABAP:false
      hasJava:true
      isAddinInstallation:undefined
      unicode:true
      workplace:undefined
      loadType:SAP
      umeConfiguration:
      doMigMonConfig:
      useParallelAbapSystemCopy:
      dir_profile:\\SAPJEP\sapmnt\JEP\SYS\profile
      os4_krnlib:
      os4_seclanglib:
      useCurrentUser:false
      noShares:false
    TRACE      2015-03-18 11:30:29.370
    NW({
      sid:JEP
      sapDrive:E:
      dbtype:ind
      sapmnt:/sapmnt
      hasABAP:false
      hasJava:true
      isAddinInstallation:undefined
      unicode:true
      workplace:undefined
      loadType:SAP
      umeConfiguration:
      doMigMonConfig:
      useParallelAbapSystemCopy:
      dir_profile:\\SAPJEP\sapmnt\JEP\SYS\profile
      os4_krnlib:
      os4_seclanglib:
      useCurrentUser:false
      noShares:false
    TRACE      2015-03-18 11:30:29.371
    NW() done
    TRACE      2015-03-18 11:30:29.371
    NWInstall() done
    TRACE      2015-03-18 11:30:29.371
    NWInstall.getSystem() done
    TRACE      2015-03-18 11:30:29.372
    NWInstall.mapUnpackTable(t_askUnpack)
    TRACE      2015-03-18 11:30:29.373
    NWInstall.mapUnpackTable() done: true
    TRACE      2015-03-18 11:30:29.377
    NWDB._get(JEP, ind)
    TRACE      2015-03-18 11:30:29.377
    NWDB._fromTable(JEP, ctor)
    TRACE      2015-03-18 11:30:29.379
    NWDB(JEP, {
      sid:JEP
      hostname:undefined
      dbsid:undefined
      installDB:false
      ORACLE_HOME:
      abapSchema:
      abapSchemaUpdate:
      javaSchema:
    TRACE      2015-03-18 11:30:29.379
    NWDB() done
    TRACE      2015-03-18 11:30:29.379
    NWDB._fromTable() done
    TRACE      2015-03-18 11:30:29.380
    NWDB._get() done
    TRACE      2015-03-18 11:30:29.380
    NWDB.getDBType(): ind
    TRACE      2015-03-18 11:30:29.390 [iaxxgenimp.cpp:283]
              CGuiEngineImp::showDialogCalledByJs()
    showing dlg d_nw_ask_unpack
    TRACE      2015-03-18 11:30:29.390 [iaxxgenimp.cpp:293]
              CGuiEngineImp::showDialogCalledByJs()
    <dialog sid="d_nw_ask_unpack">
    <title>Unpack Archives</title>
    <table enabled="true" fixedrows="true" sid="askUnpack">
    <caption>Archives to Be Unpacked</caption>
    <column enabled="true" type="check" numeric="false" upper="false" name="unpack">
      <caption>Unpack</caption>
      <helpitem id="common.Unpack"/>
      <value type="string">false</value>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="path">
      <caption>Archive</caption>
      <helpitem id="common.UnpackArchive"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="codepage">
      <caption>Codepage</caption>
      <helpitem id="common.UnpackArchiveCodepage"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="destination">
      <caption>Destination</caption>
      <helpitem id="common.UnpackArchiveDestination"/>
    </column>
    <column enabled="true" type="file" numeric="false" upper="false" name="ownPath">
      <caption>Downloaded To</caption>
      <helpitem id="common.UnpackArchiveDownload"/>
      <value type="string"/>
    </column>
    <row rowId="0">
      <boolvalue>
      <true/>
      </boolvalue>
      <value>DBINDEP\SAPEXE.SAR</value>
      <value>Unicode</value>
      <value>E:\usr\sap\JEP\SYS\exe\uc\NTAMD64</value>
      <value></value>
    </row>
    </table>
    <dialog/>
    TRACE      2015-03-18 11:30:29.391 [iaxxgenimp.cpp:1031]
              CGuiEngineImp::acceptAnswerForBlockingRequest
    Waiting for an answer from GUI
    TRACE      2015-03-18 11:30:30.617 [iaxxdlghnd.cpp:96]
              CDialogHandler::doHandleDoc()
    CDialogHandler: ACTION_NEXT requested
    TRACE      2015-03-18 11:30:30.618 [iaxxcdialogdoc.cpp:190]
              CDialogDocument::submit()
    <dialog sid="d_nw_ask_unpack">
    <table enabled="true" fixedrows="true" sid="askUnpack">
    <caption>Archives to Be Unpacked</caption>
    <column enabled="true" type="check" numeric="false" upper="false" name="unpack">
      <caption>Unpack</caption>
      <helpitem id="common.Unpack"/>
      <value type="string">false</value>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="path">
      <caption>Archive</caption>
      <helpitem id="common.UnpackArchive"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="codepage">
      <caption>Codepage</caption>
      <helpitem id="common.UnpackArchiveCodepage"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="destination">
      <caption>Destination</caption>
      <helpitem id="common.UnpackArchiveDestination"/>
    </column>
    <column enabled="true" type="file" numeric="false" upper="false" name="ownPath">
      <caption>Downloaded To</caption>
      <helpitem id="common.UnpackArchiveDownload"/>
      <value type="string"/>
    </column>
    <row rowId="0">
      <boolvalue>
      <true/>
      </boolvalue>
      <value>DBINDEP\SAPEXE.SAR</value>
      <value>Unicode</value>
      <value>E:\usr\sap\JEP\SYS\exe\uc\NTAMD64</value>
      <value/>
    </row>
    </table>
    <dialog/>
    TRACE      2015-03-18 11:30:30.646
      Call block: NW_Unpack2_ind_ind_ind_ind
        function: NW_Unpack2_ind_ind_ind_ind_DialogPhase_askUnpackValidator_default
    is validator: true
    TRACE      2015-03-18 11:30:30.646
    NWInstall.getSystem(JEP)
    TRACE      2015-03-18 11:30:30.648
    NWOption(localExeDir).value()
    TRACE      2015-03-18 11:30:30.648
    NWOption(localExeDir).value() done: true
    TRACE      2015-03-18 11:30:30.649
    NWOption(collected).value()
    TRACE      2015-03-18 11:30:30.649
    NWOption(collected).value() done: true
    TRACE      2015-03-18 11:30:30.652
    NWInstall({
      sid:JEP
      sapDrive:E:
      dbtype:ind
      sapmnt:/sapmnt
      hasABAP:false
      hasJava:true
      isAddinInstallation:undefined
      unicode:true
      workplace:undefined
      loadType:SAP
      umeConfiguration:
      doMigMonConfig:
      useParallelAbapSystemCopy:
      dir_profile:\\SAPJEP\sapmnt\JEP\SYS\profile
      os4_krnlib:
      os4_seclanglib:
      useCurrentUser:false
      noShares:false
    TRACE      2015-03-18 11:30:30.653
    NW({
      sid:JEP
      sapDrive:E:
      dbtype:ind
      sapmnt:/sapmnt
      hasABAP:false
      hasJava:true
      isAddinInstallation:undefined
      unicode:true
      workplace:undefined
      loadType:SAP
      umeConfiguration:
      doMigMonConfig:
      useParallelAbapSystemCopy:
      dir_profile:\\SAPJEP\sapmnt\JEP\SYS\profile
      os4_krnlib:
      os4_seclanglib:
      useCurrentUser:false
      noShares:false
    TRACE      2015-03-18 11:30:30.654
    NW() done
    TRACE      2015-03-18 11:30:30.654
    NWInstall() done
    TRACE      2015-03-18 11:30:30.654
    NWInstall.getSystem() done
    TRACE      2015-03-18 11:30:30.655
    NWInstall.validateUnpackTable(t_askUnpack)
    TRACE      2015-03-18 11:30:30.657
    NWInstall.validateUnpackTable(): validating row {
      codepage:Unicode
      destination:E:\usr\sap\JEP\SYS\exe\uc\NTAMD64
      ownPath:
      path:DBINDEP\SAPEXE.SAR
      sid:JEP
      unpack:true
    TRACE      2015-03-18 11:30:30.659 [tablecpp.cpp:136]
              Table(t_NW_unpack).updateRow({
      cd:UKERNEL
      codepage:Unicode
      confirmUnpack:
      destination:E:\usr\sap\JEP\SYS\exe\uc\NTAMD64
      list:
      needUnpacking:true
      ownPath:
      path:DBINDEP\SAPEXE.SAR
      selectiveExtract:false
      sid:JEP
      unpack:true
      wasUnpacked:false
    }, WHERE sid='JEP' AND path='DBINDEP\SAPEXE.SAR' AND codepage='Unicode' AND destination='E:\usr\sap\JEP\SYS\exe\uc\NTAMD64')
    updating
    TRACE      2015-03-18 11:30:30.661
    NWInstall.validateUnpackTable() done
    TRACE      2015-03-18 11:30:30.722 [iaxxgenimp.cpp:301]
              CGuiEngineImp::showDialogCalledByJs()
    <dialog sid="d_nw_ask_unpack">
    <table enabled="true" fixedrows="true" sid="askUnpack">
    <caption>Archives to Be Unpacked</caption>
    <column enabled="true" type="check" numeric="false" upper="false" name="unpack">
      <caption>Unpack</caption>
      <helpitem id="common.Unpack"/>
      <value type="string">false</value>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="path">
      <caption>Archive</caption>
      <helpitem id="common.UnpackArchive"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="codepage">
      <caption>Codepage</caption>
      <helpitem id="common.UnpackArchiveCodepage"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="destination">
      <caption>Destination</caption>
      <helpitem id="common.UnpackArchiveDestination"/>
    </column>
    <column enabled="true" type="file" numeric="false" upper="false" name="ownPath">
      <caption>Downloaded To</caption>
      <helpitem id="common.UnpackArchiveDownload"/>
      <value type="string"/>
    </column>
    <row rowId="0">
      <boolvalue>
      <true/>
      </boolvalue>
      <value>DBINDEP\SAPEXE.SAR</value>
      <value>Unicode</value>
      <value>E:\usr\sap\JEP\SYS\exe\uc\NTAMD64</value>
      <value/>
    </row>
    </table>
    <dialog/>
    TRACE      2015-03-18 11:30:30.724 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:30.725 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:30.728 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:30.776
    The step askUnpack with key |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_Unpack2|ind|ind|ind|ind|unpack|0 has been executed successfully.
    INFO      2015-03-18 11:30:30.778 [synxccuren.cpp:887]
              CSyCurrentProcessEnvironmentImpl::setWorkingDirectory(const CSyPath & C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1)
              lib=syslib module=syslib
    Working directory changed to C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1.
    TRACE      2015-03-18 11:30:30.780 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:30.781 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:30.785 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:30.802 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:30.803 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:30.806 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:30.819 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:30.842 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\keydb.xml
    TRACE      2015-03-18 11:30:30.908 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:30.909 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:30.963 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\statistic.xml
    TRACE      2015-03-18 11:30:31.139 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    INFO      2015-03-18 11:30:31.142 [csistepexecute.cpp:1024]
    Execute step setDisplayNames of component |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0
    TRACE      2015-03-18 11:30:31.177
      Call block: NW_FirstClusterNode_ind_ind_ind_ind
        function: NW_FirstClusterNode_ind_ind_ind_ind_DialogPhase_setDisplayNames_Preprocess
    is validator: false
    TRACE      2015-03-18 11:30:31.219 [csistepexecute.cpp:1079]
    Execution of preprocess block of |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|setDisplayNames returns TRUE
    TRACE      2015-03-18 11:30:31.243
      Call block: NW_FirstClusterNode_ind_ind_ind_ind
        function: NW_FirstClusterNode_ind_ind_ind_ind_DialogPhase_setDisplayNames
    is validator: false
    TRACE      2015-03-18 11:30:31.246
    NWInstance.byNumberAndRealHost(00, DCEPPRDCI)
    TRACE      2015-03-18 11:30:31.246
    NWInstance.find()
    TRACE      2015-03-18 11:30:31.248
    NWOption(localExeDir).value()
    TRACE      2015-03-18 11:30:31.248
    NWOption(localExeDir).value() done: true
    TRACE      2015-03-18 11:30:31.249
    NWOption(collected).value()
    TRACE      2015-03-18 11:30:31.249
    NWOption(collected).value() done: true
    TRACE      2015-03-18 11:30:31.252
    NWInstance._fromRow({
      sid:JEP
      name:SCS00
      number:00
      host:SAPJEP
      guid:0
      realhost:dcepprdci
      type:SCS
      installationStatus:installing
      startProfileName:JEP_SCS00_SAPJEP
      instProfilePath:
      dir_profile:
      unicode:false
      collectSource:
    TRACE      2015-03-18 11:30:31.252
    NWSCSInstance()
    TRACE      2015-03-18 11:30:31.252
    NWInstance(JEP/SCS00/SAPJEP)
    TRACE      2015-03-18 11:30:31.252
    NWInstance(JEP/SCS00/SAPJEP) done
    TRACE      2015-03-18 11:30:31.253
    NWInstanceInstall()
    TRACE      2015-03-18 11:30:31.253
    NWInstanceInstall() done
    TRACE      2015-03-18 11:30:31.253
    NWInstall.getSystem(JEP)
    TRACE      2015-03-18 11:30:31.253
    NWOption(localExeDir).value()
    TRACE      2015-03-18 11:30:31.254
    NWOption(localExeDir).value() done: true
    TRACE      2015-03-18 11:30:31.254
    NWOption(collected).value()
    TRACE      2015-03-18 11:30:31.254
    NWOption(collected).value() done: true
    TRACE      2015-03-18 11:30:31.256
    NWInstall({
      sid:JEP
      sapDrive:E:
      dbtype:ind
      sapmnt:/sapmnt
      hasABAP:false
      hasJava:true
      isAddinInstallation:undefined
      unicode:true
      workplace:undefined
      loadType:SAP
      umeConfiguration:
      doMigMonConfig:
      useParallelAbapSystemCopy:
      dir_profile:\\SAPJEP\sapmnt\JEP\SYS\profile
      os4_krnlib:
      os4_seclanglib:
      useCurrentUser:false
      noShares:false
    TRACE      2015-03-18 11:30:31.257
    NW({
      sid:JEP
      sapDrive:E:
      dbtype:ind
      sapmnt:/sapmnt
      hasABAP:false
      hasJava:true
      isAddinInstallation:undefined
      unicode:true
      workplace:undefined
      loadType:SAP
      umeConfiguration:
      doMigMonConfig:
      useParallelAbapSystemCopy:
      dir_profile:\\SAPJEP\sapmnt\JEP\SYS\profile
      os4_krnlib:
      os4_seclanglib:
      useCurrentUser:false
      noShares:false
    TRACE      2015-03-18 11:30:31.258
    NW() done
    TRACE      2015-03-18 11:30:31.258
    NWInstall() done
    TRACE      2015-03-18 11:30:31.258
    NWInstall.getSystem() done
    TRACE      2015-03-18 11:30:31.258
    NWSCSInstance() done
    TRACE      2015-03-18 11:30:31.259
    NWInstance._fromRow() done
    TRACE      2015-03-18 11:30:31.259
    NWInstance._fromRow({
      sid:JEP
      name:ERS10
      number:10
      host:SAPJEP
      guid:1
      realhost:dcepprdci
      type:ERS
      installationStatus:installing
      startProfileName:JEP_ERS10_SAPJEP
      instProfilePath:
      dir_profile:
      unicode:false
      collectSource:
    TRACE      2015-03-18 11:30:31.259
    NWERSInstance()
    TRACE      2015-03-18 11:30:31.260
    NWInstance(JEP/ERS10/SAPJEP)
    TRACE      2015-03-18 11:30:31.260
    NWInstance(JEP/ERS10/SAPJEP) done
    TRACE      2015-03-18 11:30:31.260
    NWInstanceInstall()
    TRACE      2015-03-18 11:30:31.260
    NWInstanceInstall() done
    TRACE      2015-03-18 11:30:31.261
    NWERSInstance() done
    TRACE      2015-03-18 11:30:31.261
    NWInstance._fromRow() done
    TRACE      2015-03-18 11:30:31.261
    NWInstance.find() done: 1 instances found
    TRACE      2015-03-18 11:30:31.261
    t_scs.setDisplayName([nw.progress.installInstance, SCS00], WHERE ROWNUM=0)
    TRACE      2015-03-18 11:30:31.305
    The step setDisplayNames with key |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0 has been executed successfully.
    INFO      2015-03-18 11:30:31.307 [synxccuren.cpp:887]
              CSyCurrentProcessEnvironmentImpl::setWorkingDirectory(const CSyPath & C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1)
              lib=syslib module=syslib
    Working directory changed to C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1.
    TRACE      2015-03-18 11:30:31.362
    Build Client for subcomponent |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_FinishFirstClusterNode|ind|ind|ind|ind|finishFirst|0
    TRACE      2015-03-18 11:30:31.364
    TRACE      2015-03-18 11:30:31.364 [ccdclient.cpp:22]
              CCdClient::CCdClient()
    TRACE      2015-03-18 11:30:31.385 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:31.387 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\inifile.xml
    TRACE      2015-03-18 11:30:31.390 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:31.406 [cpropertycontentmanager.hpp:220]
              CPropertyContentManager::logMissingParameters()
    The following parameters can be set via SAPINST_PARAMETER_CONTAINER_URL (inifile.xml) but not via SAPINST_INPUT_PARAMETERS_URL:
    Component 'NW_ERS_Instance': ersInstanceNumber, restartSCS
    Component 'NW_GetDomainOU': isOUInstallation, ou_delimiter, windows_domain_ous
    Component 'NW_GetSidFromCluster': clusterGroup, isUnicode, localDrive, network, networkName, sharedDrive, sid, useDHCP
    Component 'NW_GetUserParameterWindows': sapDomain, sapServiceSIDPassword, sidAdmPassword
    Component 'NW_SAPCrypto': installSAPCrypto
    Component 'NW_SCS_Instance': instanceNumber, scsMSPort, scsMSPortInternal
    Component 'Preinstall': installationMode
    INFO      2015-03-18 11:30:31.802 [synxcpath.cpp:799]
              CSyPath::createFile() lib=syslib module=syslib
    Creating file C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\summary.html.
    TRACE      2015-03-18 11:30:31.828 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:31.852 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\keydb.xml
    TRACE      2015-03-18 11:30:31.916 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:31.917 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:31.971 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\statistic.xml
    TRACE      2015-03-18 11:30:32.152 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:32.154 [sixxbsummary.cpp:320]
              CSiSummaryContainer::showDialog()
    <?xml version="1.0" encoding="UTF-8"?><sapinstgui version="1.0"><dialog version="1.0" sid="diReviewDialog"> <title>Parameter Summary</title><description>Choose 'Next' to start with the values shown. Otherwise, select the parameters to be changed and choose 'Revise'. You are then taken to the screen where you can change the parameter. You might be guided through other screens that have so far been processed.</description> <review id="review_controll"><caption><![CDATA[Parameter list]]></caption><inputgroup id="ID_1"> <caption><![CDATA[Drive for Local Instances]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_2"><pulldown edit="false" enabled="true" highlight="false" sid="localDrive">
    <caption>Destination Drive for Local Instances</caption>
    <helpitem id="mscs.SelectLocalDrive"/>
    <value>C:</value>
    <value>D:</value>
    <selectvalue>D:</selectvalue>
    </pulldown></input></inputgroup><inputgroup id="ID_3"> <caption><![CDATA[SAP System Cluster Parameters]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_4"><field enabled="true" highlight="false" sid="sid">
    <caption>SAP System ID (SAPSID)</caption>
    <helpitem id="common.SAPSystemIDCIDB"/>
    <value type="upper" maxlength="3" regexp="^[A-Z][A-Z0-9]{2}" minlength="3">JEP</value>
    </field></input><input  id="ID_5"><field enabled="true" highlight="false" sid="networkName">
    <caption>Network Name (SAP Virtual Instance Host)</caption>
    <helpitem id="common.VirtualInstanceHostWindowsMSCS"/>
    <value type="string" maxlength="13" minlength="1">SAPJEP</value>
    </field></input><input  id="ID_6"><pulldown edit="false" enabled="true" highlight="false" sid="network">
    <caption>Use Public Network</caption>
    <helpitem id="mscs.UsePublicNetwork"/>
    <value>Cluster Network 1</value>
    <selectvalue>Cluster Network 1</selectvalue>
    </pulldown></input><input  id="ID_7"><pulldown edit="false" enabled="true" highlight="false" sid="sharedDrive">
    <caption>Destination Drive for Clustered Instances</caption>
    <helpitem id="mscs.SelectSharedDrive"/>
    <value>BACKUP</value>
    <value>SAP_USR_SID_CI_NODE</value>
    <value>SAPDATA1</value>
    <value>SAPDATA2</value>
    <value>SAPDATA3</value>
    <value>SAPDATA4</value>
    <value>SAPLOG1</value>
    <value>SAPLOG2</value>
    <value>SAPTEMP_SAPDIAG_SAPSYBSYSTEM</value>
    <value>SYBASE</value>
    <selectvalue>SAP_USR_SID_CI_NODE</selectvalue>
    </pulldown></input></inputgroup><inputgroup id="ID_8"> <caption><![CDATA[DNS Domain Name]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_9"><check enabled="true" highlight="false" sid="setFQDN">
    <caption>Set FQDN for SAP system</caption>
    <helpitem id="common.FullyQualifiedDomainNameSelect"/>
    <boolvalue>
      <true/>
    </boolvalue>
    </check></input><input  id="ID_10"><field enabled="true" highlight="false" sid="FQDN" depends="setFQDN">
    <caption>DNS Domain Name for SAP System</caption>
    <helpitem id="common.FullyQualifiedDomainName"/>
    <value type="string" regexp="^[^.]+.*$" minlength="1">jnport.com</value>
    </field></input></inputgroup><inputgroup id="ID_11"> <caption><![CDATA[Windows Domain]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_12"><radiobox enabled="true" highlight="false" sid="domainType">
    <caption>Domain Model</caption>
    <helpitem id="common.DomainModelDomainLocal"/>
    <radio enabled="true" sid="global">
      <caption>Domain of Current User</caption>
      <boolvalue>
      <true/>
      </boolvalue>
    </radio>
    <radio enabled="true" sid="other">
      <caption>Different Domain</caption>
    </radio>
    </radiobox></input></inputgroup><inputgroup id="ID_13"> <caption><![CDATA[Operating System Users]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_14"><password confirm="true" enabled="true" highlight="false" sid="sidAdmPassword">
    <caption>Password of SAP System Administrator</caption>
    <helpitem id="common.PasswordCreateExistsOS"/>
    <encrvalue maxlength="63">*****</encrvalue>
    </password></input><input  id="ID_15"><password confirm="true" enabled="true" highlight="false" sid="sapServiceSIDPassword">
    <caption>Password of SAP System Service User</caption>
    <helpitem id="common.PasswordCreateExistsOS"/>
    <encrvalue maxlength="63">*****</encrvalue>
    </password></input></inputgroup><inputgroup id="ID_16"> <caption><![CDATA[Windows Domain for SAP Host Agent]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_17"><radiobox enabled="true" highlight="false" sid="domainType">
    <caption>Domain Model</caption>
    <helpitem id="common.DomainModelDomainLocal"/>
    <radio enabled="true" sid="local">
      <caption>Local Domain</caption>
      <boolvalue>
      <true/>
      </boolvalue>
    </radio>
    <radio enabled="true" sid="global">
      <caption>Domain of Current User</caption>
    </radio>
    <radio enabled="true" sid="other">
      <caption>Different Domain</caption>
    </radio>
    </radiobox></input></inputgroup><inputgroup id="ID_18"> <caption><![CDATA[Operating System Users]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_19"><password confirm="true" enabled="true" highlight="false" sid="sidAdmPassword">
    <caption>Password of SAP System Administrator</caption>
    <helpitem id="common.PasswordCreateExistsOS"/>
    <encrvalue maxlength="63">*****</encrvalue>
    </password></input></inputgroup><inputgroup id="ID_20"> <caption><![CDATA[SCS Instance]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_21"><field enabled="true" highlight="false" sid="instanceNumber">
    <caption>SCS Instance Number</caption>
    <helpitem id="common.InstanceNumberCreate"/>
    <value type="string" maxlength="2" regexp="[0-9]{2}" minlength="2">00</value>
    </field></input></inputgroup><inputgroup id="ID_22"> <caption><![CDATA[Java Message Server Port]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_23"><field enabled="true" highlight="false" sid="scsMSPortInternal">
    <caption>Internal Java Message Server Port</caption>
    <helpitem id="common.InternalMessagePort"/>
    <value type="numeric" maxlength="5" min="1025" max="65535">3900</value>
    </field></input></inputgroup><inputgroup id="ID_24"> <caption><![CDATA[Enqueue Replication Server  Instance]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_25"><field enabled="true" highlight="false" sid="ersInstanceNumber">
    <caption>Number of the ERS Instance</caption>
    <helpitem id="common.InstanceNumberCreate"/>
    <value type="numeric" maxlength="2" max="99" minlength="2">10</value>
    </field></input><input  id="ID_26"><field enabled="true" highlight="false" sid="ersVirtualHostname">
    <caption>ERS Instance - Virtual Host Name</caption>
    <helpitem id="common.virtualHostname"/>
    <value type="string" maxlength="60" minlength="1">SAPJEP</value>
    </field></input></inputgroup><inputgroup id="ID_27"> <caption><![CDATA[Unpack Archives]]></caption><selector enabled="true"><bool><boolvalue><false/></boolvalue></bool></selector><input  id="ID_28"><table enabled="true" fixedrows="true" sid="askUnpack">
    <caption>Archives to Be Unpacked</caption>
    <column enabled="true" type="check" numeric="false" upper="false" name="unpack">
      <caption>Unpack</caption>
      <helpitem id="common.Unpack"/>
      <value type="string">false</value>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="path">
      <caption>Archive</caption>
      <helpitem id="common.UnpackArchive"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="codepage">
      <caption>Codepage</caption>
      <helpitem id="common.UnpackArchiveCodepage"/>
    </column>
    <column enabled="false" type="field" numeric="false" upper="false" name="destination">
      <caption>Destination</caption>
      <helpitem id="common.UnpackArchiveDestination"/>
    </column>
    <column enabled="true" type="file" numeric="false" upper="false" name="ownPath">
      <caption>Downloaded To</caption>
      <helpitem id="common.UnpackArchiveDownload"/>
      <value type="string"/>
    </column>
    <row rowId="0">
      <boolvalue>
      <true/>
      </boolvalue>
      <value>DBINDEP\SAPEXE.SAR</value>
      <value>Unicode</value>
      <value>E:\usr\sap\JEP\SYS\exe\uc\NTAMD64</value>
      <value/>
    </row>
    </table></input></inputgroup></review><button sid="btPREV" default="false" enabled="false" ><caption>&lt; &amp;Back</caption><action>ACTION_PREV</action><tooltiphelp><![CDATA[]]></tooltiphelp></button><button sid="btNEXT" default="true"><caption>&amp;Next</caption><action>ACTION_NEXT</action><tooltiphelp><![CDATA[Continue processing]]></tooltiphelp></button><button sid="btEDIT"><caption>&amp;Revise</caption><action>ACTION_EDIT</action><tooltiphelp><![CDATA[Edit input values]]></tooltiphelp></button><button sid="btDETAIL" default="true"><caption>Show &amp;Detail</caption><action>ACTION_CBD_DETAIL</action><tooltiphelp><![CDATA[Show all input values]]></tooltiphelp></button></dialog></sapinstgui>
    TRACE      2015-03-18 11:30:32.161 [iaxxgenimp.cpp:1031]
              CGuiEngineImp::acceptAnswerForBlockingRequest
    Waiting for an answer from GUI
    TRACE      2015-03-18 11:30:33.330 [sixxbsummary.cpp:83]
              CSiSummaryContainer::setDocument(const iXMLDocument & doc)
    <?xml version="1.0" encoding="utf-8"?>
    <sapinstguiresp>
    <action>ACTION_NEXT</action>
    <dialog sid="diReviewDialog">
      <review id="review_controll"/>
    </dialog>
    </sapinstguiresp>
    TRACE      2015-03-18 11:30:33.331 [iaxxdlghnd.cpp:96]
              CDialogHandler::doHandleDoc()
    CDialogHandler: ACTION_NEXT requested
    TRACE      2015-03-18 11:30:33.331 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:33.356 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\keydb.xml
    TRACE      2015-03-18 11:30:33.375 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:33.375 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:33.429 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\statistic.xml
    TRACE      2015-03-18 11:30:33.500 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:33.773
    Status of Step |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_First_Steps|ind|ind|ind|ind|firstSteps|0|NW_Update_DLLs|ind|ind|ind|ind|dll|0|ntpatch2 is not OK. So we have to restart at this point.
    TRACE      2015-03-18 11:30:33.773
    Switch to STANDARD mode
    TRACE      2015-03-18 11:30:33.773 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:33.801 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\keydb.xml
    TRACE      2015-03-18 11:30:33.866 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:33.868 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:33.926 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\statistic.xml
    TRACE      2015-03-18 11:30:34.67 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    INFO      2015-03-18 11:30:34.67 [csistepexecute.cpp:1024]
    Execute step ntpatch2 of component |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_First_Steps|ind|ind|ind|ind|firstSteps|0|NW_Update_DLLs|ind|ind|ind|ind|dll|0
    TRACE      2015-03-18 11:30:34.084
      Call block: NW_Update_DLLs_ind_ind_ind_ind
        function: NW_Update_DLLs_ind_ind_ind_ind_SubComponentContainer_ntpatch2_Preprocess
    is validator: false
    TRACE      2015-03-18 11:30:34.115 [csistepexecute.cpp:1079]
    Execution of preprocess block of |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_First_Steps|ind|ind|ind|ind|firstSteps|0|NW_Update_DLLs|ind|ind|ind|ind|dll|0|ntpatch2 returns TRUE
    TRACE      2015-03-18 11:30:34.127
      Call block: NW_Update_DLLs_ind_ind_ind_ind
        function: NW_Update_DLLs_ind_ind_ind_ind_SubComponentContainer_ntpatch2
    is validator: false
    TRACE      2015-03-18 11:30:34.127
    NW_Update_DLLs.installDLLs(false)
    TRACE      2015-03-18 11:30:34.127
    NW_Update_DLLs.installDLLs() already run
    TRACE      2015-03-18 11:30:34.153
    The step ntpatch2 with key |NW_FirstClusterNode_Java|ind|ind|ind|ind|0|0|NW_FirstClusterNode|ind|ind|ind|ind|NW_FirstClusterNode|0|NW_First_Steps|ind|ind|ind|ind|firstSteps|0|NW_Update_DLLs|ind|ind|ind|ind|dll|0 has been executed successfully.
    INFO      2015-03-18 11:30:34.154 [synxccuren.cpp:887]
              CSyCurrentProcessEnvironmentImpl::setWorkingDirectory(const CSyPath & C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1)
              lib=syslib module=syslib
    Working directory changed to C:/Program Files/sapinst_instdir/NW740SR2/SYB/INSTALL/HA/JAVA/MSCS-NODE1.
    TRACE      2015-03-18 11:30:34.187 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:34.200 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\keydb.xml
    TRACE      2015-03-18 11:30:34.255 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    TRACE      2015-03-18 11:30:34.256 [kdxxctaco.cpp:93]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile start ...
    TRACE      2015-03-18 11:30:34.283 [kdxxctaco.cpp:121]
              CKdbTableContainerImpl::syncToContainerFile
    after creating out stream  for C:\Program Files\sapinst_instdir\NW740SR2\SYB\INSTALL\HA\JAVA\MSCS-NODE1\statistic.xml
    TRACE      2015-03-18 11:30:34.425 [kdxxctaco.cpp:155]
              CKdbTableContainerImpl::syncToContainerFile
    CKdbTableContainerImpl::syncToContainerFile stop ...
    INFO      2015-03-18 11:30:34.426 [csistepexecute.cpp:1024]
    Execute s

    Hello Arjun,
    Have you executed any other installation prior this attempt?
    It seems there are unfinished installations on this server. If this is the case, run the SWPM to uninstall the unfinished installation. Then, execute the installation once again.
    Also, make sure all IPs are corrects in the DND and host files. There were cases in which the IPs were not properly configured, or maybe a typo.
    Regards,
    Henrique Girardi
    SAP Active Global Support
    #2081285 - How to enter good search terms to an SAP search?
    https://service.sap.com/sap/support/notes/2081285
    SWPM Troubleshooting documents:
    http://scn.sap.com/docs/DOC-62646

Maybe you are looking for

  • Help Required...Web Service does not start automatically...

    I have a web service created with a EJB, everything is fine now. After restarting the Server, web service doesn't get start automatically. I have done the changes for the EJB - .ear project - application-j2ee-engine.xml file with the below. <start-up

  • Downloading and installing on admin account vs standard account

    What is the difference for downloading and installing files when: logged in on an admin account logged in on a standard account (using the admin details when prompted) I ask this, because when I am in my standard account, the downloaded files and ins

  • Error in determining receiver

    Hi Experts,    I am facing an error in one of the file to Proxy scenario. It's giving the following error. SOAP:mustUnderstand="">   <SAP:Category>XIServer</SAP:Category>   <SAP:Code area="RCVR_DETERMINATION">CX_RD_PLSRV</SAP:Code>   <SAP:P1>Problem

  • Trying to Export - Quicktime to DPX - in Linear (R709) Color space NOT LOG

    Hello, I am trying to export an image sequence through compressor from a Linear (R709) color space. When I try to do the dpx image sequence output it changes the files to LOG. Is there a way to pass the material through without any color space change

  • JDev: af:table with a large number of rows

    Hi We are developing with JDeveloper 11.1.2.1. We have a VO that returns > 2.000.000 of rows and that we display in a af:table with access mode 'scrollable' (the default) and 'in Batches of' 101. The user can select one row and do CRUD operations in