Keylogger for X server?

I'm looking for a keylogger that can be started at the beginning of an X session (.xinitrc/.xprofile) that doesn't require root and only logs stuff from that X session.

teekay wrote:
I'd call this morally questionable. I have kids, too, but never would I spy on them in such a shameless and intrusive fashion. Not even telling them "hey, be aware I log every key you press" would make this more acceptabe.
It's perfectly fine to get some basic level of protection in place, though. Running a transparent proxy with content filtering, OpenDNS and/or programs like MintNanny are sane ways to implement this. A keylogger is not. It's about prevention, not prosecution.
Yeah, this isn't something I'm doing on my own - I'm actually securing this computer per request of the parent. All the other computers are secured through Firefox plugins, but my computer has Chromium on it, so we need something more.
Furthermore, this conversation about the inner workings of my involvement in my family's efforts to guide my wayward sibling is getting rather awkward - I've discussed the various parental control options with the parent, and we've both settled on this.
Please, can someone just answer my question?
(As for the two suggestions, OpenDNS and MintNanny, I have my reasons for loathing OpenDNS. MintNanny looks interesting, but I can't count on every posible domain being blocked.)
About your concerns of every keystroke being pressed, obviously I'd only enable it at times when Certain Person was not supposed to be anywhere near my computer.

Similar Messages

  • Remote Server Admin not working from outside of network for 1 server

    Our company recently changed ISPs and I had to change our two 10.4 server's IP addresses. We have a mail server (intel xserve) and a file/web server (quicksilver g4). Both servers have two network cards in them. The problem is two-fold:
    1- I can successfully manage the Xserve machine locally on our network and from my home. However, I can only manage the Quicksilver g4 server locally. Any kind of external access is not even acknowledged.
    2- I'm not sure if I missed any steps when changing IP addresses for these server-based computers. Also, I'm not sure if I correctly set our dns names to the correct IP address.
    For some background, this is the exact IP update process I used for each server:
    Quicksilver G4 (file/web server) - Installed network card #2 and configured it with the new Public IP in the "System Preferences/Network" panel. In Server Admin I set our website to use the new public IP address. (network card #2 has no firewall device in between it and the internet.)
    Then, I configured the default network card #1 to a static, yet private IP address that's behind our DLink firewall device with the rest of our network.
    Intel Xserve (e-mail server) - Network card #1 was the only one setup before our network change. It had a static, public IP address. When we changed ISPs, I configured network card #2 to the new static, public IP address supplied to us by the new ISP in the "System Preferences/Network" panel. This new IP address is where all email traffic currently gets pointed to. (All mail is successfully sent and delivered.) Until our former ISP gets shutdown, I still have network card #1 configured to the older static public IP address. After the old ISP is shut off, I plan on changing network card #1 to a static, private address.
    Any assistance would be greatly appreciated.
      Mac OS X (10.4.8)  

    What should I check in AD?  I am by no means an expert with AD.
    Yes, I am using the same client OS.
    I am talking about RDP over the internet, like from home to the office.  We have a static IP assigned to the router from ISP.  A static internal IP assigned to the server on the LAN.  And the router port forwards 3389 to the assigned IP. 
    It was working fine before we reinstalled Server 2012.  These are the steps I took when reinstalling:
    1. format drive and install OS
    2. rename the server
    3. install SQL server
    4. Install TFS and SharePoint
    5. Add Active Directory role and promote to Domain Controller
    6. Add domain users
    7. Enable remote access on the server and add users to remote access list

  • Increase Performance and ROI for SQL Server Environments

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

  • Set up Search Service App For SharePoint server 2013 on Windows server 2012 R2 not working

    Hi all,
    I installed SharePoint server 2013 on Windows  server 2012 R2 using VirtualBox.  I created a DC(domain controller) server with a domain set up on one VM and it has SQL server 2012 SP1 installed. Then SharePoint 2013 on another VM was set up to access
    the DC server.  Everything seems working except Search Service App which cannot be sucessfully set up. Creation process for Search service app says Successful and 4 search databases were created and look fine. But when I navigate to search service app
    admin page, it gives error info:
    System status:  The search service is not able to connect to the machine that hosts the administration component. Verify that the administration component '386f2cd6-47ca-4b3a-aeb5-d9116772ef16' in search application 'Search Service Application 1' is in
    a good state and try again.
    Search Application Topology:  Unable to retrieve topology component health states. This may be because the admin component is not up and running.
    From event viewer, I see following errors:
    (1) Error From source: SharePoint Server
    Application Server Administration job failed for service instance  Microsoft.Office.Server.Search.Administration.SearchServiceInstance
    (b7c72eb8-cbaf-435e-b4c9-963cb6e4e745).
    Reason: The object you are trying to create already exists. Try again using a different name.  
    Technical Support Details:
    System.Runtime.InteropServices.COMException (0x80040D02): The object you are trying to create already exists. Try again using a different name.  
       at Microsoft.Office.Server.Search.Administration.SearchServiceInstance.Synchronize()
       at Microsoft.Office.Server.Administration.ApplicationServerJob.ProvisionLocalSharedServiceInstances(Boolean
    isAdministrationServiceJob)
    (2) Error From source: SharePoint Server Search
    Could not access the Search database. A generic error occurred while trying to access the database to obtain the schema version info.
    Context: Application '386f2cd6-47ca-4b3a-aeb5-d9116772ef16'
    (3) Warning from source: SharePoint Server Search
    A database error occurred. Source: .Net SqlClient Data Provider Code: 8169 occurred 0 time(s) Description:  Error ordinal: 1 Message:
    Conversion failed when converting from a character string to uniqueidentifier., Class: 16, Number: 8169, State: 2    at
    System.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection, Action`1 wrapCloseInAction)
    (4) Error From source: SharePoint Server
    Application Server Administration job failed for service instance Microsoft.Office.Server.Search.Administration.SearchServiceInstance
    (b7c72eb8-cbaf-435e-b4c9-963cb6e4e745).
    Reason: The gatherer application could not be mounted because the search administration database schema version does not match the expected backwards compatibility schema version. The database might not have been upgraded.  
    Technical Support Details:
    System.Runtime.InteropServices.COMException (0xC0041235): The gatherer application could not be mounted because the search administration database schema version does not match the expected backwards compatibility schema version. The database might not have
    been upgraded.  
    Since separate DC server and SharePoint server do not work, I installed SharePoint 2013 on DC server ( so DC server has everything on it now ) but it gives exactly same result. Later I installed SharePoint 2013 SP1 and still have the same problem with Search
    Service app. I spent two weeks tried all suggestions available from Web and Google but SharePoint Search Service simply does not work. Config and other databases work but why Search Service has this issue seemingly related to search DB.
    Could anybody please help out? You deserve a top SharePoint consultant award if you could find a solution. I am so frustrated and so tired by this issue.    This seems also to be a SP set up issue.
    Thanks a lot.

    Using new Search Service App wizard to create SSA is always a success. I could delete existing SSA and recreate it and no problem. It says successful but when I open Search Admin page from CA, it gives me errors as mentioned.
    Now I used the following PS script for creating SSA from Max Mercher, but it stays at the last setps in following script:
    Add-PsSnapin Microsoft.SharePoint.PowerShell -ErrorAction SilentlyContinue
    $IndexLocation = "C:\Search"  #Location must be empty, will be deleted during the process!
    $SearchAppPoolName = "SSAPool"
    $SearchAppPoolAccountName = "mydomain\admin"
    $SearchServiceName = "SSA"
    $SearchServiceProxyName = "SSA Proxy"
    $DatabaseServer = "W12R2DC1"
    $DatabaseName = "SSA"
    $spAppPool = Get-SPServiceApplicationPool -Identity $SearchAppPoolName -ErrorAction SilentlyContinue
    if (!$spAppPool)
     $spAppPool = New-SPServiceApplicationPool -Name $SearchAppPoolName -Account $SearchAppPoolAccountName -Verbose
    $ServiceApplication = Get-SPEnterpriseSearchServiceApplication -Identity $SearchServiceName -ErrorAction SilentlyContinue
    if (!$ServiceApplication)
    # process stays at the following step forever, already one hour now.  
    $ServiceApplication = New-SPEnterpriseSearchServiceApplication -Name $SearchServiceName -ApplicationPool $spAppPool.Name -DatabaseServer  $DatabaseServer -DatabaseName $DatabaseName
    Account mydomain\admin is an farm managed account, domain admin account, in WG_ADMIN role, It is in all SQL server roles and is DBO. I see search DBs are already on SQL server. From Event viewer, I got following errors in sequence:
    (1) Crawler:Content Plugin under source Crawler:Content Plugin 
    Content Plugin can not be initialized - list of CSS addresses is not set.
    (2) Warning for SharePoint Server Search
    A database error occurred. Source: .Net SqlClient Data Provider Code: 8169 occurred 0 time(s) Description:  Error ordinal: 1 Message: Conversion failed when converting from a character string to uniqueidentifier., Class: 16, Number: 8169, State: 2   
    at System.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection, Action`1 wrapCloseInAction)
    (3) Error for SharePoint Server Search
    Could not access the Search database. A generic error occurred while trying to access the database to obtain the schema version info.
    Context: Application 'cbc5a055-996b-44a7-9cbc-404322f9cfdf'
    (4) Error for SharePoint Server
    Application Server Administration job failed for service instance Microsoft.Office.Server.Search.Administration.SearchServiceInstance (b7c72eb8-cbaf-435e-b4c9-963cb6e4e745).
    Reason: The gatherer application could not be mounted because the search administration database schema version does not match the expected backwards compatibility schema version. The database might not have been upgraded. 
    (5) Error Shared Services for SharePoint Server Search 
    Application Server Administration job failed for service instance Microsoft.Office.Server.Search.Administration.SearchServiceInstance (b7c72eb8-cbaf-435e-b4c9-963cb6e4e745).
    Reason: The object you are trying to create already exists. Try again using a different name. 
    Technical Support Details:
    System.Runtime.InteropServices.COMException (0x80040D02): The object you are trying to create already exists. Try again using a different name. 
       at Microsoft.Office.Server.Search.Administration.SearchServiceInstance.Synchronize()
       at Microsoft.Office.Server.Administration.ApplicationServerJob.ProvisionLocalSharedServiceInstances(Boolean isAdministrationServiceJob
    Above errors keep being generated. Last step for SSA creation stay there forever.  Any clue what is really going on?  Thanks.

  • I need to Health check for Web server and need report on daily basis

    Hello,
    How to monitor Web servers  ?
    It would be great for administrators to have a daily alert using power shell script for web server
    Can anyone have any scripts that triggers email like IIS services , etc.. are up and running.
    Thanks
    Pushparaj
    Thanks, Pushparaj

    Hi Pushparaj, this forum is not for sharing or building script based on request
    You check the below code and build your script
    $Service = Get-WmiObject -Class Win32_Service -ComputerName SERVERNAME | ?{$_.Name -eq 'IISADMIN'}
    Send-MailMessage -From 'mailID' -To 'mailID' -SmtpServer 'XXXX.DOMAIN.XXX' -Body $Service.State `
    -Subject 'IIS Service'
    Search Script
    Request Script
    See this script and modify as per your needs
    Regards Chen V [MCTS SharePoint 2010]

  • Verification Failed. The certificate for this server is invalid

    I am trying to backup my iPad to iCloud.  I get:
    Verification Failed.  The certificate for this server is invalid. You might be connecting to a server that is pretending to be "setup.icloud.com" which could put your confidential information at risk.
    I am logged onto the wifi at work and I can surf the Internet so the connection is OK.
    How do I begin to debug this issue?

    Hey everyone,
    I had this issue for few months already! I have a @me.com mail address, theoretically should be compatible with iCloud, just for some reason, it's not working!!
    The mail all work fine on my macbook and iPhone, just the ipad kept showing the error message, the "verification failed".
    The internet is wi-fi and it all working fine.
    Now here is the solution.
    Make sure you got internet connected fine.
    Go Setting --> Mail, Contacts, Calendars --> Add Account --> Other
    --> Add Mail Address -->
    in the description, it will auto showing "Me", don't change it!
    Go "Next" -->
    in the Incoming Mail Server
    Host Name --> Type "imap.mail.me.com"
    Then your user name and password
    in the Outgoing Mail Server
    Host Name --> Type "smtp.mail.me.com"
    Then your user name and password
    Then next and save it.
    This should work, it works for me,
    Hope this helps!!

  • Xcode - "The certificate for this server is invalid"

    Hi,
    I'm trying to log into my Xcode Server (with it's own self-signed certificate) from my work iMac, I receive the following message:
    "The certificate for this server is invalid. You might be connecting to a server that is pretending to be “[server name]” which could put your confidential information at risk. Would you like to connect to the server anyway?"
    The only option is "OK" which just prompts the login screen again, it doesn't allow me to "connect anyway".  How do I get around this? :/

    What account are you trying to access? Skype or some other site and you are being redirected to Skype?

  • Verification failed certificate for this server is invalid

    I attempt to log into the iCloud on my iPod over a WiFi connection and it gives me the Verification Failed. The certificate for this server is invalid. You might be connecting to a server that is pretending to be "setup.icloud.com" which could put your confidential information at risk.
    Why am I getting this and how can I access the iCloud to backup my files?

    This happened to me because my router was configured to use Open DNS servers (Primary 208.67.222.222, Secondary 208.67.220.220). Temporarily adjusting the DNS settings on the router to 4.2.2.1 and 4.2.2.2 did the trick. Open DNS does weird things with certain certificates. Even if your router is configured to used something else, try the ones listed above to see if it makes a difference.

  • The certificate for this server is invalid skype

    I am attempting at access My Account on an iPhone 5, I am getting a message saying the certificate for this server is invalid.  You might be connecting to a server that is pretending to be "secure.skype.com" which could put your confidential information at risk.
    My date, time and time zone are all correct.  Any suggestions?

    What account are you trying to access? Skype or some other site and you are being redirected to Skype?

  • Mobile Me Verification Failed - it says the certificate for this server is invalid.  What does that mean?

    Trying to set up a 2nd device on Mobile Me and it's telling me "Verification Failed, Certificate for this server is invalid, I might be connecting to a server that is pretending to be setup.me.com which can put my confidential information at risk."   Please tell me what this means and how to correct it.  My 1st device set up with Mobile Me just fine. 

    Check your date/time settings if correct.
    Happened to me when I changed the battery then I found out that the settings is not yet set.
    Works flawless now!
    Good luck!

  • My 4th generation iPod Touch won't let me get on to the App Store. When I log on to iTunes, an alert pops up that says the certificate for the server is invalid, and that it may be a server pretending to be iTunes. What should I do?

    My iPod won't let me on to the App Store, and whenever I go on to ITunes, an alert pops up that the certificate for the server is invalid, and that I may be connecting to a server that is only pretending to be iTunes.apple.com and my personal info may be at risk. I downloaded an emulator yesterday from coolroms.com but deleted the app this afternoon. I cleared my safari search data, my cookies and data, and web inspector, which still didn't work. I then proceeded to reset my iPod and then download the newest version of IOS 6.1.5 but yet still am having problems. Also to the App Store and iTunes, several other apps aren't working. Any help here?

    Also, when I go on to safari, another alert pops up that safari cannot verify the identity of the website, anything that I type in to as common as google.com. It gives me 3 options to either cancel, look at details, and continue. I've looked at the details of the website of Google and it is legitimate the site. Any help?

  • How to set up different realms for each server ?

    I am using weblogic 6 sp1.I have a domain with several servers. i want to assoicate or use different custom security realm for each server.However -on weblogic console- there is no 'Target' tab in the Security node ,and no way to set up multiple Security nodes.How ca i do this (or can i)?

    Main reason why is this:
    Imagine situation I have 2 servers A and B both of them bellongs to domain SomeDomain.
    Server A is as WebServer and server B hosts all EJBs. And server A is outside
    firewall (does not have DB access). But i would like to use RDBMS realm so what
    to do in this situation. Place both servers in different domains and in server
    A install proxy realm for RDBS realm ?
    "Tom Moreau" <[email protected]> wrote:
    >
    You can't have a different realm for each
    server. There is only one realm for all
    servers - think of it as there is only
    one set of authentication & authorization
    information (that is users/groups/permissions)
    and it applies to all servers.
    Why do you want each server to have its
    own realm? If someone tries to log in,
    do you want WLS to automatically route
    them to a server who is capable of logging
    them in? I'm having trouble understanding
    why you want this feature.
    Thanks,
    -Tom Moreau
    Rachel <[email protected]> wrote:
    I am using weblogic 6 sp1.I have a domain with several servers. i want
    to assoicate or use different custom security realm for each server.However
    -on weblogic console- there is no 'Target' tab in the Security node
    ,and no way to set up multiple Security nodes.How ca i do this (or can
    i)?

  • How to access the datasource window in SSRS for sql server 2008 R2 for writing my query without having to go through the wizard?

    I have used SSRS a lot years ago with Sql Server 2000 and Sql Server 2005. I have written external assemblies, ... But now I have to do this with Sql Server 2008 (R2 -- which I realize I am way behind the times already but ...)  in sql server 2000 and
    2005 there was a tab for datasource to the left of the tab for design which was to the left of the preview tab. How do I get to the datasource window in sql server 2008 (r2) ?
    I see that datasource explorer.  But where can I get to the datasource window to edit my queries and so forth for sql server 2008 (R2)?
    Thanks
    Rich P

    I think I found the answer to my question --- just right-click on the the Data Sources or Datasets for editing connections and dataset queries.  I'm guessing it gets even fancier with Sql Svr 2012 - 2014.    Man, that's the one thing
    about coding platforms -- you let it go for a few years and come back, and everything has changed (well, a lot of things).  Now I need to figure out how to add an external assembly to SSRS 2008 (R2).
    Rich P

  • Error on clone database, oracle 10g release 2 for windows server 2008

    Hi,
    OS: Windows Server 2008 X64
    Oracle: oracle 10g release r2 for windows server 2008
    An error occurred when creating a database, stop at last step, Screenshot:
    [http://temp.wsria.com/oracle_create_db.png|http://temp.wsria.com/oracle_create_db.png]
    the trace log content:
    [main] [9:56:35:982] [CommandLineArguments.process:563] CommandLineArguments->process: number of arguments = 32
    [main] [9:56:35:982] [CommandLineArguments.process:738] CommandLineArguments->process: Create Database argument is specified
    [main] [9:56:35:982] [CommandLineArguments.process:910] CommandLineArguments->process: template Name argument is specified
    [main] [9:56:35:982] [CommandLineArguments.process:960] CommandLineArguments->process: db name argument is specified
    [main] [9:56:35:982] [CommandLineArguments.process:3074] CommandLineArguments->process: in Operation Type is Creation/GenerateScripts Mode condition
    [main] [9:56:35:997] [CommandLineArguments.process:3298] CommandLineArguments->process: Successfully process command line arguments
    [main] [9:56:36:668] [Host.checkOPS:2055] Inside checkOPS
    [main] [9:56:36:668] [Host.checkOPS:2067] Trying to check cluster existence
    [main] [9:56:36:715] [Library.getInstance:97] Created instance of Library.
    [main] [9:56:36:715] [Library.load:197] Loading orauts.dll...
    [main] [9:56:36:715] [Library.load:203] oracleHome D:\oracle\product\10.2.0\db_1
    [main] [9:56:36:715] [sPlatform.isHybrid:63] osName=Windows Vista osArch=amd64 rc=false
    [main] [9:56:36:715] [Library.load:223] Loading library D:\oracle\product\10.2.0\db_1\bin\orauts.dll
    [main] [9:56:36:715] [Library.load:247] Loaded library D:\oracle\product\10.2.0\db_1\bin\orauts.dll from path=
    D:\oracle\product\10.2.0\db_1\bin
    [main] [9:56:36:715] [Library.load:197] Loading MSVCRT.dll...
    [main] [9:56:36:715] [Library.load:203] oracleHome D:\oracle\product\10.2.0\db_1
    [main] [9:56:36:715] [sPlatform.isHybrid:63] osName=Windows Vista osArch=amd64 rc=false
    [main] [9:56:36:715] [Library.load:223] Loading library D:\oracle\product\10.2.0\db_1\bin\MSVCRT.dll
    [main] [9:56:36:731] [Library.load:247] Loaded library D:\oracle\product\10.2.0\db_1\bin\MSVCRT.dll from path=
    D:\oracle\product\10.2.0\db_1\bin
    [main] [9:56:36:731] [Library.load:197] Loading orawsec10.dll...
    [main] [9:56:36:731] [Library.load:203] oracleHome D:\oracle\product\10.2.0\db_1
    [main] [9:56:36:731] [sPlatform.isHybrid:63] osName=Windows Vista osArch=amd64 rc=false
    [main] [9:56:36:731] [Library.load:223] Loading library D:\oracle\product\10.2.0\db_1\bin\orawsec10.dll
    [main] [9:56:36:731] [Library.load:247] Loaded library D:\oracle\product\10.2.0\db_1\bin\orawsec10.dll from path=
    D:\oracle\product\10.2.0\db_1\bin
    [main] [9:56:36:731] [Library.load:197] Loading orasrvm10.dll...
    [main] [9:56:36:731] [Library.load:203] oracleHome D:\oracle\product\10.2.0\db_1
    [main] [9:56:36:731] [sPlatform.isHybrid:63] osName=Windows Vista osArch=amd64 rc=false
    [main] [9:56:36:731] [Library.load:223] Loading library D:\oracle\product\10.2.0\db_1\bin\orasrvm10.dll
    [main] [9:56:36:731] [Library.load:247] Loaded library D:\oracle\product\10.2.0\db_1\bin\orasrvm10.dll from path=
    D:\oracle\product\10.2.0\db_1\bin
    [main] [9:56:36:731] [Version.isPre10i:189] isPre10i.java: Returning FALSE
    [main] [9:56:36:731] [WindowsSystem.regKeyExists:1006] WindowsSystem.regKeyExists: mainkey= HKEY_LOCAL_MACHINE subkey = Software\Oracle\Ocr
    [main] [9:56:36:746] [WindowsSystem.getCSSConfigType:1163] configType=null
    [main] [9:56:36:746] [Host.checkOPS:2073] cluster existence:false
    [main] [9:56:36:746] [Host.checkOPS:2111] Cluster installed=false
    [main] [9:56:36:902] [InitParamHandler.endElement:506] CustomSGA flag: false
    [main] [9:56:36:902] [InitParamHandler.endElement:507] Database Type: MULTIPURPOSE
    [main] [9:56:36:918] [InitParamHandler.endElement:508] Mem Percentage: 40
    [main] [9:56:36:918] [InitParamHandler.endElement:526] distributing Memory: 13737443328
    [main] [9:56:36:918] [MemoryCalculator.calculateMemory:122] Setting SGA to MAX_SGA 1610612736
    [main] [9:56:36:918] [StorageAttributes.setAttribute:232] IN threadID:1 group#=1
    [main] [9:56:36:918] [StorageAttributes.setAttribute:232] IN threadID:1 group#=2
    [main] [9:56:36:918] [StorageAttributes.setAttribute:241] Current threadID=1
    [main] [9:56:36:918] [StorageAttributes.setAttribute:248] Current threadID=1 ==> redoGroups[0]=1
    [main] [9:56:36:918] [StorageAttributes.setAttribute:258] vRedoGroups:[1]
    [main] [9:56:36:918] [StorageAttributes.setAttribute:288] setAttribute: bExists=false
    [main] [9:56:36:918] [StorageAttributes.setAttribute:232] IN threadID:1 group#=3
    [main] [9:56:36:918] [StorageAttributes.setAttribute:241] Current threadID=1
    [main] [9:56:36:918] [StorageAttributes.setAttribute:248] Current threadID=1 ==> redoGroups[0]=1
    [main] [9:56:36:918] [StorageAttributes.setAttribute:248] Current threadID=1 ==> redoGroups[1]=2
    [main] [9:56:36:918] [StorageAttributes.setAttribute:258] vRedoGroups:[1, 2]
    [main] [9:56:36:933] [StorageAttributes.setAttribute:288] setAttribute: bExists=false
    [main] [9:56:36:933] [TemplateManager.parseCloneTemplate:1477] See for any transportable datafiles in TemplateManager.....
    [main] [9:56:36:933] [TemplateManager.isInstallTemplate:2178] Selected Template by user:=General Purpose
    [main] [9:56:36:933] [TemplateManager.isInstallTemplate:2185] The Message Id to be searched:=GENERAL_PURPOSE
    [main] [9:56:36:933] [TemplateManager.parseCloneTemplate:1489] create new clone data file for tp file.......
    [main] [9:56:36:933] [Host.setupOIDCommandlineParameters:7184] setupOIDCommandlineParameters:
    [main] [9:56:36:933] [Host.setupOIDCommandlineParameters:7185] m_regWithdirService: false
    [main] [9:56:36:933] [Host.setupOIDCommandlineParameters:7186] m_unregWithdirService: false
    [main] [9:56:36:933] [Host.setupOIDCommandlineParameters:7187] m_updateDirService: false
    [main] [9:56:36:933] [Verifier.processRawConfigFile:3523] StorageType == 0
    [main] [9:56:36:933] [Verifier.setOradataDest:4349] setOradataDest:dfDest=D:\oracle\product\10.2.0\oradata
    [main] [9:56:36:933] [TemplateManager.updateDatafileDestination:1957] updateDatafiles:datafileDir=D:\oracle\product\10.2.0\oradata
    [main] [9:56:36:933] [TemplateManager.updateDatafileDestination:2103] From template, RedoLogGrName=1
    [main] [9:56:36:965] [TemplateManager.updateDatafileDestination:2118] new file name redo01.log
    [main] [9:56:36:965] [TemplateManager.updateDatafileDestination:2103] From template, RedoLogGrName=2
    [main] [9:56:36:965] [TemplateManager.updateDatafileDestination:2118] new file name redo02.log
    [main] [9:56:36:965] [TemplateManager.updateDatafileDestination:2103] From template, RedoLogGrName=3
    [main] [9:56:36:965] [TemplateManager.updateDatafileDestination:2118] new file name redo03.log
    [main] [9:56:36:965] [ProgressOnlyHost.performOperation:162] processRawConfigFile=false
    [main] [9:56:36:965] [Verifier.validateTemplate:1629] StorageType == 0
    [main] [9:56:36:965] [ProgressOnlyHost.performOperation:178] validateTemplate=true
    [main] [9:56:36:965] [OracleHome.isRacEnabled:149] bRacOn = false
    [main] [9:56:36:980] [Verifier.validateTemplate:1629] StorageType == 0
    [main] [9:56:36:980] [Verifier.calculateCloneDatafilePathsAndSizes:2951] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateCloneDatafilePathsAndSizes:2951] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateCloneDatafilePathsAndSizes:2951] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateCloneDatafilePathsAndSizes:2951] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateCloneDatafilePathsAndSizes:2951] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateRedoLogGroupFileSizes:3083] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateRedoLogGroupFileSizes:3083] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.calculateRedoLogGroupFileSizes:3083] canonicalPath=D:\oracle\product\10.2.0\
    [main] [9:56:36:980] [Verifier.getControlfFileSizes:3001] No. of Control files:=3
    [main] [9:56:37:105] [Host.executeSteps:4044] Executing steps....
    [main] [9:56:37:105] [Host.setUpForOperation:2920] setUpForOperation: Mode = 128
    [main] [9:56:37:121] [Host.executeSteps:4186] setupForOperation returned: true
    [main] [9:56:37:121] [Host.createStepSQLInterface:5948] sid =ora10g
    [main] [9:56:37:136] [SQLEngine.initialize:242] Execing SQLPLUS/SVRMGR process...
    [main] [9:56:37:136] [SQLEngine.initialize:270] m_bReaderStarted: false
    [main] [9:56:37:136] [SQLEngine.initialize:274] Starting Reader Thread...
    [Thread-4] [9:56:37:355] [StepContext$ModeRunner.run:2478] ---- Progress Needed:=true
    [Thread-4] [9:56:37:464] [BasicStep.execute:202] Executing Step : CLONE_DB_CREATION_RMAN_RESTORE
    [Thread-4] [9:56:37:464] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-01092
    [Thread-4] [9:56:37:464] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-01034
    [Thread-4] [9:56:37:464] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-03114
    [Thread-4] [9:56:37:464] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-12560
    [Thread-4] [9:56:37:464] [StepErrorHandler.setIgnorableErrors:250] setting Ignorable Error: ORA-01109
    [Thread-4] [9:56:37:464] [BasicStep.configureSettings:304] messageHandler being set=null
    [Thread-4] [9:56:37:464] [BasicStep.execute:202] Executing Step : INSTANCE_CREATION
    [Thread-4] [9:56:37:464] [BasicStep.configureSettings:304] messageHandler being set=null
    [Thread-4] [9:56:37:479] [InitParamAttributes.sortParams:3532] m_sortOn:-1 sortOn:4
    [Thread-4] [9:56:37:620] [OracleHome.isRacEnabled:149] bRacOn = false
    [Thread-4] [9:56:37:667] [Host.noEntryinOratab:5115] Check made for oratab arg passed............
    [Thread-4] [9:56:37:667] [Oradim.getAddEntryCommand:353] AddEntry=[D:\oracle\product\10.2.0\db_1\bin\oradim.exe, -new, -sid, ORA10G, -startmode, manual, -spfile]
    [Thread-4] [9:56:38:899] [Oradim.getEditEntryCommand:422] getEditEntry cmd=[D:\oracle\product\10.2.0\db_1\bin\oradim.exe, -edit, -sid, ORA10G, -startmode, auto, -srvcstart, system]
    [Thread-4] [9:56:39:55] [Oradim.addSidToRegistry:871] oracleHomeKey: SOFTWARE\ORACLE\KEY_OraDb10g_home1
    [Thread-4] [9:56:39:117] [BasicStep.configureSettings:304] messageHandler being set=oracle.sysman.assistants.util.UIMessageHandler@c73f0d8
    [Thread-4] [9:56:39:117] [CloneRmanRestoreStep.executeImpl:217] Instance Creation went fine..........
    [Thread-4] [9:56:39:117] [CloneRmanRestoreStep.executeImpl:224] db_recovery_file_dest=D:\oracle\product\10.2.0\flash_recovery_area
    [Thread-4] [9:56:39:117] [CloneRmanRestoreStep.executeImpl:227] db_recovery_file_dest_size=2147483648
    [Thread-4] [9:56:39:991] [SQLEngine.setSpool:1750] old Spool = null
    [Thread-4] [9:56:39:991] [SQLEngine.setSpool:1751] Setting Spool = D:\oracle\product\10.2.0\db_1\cfgtoollogs\dbca\ora10g\CloneRmanRestore.log
    [Thread-4] [9:56:39:991] [SQLEngine.setSpool:1752] Is spool appendable? --> true
    [Thread-4] [9:56:39:991] [CloneRmanRestoreStep.executeImpl:320] starting with pfile=D:\oracle\product\10.2.0\admin\ora10g\pfile\init.ora

    Check your SQLNET.AUTHENTICATION_SERVICES settings in sqlnet.ora file.
    ORA-28547 : Connection to server failed, probable Net8 admin error
    thanks
    http://swervedba.wordpress.com/

  • Carrier Type In DMS for content server

    Dear All,
    Can anyone help me in telling which carrier type i have to define in dc20 for Content server, if any required.
    i have configured content server in Oac0,oact, csadmin, but while saving file on server (check in) it is giving me this error
    Error while checking in and storing:  Message no. 26253 .
    Regards
    Kashif

    This error usually comes beacuse of improper settings maintained in CSADMIN for repository:
    Do few changes in CSADMIN Settings for repository:
    ContentStorage Host : (Enter name instead of IP)
    ContentStorageName : WCC
    Storage : ContentStorage.dll
    Security : 1
    ContRepDescription : (Enter descreption as given in OAC0)
    driver : LiveCache or MaxDB (Whatever you are using)
    Incase of DB, maintain also
    SAPDMSUser :
    SAPDBPassword :
    Please check after making these changes and revert.
    /Tilak

Maybe you are looking for

  • Could not open log file in Win2000 sp3 Message in console

    I have configured log file for ias. Then restarted my Application server. It is giving Error message "Could not open log file ''logs/logVal.10817" in kas window. But logs files are there (created by the server itself) in the logs floder. I have confi

  • Using Threads And Sockets

    Hello, I want to create a program that can send 2 or more files through the socket to the same client at the same time. Is it possible? I am trying to use threads to do this but I am not sure what to do in those threads. Should I send the socket to t

  • HP 250 G1 wakes up every night

     After work i just closing my netbook and it goes to sleep But every night it wakes up for one two minute, fun stars work and then goes sleep.  This problem was always.

  • Digidesign 002 Rack with Logic Studio

    Hi, I have recently bought a macbook pro and am wanting to record at my drum teachers house. He uses an iMac running Pro Tools with the 002 Rack. As i am wanting to record and mix everything in logic, not pro tools, can i use a firewire 800 to 400 ca

  • OCM, the Balanced Scorecard, and Fostering Collaboration

    Many organizations try to implement a management methodology, like the Balanced Scorecard, in order to improve performance and meet the shifting needs of the marketplace. When developing a scorecard, the organization identifies what it believes are t