PowerCLI script to get Resource Pool Path

Hi
I was wondering if someone could help me, I am looking for a PowerShell script that would give me the specific path of a specific Resource Pool, but I need to base the filter on the path and not the name of the resource pool.
The reason for this is that our guys created a "Infrastructure Services" resource pool in 4 different places so when I query the resource pool I get 4 results but in 4 different paths.
Now what I would like to do is when I create the VM from PowerShell I would need to place the VM into one of the 4 resource pools but specifying a specific path.
Is this possible and could someone please give me an example code for this.
I hope this makes sense :)
Thanks in advance

Hi Ravi0211,
The script below may be helpful for you, which can filter the process based on the property "idprocess" and the services based on the service "name":
$pids=@()
$output=@()
$output1=@()
get-content d:\processid.txt|foreach{
$pids+=$_} #store the processid as an array
Get-WmiObject -class Win32_PerfFormattedData_PerfProc_Process | foreach{
if ($pids -contains $_.idprocess){#filter the processid listed in the file d:\processid.txt
$Obj = New-Object -TypeName PSObject
$Obj | Add-Member -MemberType NoteProperty -Name idprocess -Value $_.idprocess
$Obj | Add-Member -MemberType NoteProperty -Name Memory -Value $_.WorkingSetPrivate
$output+=$obj}
$services = get-content d:\service.txt
Foreach($service in $services){
Get-service | foreach{
If ($_.name –like “*$service*”){#filter the service based on the service name stored in d:\service.txt
$Obj1 = New-Object -TypeName PSObject
$Obj1 | Add-Member -MemberType NoteProperty -Name Service-Name -Value $_.name
$Obj1 | Add-Member -MemberType NoteProperty -Name Service-Status -Value $_.status
$output1+=$obj1}
$output
$output1
I hope this helps.

Similar Messages

  • Problems creating file. Can't get resource path.

    Hey, I'm new to Java with an Actionscript background and I'm using NetBeans IDE.
    Now this comes with my basic lack of understanding for how Java works but I'm getting confused a lot so if someone could just steer me in the right direction I'd be very grateful.
    Essentially I'm checking for a config file and if it's not there, I want to create a default one but the paths are messing me up. class.getResource seems to use the src or the build/classes dir? What's the relationship of these? If I put assets in the src folder should they be moved to the build folder or something like that?
    The actual problem in code is that if I create a file, (config.xml) and put it in my src folder I can grab it using Main.class.getResource("config.xml"); Which is great but if the file's not there it throws an exception..
    Now In actionscript I'd use the File class, which seems very similar to Java's. I'd pass in a literal for the filename "config.xml" and prepend a constant that gives the path of the apps storage directory. If the file's not there then the File.exists property == false.
    What I need is to get the same path that getReresource("config.xml") outputs, but in a File and I can't do that with the exception being thrown because the file's not there.
    Any help appreciated.. I can't read another blog full of different setups or face another Google page of results pertaining to different problems.
    - Chris W

    RoninStretch wrote:
    Essentially I'm checking for a config file and if it's not there, I want to create a default one but the paths are messing me up. class.getResource seems to use the src or the build/classes dir?It looks relative to the classpath, as I recall.
    What's the relationship of these? If I put assets in the src folder should they be moved to the build folder or something like that?NetBeans might do that for you. It's a reasonable thing for a build system to do.
    You might be better off working without an IDE until you're more familiar with Java.
    The actual problem in code is that if I create a file, (config.xml) and put it in my src folder I can grab it using Main.class.getResource("config.xml"); Which is great but if the file's not there it throws an exception..Actually I think getResource will return null if the resource isn't there. Are you not checking for null?
    Now In actionscript I'd use the File class, which seems very similar to Java's. I'd pass in a literal for the filename "config.xml" and prepend a constant that gives the path of the apps storage directory. If the file's not there then the File.exists property == false.The Java File object refers to a path on the filesystem, not a virtual one in the manner of Class.getResource. Check the result of getResource to see if it's null.
    What I need is to get the same path that getReresource("config.xml") outputs, but in a File and I can't do that with the exception being thrown because the file's not there.Can you show us the actual exception and actual code?
    Anyway, if you're packaging all this up in a jar, you wouldn't put the new config.xml that you create in the jar anyway.
    More likely you'd do something like:
    1) check user's home directory for config.xml
    2) if it's not present, check to see if there's a config.xml packaged in your jar.
    3) if the user has any config changes, put them in a config.xml in the user's home directory.
    So you wouldn't be trying to get the path of the default config.xml anyway.
    Or maybe you can use the Preferences framework.

  • Getting complete file path in Acro Java Script

    Hi All,
    I have a requirement where I have a field ZPATH1 on Adobe Interactive Form. I am putting a BROWSE button and writing following JavaScript on CLICK event on this button for populating the chosen file path into field ZPATH1.
    *Start of JavaScript
    var sFile;
    event.target.importDataObject(sFile);
    var oFile;
    oFile = event.target.getDataObject(sFile);
    this.parent.parent.Sub_Form_Fields_Mass.ZPATH1.rawValue = oFile.path;
    *End of JavaScript
    1. event.target.importDataObject(sFile) --> Giving me the pop up to select the file.
    2. this.parent.parent.Sub_Form_Fields_Mass.ZPATH1.rawValue = oFile.path; --> Putting the oFile path in field ZPATH1 on the form.
    3. This way is only giving me the selected FILE NAME whereas I need the complete path.
    I have tried below way also:
    this.parent.parent.Sub_Form_Fields_Mass.ZPATH1.rawValue = even.target.path;
    This function returns me the complete path but in URL form. It does not give me the presentation server path.
    Please suggest me how to get the complete path of presentation server to the file.
    Appreciation would be honored in form of points!!!
    Regards,
    Ashish

    I don't recognize the tech - is it something related to BusinessObjects Enterprise?
    Sincerely,
    Ted Ueda

  • Hyper-V Resource Pools for Memory and CPU

    Hi all,
    I'm trying to understand the concepts and details of resource pools in Hyper-V in Windows Server 2012. It seems as if there is almost no documentation on all that. Perhaps somebody can support me here, maybe I've not seen some docs yet.
    So far, I learned that resource pools in their current implementation serve mainly for metering purposes. You can create pools per tenant and then group VM resources into those pools to facilitate resource metering per tenant. That is, you enable metering
    once per pool and get all the data necessary to bill that one customer for all their resources (without metering individual VMs). Is that correct?
    Furthermore, it seems to me that an ethernet pool goes one step further by providing an abstraction level for virtual switches. As far as I've understood you can add multiple vSwitches to a pool and then connect a VM to the pool. Hyper-V then decides which
    actual switch to use. This may be handy in a multi-host environment if vSwitches on different hosts use different names although they connect to the same network. Is that correct?
    So - talking about actually managing that stuff I've learned how to create a pool and how to add VHD locations and virtual switches to a pool. Enabling resource metering for a pool then collects usage data from all the resources inside that pool.
    But now: I can create a pool for memory and a pool for CPU. But I cannot add resources to those. Neither can I add a complete VM to a pool. Now I'm launching a VM that belongs to a customer whose resources I'm metering. How will Hyper-V know that it's
    supposed to collect data on CPU and memory usage for that VM?
    Am I missing something here? Or is pool-based metering only good for ethernet and VHD resources, and CPU and memory still need to be metered per VM?
    Thanks for clarification,
    Nils
    Nils Kaczenski
    MVP Directory Services
    Hannover, Germany

    Thank you for the links. I already knew those, and unfortunately they are not matching my question. Two of them are about Windows Server 2008/R2, and one only lists a WMI interface. What I'm after is a new feature in Windows Server 2012, and I need conceptional
    information.
    Thanks for the research anyway. I appreciate that a lot!
    In the meantime I've gotten quite far in my own research. See my entry above of January 7th. Some additions:
    In Windows Server 2012, Hyper-V resource pools are mainly for metering purposes. You cannot compare them to resource pools in VMware.
    A resource pool in Hyper-V (2012) facilitates resource metering and billing for VM usage especially in hosting scenarios. You can either measure resource usage for single VMs, or you can group existing resources (such as CPU power, RAM, virtual hard disk
    storage, Ethernet traffic) into pools. Those pools will mostly be assigned to one customer each. That way you can bill the customer for their resource usage in a given time period by just querying the customer's pool.
    Metering only collects aggregated data with one value per resource (i.e. overall CPU usage, maximum VHD storage, summed Ethernet traffic and so on). You can control the time period by explicitly resetting the counter at any given time (a day, a week, a
    month or what you like).
    There is no detailed data. The aggregate values serve as a basis for billing, not as monitoring data. If you need detailed monitoring data use Performance Monitor.
    There is currently only one type of resource pool that adds an abstraction layer to a virtualization farm, and that is the Ethernet type. You can use that type for metering, but you can also use it to group a number of virtual switches (that connect to
    the same network segment) and then a VM connected to that pool will automatically use an appropriate virtual switch from the pool. You need no longer worry about virtual switch names across multiple hosts as long as all equivalent virtual switches are
    added to the pool.
    While you can manage two types of pool resources in the GUI (VHD pools and Ethernet pools) you should only manage resource pools via PowerShell. Only there will you be able to control what happens. And only PowerShell provides a means to start, stop, and
    reset metering and query metering data.
    The process to use resource pools in Hyper-V (2012) in short:
    First create a new pool via PowerShell (New-VMResourcePool). (In case of a VHD pool you must specify the VHD storage paths to add to the pool in the moment you create the pool.)
    In case of an Ethernet pool add existing virtual switches to the pool (Add-VMSwitch).
    Reconfigure existing VMs that you want to measure so that they use resources from the pool. The PowerShell
    Set-VM* commands accept a parameter -ResourcePoolName to do that. Example:
    Set-VMMemory -VMName APP-02 -ResourcePoolName MyPool1
    Start measuring with Enable-VMResourceMetering.
    Query collected data as often as you need with Measure-VMResourcePool.
    Note that you should specify the pool resource type in the command to get reliable data (see my post above, Jan 7th).
    When a metering period (such as a week or a month) has passed, reset the counter to zero with
    Reset-VMResourceMetering.
    Hope that helps. I consider this the answer to my own question. ;)
    Here's some links I collected:
    http://itproctology.blogspot.ca/2012/12/hyper-v-resource-pool-introduction.html
    http://www.ms4u.info/2012/12/configure-ethernet-resource-pool-in.html
    http://blogs.technet.com/b/virtualization/archive/2012/08/16/introduction-to-resource-metering.aspx
    http://social.technet.microsoft.com/Forums/en-US/winserverhyperv/thread/1ce4e2b2-8fdd-4f16-8ab6-e1e1da6d07e3
    Best wishes, Nils
    Nils Kaczenski
    MVP Directory Services
    Hannover, Germany

  • Resource pool selection enhancement in UCCX 7.0

    Good day. The issue is with our current CSQ setup and available resource pool selection criteria. Just for your record here is how its currently configured.
    We have five languages based CSQ’s operating in our call centre and following is how they are currently configured.
    I. English
    Contains all agents in the call centre
    Resource pool selection mode – Resource Group
    Resource Selection Criteria – Circular
    II. Italian / French / German / Spanish
    Resource pool selection mode – Resource Skills
    Resource Selection Criteria – Most Skilled
    Minimum Competence Level - 5 for respective CSQ
    Agents skill ratings are set currently in following manner
    Italian Agents – Italian (10), German (5), French (5), Spanish (5)
    French Agents – French (10), German (5), Italian (5), Spanish (5)
    German Agents – German (10), French (5), Italian (5), Spanish (5)
    Spanish Agents – Spanish (10), French (5), Italian (5), German (5)
    Now , all non- English CSQ are setup based on “Resource Skills” and “ Most Skilled” algorithm being set as resource selection criteria. Which currently defaults to “ Longest Available” if two or more agents have equal competency level. Things are nice and smooth until calls routed to Longest Available agent and they go for shorter smoke or toilet breaks. Calls then get routed to the next longest available agent who in our case have already answered the previous call coming to same CSQ. So what’s happening here is, agents who doesn’t go for these shorter breaks get punished with more calls.
    We don’t have this problem on English CSQ as we have different resource pool selection set in there. Which is Set as Resource group with “Circular” as resource selection criteria. This works fine because there’s no fall-back to “ Longest Available”. We have tried every other selection criteria but none works for us unfortunately. Raising this with you hoping you can advise different technique or guide us to the right path to re-design our non-English CSQs. Can you tell if we can create additional resource selection criteria other than the default ones? Can we add something like followings
    •1. Lowest total talk time
    •2. Least Handled contacts
    •3. Circular ( like you can select when agents are in recourse group )
    Also , is there a way to change the default selection criteria “ Longest Available” to something else. E.g. Shortest Average Handle Time or anything else.
    Any help would be highly appriciated.
    Thanks
    mschowdhury

    If you're using skills-based routing it's going to work the way you've described it above: the most skilled agent who has been in the READY state the longest will be offered the next call. If you want to use Circular or Most Handled Contacts (Least Handled Contacts is not an option sadly) you must use a Resource Group on the CSQ instead of a Skill. The problem with that is that you can only assign an agent to one Resource Group.
    Sorry but this is sort of an either/or decision. You can't mix the behaviors.
    Please remember to rate helpful responses and identify helpful or correct answers.

  • SQL2012 Express Error: "insufficient system memory in resource pool 'internal' to run this query"

    I am running the SQL2012 express (version shown below) and get the following error after a day or two of operation. This is a new server 2012 installation.
    source: SQL2012Express  "There is insufficient system memory in resource pool 'internal' to run this query."
    The installed version of SQL2012 express is:
    Microsoft SQL Server 2012 (SP1) - 11.0.3128.0 (X64)   Dec 28 2012 20:23:12  Copyright (c) Microsoft Corporation
    Express Edition (64-bit) on Windows NT 6.2 <X64> (Build 9200: ) (Hypervisor)
    I read a few of the postings but couldn't determine the exact hot fix or update to run or if I am running the latest already.
    Which .exe file should I run for the update? I downloaded a file called 473913_intl_x64_zip.exe but it said it wasn't tested yet?
    Can I uninstall this path or hot fix if it breaks something?
    What can I do to correct this internal memory error and prevent from happening.
    Thanks
    Morris
    Thank you Morris

    Hi MoCoder,
    According to your description, the error may be occurred by the following possible reasons, for example,
     the ram was be completely used, reached the maximum memory allocation configured values in SQL Server or virtual memory is full and so on. Before you install the related cumulative update of SQL Server 2012, I recommend you do the following
    steps for fixing this error. For example, allocate more memory to SQL Server, kill the unnecessary idle sessions, increase the RAM and virtual memory and reduce the number of users and so on.
    For more information, you can review the following article,
    http://sqlserverlearner.com/tag/there-is-insufficient-system-memory-in-resource-pool-internal-to-run-this-query
    In addition, you also need to check if there is the SQL Server memory leak, most of the memory leaks in SQL Server is caused by 3rd party Dll’s which are loaded in SQL Server process,
     or you enable the Auto Update Statistics Asynchronously statistics option in a database of Microsoft SQL Server 2012.
    For more information, see:
    http://support.microsoft.com/kb/2778088/en-us
    http://mssqlwiki.com/2012/12/04/sql-server-memory-leak/
    Regards,
    Sofiya Li
    If you have any feedback on our support, please click here.
    Sofiya Li
    TechNet Community Support

  • Critical error All management server resource pool

    I just receive critical error: All management server resource pool in scom 2012 management console.
    Alert detail:
    Alert subscription data source module encountered errors while running: Alert subscription data source module was unable to find alerts that match the subscription because of database errors.
    The following error(s) were encountered:
    Exception Message: ExecuteScalar requires an open and available Connection. The connection's current state is closed.
    One or more workflows were affected by this.
    Workflow name: Subscription3f818422_f4a1_42aa_b1b3_8bc14eb54cd8
    Instance name: Alert Notification Subscription Server
    Instance ID: {E07E3FAB-53BC-BC14-1634-5A6E949F9230}
    Management group: SCOM_2012
    What is problem and how to fix it ? I haven't changed any setting in scom !!

    Only found some of error and warning in operation manager event log.
    Error:
    Source: Health Service Modules
    Event ID: 22406
    Task Category: NOne
    The PowerShell script failed with below exception
    System.Management.Automation.CmdletInvocationException: The following error occurred while loading the extended type data file:
    Microsoft.PowerShell, C:\Program Files\System Center Operations Manager 2012\Powershell\OperationsManager\.\OM10.CoreCommands\Microsoft.EnterpriseManagement.Core.Cmdlets.Types.ps1xml : File skipped because it was already present from "Microsoft.PowerShell".
    At line:8 char:66
    +               if (!(Get-Module OperationsManager)) {Import-module <<<<  $SCOMModulePath}
       at System.Management.Automation.Internal.PipelineProcessor.SynchronousExecuteEnumerate(Object input, Hashtable errorResults, Boolean enumerate)
       at System.Management.Automation.PipelineNode.Execute(Array input, Pipe outputPipe, ArrayList& resultList, ExecutionContext context)
       at System.Management.Automation.StatementListNode.ExecuteStatement(ParseTreeNode statement, Array input, Pipe outputPipe, ArrayList& resultList, ExecutionContext context)
    Script Name: Discover Agent Versions
    One or more workflows were affected by this. 
    Workflow name: ManagementGroupDiscoveryAgentVersions
    Instance name: Operations Manager Management Group
    Instance ID: {6B1D1BE8-EBB4-B425-08DC-2385C5930B04}
    Management group: SCOM_2012
    Error:
    Source: HealthService
    Event ID: 15002
    Task Category: Pool Manager
    The pool member cannot send a lease request to acquire ownership of managed objects assigned to the pool because half
    or fewer members of the pool acknowledged the most recent initialization check request. The pool member will continue
    to send an initialization check request.
    Management Group: SCOM_2012
    Management Group ID: {56D0B0FE-FB03-2A6C-E3F6-54287D96855D}
    Pool Name: AD Assignment Resource Pool
    Pool ID: {529CF61E-A357-5AED-73CC-81D48E4327CA}
    Pool Version: 7607127134982980644
    Number of Pool Members: 2
    Number of Observer Only Pool Members: 1
    Number of Instances: 1
    Error:
    Source: HealthService
    Event ID: 15002
    Task Category: Pool Manager
    The pool member cannot send a lease request to acquire ownership of managed objects assigned to the pool because half
    or fewer members of the pool acknowledged the most recent initialization check request. The pool member will continue
    to send an initialization check request.
    Management Group: SCOM_2012
    Management Group ID: {56D0B0FE-FB03-2A6C-E3F6-54287D96855D}
    Pool Name: Notifications Resource Pool
    Pool ID: {3F9F97EA-E584-B0EE-A4C9-78A86FB4F21F}
    Pool Version: 7607127134982980644
    Number of Pool Members: 2
    Number of Observer Only Pool Members: 1
    Number of Instances: 1
    Error:
    Source: HealthService
    Event ID: 15002
    Task Category: Pool Manager
    The pool member cannot send a lease request to acquire ownership of managed objects assigned to the pool because half
    or fewer members of the pool acknowledged the most recent initialization check request. The pool member will continue
    to send an initialization check request.
    Management Group: SCOM_2012
    Management Group ID: {56D0B0FE-FB03-2A6C-E3F6-54287D96855D}
    Pool Name: All Management Servers Resource Pool
    Pool ID: {4932D8F0-C8E2-2F4B-288E-3ED98A340B9F}
    Pool Version: 7607127134982980644
    Number of Pool Members: 2
    Number of Observer Only Pool Members: 1
    Number of Instances: 116
    Warning:
    Source: Health Service Modules
    Event ID: 10103
    Task Category: None
    The Windows Event Log Provider has resumed processing the Operations Manager event log on computer 'HKWOMM01.ap.zurich.com' after recovering from errors.
    One or more workflows were affected by this. 
    Workflow name: Microsoft.SystemCenter.CM.AEM.Modules.EventCollection
    Instance name: HKWOMM01.ap.zurich.com
    Instance ID: {BF86169C-4A71-C728-9649-338260E86093}
    Management group: SCOM_2012
    Warning:
    Source: Health Service Modules
    Event ID: 10103
    Task Category: None
    In PerfDataSource, could not resolve counter instance OpsMgr DW Writer Module, Avg. Batch Processing Time, ms, All Instances. Module will not be unloaded.
    One or more workflows were affected by this. 
    Workflow name: Microsoft.SystemCenter.DataWarehouse.CollectionRule.Performance.Writer.AvgBatchProcessingTime
    Instance name: HKWOMM01.ap.zurich.com
    Instance ID: {BF86169C-4A71-C728-9649-338260E86093}
    Management group: SCOM_2012
    Warning:
    Source: Health Service Modules
    Event ID: 10103
    Task Category: None
    In PerfDataSource, could not resolve counter instance OpsMgr DW Writer Module, Batches/sec, All Instances. Module will not be unloaded.
    One or more workflows were affected by this. 
    Workflow name: Microsoft.SystemCenter.DataWarehouse.CollectionRule.Performance.Writer.BatchesPerSecond
    Instance name: HKWOMM01.ap.zurich.com
    Instance ID: {BF86169C-4A71-C728-9649-338260E86093}
    Management group: SCOM_2012
    Warning:
    Source: Health Service Modules
    Event ID: 10103
    Task Category: None
    In PerfDataSource, could not resolve counter instance LogicalDisk, Current Disk Queue Length,
    \\?\Volume{9470dbf8-09e8-11e1-887f-806e6f6e6963}. Module will not be unloaded.
    One or more workflows were affected by this. 
    Workflow name: Microsoft.Windows.Server.2008.LogicalDisk.CurrentDiskQueueLength.Collection
    Instance name:
    \\?\Volume{9470dbf8-09e8-11e1-887f-806e6f6e6963}
    Instance ID: {80A7A920-B571-987C-9B3B-DE2164998AB9}
    Management group: SCOM_2012
    Warning:
    Source: Health Service Modules
    Event ID: 10103
    Task Category: None
    In PerfDataSource, could not resolve counter instance OpsMgr DW Synchronization Module, Avg. Batch Processing Time, ms, All Instances. Module will not be unloaded.
    One or more workflows were affected by this. 
    Workflow name: Microsoft.SystemCenter.DataWarehouse.CollectionRule.Performance.Synchronization.AvgBatchProcessingTime
    Instance name: b9e5cd4c-1b78-4530-b93f-69806d14d7f3
    Instance ID: {E6A569EF-6527-2B26-8345-2E16F0979064}
    Management group: SCOM_2012

  • Is there a limit of how many resources can be shared via a single Resource Pool in Project 2010 Professional standalone?

    We have a single master project resource pool of over 150 resources that are shared across let's say 10-15 sub projects or sharer files.
    We are using Microsoft Project Professional (standalone no project server)
    Are you aware of any limitations outside of project server with resource pool and sharing, or should we be using Project server (as I have read a lot of articles about resource pool sharing working better in project server)
    I'd like to get clarification on:
    1. I read that the calendars for resources (of which there could be many shift patterns etc.) should always be created in the master project only, not in the sharer files. In my testing I see that a shared resource also brings the additional calendars from
    the Master Project file - and can be chosen in the calendar column from Resource sheet.
    2. Each sharer file can have it's own specific project calendar, but the resource calendar (used by the assigned resources from the pool) will drive the scheduling (as per normal), unless;
    3. A specific task has a 'Task Calendar' applied to it AND "scheduling ignores resource calendars" is ticked.
    I am trying to find an article that addresses, in detail, the limitations or 'look out fors' in this area.  I have found plenty about how to create and share resources, but this is more troubleshooting.
    Any advice or direction that can be given would be appreciated. Thanks.

    No limit to resources, but linked master and resource pools are at high risk of file corruption if you rename, over-write or move a linked file without removing it from the master and pool first. Files will corrupt, its when not if.
    Master files should never have any resources or tasks in them, only inserted projects.
    I much prefer creating master files with the Link option deselected. This copies all data and consolidates resource info (so no pool needed). I record a macro to create new masters each week. This method has no corruption risk.
    Your alternative now is to experiment with Project Server. By far the best way to prototype a system is to get Project Online through Microsoft, ProjectHosts or Bemo. It will give you access to a vanilla project server very quickly and cost for a few months
    testing is trivial for a few users.
    If you go ahead with a full implementation I strongly recommend getting a qualified Project Server partner to help you implement the full system.
    Rod Gill
    Author of the one and only Project VBA Book
    www.project-systems.co.nz

  • Open script cannot get connection from the brower helper after 15 seconds.

    Error:
    ===
    Open script cannot get connection from the brower helper after 15 seconds. Do you want to continue waiting for the browser to load?
    Please Note:
    ========
    1. I have tried this only on IE
    2. I am running OATS on a Remote desktop
    Situation:
    ======
    Trying to stop the recording
    Try to get xpath of an object using Inspect Path
    Setup details
    ========
    Windows XP 5.1 Service Pack 3, x86
    OpenScript 12.1.0.1.383
    Internet Explorer 8.0.6001.18702
    FireFox 13.0.1
    Mitigation steps done till now:
    ==================
    1. Disabled windows firewall
    2. Disable XSS filter setting
    3. Restarted the ATS services (3 of them)
    4. Run the Open Script Diagnosis Tool (PS: There are 3 errros even after running it. The 3 errros are listed in the workspace_log log file snippet below...)
    Error in worspace_log:
    =============
    To Change setting:
    Go to Tools > Internet Options and Choose Security Tab
    Select the Zone to modify and Press Custom level
    Find Enable XSS filter Setting - Select Disable and click Ok
    !ENTRY oracle.oats.scripting.diagnosisTool.api.DiagnosisExecutor 4 0 2012-07-09 17:08:52.594
    !MESSAGE Failure found when diagnosing Oracle EBS/Forms Load Testing Forms LT Diagnoser
    !ENTRY oracle.oats.scripting.diagnosisTool.api.DiagnosisExecutor 4 0 2012-07-09 17:08:52.594
    !MESSAGE Did not auto-fix the problem.
    !ENTRY oracle.oats.scripting.diagnosisTool.api.DiagnosisExecutor 4 0 2012-07-09 17:08:52.594
    !MESSAGE Suggestion for fixing: Please change your Java proxy setting to Use Browser Settings
    Aprreciate help on this.

    To resolve this, you need to reconfigure the "Oracle Application Testing Suite Helper Service" (OATSHelperSvr) to start as a user who has privledges to run open script tests rather than the default SYSTEM user.
    Reconfiguring the OATSHelperSvr Service:
    1. Open the services panel (Start > Run > services.msc)
    2. Find the Oracle Application Testing Suite Helper Service
    3. Right Click > Properties then select the Log On Tab
    4. Specify an interactive user that has rights to run OpenScript (test by logging in as that user and running tests):
    5. Click OK
    6. Restart the service after dialogs are closed by Right Click > Restart
    7. You should now repeat this process for the "Oracle Application Testing Suite Agent Service" (eLoadAgentMon) Service (Two services in
    total)
    You should now retry running the test in Oracle Test Manager

  • Project Server 2010 - Active Directory Enterprise Resource Pool Synchronization limitations

    Greetings again.
    I have a quick question about the limitations of Active Directory Enterprise Resource Pool Synchronization.  Specifically, what has your experience
    been with extremely large numbers of users (10k plus).  Is anyone aware of a
    practical limit of users in your AD group you would recommend when using the
    Schedule Synchronization feature on a nightly or weekly basis?
    There is a caveat to this question however.  The client has decided (perhaps for some misinformed reasons) to allow access for every user, to every
    project site, within their PWA environment.  They’ve selected the View Project Site option within the
    Categories for the Team Members PWA Group for which 90% of their intended users reside.  So when we ran a couple test syncs in DEV with a smaller AD group of about 8,000 users, the sync understandably
    lasted upwards of 18 hours.  Obviously unacceptable for a PROD environment on a nightly basis and not necessarily ideal for a weekly sync either.
    Experience in addition to documents like these, “Best
    practices for managing a large number of resources in Project Server 2010” tell me that we are way over the practical limit of a scheduled resource pool synchronization...IF
    the client really desires that all users access all their sites.  But before I submit my recommendations, I wanted to check with the community just in case others may have found a way to synchronize large numbers of users (10k plus) on a nightly or weekly
    basis, within a reasonable time frame AND allowed all users to access all sites within PWA.
    What do you think?
    As always, thanks for your help.
    Chris Addis - MCTS

    Hello Hrishi.
    My delayed response has been due to a large amount of testing we have been performing on this particular topic.  Here is an update, please feel free to provide feedback.
    We went back to testing and spent more time reviewing the documents: 
    Best practices for managing a large number of resources in Project Server 2010 and
    SharePoint Server 2010 capacity management: Software boundaries and limits.  Our team interpreted those documents as saying, 1,000 security
    scopes per site is a recommended limit.  It does not say it’s a hard limit, just a recommended limit.  “When the recommended unique security scope boundaries are exceeded, performance issue can occur.”
    So we decided to perform some tests (31 in total) to try and get a gage of what we are seeing.
    We needed to establish a baseline first.  So we performed a series of 23 Active Directory Resource Pool Synchronizations with various settings in a clean, Out of the Box, environment in order to see some consistent numbers.  Here is it’s summary:
    Our AD group of 8,000 users took about 32 minutes, on average, every time to sync.  The difference between the first sync and last sync differed only by 1-3 minutes.
    Adding 40 project sites increased the average sync time from 32 to 120 minutes.
    Adding 100 users to each of those 40 project sites, did not increase the sync times.
    One setting (identified at this time) reduced the synchronization time.  It was the
    Project Site Permissions check box found within the Project Web App > Server Settings > Project Site Provisioning Settings area.  By deselecting this check box we reduced or synchronization time back to the 32 minute
    average.
    The View Project Site check box within the Project Web App > Server Settings > Manage Groups > Team Members group had no apparent effect on the sync times besides what we had gleaned from the Microsoft documentation.
    Naturally, this left us with a problem.  As I’m sure you know, by deselecting the check box (Project Site Permissions), our project sites are now (figuratively speaking) orphans with no connection to the parent site.  This generates
    a new set of issues.  For example:
    Newly created project sites cannot be accessed by the owner and team members.  They will require someone like the farm admin to come in behind them and add the intended users to the project site along with their required permissions.
    All current and futures sites will no longer have users added via the standard method of building a team and publishing the project, but will have to be added manually.
    You can use the Synchronize option found within Project Web App > Server Settings > Project Sites page, but that kinda defeats the purpose.  It would require constant updating on a per site basis to keep up with
    PM changes.  Not very sensible, but it does work.
    With this baseline information, we moved our tests into our DEV environment which somewhat mimics our PROD environment.  This environment has 352 project sites and we performed 8 tests.  This is where we had some large sync time numbers. 
    Here is the DEV test summary:
    With the Project Site Permissions check box cleared, our AD group of 8,000 users took on average 30 minutes to sync.  This was in line with our baseline times.  With the exception of one test sync that took 99 minutes to complete. 
    That anomaly is acquiescent with what I’ve seen over the years.  Sometimes syncs do some unusual things.
    With the Project Site Permissions check box selected, our AD group of 8,000 users took on average 690 minutes (11.5 hours) to sync.  Unacceptable of course.
    So here’s what we’ve learned when dealing with extremely large numbers of active directory users in your Resource Pool sync:
    We did not see a decrease in subsequent sync times after the initial Active Directory Resource Pool Synchronization as some might expect.
    Our attempt to decrease sync time via the option of removing the View Project Sites was not successful.  (Unless we interpreted Microsoft's document incorrectly.)
    However, our interpretation of the recommended software boundaries and limits of SharePoint Server 2010 as it pertains to security scopes per site at 1,000, appears to be correct.
    Using the option of clearing the Project Site Permissions does produce a reduction in AD sync time, but at a cost segregating your project sites and thus the creation of new processes of maintaining them.
    I’ll remind others that these results are particular to our environments, there may still be exceptions yet to discover.  Others may see numbers contrary to ours.
    The biggest surprise to some members on our team (myself excluded) was that we did not see a reduced sync time after any of our initial syncs.  Some are under the impression that after your initial sync, you should see reduced sync times.  I haven’t
    found that to be the absolute case in all situations, just in some situations.  The reason for this still eludes me.  Any thoughts would be appreciated.
    I’ll let this sit a bit longer, but if no one disagrees with the results, I think we have our answer:
    The number of project sites directly affects your Active Directory Resource Pool Synchronizations if you are using the
    Project Site Permissions option.  If you plan on synchronizing over 1,000 users
    and you have a large number of project sites, proceed with the knowledge that you may have performance issues and long sync times.
    As always, I’d love to hear from you or others just in case I’m missing something.
    Chris Addis - MCTS

  • How can i get the local path to my cached app in jws

    Hi,
    sorry, i think it is a stupid question of me. but i'm a newbie in java and jws...
    So my question is...
    is it possible to get the absolute path to my downloaded an d cached jarfile? (e.g. c:\program files\java web start\.javaws\cache\http\Dmy.server.com\P80\RMmyapp.jar)
    thanks a lot for any help.

    I just answered this elsewhere, but here it is again.
    This code seems to work. I found it last year in one of the forums.
    I use it to find resources in Jars or to find images when I am in an IDE
    The "SetIcon.class" below should be changed to the class you are running from.
    Note: If you are running a class in a package, this returns the root of your application instead of the "folder" containing your package. This would mean that the "images" folder I refer to below is under the root folder, not the package folder.
    javax.swing.ImageIcon img;
    URL imgPath1;
    String imgPath2;
    imgPath1 = SetIcon.class.getProtectionDomain().getCodeSource().getLocation();
    imgPath2 = imgPath1.toString();
    imgPath2 = imgPath2.substring(6); // strip off the "file:"
    imgPath2 = imgPath2 + "images/sample.gif";
    Hope this helps,
    Hale

  • Master Projects, Subprojects, Resource Pool

    I have reviewed the previous threads and have not found a good match with my issue.  Here goes.
    I have a project plan (Plan1) stored on OneDrive.  I have a second plan that is my resource pool stored in the same location.  I open both plans by clicking on them in OneDrive.  When I go into Plan1, I go to Resource Pool -> Share Resources. 
    The pop-up box comes up with the 'Use Own Resources' selected.  The other option of 'Use Resources (requires at least one open resource pool)' is greyed out.  How do I get it such that I can select the second option and enter the name of my resource
    pool. 
    The resource pool is a .mpp file that has no tasks and it has about 70 resources defined.  Is there something I must toggle to show that it is a resource pool file? 
    I am using MS Project 2013 Professional.  I am not using Project Server or PWA. 
    Thanks in advance for your thoughts and assistance!
    Hillary
    Clarification:  I have both plans (which are both on OneDrive) opened at the same time - and the Plan1 file does not recognize the the other file (pool) as a potential candidate for a resource pool.
    Clarification #2:  When I copy the resource pool to my desktop, and then open it, Plan1 (on OneDrive) is able to see the resource pool and share it...

    HRDavidson,
    Convolved it is and you are also setting yourself up for corruption as I've indicated in my initial response. This is what I suggest. Since each element of the overall structure (i.e. resource pool, sharer files, master file, network, multi-user environment)
    contributes to the probability of corruption, your best chance for minimizing corruption is to eliminate and/or control each element.
    The ideal configuration is for all files to be resident on a local drive with one person in control but that probably isn't going to work in your scenario. However, there are some things you can do. First, I would not maintain a dynamic master file. You
    can still see the whole plan with all 6 individual projects whenever you need by creating a static master (i.e. uncheck the "link to project" option in the lower right corner of the Insert Project window). That will produce a new single file that
    is a snapshot in time of all the individual projects. The main advantage is that it does not create an additional link structure. Second, develop a set of groundrules on file maintenance and make sure all users are trained. The basic groundrules are: never
    move, rename, replace, or save off any files in the structure. You may wish to add other rules concerning what users can and cannot change with regard to data.
    Using the above approach you will still have a resource pool/sharer file linked structure on a network and it will still be prone to corruption but your chances for success will be increased. If you want to totally eliminate, (although nothing is really
    "total"), the chance of corruption, combine the whole structure into a single file with either one person in charge or a defined set of access rules (e.g. time) for each user. Filtering, grouping or other techniques can facilitate easier maintenance
    by each user.
    The other option of course is to move to Project Server.
    Those are my thoughts. Perhaps Dale and Rod will drop in with their comments.
    John

  • How to get the output path in Java?

    Hi all,
    is there a way (method) to get the output path (where compiled classes are put) in Java?
    thx a lot!
    Michele

    If you have already successfully loaded the classes into memory, and you want to find out where the classes are physically stored, then you can use Class.getResource() to retrieve the location of the file.
    import java.net.URL;
    public class Find
      private void run(String obj) {
        try {             
          Class cls = Class.forName(obj);
          //Here is the change to input correct resource path
          //instead of class name 
          String resourcePath = "/"+obj.replace('.','/')+".class";
          URL url = cls.getResource(resourcePath);
          System.out.println(url);
        catch (Exception e) {
          e.printStackTrace();
      public static void main(String[] args) {
        Find find = new Find();
        find.run(args[0]);   
    }java Find java.lang.String
    jar:file:/usr/local/j2sdk1.4.2_13/jre/lib/rt.jar!/java/lang/String.class
    Edited by: Jin on Oct 23, 2007 10:38 AM

  • JDBC Receiver Error: Resource Pool Exception

    Hi All,
    I am having a scenario where my messages are flowing from SAP to legacy system using JDBC as the receiver adapter and at the legacy system it is updating the Oracle Database. It was all running succesfully from long back. Now I got the following error (in every communicaton channel which using the JDBC adapter, I chcekd this in communication channel monitoring) in Runtime Workbench --> Message Monitoring --> Adapter Engine.
    Error Category XI_J2EE_ADAPTER_JDBC
    Error Code PROCESSING_ERROR
    2009-09-16 22:03:35 Success Receiver JDBC adapter: processing started; QoS required: ExactlyOnce
    2009-09-16 22:03:59 Error Exception caught by adapter framework: Error when attempting to get processing resources: com.sap.aii.af.service.util.concurrent.ResourcePoolException: Unable to create new pooled resource: DriverManagerException: Cannot establish connection to URL 'jdbc:oracle:thin:@XX.XX.XX.XX:XXXX:XXXX': SQLException: ORA-04031: unable to allocate 3968 bytes of shared memory ("shared pool","unknown object","sga heap(1,0)","transaction")
    I have checked in SDN where I get information about : Cannot establish connection to URL 'jdbc:oracle:thin:@XX.XX.XX.XX:XXXX:XXXX':. But I am not getting any information about SQLException: ORA-04031: unable to allocate 3968 bytes of shared memory ("shared pool","unknown object","sga heap(1,0)","transaction").
    Can any body please let me know about the error which I am getting and how can I resolve it.
    Thanks in advance.

    Thanks Abhishek for ur reply.
    This means that it not the technical job in which I can provide any help. Our DBA they all are client.
    I have read in SDN that it cannot establish connection to URL because either Oracle is down or JDBC parameters have been changed (like user name/password).
    If enough resource pools are not available...enough space / resources seems to be unavailable then what needs to be done from my side (as the client always ask me what needs to be done then and by whom).
    Thanks
    Zeba

  • Powershell script to get the low disk space on the servers...

    Hi,
    We have many SQL & Sharepoint servers running on windows server. I need some Powershell script to get the low disk space on those servers and it should send an mail alert saying "Disc d:\ have less than threshold 15% free space".
    Awaiting for response.
    Thanks
    Anil

    you should be able to use this powershell script which you can run as a scheduled task.
    just change the details at the start of the script to point the script to the right location to create log files. you will need to have a plain text file containing a list of server names or ip addresses that you want to check drive space on.
    # Change the following variables based on your environment
    #specify the path to the file you want to generate containing drive space information
    $html_file_dir = "\\server\share\DriveSpace"
    #specify the path and file name of the text file containing the list of servers names you want to check drive space on - each server name on a seperate line in plain text file
    $server_file = "\\server\share\servers.txt"
    #speicfy the from address for the email
    $from_address = "[email protected]"
    #specify the to address for the email
    $to_address = "[email protected]"
    #specify the smtp server to use to send the email
    $email_gateway = "smtp.domain.com" # Can either be DNS name or IP address to SMTP server
    # The seventh line from the bottom (line 167) is used if your smtp gateway requires authentication. If you require smtp authentication
    # you must uncomment it and set the following variables.
    $smtp_user = ""
    $smtp_pass = ""
    # Change the following variables for the style of the report.
    $background_color = "rgb(140,166,193)" # can be in rgb format (rgb(0,0,0)) or hex format (#FFFFFF)
    $server_name_font = "Arial"
    $server_name_font_size = "20px"
    $server_name_bg_color = "rgb(77,108,145)" # can be in rgb format (rgb(0,0,0)) or hex format (#FFFFFF)
    $heading_font = "Arial"
    $heading_font_size = "14px"
    $heading_name_bg_color = "rgb(95,130,169)" # can be in rgb format (rgb(0,0,0)) or hex format (#FFFFFF)
    $data_font = "Arial"
    $data_font_size = "11px"
    # Colors for space
    $very_low_space = "rgb(255,0,0)" # very low space equals anything in the MB
    $low_space = "rgb(251,251,0)" # low space is less then or equal to 10 GB
    $medium_space = "rgb(249,124,0)" # medium space is less then or equal to 100 GB
    #### NO CHANGES SHOULD BE MADE BELOW UNLESS YOU KNOW WHAT YOU ARE DOING
    # Define some variables
    $ErrorActionPreference = "SilentlyContinue"
    $date = Get-Date -UFormat "%Y%m%d"
    $html_file = New-Item -ItemType File -Path "$html_file_dir\DiskSpace_$date.html" -Force
    # Create the file
    $html_file
    # Function to be used to convert bytes to MB or GB or TB
    Function ConvertBytes {
    param($size)
    If ($size -lt 1MB) {
    $drive_size = $size / 1KB
    $drive_size = [Math]::Round($drive_size, 2)
    [string]$drive_size + ' KB'
    }elseif ($size -lt 1GB){
    $drive_size = $size / 1MB
    $drive_size = [Math]::Round($drive_size, 2)
    [string]$drive_size + ' MB'
    }ElseIf ($size -lt 1TB){
    $drive_size = $size / 1GB
    $drive_size = [Math]::Round($drive_size, 2)
    [string]$drive_size + ' GB'
    }Else{
    $drive_size = $size / 1TB
    $drive_size = [Math]::Round($drive_size, 2)
    [string]$drive_size + ' TB'
    # Create the header and footer contents of the html page for output
    $html_header = '<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
    <html lang="en-US" xml:lang="en-US" xmlns="http://www.w3.org/1999/xhtml">
    <head>
    <title>Server Drive Space</title>
    <style type="text/css">
    .serverName { text-align:center; font-family:"' + $server_name_font + '"; font-size:' + $server_name_font_size + `
    '; font-weight:bold; background-color: ' + $server_name_bg_color + '; border: 1px solid black; width: 150px; }
    .headings { text-align:center; font-family:"' + $heading_font + '"; font-size:' + $heading_font_size + `
    '; font-weight:bold; background-color: ' + $heading_name_bg_color + '; border: 1px solid black; width: 150px; }
    .data { font-family:"' + $data_font + '"; font-size:' + $data_font_size + '; border: 1px solid black; width: 150px; }
    #dataTable { border: 1px solid black; border-collapse:collapse; }
    body { background-color: ' + $background_color + '; }
    #legend { border: 1px solid black; ; right:500px; top:10px; }
    </style>
    <script language="JavaScript" type="text/javascript">
    <!--
    function zxcWWHS(){
    if (document.all){
    zxcCur=''hand'';
    zxcWH=document.documentElement.clientHeight;
    zxcWW=document.documentElement.clientWidth;
    zxcWS=document.documentElement.scrollTop;
    if (zxcWH==0){
    zxcWS=document.body.scrollTop;
    zxcWH=document.body.clientHeight;
    zxcWW=document.body.clientWidth;
    else if (document.getElementById){
    zxcCur=''pointer'';
    zxcWH=window.innerHeight-15;
    zxcWW=window.innerWidth-15;
    zxcWS=window.pageYOffset;
    zxcWC=Math.round(zxcWW/2);
    return [zxcWW,zxcWH,zxcWS];
    window.onscroll=function(){
    var img=document.getElementById(''legend'');
    if (!document.all){ img.style.position=''fixed''; window.onscroll=null; return; }
    if (!img.pos){ img.pos=img.offsetTop; }
    img.style.top=(zxcWWHS()[2]+img.pos)+''px'';
    //-->
    </script>
    </head>
    <body>'
    $html_footer = '</body>
    </html>'
    # Start to create the reports file
    Add-Content $html_file $html_header
    # Retrieve the contents of the server.txt file, this file should contain either the
    # ip address or the host name of the machine on a single line. Loop through the file
    # and get the drive information.
    Get-Content $server_file |`
    ForEach-Object {
    # Get the hostname of the machine
    $hostname = Get-WmiObject -Impersonation Impersonate -ComputerName $_ -Query "SELECT Name From Win32_ComputerSystem"
    $name = $hostname.Name.ToUpper()
    Add-Content $html_file ('<Table id="dataTable"><tr><td colspan="3" class="serverName">' + $name + '</td></tr>
    <tr><td class="headings">Drive Letter</td><td class="headings">Total Size</td><td class="headings">Free Space</td></tr>')
    # Get the drives of the server
    $drives = Get-WmiObject Win32_LogicalDisk -Filter "drivetype=3" -ComputerName $_ -Impersonation Impersonate
    # Now that I have all the drives, loop through and add to report
    ForEach ($drive in $drives) {
    $space_color = ""
    $free_space = $drive.FreeSpace
    $total_space = $drive.Size
    $percentage_free = $free_space / $total_space
    $percentage_free = $percentage_free * 100
    If ($percentage_free -le 5) {
    $space_color = $very_low_space
    }elseif ($percentage_free -le 10) {
    $space_color = $low_space
    }elseif ($percentage_free -le 15) {
    $space_color = $medium_space
    Add-Content $html_file ('<tr><td class="data">' + $drive.deviceid + '</td><td class="data">' + (ConvertBytes $drive.size) + `
    '</td><td class="data" bgcolor="' + $space_color + '">' + (ConvertBytes $drive.FreeSpace) + '</td></tr>')
    # Close the table
    Add-Content $html_file ('</table></br><div id="legend">
    <Table><tr><td style="font-size:12px">Less then or equal to 5 % free space</td><td bgcolor="' + $very_low_space + '" width="10px"></td></tr>
    <tr><td style="font-size:12px">Less then or equal to 10 % free space</td><td bgcolor="' + $low_space + '" width="10px"></td></tr>
    <tr><td style="font-size:12px">Less then or equal to 15 % free space</td><td bgcolor="' + $medium_space + '" width="10px"></td></tr>
    </table></div>')
    # End the reports file
    Add-Content $html_file $html_footer
    # Email the file
    $mail = New-Object System.Net.Mail.MailMessage
    $att = new-object Net.Mail.Attachment($html_file)
    $mail.From = $from_address
    $mail.To.Add($to_address)
    $mail.Subject = "Server Diskspace $date"
    $mail.Body = "The diskspace report file is attached."
    $mail.Attachments.Add($att)
    $smtp = New-Object System.Net.Mail.SmtpClient($email_gateway)
    #$smtp.Credentials = New-Object System.Net.NetworkCredential($smtp_user,$smtp_pass)
    $smtp.Send($mail)
    $att.Dispose()
    # Delete the file
    Remove-Item $html_file
    Regards,
    Denis Cooper
    MCITP EA - MCT
    Help keep the forums tidy, if this has helped please mark it as an answer
    My Blog
    LinkedIn:

Maybe you are looking for

  • FI-TV Scheme and Field control in Travel Manager

    Hello! Question: The question is - Is it possible to somehow customize field control for TRIP selective screen by means of SPRO IMG Scheme and field control or transaction FITVFELD (i.e. hide some fields or add new fields)? Or fields can be added or

  • Will Thunderbolt allow multiple external displays?

    The news about Thunderbolt mentions the ability to "daisy chain" up to 6 devices to the port. Have there been any details released about the potential of driving multiple external monitors with Thunderbolt? I have two 24" monitors that I'd like to us

  • Tee nodes with custom MFTs as outputs

    Hi guys, Hopefully you have an idea on what I am observing. To give you guys a background on my application. I originally setup a pipeline that would take some user video editing requests and applies them to the frames. For example, users want to cho

  • Decompiled Sun's Active X Bridge Bean Examples

    The examples you can download, JCalendarPanel and JellyBean, I unzipped the jar files and then ran each class file through a decompiler which produced some decent code to look at. It at least gave me one perspective to see what was going on...on the

  • Error message A12E1-what is this?

    Error message A12E1 - what is this?