CProjects 4.0 :: Task and subproject connection table

Hello!
Could u please name tables where connection task and it's subproject is recorded in data base?
I'm using multi-project management with subprojects but I haven't found any connection in DPR_TASK table. I wounder where it records how master project task connecteed with subprojects.
Best regards,
Valerie

Hi Valerie,
Probably CGPL_RELATION and CGPL_HIERARCHY might help.
Let me know if this works.
Regards,
Vivek Pandey

Similar Messages

  • CProjects 4.0 : Task and subproject connection table record

    Hello!
    Could u please name tables where connection task and it's subproject is recorded in data base?
    I'm using multi-project management with subprojects but I haven't found any connection in DPR_TASK table. I wounder where it records how master project task connecteed with subprojects.
    Best regards,
    Valerie

    Answered in the other topic.

  • Task to add new row and values to Table?

    Hi, I am new to SSIS, I have a table that I would like to capture daily currency conversion rates. In my Control Flow
    pane, I've been able to use a web service task to get a string value of the conversion rate pass it to an XML task, and connect it to at Data Flow Task. So now I'm not sure what to do from here. I would like to input the daily values into my table:
    [CurrencyID] [int] IDENTITY(1,1) NOT NULL,
    [CurrXDate] [datetime] NOT NULL,
    [Country] [nvarchar](50) NOT NULL,
    [CurrXRate] [float] NOT NULL,
    insert into tbl_CurrXrates
    (CurrXDate, Country, CurrXRate)
    values
    (getdate(), 'USD', @[User::ConversionRate])
    What data tasks and order would I use to add a new row to my table which I think would be a simple insert statement. Hope my
    question makes sense.

    You can put into a package variable and use as part of the insert statement via a SSIS expression to drive a Execute SQL Command to insert it or
    Have the value off the webservice dumped to a XML file and then consume it with XML Source which will allow the values to be inserted into the table (e.g. OLEDB Destination)
    Arthur
    MyBlog
    Twitter

  • SQL Server 2012 Management Studio:In the Database, how to print out or export the old 3 dbo Tables that were created manually and they have a relationship for 1 Parent table and 2 Child tables?How to handle this relationship in creating a new XML Schema?

    Hi all,
    Long time ago, I manually created a Database (APGriMMRP) and 3 Tables (dbo.Table_1_XYcoordinates, dbo.Table_2_Soil, and dbo.Table_3_Water) in my SQL Server 2012 Management Studio (SSMS2012). The dbo.Table_1_XYcoordinates has the following columns: file_id,
    Pt_ID, X, Y, Z, sample_id, Boring. The dbo.Table_2_Soil has the following columns: Boring, sample_date, sample_id, Unit, Arsenic, Chromium, Lead. The dbo.Table_3_Water has the following columns: Boring, sample_date, sample_id, Unit, Benzene, Ethylbenzene,
    Pyrene. The dbo.Table_1_XYcoordinates is a Parent Table. The dbo.Table_2_Soil and the dbo.Table_3_Water are 2 Child Tables. The sample_id is key link for the relationship between the Parent Table and the Child Tables.
    Problem #1) How can I print out or export these 3 dbo Tables?
    Problem #2) If I right-click on the dbo Table, I see "Start PowerShell" and click on it. I get the following error messages: Warning: Failed to load the 'SQLAS' extension: An exception occurred in SMO while trying to manage a service. 
    --> Failed to retrieve data for this request. --> Invalid class.  Warning: Could not obtain SQL Server Service information. An attemp to connect to WMI on 'NAB-WK-02657306' failed with the following error: An exception occurred in SMO while trying
    to manage a service. --> Failed to retrieve data for this request. --> Invalid class.  .... PS SQLSERVER:\SQL\NAB-WK-02657306\SQLEXPRESS\Databases\APGriMMRP\Table_1_XYcoordinates>   What causes this set of error messages? How can
    I get this problem fixed in my PC that is an end user of the Windows 7 LAN System? Note: I don't have the regular version of Microsoft Visual Studio 2012 in my PC. I just have the Microsoft 2012 Shell (Integrated) program in my PC.
    Problem #3: I plan to create an XML Schema Collection in the "APGriMMRP" database for the Parent Table and the Child Tables. How can I handle the relationship between the Parent Table and the Child Table in the XML Schema Collection?
    Problem #4: I plan to extract some results/data from the Parent Table and the Child Table by using XQuery. What kind of JOIN (Left or Right JOIN) should I use in the XQuerying?
    Please kindly help, answer my questions, and advise me how to resolve these 4 problems.
    Thanks in advance,
    Scott Chang    

    In the future, I would recommend you to post your questions one by one, and to the appropriate forum. Of your questions it is really only #3 that fits into this forum. (And that is the one I will not answer, because I have worked very little with XSD.)
    1) Not sure what you mean with "print" or "export", but when you right-click a database, you can select Tasks from the context menu and in this submenu you find "Export data".
    2) I don't know why you get that error, but any particular reason you want to run PowerShell?
    4) If you have tables, you query them with SQL, not XQuery. XQuery is when you query XML documents, but left and right joins are SQL things. There are no joins in XQuery.
    As for left/right join, notice that these two are equivalent:
    SELECT ...
    FROM   a LEFT JOIN b ON a.col = b.col
    SELECT ...
    FROM   b RIGHT JOIN a ON a.col = b.col
    But please never use RIGHT JOIN - it gives me a headache!
    There is nothing that says that you should use any of the other. In fact, if you are returning rows from parent and child, I would expect an inner join, unless you want to cater for parents without children.
    Here is an example where you can study the different join types and how they behave:
    CREATE TABLE apple (a int         NOT NULL PRIMARY KEY,
                        b varchar(23) NOT NULL)
    INSERT apple(a, b)
       VALUES(1, 'Granny Smith'),
             (2, 'Gloster'),
             (4, 'Ingrid-Marie'),
             (5, 'Milenga')
    CREATE TABLE orange(c int        NOT NULL PRIMARY KEY,
                        d varchar(23) NOT NULL)
    INSERT orange(c, d)
       VALUES(1, 'Agent'),
             (3, 'Netherlands'),
             (4, 'Revolution')
    SELECT a, b, c, d
    FROM   apple
    CROSS  JOIN orange
    SELECT a, b, c, d
    FROM   apple
    INNER  JOIN orange ON apple.a = orange.c
    SELECT a, b, c, d
    FROM   apple
    LEFT   OUTER JOIN orange ON apple.a = orange.c
    SELECT a, b, c, d
    FROM   apple
    RIGHT  OUTER JOIN orange ON apple.a = orange.c
    SELECT a, b, c, d
    FROM   apple
    FULL OUTER JOIN orange ON apple.a = orange.c
    go
    DROP TABLE apple, orange
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Assigning Role to a Task and Resource to a Task

    Hello Friends,
    In cProjects, I go to the below tab Projects > Resources > Tasks.
    Here I do an assignment of Role to a Task and assignment of Resource to a Task.
    Could you please tell me as to where does this information gets stored.(ie which tables).
    And could you pleae tell me as to how do we differentiate between the Task, Roles and Resources in the internal table created.
    Because the internal table consists of Task Guid, Participant Guid and Entity Guid.
    Thanks,
    Salil

    I can tell you the table where the the role assignments get stored . It is
    DPR_ENTITY_LINK.

  • ODI-1228: Task PRD-create-populate-table (Procedure) fails on the target

    I get this error when trying to run PRD-create-populate-table from Oracle by example ODI
    ODI-1228: Task PRD-create-populate-table (Procedure) fails on the target ORACLE connection ODI_DATA_SERVER.
    Caused By: java.sql.SQLSyntaxErrorException: ORA-02264: name already used by an existing constraint
    ODI-1228: Task PRD-create-populate-table (Procedure) fails on the target ORACLE connection ODI_DATA_SERVER.
    Caused By: java.sql.SQLSyntaxErrorException: ORA-02264: name already used by an existing constraint
         at oracle.jdbc.driver.SQLStateMapping.newSQLException(SQLStateMapping.java:91)
         at oracle.jdbc.driver.DatabaseError.newSQLException(DatabaseError.java:133)
         at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:206)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:455)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:413)
         at oracle.jdbc.driver.T4C8Oall.receive(T4C8Oall.java:1035)
         at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:194)
         at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:953)
         at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1224)
         at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3386)
         at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3467)
         at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeUpdate(OraclePreparedStatementWrapper.java:1350)
         at com.sunopsis.sql.SnpsQuery.executeUpdate(SnpsQuery.java:665)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.executeUpdate(SnpSessTaskSql.java:3218)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execStdOrders(SnpSessTaskSql.java:1785)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java:2805)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2515)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:534)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:449)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1954)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:322)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:224)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:246)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:237)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:794)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:114)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
         at java.lang.Thread.run(Thread.java:662)

    the specified constraints name is still used. Please check if you are having the order correctly ie drop table , create and insert and check if the table is dropped , if still the constraints is there then drop it from database

  • Parallel Script Tasks Using Same Connection Manager are Corrupted

    We're seeing a bizarre issues in our SSIS package.  We are using an HTTP connection manager in multiple script tasks running in parallel and the connections seem to be getting mismatched.  Here's the basic flow.
    We have 6 containers that run in parallel.  Each container does the exact same work (they are copy/pasted) except we change the value of a variable representing the report to generate that is scoped to each container.  The container runs a query
    (using a SQL connection manager defined in the package designer) and then uses a foreach container to enumerate the rows.  The variables inside the foreach container are scoped to it so we don't have a scoping issue.  Within the loop we write
    the variable values to a log table, use a script task to generate an SSRS report and then write all the data and the report to an external system.  Everything is working fine except for the generation of the report.
    Within the report script block we acquire a connection to a HTTP connection manager that is defined in the package (just like the SQL connection manager).  We then create a new HttpClientConnection with it and update the URL to include the necessary
    parameters that SSRS needs to generate the report.  We then download the results and return it back to the container.  Here's the basic code we use
    var httpConn = Dts.Connections["ReportServerUrl"];
    var clientConn = new HttpClientConnection(httpConn.AcquireConnection(null));
    var uri = new UriBuilder(clientConn.ServerURL) { Query = reportQueryInfo };
    clientConn.ServerURL = uri.Uri.ToString();
    clientConn.DownloadData();
    The issue we are seeing is that when 2 of the scripts run at the same time the ServerURLs are getting cross wired such that we see 2 of the same report being requested rather than 1 of each type that we had requested (based upon the SSRS execution logs). 
    It's almost like we are getting a shared HTTP client connection.  Provided the scripts don't run at the same time everything works correctly.  We are confident our variables are set correctly because we are storing them both before and after in external
    tables and they are always correct.  The URL calculation logic is also correct because this problem only occurs when 2 requests are sent in parallel.
    What am I missing here?
    Michael Taylor
    http://blogs.msmvps.com/p3net

    Thank you for the feedback.  Before I go to Connect I'd like to provide some more info that may provide a better idea of what the problem is.
    We broke things down trying to replicate the issue and it isn't a conn mgr problem after all I believe.  To clarify our package, we have 4 containers that execute the same logic (they were copy/pasted) except for the variable values that we set. The
    containers are not linked together so they run in parallel. Each container executes a query using the ADO.NET conn mgr.  A foreach loop captures the value for the current row and then calls SSRS using a script block, forwards the report onto an external
    resource and then updates a table indicating the row has been processed.  The only difference between them is the parameter we pass to the query and the value of the variables (that are scoped to the container or foreach loop).
    The problem we have found is that the very first time through the foreach loop ALL containers are using the same row (from the first container's query).  Second and subsequent rows are fine.  We initially thought the queries were fine because of
    our update at the end of the foreach loop but it is now clear that isn't the case.
    We saw this problem when we moved our package to production.  We went back to Dev and was able to replicate it as well.  It happens every time and it is always the same row when given the same input set.   We started eliminating containers
    and we found it only happens when you get to 3 containers or more.  2 containers had no problems.
    I'll try to replicate it on a simpler solution but any thoughts on whether there is a race condition or something in SSIS would be helpful.

  • What is the better way to open and close connection in bean

    Hello, may i ask about the best way to open and close the connection?
    At the first, the connection code is store in Conn class bean. The JSP page will call a method in Process bean to process the query and update of database. the Process bean is using th Conn to connect to database.
    In my process bean, it have a method for each type of sql like select, insert, update, and delete for each tables. Each method will call the Conn to open and close the connection. The jsp page can simply call a method retrieve or update records.
    However, when calling the processUpdate method, it is a method that update a group of tables base on the query result of another group of tables and calculation. If I calling the query method in the same bean file, it will open and close the connection many times (more than 1000 times). After open and close the connection several times, the connection to mySQL will be fail.
    Currently, i include 2 method in Process that is only calling the Conn to open the connection and close connection. When calling the openConnection method, it will also set the boolean connected in the class to true and call Conn to open connection. When calling the closeConnection method, it will also set the boolean connected in the class to false and call Conn to close connection.
    After that, i add some change to the query method which is when the connected is false, it will open the connection by calling the Conn and close the connection themself. If the connected is true, it will not call the Conn to open and close the connection because it already connected to the database.
    Now, i when i calling the query method in jsp, it can open the connection and close the connection individually. When i call the processUpdate in the jsp, this method will calling the openConnection before calling the query and calling closeConnection at the end. It not only faster the process(less open and close), it also no cause the connect to mySQL fail error again. However, it also make the coding in the bean file more longer and complex.
    Is there have any better solution?

    Use connection pooling. This will put the connection management in the hands of the webserver, not your code. Check out this post:
    http://forum.java.sun.com/thread.jspa?threadID=741788&messageID=4252932#4252932

  • PHP and MySQL Connection problem

    I am trying to make a PHP MySQL on a remote server connection in Dreamwesaver CS3.
    When I enter the information (host, user, password, etc.) and hit TEST, I get the following error message.
    "Access Denied. The file may not exist, or there could be a permission problem. Make sure you have proper authorization on the server and the server is properly configured."
    I know the user, password, etc. work, as I can access the MySQL database with other software.
    I checked with my host and they say there are no permission problems on their end.
    I have checked the settings on the Test Server page in Site Setup and everything looks correct and works there.
    I have not seen this particular problem described in other forum postings (although there are a LOT of postings on this topic!).
    Any help would be appreciated.

    I thought my testing server was the remote server and is always on as far as I know. I don't know how to turn it on or off.
    Is there some other testing server that I should be aware of?
    Frank
    Date: Wed, 3 Jun 2009 15:43:02 -0600
    From: [email protected]
    To: [email protected]
    Subject: PHP and MySQL Connection problem
    I know you are using remote, but is your testing server on? if not turn that on and see if that does it. it just happened to me working on remote server and could not get mysql conn to work, until I turn on my testing (developer server), then I was able to connect to mysql and the tables that I created in phpmyadmin remotly was downloaded.
    >

  • Performance with MySQL and Database connectivity toolbox

    Hi!
    I'm having quite some problems with the performance of MySQL and Database connectivity toolbox. However, I'm very happy with the ease of using database connectivity toolbox. The background is:
    I have 61 variables (ints and floats) which I would like to save in the MySQL-database. This is no problem, however, the loop time increases from 8ms to 50ms when using the database. I have concluded that it has to do with the DB Tools Insert Data.vi and I think that I have some kind of performance issue with this VI. The CPU never reach more the 15% of its maximum performance. I use a default setup and connect through ODBC.
    My questions are:
    1. I would like to save 61 variables each 8-10ms, is this impossible using this solution?
    2. Is there any way of increasing the performance of the DB Tools Insert Data.vi or use any other VI?
    3. Is there any way of adjusting the MySQL setup to achieve better performance?
    Thank you very much for your time.
    Regards,
    Mattias

    First of all, thank you very much for your time. All of you have been really good support to me.
    >> Is your database on a different computer?  Does your loop execute 61 times? 
    Database is on the same computer as the MySQL server.
    The loop saves 61 values at once to the database, in one SQL-statement.
    I have now added the front panel and block diagram for my test-VI. I have implemented the queue system and separate loops for producer and consumer. However, since the queue is building up faster then the consumer loop consumes values, the queue is building up quite fast and the disc starts working.
    The test database table that I add data to is created by a simple:
    create table test(aa int, bb char(15));
    ...I'm sure that this can be improved in some way.
    I always open and close the connection to the database "outside the loop". However, it still takes some 40-50 ms to save the data to the database table - so, unfortunatly no progress to far. I currently just want to save the data.
    Any more advise will be gratefully accepted.
    Regards,
    Mattias
    Message Edited by mattias@hv on 10-23-2007 07:50 AM
    Attachments:
    front panel 2.JPG ‏101 KB
    block diagram.JPG ‏135 KB

  • Email Notifications for UWL Tasks and Collaboration Tasks

    Hi,
    I want to configure Email Notifications for both UWL Tasks and Collaboration Tasks.
    I done with the all the Configurations suggested in the below link:
    [https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/706d054d-da24-2b10-f18a-fc82faf6468e]
    But Email Notifications are not reaching the users Lotus Mail ID. Could any one please suggest me that i am missing any Configurations.
    Thanks in Advance.
    Regards,
    Sridhar.

    Hi Sridhar,
    I created a Wiki page with usefull links to troubleshoot problems with the Email connectivity from KM:
    https://wiki.sdn.sap.com/wiki/x/g4KmAw
    Please see if it helps,
    Robert

  • Is there ONE single app available for TASK and CALENDAR both which can be synched with Google account for OSX, iPhone, iPad

    Hi guys,
    I have been searching to death to find ONE great app which:
    1- handles both calendar (events) & To-Do list (task).
    2- connects to googla account (both calendar & task)
    3- should be availabe in Mac OS (OSX), iPhone & iPad (iOS) & Android.
    For instance, Pocket Informant, Calengoo or BusyCal.
    All of them takes care of the first, but fail in at least one of the remaining features.
    Pocket informant:
    1- Yes
    2- Calender: Yes, Task: No
    3- No Mac OS X version but it gives to web-based app (15$).
    Calengoo:
    1- Yes
    2- Yes
    3- No Mac OS X version.
    BusyCal:
    1- Yes
    2- Calender: Yes, Task: No
    3- No Portable Version. (iOS & Android).
    There is no one or combined solution for this!!! If there is no single software to take care of both, please give me a solution  but keep in your mind that ONE app is very important to me on a single device not two because I know already that!
    At least
    HELP me plz!!

    Hi,
    here is a schematic that I *Think* should work, but I haven't bought BusyCal 2 yet.  Reminders syncs with Toodledo, which syncs with Pocket Informant (I've tested this, and can now create PI tasks by talking to Siri); Reminders also syncs w/ BusyCal 2, so Busycal *should* be able to create, edit and delete tasks that will appear on Pocket Informant, and the reverse *should* be true.  Likewise, iCal syncs w/ both BusyCal 2 and Pocket Informant's appointments.  Thus BusyCal becomes your desktop hub, and everything *should* be sync'd up with PI on your iOS device.  I don't know anything about Android. 
      Siri —> "Reminders" <—> "Toodledo" <—> "Pocket Informant" tasks
                        I
                  "BusyCal 2" (desktop)
                        I
                     "iCal" <—> "Pocket Informant" appointments

  • How to export and import dependent tables from 2 different schema

    I have a setup where schema1 has table 1 and schema 2 has table2. And table 1 from schema1 depends upon table 2 from schema 2.
    I would like to export and import these tables only and not any other tables from these 2 schemas with all information like grants,constraints.
    Also will there be same method for Oracle 10g R1,R2 and Oracle 11G.
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10825/dp_export.htm#i1007514
    Looking at this For table mode it says
    Also, as in schema exports, cross-schema references are not exported
    Not sure what this means.
    As I am interested in only 2 tables I think I need to use table mode. But if I try to run export with both tables names, it says table mode support only one schema at a time. Not sure then How would the constraints would get exported in that case.
    -Rohit

    worked for my 1st time I tried
    exp file=table2.dmp tables="dbadmin.temp1,scott.emp"
    Export: Release 10.2.0.1.0 - Production on Mon Mar 1 16:32:07 2010
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Username: / as sysdba
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P1 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    Current user changed to DBADMIN
    . . exporting table                          TEMP1         10 rows exported
    EXP-00091: Exporting questionable statistics.
    Current user changed to SCOTT
    . . exporting table                            EMP         14 rows exported
    EXP-00091: Exporting questionable statistics.
    EXP-00091: Exporting questionable statistics.
    Export terminated successfully with warnings.

  • Unable to retrieve clouds on VMM server. Please make sure the server name is correct and try connecting again

    I getting this annoying error "Unable to retrieve clouds on VMM server. Please make sure the server name is correct and try connecting again."  I've followed the recommended documentation for setting up SPF and Azure using the following:
    http://www.hyper-v.nu/archives/mvaneijk/2013/01/installing-and-configuring-system-center-service-provider-foundation/
    I've gone thru the troubleshoot steps from the following:
    http://blogs.technet.com/b/scvmm/archive/2014/03/04/kb-the-windows-azure-pack-service-management-portal-does-not-retrieve-cloud-settings.aspx
    http://social.technet.microsoft.com/forums/windowsazure/en-US/363e7dae-7370-4605-bc8d-b98fa6dc9d8e/unable-to-register-vmm-server
    So you don't need to send me those links.
    I found a few others which pointed me to the event logs WindowsAzurePack\MGmtSVC-AdminAPi and AdminSite.  In these event logs, I see a lot of event ID 12 and 1046.  The contents of event id 12 is the following:
    Log Name:      Microsoft-WindowsAzurePack-MgmtSvc-AdminSite/Operational
    Source:        Microsoft-WindowsAzurePack-MgmtSvc-AdminSite
    Date:          6/19/2014 3:27:55 PM
    Event ID:      12
    Task Category: (65522)
    Level:         Error
    Keywords:      None
    User:          IIS APPPOOL\MgmtSvc-AdminSite
    Computer:      usw-azure-01.uswired.com
    Description:
    Error:ManagementClientException: Resource not found.
    <Exception>
      <Type>ManagementClientException</Type>
      <Message>Resource not found.</Message>
      <StackTrace><![CDATA[
       at Microsoft.WindowsAzure.Server.Management.ManagementClientBase.<ThrowIfResponseNotSuccessful>d__2b.MoveNext()
       at Microsoft.WindowsAzure.Management.TaskSequencer.<>c__DisplayClass1e`1.<RunSequenceAsync>b__1d(Task previousTask)
    --- End of stack trace from previous location where exception was thrown ---
       at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
       at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
       at Microsoft.WindowsAzure.Server.SystemCenter.AdminExtension.Controllers.SystemCenterAdminController.<GetResourceProviderAsync>d__2e7.MoveNext()
    --- End of stack trace from previous location where exception was thrown ---
       at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
       at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
       at Microsoft.WindowsAzure.Server.SystemCenter.AdminExtension.Controllers.SystemCenterAdminController.<GetAutomationResourceProvider>d__2b6.MoveNext()
    --- End of stack trace from previous location where exception was thrown ---
       at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
       at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
       at lambda_method(Closure , Task )
       at System.Web.Mvc.Async.TaskAsyncActionDescriptor.EndExecute(IAsyncResult asyncResult)
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass3f.<BeginInvokeAsynchronousActionMethod>b__3e(IAsyncResult asyncResult)
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass37.<>c__DisplayClass39.<BeginInvokeActionMethodWithFilters>b__33()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass4f.<InvokeActionMethodFilterAsynchronously>b__49()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass4f.<InvokeActionMethodFilterAsynchronously>b__49()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass4f.<InvokeActionMethodFilterAsynchronously>b__49()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass37.<BeginInvokeActionMethodWithFilters>b__36(IAsyncResult asyncResult)
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass25.<>c__DisplayClass2a.<BeginInvokeAction>b__20()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.<>c__DisplayClass25.<BeginInvokeAction>b__22(IAsyncResult asyncResult)]]></StackTrace>
      <HttpContext>
        <User IsAuthenticated="true" Name="USWIRED\Administrator" />
        <Request>
          <RawUrl>/SystemCenterAdmin/GetAutomationResourceProvider</RawUrl>
          <UserHostAddress>fe80::c179:19c8:c5f2:7309%12</UserHostAddress>
          <Headers>
            <Header Key="Cache-Control" Value="no-cache" />
            <Header Key="Connection" Value="Keep-Alive" />
            <Header Key="Content-Length" Value="2" />
            <Header Key="Content-Type" Value="application/json" />
            <Header Key="Accept" Value="application/json, text/javascript, */*; q=0.01" />
            <Header Key="Accept-Encoding" Value="gzip, deflate" />
            <Header Key="Accept-Language" Value="en-US" />
            <Header Key="Host" Value="usw-azure-01.uswired.com:30091" />
            <Header Key="Referer" Value="https://usw-azure-01.uswired.com:30091/" />
            <Header Key="User-Agent" Value="Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko" />
            <Header Key="x-ms-client-session-id" Value="c343c29d-20b1-412b-be94-879409205700" />
            <Header Key="x-ms-client-antiforgery-id" Value="0crwi2qNk20bMr6C04/vKd4L28WG1nd8JKBdOlCCp+xlph40YeL2oTdmeiTkPmE272S2dqmGQMNo+FElKkG2MdR0Xhq41RU/t8PrFjcG0Vgx4a8dgZLbfKVURo9n84F1XWtJOkTTUK/OeNVWBrK1+w=="
    />
            <Header Key="x-ms-client-request-id" Value="dcc427c7-40fc-4fdd-9a70-7622969ff40d-2014-06-19 22:27:55Z" />
            <Header Key="X-Requested-With" Value="XMLHttpRequest" />
            <Header Key="DNT" Value="1" />
            <Cookies>
              <Cookie Name="AdminSiteFedAuth" Secure="false" Expires="0001-01-01T00:00:00Z" Domain="" Path="/" Value="Redacted (2000 characters)" />
              <Cookie Name="AdminSiteFedAuth1" Secure="false" Expires="0001-01-01T00:00:00Z" Domain="" Path="/" Value="Redacted (424 characters)" />
              <Cookie Name="__RequestVerificationToken_Admin" Secure="false" Expires="0001-01-01T00:00:00Z" Domain="" Path="/" Value="Redacted (128 characters)" />
            </Cookies>
          </Headers>
        </Request>
      </HttpContext>
    </Exception>, operationName:SystemCenterAdmin.GetAutomationResourceProvider, version:, accept language:en-US, subscription Id:, client request Id:dcc427c7-40fc-4fdd-9a70-7622969ff40d-2014-06-19 22:27:55Z, principal Id:USWIRED\Administrator, page request
    Id:c343c29d-20b1-412b-be94-879409205700, server request id:
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="Microsoft-WindowsAzurePack-MgmtSvc-AdminSite" Guid="{5E78D550-1384-5A96-C12A-CB6DA7BC6365}" />
        <EventID>12</EventID>
        <Version>0</Version>
        <Level>2</Level>
        <Task>65522</Task>
        <Opcode>0</Opcode>
        <Keywords>0x0</Keywords>
        <TimeCreated SystemTime="2014-06-19T22:27:55.971935000Z" />
        <EventRecordID>399</EventRecordID>
        <Correlation ActivityID="{FAA512AF-E0A9-4F2A-80EF-09F72D2F7918}" />
        <Execution ProcessID="3424" ThreadID="2308" />
        <Channel>Microsoft-WindowsAzurePack-MgmtSvc-AdminSite/Operational</Channel>
        <Computer>usw-azure-01.uswired.com</Computer>
        <Security UserID="S-1-5-82-4118156790-2181488624-806685255-2797011695-3958995103" />
      </System>
      <EventData>
        <Data Name="message">ManagementClientException: Resource not found.
    &lt;Exception&gt;
      &lt;Type&gt;ManagementClientException&lt;/Type&gt;
      &lt;Message&gt;Resource not found.&lt;/Message&gt;
      &lt;StackTrace&gt;&lt;![CDATA[
       at Microsoft.WindowsAzure.Server.Management.ManagementClientBase.&lt;ThrowIfResponseNotSuccessful&gt;d__2b.MoveNext()
       at Microsoft.WindowsAzure.Management.TaskSequencer.&lt;&gt;c__DisplayClass1e`1.&lt;RunSequenceAsync&gt;b__1d(Task previousTask)
    --- End of stack trace from previous location where exception was thrown ---
       at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
       at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
       at Microsoft.WindowsAzure.Server.SystemCenter.AdminExtension.Controllers.SystemCenterAdminController.&lt;GetResourceProviderAsync&gt;d__2e7.MoveNext()
    --- End of stack trace from previous location where exception was thrown ---
       at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
       at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
       at Microsoft.WindowsAzure.Server.SystemCenter.AdminExtension.Controllers.SystemCenterAdminController.&lt;GetAutomationResourceProvider&gt;d__2b6.MoveNext()
    --- End of stack trace from previous location where exception was thrown ---
       at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
       at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
       at lambda_method(Closure , Task )
       at System.Web.Mvc.Async.TaskAsyncActionDescriptor.EndExecute(IAsyncResult asyncResult)
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass3f.&lt;BeginInvokeAsynchronousActionMethod&gt;b__3e(IAsyncResult asyncResult)
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass37.&lt;&gt;c__DisplayClass39.&lt;BeginInvokeActionMethodWithFilters&gt;b__33()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass4f.&lt;InvokeActionMethodFilterAsynchronously&gt;b__49()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass4f.&lt;InvokeActionMethodFilterAsynchronously&gt;b__49()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass4f.&lt;InvokeActionMethodFilterAsynchronously&gt;b__49()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass37.&lt;BeginInvokeActionMethodWithFilters&gt;b__36(IAsyncResult asyncResult)
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass25.&lt;&gt;c__DisplayClass2a.&lt;BeginInvokeAction&gt;b__20()
       at System.Web.Mvc.Async.AsyncControllerActionInvoker.&lt;&gt;c__DisplayClass25.&lt;BeginInvokeAction&gt;b__22(IAsyncResult asyncResult)]]&gt;&lt;/StackTrace&gt;
      &lt;HttpContext&gt;
        &lt;User IsAuthenticated="true" Name="USWIRED\Administrator" /&gt;
        &lt;Request&gt;
          &lt;RawUrl&gt;/SystemCenterAdmin/GetAutomationResourceProvider&lt;/RawUrl&gt;
          &lt;UserHostAddress&gt;fe80::c179:19c8:c5f2:7309%12&lt;/UserHostAddress&gt;
          &lt;Headers&gt;
            &lt;Header Key="Cache-Control" Value="no-cache" /&gt;
            &lt;Header Key="Connection" Value="Keep-Alive" /&gt;
            &lt;Header Key="Content-Length" Value="2" /&gt;
            &lt;Header Key="Content-Type" Value="application/json" /&gt;
            &lt;Header Key="Accept" Value="application/json, text/javascript, */*; q=0.01" /&gt;
            &lt;Header Key="Accept-Encoding" Value="gzip, deflate" /&gt;
            &lt;Header Key="Accept-Language" Value="en-US" /&gt;
            &lt;Header Key="Host" Value="usw-azure-01.uswired.com:30091" /&gt;
            &lt;Header Key="Referer" Value="https://usw-azure-01.uswired.com:30091/" /&gt;
            &lt;Header Key="User-Agent" Value="Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko" /&gt;
            &lt;Header Key="x-ms-client-session-id" Value="c343c29d-20b1-412b-be94-879409205700" /&gt;
            &lt;Header Key="x-ms-client-antiforgery-id" Value="0crwi2qNk20bMr6C04/vKd4L28WG1nd8JKBdOlCCp+xlph40YeL2oTdmeiTkPmE272S2dqmGQMNo+FElKkG2MdR0Xhq41RU/t8PrFjcG0Vgx4a8dgZLbfKVURo9n84F1XWtJOkTTUK/OeNVWBrK1+w=="
    /&gt;
            &lt;Header Key="x-ms-client-request-id" Value="dcc427c7-40fc-4fdd-9a70-7622969ff40d-2014-06-19 22:27:55Z" /&gt;
            &lt;Header Key="X-Requested-With" Value="XMLHttpRequest" /&gt;
            &lt;Header Key="DNT" Value="1" /&gt;
            &lt;Cookies&gt;
              &lt;Cookie Name="AdminSiteFedAuth" Secure="false" Expires="0001-01-01T00:00:00Z" Domain="" Path="/" Value="Redacted (2000 characters)" /&gt;
              &lt;Cookie Name="AdminSiteFedAuth1" Secure="false" Expires="0001-01-01T00:00:00Z" Domain="" Path="/" Value="Redacted (424 characters)" /&gt;
              &lt;Cookie Name="__RequestVerificationToken_Admin" Secure="false" Expires="0001-01-01T00:00:00Z" Domain="" Path="/" Value="Redacted (128 characters)" /&gt;
            &lt;/Cookies&gt;
          &lt;/Headers&gt;
        &lt;/Request&gt;
      &lt;/HttpContext&gt;
    &lt;/Exception&gt;</Data>
        <Data Name="requestId">
        </Data>
        <Data Name="subscriptionId">
        </Data>
        <Data Name="clientRequestId">dcc427c7-40fc-4fdd-9a70-7622969ff40d-2014-06-19 22:27:55Z</Data>
        <Data Name="principalId">USWIRED\Administrator</Data>
        <Data Name="version">
        </Data>
        <Data Name="pageRequestId">c343c29d-20b1-412b-be94-879409205700</Data>
        <Data Name="acceptLanguage">en-US</Data>
        <Data Name="operationName">SystemCenterAdmin.GetAutomationResourceProvider</Data>
      </EventData>
    </Event>
    Event ID 1046 is 
    Log Name:      Microsoft-WindowsAzurePack-MgmtSvc-AdminSite/Operational
    Source:        Microsoft-WindowsAzurePack-MgmtSvc-AdminSite
    Date:          6/19/2014 3:32:56 PM
    Event ID:      1046
    Task Category: (116)
    Level:         Error
    Keywords:      None
    User:          IIS APPPOOL\MgmtSvc-AdminSite
    Computer:      usw-azure-01.uswired.com
    Description:
    Could not connect to Resource Provider. Error: Resource not found.
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="Microsoft-WindowsAzurePack-MgmtSvc-AdminSite" Guid="{5E78D550-1384-5A96-C12A-CB6DA7BC6365}" />
        <EventID>1046</EventID>
        <Version>0</Version>
        <Level>2</Level>
        <Task>116</Task>
        <Opcode>0</Opcode>
        <Keywords>0x0</Keywords>
        <TimeCreated SystemTime="2014-06-19T22:32:56.854137500Z" />
        <EventRecordID>413</EventRecordID>
        <Correlation ActivityID="{C20DBEAD-388F-4CF4-80FD-21C6419CC2F1}" />
        <Execution ProcessID="3424" ThreadID="2768" />
        <Channel>Microsoft-WindowsAzurePack-MgmtSvc-AdminSite/Operational</Channel>
        <Computer>usw-azure-01.uswired.com</Computer>
        <Security UserID="S-1-5-82-4118156790-2181488624-806685255-2797011695-3958995103" />
      </System>
      <EventData>
        <Data Name="errorMessage">Resource not found.</Data>
      </EventData>
    </Event>
    The other thing I notice is that when trying to connect to the VMM. It's trying to use domain\username, yet when I try logging in with domain\username via the SCVMM Console. I get the  You cannot access VMM management server (Error ID: 1604).
     I'm not sure if this is why I can not add the VMM to WAP.  Any help would be appreciated.  Thanks
    Nick

    Hello,
    I am not sure about the answer but I guess these steps will solve your problem:
    1) Log in to the SPF server by using an account that is a member of the SPF_Admin local group.
    2) Open an elevated Windows PowerShell window.
    3) Type the following commands, and press Enter after you type each command:
    Import-module spfadmin
    New-SCSPFTenant -Name Administrator -SubscriptionId 00000000-0000-0000-0000-000000000000
    4) Close the WAP Server Management Portal, and then reconnect to it to update all items.

  • JMQ cluster and unstable connections

    Hello all.
    I have a few architectural questions about building an OpenMQ message-passing infrastructure between multiple offices which do not always have on-line internet connections. We also need to distribute the MQ mesh configuration info.
    From the scale of my questions it seems, that I or our developers don't fully understand MQ, because I think that many of our problems and/or solution ideas (below) should already be implemented within the MQ middleware, and not by us from outside it.
    The potential client currently has a (relatively ugly) working solution which they wanted to revise for simplification, if possible, but this matter is not urgent and answers are welcome at any timeframe :)
    I'd welcome any insights, ideas and pointers as to why our described approach may be plain wrong :)
    To sum this post up, here's my short questionnaire:
    1) What is a good/best way to distribute MQ mesh config when not all nodes are available simultaneously?
    2) What are the limitations on number of brokers and queues in one logical mesh?
    3) Should we aim for separate "internal" and "external" MQ networks, or can they be combined into one large net?
    4) Should we aim for partial solution external to OpenMQ (such as integration with SMTP for messaging, or SVN for config distribution), or can this quest be solved within OpenMQ functionality?
    5) Can a clustered broker be forced to fully start without available master broker connection?
    6) Are broker clusters inherently local-network, or is there some standard solution (pattern) for geographically disperse MQ clusters?
    7) How to enforce pushing of the messages from one broker to another? Are any priority assignments available for certain brokers and "their" queues?
    Detailed rumblings follow below...
    We are thinking about implementing JMQ in a geographically disperse project, where it will be used for asynchronous communications to connect application servers in different branch offices with a central office. The problematic part is, that the central and especially branch offices are not expected to be always on-line, hence the MQ - whenever a connection is available, queued messages (requests, responses, etc.) are to be pushed to the other side's MQ broker. And if all goes well with the project, there may eventually be hundreds of such branch offices and more than one central office for failover, and a mesh of interconnection MQ agreements.
    The basic idea is simple: an end-user of the app server in a branch generates a request, this request is passed via message queue to another branch or to a central office, then another app server processes it to generate a response and the answer is queued back to the requesting app server. At some time after the initial request, the end-user would see in his web-page that the request's status has been updated with a response value. A branch office's app server and MQ broker may be an appliance-server distributed as a relatively unmaintained "black box".
    During the POC we configured several JMQ broker instances in this manner and it worked. From what I gather from our developers, each branch office's request and response queues are separate destinations in the system, and requests (from a certain branch) may be subscribed by any node, and responces (to a certain branch) may be submitted by any node. This may be restricted by passwords and/or certificate-based SSL tunnel channels, for example (suggestions welcome, though).
    However, we also wanted to simplify spreading the configuration of the MQ nodes' network by designating "master brokers" (as per JMQ docs) which keep track of the config and each other broker downloads the cluster config from its master. Perhaps it was wrong on our side, and a better idea is available to avoid manual reconfiguration of each MQ broker whenever another broker or a queue destination is added?
    Problem here is: it seems an "MQ cluster" is a local-network oriented concept. When we have a master broker in a central office, and the inter-connection is not up, branch offices loop indefinitely waiting for connection to a master, and reject client connections (published JMS port remains 0, and appropriate comments in the log files). In this case the branch office can not function until its JMQ broker connects to a central office, updates the MQ config, and permits client connections to itself.
    Also we are not certain (and it seems to be a popular question on Google, too) how to enforce a queued message to be pushed to another side - to a broker "nearest" to the target app server? Can this be done within OpenMQ config, or does this require an MQ client application to read and manipulate such messages somehow? For example, when a branch office's "request" queue has a message, and a connection to central office comes online, this request data should end up in the central office's broker. Apparently, a message which physically remains in the branch office broker when the interconnection goes offline, is of little use to the central appserver...
    I was thinking along the lines of different-priority brokers for a certain destinations, so that messages would automatically flow from further brokers to neares ones - like water flows from higher ground to lower ground in an aqueduct. It would then be possible to easily implement transparent routing between branch offices (available at non-intersecting times) via central office (always up).
    How many brokers and destination can be interconnected at all (practically or theoretically/hardcoded)?
    Possibly, there are other means to do some or all of this?
    Ideas we've discussed internally include:
    * Multiple networks of MQ brokers:
    Have an "internal" broker (cluster) in each branch office which talks to the app server, and a separate "external" broker which is clustered with the central office's "master broker". Some branch office application transfers messages between two brokers local to its branch. Thus the local appserver works okay, and remote queuing works whenever network is available.
    Possibly, the central office should also have separate internal and external broker setups?
    * Multi-tiered net of MQ brokers:
    Perhaps there can be "clusters of clusters" - with "external" tier-1 brokers being directly master brokers for local "internal" tier-2 clusters? Otherwise the idea of "miltiple networks of MQ brokers" above, without an extra app to relay messages between MQ brokers local to this app.
    * Multi-protocol implementation of MQ+SMTP(+POP3/IMAP)
    Many of our questions are solvable by SMTP. That is, we can send messages to a mailbox residing on a specific server (local in each office), and local appserver clients retrieve them by POP3 from the local mailbox server, and then submit responses over SMTP. This is approximately how the client currently solves this task now.
    We don't really want to invent a bicycle, but maybe this approach can also be applied to JMQ (asynch traffic not over MQ protocol, but over SMTP like in SOAP-SMTP vs. SOAP-HTTP webservices)?
    * HTTP/RCS-based config file:
    The OpenMQ config allows for the detailed configuration file to be available in local filesystem or on a web server. It is possible to fetch the config file from a central office whenever the connection is up (wget, svn/cvs/etc.) and restart the branch broker.
    Why is this approach good or bad? Advocates welcome :)
    Thanks for reading up to the end,
    and thanks in advance for any replies,
    //Jim Klimov

    Hello all.
    I have a few architectural questions about building an OpenMQ message-passing infrastructure between multiple offices which do not always have on-line internet connections. We also need to distribute the MQ mesh configuration info.
    From the scale of my questions it seems, that I or our developers don't fully understand MQ, because I think that many of our problems and/or solution ideas (below) should already be implemented within the MQ middleware, and not by us from outside it.
    The potential client currently has a (relatively ugly) working solution which they wanted to revise for simplification, if possible, but this matter is not urgent and answers are welcome at any timeframe :)
    I'd welcome any insights, ideas and pointers as to why our described approach may be plain wrong :)
    To sum this post up, here's my short questionnaire:
    1) What is a good/best way to distribute MQ mesh config when not all nodes are available simultaneously?
    2) What are the limitations on number of brokers and queues in one logical mesh?
    3) Should we aim for separate "internal" and "external" MQ networks, or can they be combined into one large net?
    4) Should we aim for partial solution external to OpenMQ (such as integration with SMTP for messaging, or SVN for config distribution), or can this quest be solved within OpenMQ functionality?
    5) Can a clustered broker be forced to fully start without available master broker connection?
    6) Are broker clusters inherently local-network, or is there some standard solution (pattern) for geographically disperse MQ clusters?
    7) How to enforce pushing of the messages from one broker to another? Are any priority assignments available for certain brokers and "their" queues?
    Detailed rumblings follow below...
    We are thinking about implementing JMQ in a geographically disperse project, where it will be used for asynchronous communications to connect application servers in different branch offices with a central office. The problematic part is, that the central and especially branch offices are not expected to be always on-line, hence the MQ - whenever a connection is available, queued messages (requests, responses, etc.) are to be pushed to the other side's MQ broker. And if all goes well with the project, there may eventually be hundreds of such branch offices and more than one central office for failover, and a mesh of interconnection MQ agreements.
    The basic idea is simple: an end-user of the app server in a branch generates a request, this request is passed via message queue to another branch or to a central office, then another app server processes it to generate a response and the answer is queued back to the requesting app server. At some time after the initial request, the end-user would see in his web-page that the request's status has been updated with a response value. A branch office's app server and MQ broker may be an appliance-server distributed as a relatively unmaintained "black box".
    During the POC we configured several JMQ broker instances in this manner and it worked. From what I gather from our developers, each branch office's request and response queues are separate destinations in the system, and requests (from a certain branch) may be subscribed by any node, and responces (to a certain branch) may be submitted by any node. This may be restricted by passwords and/or certificate-based SSL tunnel channels, for example (suggestions welcome, though).
    However, we also wanted to simplify spreading the configuration of the MQ nodes' network by designating "master brokers" (as per JMQ docs) which keep track of the config and each other broker downloads the cluster config from its master. Perhaps it was wrong on our side, and a better idea is available to avoid manual reconfiguration of each MQ broker whenever another broker or a queue destination is added?
    Problem here is: it seems an "MQ cluster" is a local-network oriented concept. When we have a master broker in a central office, and the inter-connection is not up, branch offices loop indefinitely waiting for connection to a master, and reject client connections (published JMS port remains 0, and appropriate comments in the log files). In this case the branch office can not function until its JMQ broker connects to a central office, updates the MQ config, and permits client connections to itself.
    Also we are not certain (and it seems to be a popular question on Google, too) how to enforce a queued message to be pushed to another side - to a broker "nearest" to the target app server? Can this be done within OpenMQ config, or does this require an MQ client application to read and manipulate such messages somehow? For example, when a branch office's "request" queue has a message, and a connection to central office comes online, this request data should end up in the central office's broker. Apparently, a message which physically remains in the branch office broker when the interconnection goes offline, is of little use to the central appserver...
    I was thinking along the lines of different-priority brokers for a certain destinations, so that messages would automatically flow from further brokers to neares ones - like water flows from higher ground to lower ground in an aqueduct. It would then be possible to easily implement transparent routing between branch offices (available at non-intersecting times) via central office (always up).
    How many brokers and destination can be interconnected at all (practically or theoretically/hardcoded)?
    Possibly, there are other means to do some or all of this?
    Ideas we've discussed internally include:
    * Multiple networks of MQ brokers:
    Have an "internal" broker (cluster) in each branch office which talks to the app server, and a separate "external" broker which is clustered with the central office's "master broker". Some branch office application transfers messages between two brokers local to its branch. Thus the local appserver works okay, and remote queuing works whenever network is available.
    Possibly, the central office should also have separate internal and external broker setups?
    * Multi-tiered net of MQ brokers:
    Perhaps there can be "clusters of clusters" - with "external" tier-1 brokers being directly master brokers for local "internal" tier-2 clusters? Otherwise the idea of "miltiple networks of MQ brokers" above, without an extra app to relay messages between MQ brokers local to this app.
    * Multi-protocol implementation of MQ+SMTP(+POP3/IMAP)
    Many of our questions are solvable by SMTP. That is, we can send messages to a mailbox residing on a specific server (local in each office), and local appserver clients retrieve them by POP3 from the local mailbox server, and then submit responses over SMTP. This is approximately how the client currently solves this task now.
    We don't really want to invent a bicycle, but maybe this approach can also be applied to JMQ (asynch traffic not over MQ protocol, but over SMTP like in SOAP-SMTP vs. SOAP-HTTP webservices)?
    * HTTP/RCS-based config file:
    The OpenMQ config allows for the detailed configuration file to be available in local filesystem or on a web server. It is possible to fetch the config file from a central office whenever the connection is up (wget, svn/cvs/etc.) and restart the branch broker.
    Why is this approach good or bad? Advocates welcome :)
    Thanks for reading up to the end,
    and thanks in advance for any replies,
    //Jim Klimov

Maybe you are looking for

  • How can I move music from an iPod classic to my computer?

    I need to know how to transfer my 1800 songs in my iTunes on my iPod classic to my computer.

  • MacBook Pro terribly slow unknown reason. Please help.

    Even while not running any app at all,  (and with 20%+ free disk-space, and nothing noticeable shown in Activiy Monitor). Disk utility neither finds anything, but the computer is taking about 5min to open a window, and if I open an App, my MacBook fr

  • How to debug a BAPI?

    Hi ABAPERs, i wanted to know that how you can debug a BAPI. My problem is i m working on one utility(report) in which BAPI is called to pull data from other system. I am exporting header level data and importing item data, reference data and selectio

  • Printer Recommendation Needed

    My husband needs a new low-end desktop Mac, and I have a MacBook. We share a printer connected to his desktop now, but he is unhappy with his inkjet printer and wants something better with his new system. Any recommendations for a laser - black and w

  • Cube creation in OBIEE

    Hi experts I am new to cube concepts. So please tell me the steps to create cube in OBIEE. Is cube is created on database level(Tables) or we can create at Olap level. Thanks in advance Regards frnd