Automated Duplication for Role Playing Dimensions

Im wondering if there is way to automate (short cut method) to duplicate the entire Dimension.
Simply copying and pasting or Duplicating will not create the Hierarchies.
For Example:
I have Fact_Transaction (with columns Trx_Id, Trx_Date, Create_Date, Update_Date, Invoice_Date, Amount)
I have a date dimension joined to this fact in physical layer on the Trx_Date.
Problem:
Now if I get a requirement to show the other dates 2 as dimension,
I need to create alias for the same time dim and manually create the dimensions and hierarchies in the BMM layer.
Is there any much better way we can do this using some scripts (xudml, /commands, ..etc)
Thanks
Kaushik

Hi,
I think you have to create it manually...

Similar Messages

  • Many to many relationship does not work on certain role played dimension

    Hi,
    I have a cube with role playing Account dimension, used once as “Account” and once as “Other Account” dimension.
    I’ve added Trip dimension related to the account dimension by a bridge table AccountTrip 
    making a many to many relationship (many accounts can be in one trip and one account can go to many trips).
    When I added the bridge measure group to the cube, I saw in the dimension usage tab that the relationship between the bridge and the Account dimension was done automatically.
     I added the trip dimension with regular relationship to the bridge measure group and many to many relationship to the main fact table using the bridge group.
    All compiled fine and gave correct results. Then I noticed that when slicing the fact table by the Other account dimension I get correct results,
     but when I add the Trip dimension I get the results as if I was slicing by the Account dimension, not by the “Other account” dimension.
    I checked the cube and noticed I did not set the relationship between the bridge to the “Other Account” dimension (role playing). I set it now (regular relationship, same as the account dimension), but still getting the same results.
    Conclusion-  when slicing by the new Trip (M2M) dimension and the “Other Account”
     (role played) dimension I get the results of the “Account” dimension, not those of the “Other Account” dimension.
    I checked the relationships of the “Other Account” dimension in the cube but it looks correctly set (to the external account of the main fact table and to account of the bridge table).
    (Just to note I have two other bridges on the cube which are not related and don’t look like they need to be related, plus two other measure groups of the main fact table used for distinct count which are related appropriately).
    What else should be done???
    I would greatly appreciate your help!
    Thanks
    Namnami

    The update server is down; try this temporary workaround
    App Store>Purchased>Select "All"
    Note: Look out for apps that have the word "Update"
    http://i1224.photobucket.com/albums/ee374/Diavonex/9c256282736869f322d4b3071bbb2 a82_zps51a6f546.jpg

  • Distinct count of role-played dimension attribute

    Needed distinct count of AccountGroup attribute of AccountB dimension which is a role played dimension of Account.
    Added measure distinct count of the dimension attribute (AccountGroup). In dimension usage added the main fact table as intermediate for other dimensions with many2many relationships.
    The two role played Account dimensions must be related to the new AccountFact table with fact relationship.
    But then how will I be able to count distinct attribute of
    certain Account dimension (out of the two role played dimensions)???
    Namnami

    Thanks for the links, they are useful. But still they do not explain how to manage a distinct count of
    certain role played dimension attribute. 
    I gave up on this and added that dimension attribute to the fact table so now I do a regular distinct count on a cube fact measure. 
    Thanks
    Namnami

  • Bridge table to Role Playing Dimension

    Hi,
    Fact Sales table has role playing Customer dimension table related to it. For each role it had a relationship in the DSV. One on CustomerA the other on CustomerB.
    Now I have a trip dimension which should have a many to many relationship to the customers table. So I set a bridge table relating trip to customer.
    But when I slice by trip it only displays the sales that have CustomerA in the trip.
    I want it to display the sales that have CustomerB in the trip.
    How can I affect that?
    Thank you
    Namnami

    When you create a M2M relationship, you specify 2 dimensions and one intermediate measure Group.
    Even if the bridge table is loaded once in the DSV (even if you can just duplicate it to simplify its naming), you define two measure groups for the bridge table, one connected to Customer A and Trip A, and the other to Customer B and Trip B. Then, you define
    M2M relationship between Trip A and Customer A using bridgeA as intermediate, and do the same for B.
    Again, please, since M2M require understanding how it works internally, I suggest you to read the White paper I mentioned, because you have all the information there that will allow you to design the model you need. I think it's not a discussion that is
    easy to complete in a forum thread, because M2M involves many side consequences in calculation (e.g. what happen to unrelated dimensions involved in calculation?).
    Marco Russo (Blog,
    Twitter,
    LinkedIn) - sqlbi.com:
    Articles, Videos,
    Tools, Consultancy,
    Training
    Format with DAX Formatter and design with
    DAX Patterns. Learn
    Power Pivot and SSAS Tabular.

  • How to create a role-playing Date dimension?

    I have created a date dimension Date_Dim with three levels DayL, MonthL, YearL and one hierarchy YearL-Monthy-DayL. In the fact table, I need to refer to two date dimension: Start_Date and End_Date. For the Start_Date, we can refer it to the created physical date dimension Date_Dim. For the End_date, Ralph Kimball recommended to use a role-playing dimension. That is, create a view End_Date_Vw from the dimension table Date_Dim. My question is how to create it in OWB? Should we need to define the levels and hierarchies for the view in order to leverage the query-rewrite power? How to do that? Any advice or information is appreciated.
    Lushu

    Balaji,
    You can refer to the same dimension more than once in the design of a cube (fact table). In a star schema implementation, you just end up with multiple foreign keys to the same times dimension (table). Once you integrate with a reporting tool, you refer to the same dimension using different names. Warehouse Builder supports this in the integration with Oracle Discoverer. The user's guide tells you how to do this (see http://download-west.oracle.com/docs/html/B12146_01/integrating.htm#sthref3659).
    Thanks,
    Mark.

  • Difference between Conformed dimnesion and Role play dimesion

    hi,
    I am not able to understand difference between Conformed dimension and Role play dimension.. Can anyone please explain the diff between both with real time example.
    I am confused with Date Dimension.
    Here i encountered, Cube dimension and Database dimension,what is that actually.
    Thanks

    Hi DevBi,
    According to your description, you want to know the difference between database dimension and cube dimension. Right?
    A database dimension is a collection of related objects, called attributes, which can be used to provide information about fact data in one or more cubes. What we created in BIDS is database dimension. For example, typical attributes in a product dimension
    might be product name, product category, product line, product size, and product price. These objects are bound to one or more columns in one or more tables in a data source view.
    Cube dimension is an instance of a database dimension in a cube is called a cube dimension and relates to one or more measure groups in the cube. A database dimension can be used multiple times in a cube. Those referenced and renamed for purpose database dimension
    is role-playing dimension. For example, a fact table can have multiple time-related facts, and a separate cube dimension can be defined to assist in analyzing each time-related fact. However, only one time-related database dimension needs to exist, which
    also means that only one time-related relational database table needs to exist to support multiple cube dimensions based on time.
    Reference:
    Role-playing Dimensions
    If you have any question, please feel free to ask.
    Simon Hou
    TechNet Community Support

  • CUP 5.3: Automated Provisioning for UME Roles

    Hi,
    Does CUP 5.3 provide automated provisioning for UME roles or just for R/3 roles?
    Thanks in advance. Best regards,
       Imanol

    In order to use UME provisioning though, the WebAS must have Portal components installed (not necessarily used). The Portal RTA uses the Portal's SPML interface that is installed with portal components.
    And please make sure to check the PAM - I think the Portal RTA only works for 7.00, if I'm not mistaken!
    Frank.

  • Why is the Star Transformation using two indexes for the same dimension?

    Hi,
    Recently, I have made an investigation about the Star Transformation feature. I have found a strange test case, which plays an important role in my strategy for our overall DWH architecture. Here it is:
    The Strategy:
    I would like to have the classical Star Transformation approach (single column Bitmap Indexes for each dimension foreign key column in the fact table), together with additional Bitmap Join Indexes for some of the dimension attributes, which would benefit from the materialization of the join (bitmap merge operation will be skipped/optimized).
    The query:
    select dp.brand, ds. region_name, dc.region_name
         , count(*), sum(f.extended_price)
      from fact_line_item  f
         , dim_part       dp
         , dim_supplier   ds
         , dim_customer   dc
    where dp.mfgr        = 10                 -- dimension selectivity = 1/10 --> acttual/fact selectivity = 6/10
       and f.part_dk      = dp.dk
       and ds.region_name = 'REGION #1' -- dimension selectivity = 1/9
       and f.supplier_dk  = ds.dk
       and dc.region_name = 'REGION #1' -- dimension selectivity = 1/11
       and f.customer_dk  = dc.dk
    group by dp.brand, ds. region_name, dc.region_name
    The actual plan:
    | Id  | Operation                              | Name                        | Starts | E-Rows | Cost (%CPU)| A-Rows |   A-Time   | Buffers | Reads  |
    |   0 | SELECT STATEMENT                       |                             |      1 |        |  3247 (100)|      1 |00:01:42.05 |     264K|    220K|
    |   1 |  HASH GROUP BY                         |                             |      1 |      2 |  3247   (1)|      1 |00:01:42.05 |     264K|    220K|
    |*  2 |   HASH JOIN                            |                             |      1 |  33242 |  3037   (1)|    217K|00:01:29.67 |     264K|    220K|
    |*  3 |    TABLE ACCESS FULL                   | DIM_SUPPLIER                |      1 |   1112 |   102   (0)|   1112 |00:00:00.01 |     316 |      4 |
    |*  4 |    HASH JOIN                           |                             |      1 |  33245 |  2934   (1)|    217K|00:01:29.10 |     264K|    220K|
    |*  5 |     TABLE ACCESS FULL                  | DIM_CUSTOMER                |      1 |    910 |   102   (0)|    910 |00:00:00.08 |     316 |      8 |
    |*  6 |     HASH JOIN                          |                             |      1 |  33248 |  2831   (1)|    217K|00:01:28.57 |     264K|    220K|
    |*  7 |      TABLE ACCESS FULL                 | DIM_PART                    |      1 |     10 |     3   (0)|     10 |00:00:00.01 |       6 |      0 |
    |   8 |      PARTITION RANGE ALL               |                             |      1 |  36211 |  2827   (1)|    217K|00:01:28.01 |     264K|    220K|
    |   9 |       TABLE ACCESS BY LOCAL INDEX ROWID| FACT_LINE_ITEM              |      6 |  36211 |  2827   (1)|    217K|00:01:33.85 |     264K|    220K|
    |  10 |        BITMAP CONVERSION TO ROWIDS     |                             |      6 |        |            |    217K|00:00:07.09 |   46980 |   3292 |
    |  11 |         BITMAP AND                     |                             |      6 |        |            |     69 |00:00:08.33 |   46980 |   3292 |
    |  12 |          BITMAP MERGE                  |                             |      6 |        |            |    193 |00:00:02.09 |    2408 |   1795 |
    |  13 |           BITMAP KEY ITERATION         |                             |      6 |        |            |   4330 |00:00:04.66 |    2408 |   1795 |
    |  14 |            BUFFER SORT                 |                             |      6 |        |            |     60 |00:00:00.01 |       6 |      0 |
    |* 15 |             TABLE ACCESS FULL          | DIM_PART                    |      1 |     10 |     3   (0)|     10 |00:00:00.01 |       6 |      0 |
    |* 16 |            BITMAP INDEX RANGE SCAN     | FACT_LI__P_PART_DIM_KEY_BIX |     60 |        |            |   4330 |00:00:02.11 |    2402 |   1795 |
    |* 17 |          BITMAP INDEX SINGLE VALUE     | FACT_LI__P_PART_MFGR_BJX    |      6 |        |            |   1747 |00:00:06.65 |     890 |    888 |
    |  18 |          BITMAP MERGE                  |                             |      6 |        |            |    169 |00:00:02.78 |   16695 |    237 |
    |  19 |           BITMAP KEY ITERATION         |                             |      6 |        |            |   5460 |00:00:01.56 |   16695 |    237 |
    |  20 |            BUFFER SORT                 |                             |      6 |        |            |   5460 |00:00:00.02 |     316 |      0 |
    |* 21 |             TABLE ACCESS FULL          | DIM_CUSTOMER                |      1 |    910 |   102   (0)|    910 |00:00:00.01 |     316 |      0 |
    |* 22 |            BITMAP INDEX RANGE SCAN     | FACT_LI__P_CUST_DIM_KEY_BIX |   5460 |        |            |   5460 |00:00:02.07 |   16379 |    237 |
    |  23 |          BITMAP MERGE                  |                             |      6 |        |            |    170 |00:00:03.65 |   26987 |    372 |
    |  24 |           BITMAP KEY ITERATION         |                             |      6 |        |            |   6672 |00:00:02.23 |   26987 |    372 |
    |  25 |            BUFFER SORT                 |                             |      6 |        |            |   6672 |00:00:00.01 |     316 |      0 |
    |* 26 |             TABLE ACCESS FULL          | DIM_SUPPLIER                |      1 |   1112 |   102   (0)|   1112 |00:00:00.01 |     316 |      0 |
    |* 27 |            BITMAP INDEX RANGE SCAN     | FACT_LI__S_SUPP_DIM_KEY_BIX |   6672 |        |            |   6672 |00:00:02.74 |   26671 |    372 |
    The Question:
    Why is the Star Transformation using both indexes FACT_LI__P_PART_DIM_KEY_BIX and FACT_LI__P_PART_MFGR_BJX for the same dimension criteria (dp.mfgr = 10)?? The introduction of the additional Bitmap Join Index actually make Oracle to do the work twice !!!
    Anybody, any idea ?!?

    Dom, here it is the plan with the predicates:
    | Id  | Operation                              | Name                        | Starts | E-Rows | Cost (%CPU)| A-Rows |   A-Time   | Buffers | Reads  |
    |   0 | SELECT STATEMENT                       |                             |      1 |        |  3638 (100)|      1 |00:06:41.17 |     445K|    236K|
    |   1 |  HASH GROUP BY                         |                             |      1 |      2 |  3638   (1)|      1 |00:06:41.17 |     445K|    236K|
    |*  2 |   HASH JOIN                            |                             |      1 |  33242 |  3429   (1)|    217K|00:08:18.02 |     445K|    236K|
    |*  3 |    TABLE ACCESS FULL                   | DIM_SUPPLIER                |      1 |   1112 |   102   (0)|   1112 |00:00:00.03 |     319 |    313 |
    |*  4 |    HASH JOIN                           |                             |      1 |  33245 |  3326   (1)|    217K|00:08:17.47 |     445K|    236K|
    |*  5 |     TABLE ACCESS FULL                  | DIM_CUSTOMER                |      1 |    910 |   102   (0)|    910 |00:00:00.01 |     319 |    313 |
    |*  6 |     HASH JOIN                          |                             |      1 |  33248 |  3223   (1)|    217K|00:08:16.63 |     445K|    236K|
    |*  7 |      TABLE ACCESS FULL                 | DIM_PART                    |      1 |     10 |     3   (0)|     10 |00:00:00.01 |       6 |      0 |
    |   8 |      PARTITION RANGE ALL               |                             |      1 |  36211 |  3219   (1)|    217K|00:08:16.30 |     445K|    236K|
    |   9 |       TABLE ACCESS BY LOCAL INDEX ROWID| FACT_LINE_ITEM              |      6 |  36211 |  3219   (1)|    217K|00:08:40.89 |     445K|    236K|
    |  10 |        BITMAP CONVERSION TO ROWIDS     |                             |      6 |        |            |    217K|00:00:32.00 |   46919 |  19331 |
    |  11 |         BITMAP AND                     |                             |      6 |        |            |     69 |00:00:34.50 |   46919 |  19331 |
    |  12 |          BITMAP MERGE                  |                             |      6 |        |            |    193 |00:00:00.58 |    2353 |      1 |
    |  13 |           BITMAP KEY ITERATION         |                             |      6 |        |            |   4330 |00:00:00.10 |    2353 |      1 |
    |  14 |            BUFFER SORT                 |                             |      6 |        |            |     60 |00:00:00.01 |       6 |      0 |
    |* 15 |             TABLE ACCESS FULL          | DIM_PART                    |      1 |     10 |     3   (0)|     10 |00:00:00.01 |       6 |      0 |
    |* 16 |            BITMAP INDEX RANGE SCAN     | FACT_LI__P_PART_DIM_KEY_BIX |     60 |        |            |   4330 |00:00:00.07 |    2347 |      1 |
    |* 17 |          BITMAP INDEX SINGLE VALUE     | FACT_LI__P_PART_MFGR_BJX    |      6 |        |            |   1747 |00:01:23.64 |     882 |    565 |
    |  18 |          BITMAP MERGE                  |                             |      6 |        |            |    169 |00:00:09.14 |   16697 |   7628 |
    |  19 |           BITMAP KEY ITERATION         |                             |      6 |        |            |   5460 |00:00:02.19 |   16697 |   7628 |
    |  20 |            BUFFER SORT                 |                             |      6 |        |            |   5460 |00:00:00.01 |     316 |      0 |
    |* 21 |             TABLE ACCESS FULL          | DIM_CUSTOMER                |      1 |    910 |   102   (0)|    910 |00:00:00.01 |     316 |      0 |
    |* 22 |            BITMAP INDEX RANGE SCAN     | FACT_LI__P_CUST_DIM_KEY_BIX |   5460 |        |            |   5460 |00:00:08.78 |   16381 |   7628 |
    |  23 |          BITMAP MERGE                  |                             |      6 |        |            |    170 |00:00:21.46 |   26987 |  11137 |
    |  24 |           BITMAP KEY ITERATION         |                             |      6 |        |            |   6672 |00:00:10.29 |   26987 |  11137 |
    |  25 |            BUFFER SORT                 |                             |      6 |        |            |   6672 |00:00:00.01 |     316 |      0 |
    |* 26 |             TABLE ACCESS FULL          | DIM_SUPPLIER                |      1 |   1112 |   102   (0)|   1112 |00:00:00.01 |     316 |      0 |
    |* 27 |            BITMAP INDEX RANGE SCAN     | FACT_LI__S_SUPP_DIM_KEY_BIX |   6672 |        |            |   6672 |00:00:20.94 |   26671 |  11137 |
    Predicate Information (identified by operation id):                                                                                                  
       2 - access("F"."SUPPLIER_DK"="DS"."DK")                                                                                                           
       3 - filter("DS"."REGION_NAME"='REGION #1')                                                                                                        
       4 - access("F"."CUSTOMER_DK"="DC"."DK")                                                                                                           
       5 - filter("DC"."REGION_NAME"='REGION #1')                                                                                                        
       6 - access("F"."PART_DK"="DP"."DK")                                                                                                               
       7 - filter("DP"."MFGR"=10)                                                                                                                        
      15 - filter("DP"."MFGR"=10)                                                                                                                        
      16 - access("F"."PART_DK"="DP"."DK")                                                                                                               
      17 - access("F"."SYS_NC00017$"=10)                                                                                                                 
      21 - filter("DC"."REGION_NAME"='REGION #1')                                                                                                        
      22 - access("F"."CUSTOMER_DK"="DC"."DK")                                                                                                           
      26 - filter("DS"."REGION_NAME"='REGION #1')                                                                                                        
      27 - access("F"."SUPPLIER_DK"="DS"."DK")                                                                                                           
    Note                                                                                                                                                 
       - star transformation used for this statement                                                                                                     

  • How to prevent BPC from automated calculating for hierarchies / nodes?

    Hi experts,
    I am looking for a practicable way to prevent the system from automated calculating for hierarchies, and especially for nodes within hierarchies?
    Let's say, I have ENTITIES (AS A HIERACHY) in rows and ACCOUNTS in COLUMNS. Now I want the system to block adding up the values for one specific account on node XY. Instead of the sum of all base member entities, the cell for account XY should be left empty.
    Is there a practicable way to deal with this?

    Hi Stefan,
    you can prevent the system to calculate a node by editing the Formula Property  in the Dimension.
    You can insert 0 or null to the Formula field for the specific element.
    But in this case the values from your leaf elements lying under your node wont be calculated  even in higher hierarchy nodes.
    For ex.: with the following structure
    -A
    ---A1
    A11
    A12
    ---B1
    B11
    B12
    If you set A1 to 0, the top node A will only be calculated by values of B1
    Regards
    Jörg

  • Ignoring role playing relationship

    As is the case in most cubes, I have a role playing Date dimension.  The dimensions are [Date Created] and [Date Snapshot].  [Date Created] is snowflaked off the Ticket dimension using the DateCreatedKey.  [Date Snapshot] is related to my
    fact table using the DateSnapshotKey (the day the snapshot of the fact was taken).
    With this schema, I can find the number of tickets that were created in January 2014.
    SELECT
    [Number of Tickets] ON COLUMNS,
    [Date Created].[Month Year].&[201401] ON ROWS
    FROM
    [MyCube]
    I can also find the number of tickets that were in an Open status in January 2014
    SELECT
    [Number of Tickets] ON COLUMNS,
    [Date Snapshot].[Month Year].&[201401] ON ROWS
    FROM
    (SELECT [Ticket Status].[Ticket Status].&[1]
    ON COLUMNS FROM [MyCube])
    Now what I need to do is the get the number of open tickets in January 2014 (regardless of when they were created) and divide that by the number of new tickets in January 2014.  Here's what I tried
    WITH
    MEMBER OpenTickets AS
    ([Number of Tickets]
    , [Date Snapshot].[Month Year].&[201401])
    SELECT
    {[Number of Tickets], OpenTickets]}
    ON COLUMNS,
    [Date Created].[Month Year].&[201401] ON ROWS
    FROM
    [MyCube]
    However, what I get is the number of tickets created in January 2014 and the number of open tickets that were also created in January 2014.  All other open tickets are ignored.  How do I break that connection between dates so that the open tickets
    doesn't look at the date created?
    Stephanie

    In ssas, you can use the Dimension Usage tab in Cube Designer to view and edit the dimension relationships between cube dimensions and measure groups in the cube. If you break that connection between dates, you can to to the Dimension Usage
    tab to set it.
    http://msdn.microsoft.com/en-us/library/ms183558.aspx

  • Initiater for Role removal.

    Hi,
       I need some update/input w.r.t Role removal Initiator. While configuring the role removal is it possible to use the role status in the initiator?    If not how to identify this role is only for the role removal.
    Normally we use to put only one stage for Role removal. In the config, no where we are having automatic check for the request is only for the Role removal. So we have to trust that particular stage owners. As per the CUP automation check is it possible to validate this?
    Thanks in advance.
    Regards,
    Vasantha Kumar.

    Hi Justin
    I'm assuming you are involved in or victim of a security access review. I'm usually one of those security guys asking for role or transaction removal and you are the main contact in the business coordinating the changes.
    The process of remediation will possibly consist of checking which transactions are causing segregation of duties conflict, if they are used or not and removing one side of the conflict by removing an unused transaction.
    It shouldn't require the entire contents of a role to be removed - rather swapping role A for role B without a transaction or two.
    Removing transactions that aren't used can have more subtle implications which hopefully are found during UAT but is usually missed until used in anger. This what support is for after go live.
    Saying all that and depending on your time and skills, you could ask for access to the security person's test user in dev or qas where they are working to run transaction SUIM on transaction for user following the proposed changes and compare that to the actual access of the real affected user in prod. If you can get access to the informer tab in virsa you can use the standard simulation reports to also check the resulting conflicts which will help you talk to the business and advise on actions available. There should be role owners involved in all this as they have to owner the result: expect a request for these for CUP later on
    If you can retain control and approval of the (controlled) changes being made to users you will have a better understanding of what is happening, catch potential errors and mediate between security and the business - you have an important task!
    Ask for some basic training in standard SAP reports - the security team should be more than grateful for your input
    Crikey that was hard typing on an iPhone!
    Cheers
    Edited by: David Berry on Jan 11, 2011 8:17 PM

  • Would like feedback on role play simulation created in Captivate 3

    I created this simulation in Captivate 3. It is based on a soft skill role play simulation that was initially created for the Harvard Business School using Javascript.
    I was wondering if anyone out there is doing anything similar. I'd like to share ideas with them.
    http://progressive1.acs.playstream.com/ADPNAS//progressive/Walter_Test/Harvard%20Scenario. swf

    I just tried it. The look and feel of the project are wonderful.
    I noticed that many (most) of the choices weren't hot, that is to say a hand cursor didn't display when I put the cursor over them. On many slided I could only choose one option. Is the program still in mock-up? I  viewed it in the Chrome browser.
    I choose the "wrong" answer on a slide and liked the way it let me try again. But would the program be more effective in teaching if the feedback on the wrong answer slide gave information on how the answer is deficient? In this example the Director said I was being defensive and that is feedback on the choice I made. So I know the answer is "wrong" but not what specifically is wrong about it. Something like "Your response focused on (fill in  blank A) while the strategy we are teaching focuses on (fill in blank B). Go back and see what other response does a better job of directing her attention to (blank B).
    Similarly, the Directors response to correct answers tells me that I got the answer correct, but not why. Some sort of audio feedback (I'm inventing a narrartor here) who says something like "That was an effective response. Your response focuses on (B) which is a key aspect of our strategy.
    For what its worth, thanks I enjoyed the simulation.
    Doug

  • Error while updating a custom Windows Azure Diagnostics configuration xml from powershell. "Invalid update to extension reference for role"

    I am attempting to upload a manually edited WADConfig xml to my VM. The WAD service is functioning correctly, I needed to add some custom WinEventLogs. The prescribed steps result in an error.
    What am I overlooking?
    I am following these instructions:
    Step 5: Remotely install Diagnostics on your Azure Virtual Machine
    azure.microsoft.com/en-in/documentation/articles/cloud-services-dotnet-diagnostics/#virtual-machine
    $storage_name = "wadexamplevm"
    $key = "<StorageAccountKey>"
    $config_path="c:\users\<user>\documents\visual studio 2013\Projects\WadExampleVM\WadExampleVM\WadExample.xml"
    $service_name="wadexamplevm"
    $vm_name="WadExample"
    $storageContext = New-AzureStorageContext
    -StorageAccountName $storage_name -StorageAccountKey $key
    $VM1 = Get-AzureVM
    -ServiceName $service_name -Name $vm_name
    $VM2 = Set-AzureVMDiagnosticsExtension
    -DiagnosticsConfigurationPath $config_path
    -Version "1.*"
    -VM $VM1 -StorageContext $storageContext
    $VM3 = Update-AzureVM
    -ServiceName $service_name -Name $vm_name
    -VM $VM2.VM
    Unfortunately, I am receiving this error:
    Update-AzureVM : BadRequest: Invalid update to extension reference for role: XXXXXX and reference: IaaSDiagnostics.
    What's missing from the above script?

    Hi,
    Since Azure SDK 2.5 uses the extension model the diagnostics extension, the configuration and the connection string to the diagnostic storage are no longer part of the deployment package and cscfg. All the diagnostics configuration is contained within the
    wadcfgx. The advantage with this approach is that diagnostics agent and settings are decoupled from the project and can be dynamically enabled and updated even after your application is deployed. 
    Due to this change some existing workflows need to be rethought – instead of configuring the diagnostics as part of the application that gets deployed to each environment you can first deploy the application to the environment and then apply the diagnostics
    configuration for it.  When you publish the application from Visual Studio this process is done automatically for you. However if you were deploying your application outside of VS using PowerShell then you have to install the extension separately through
    PowerShell.
    There PowerShell cmdlets for managing the diagnostics extensions on a Cloud Service are -
    Set-AzureServiceDiagnosticsExtension
    Get-AzureServiceDiagnosticsExtension
    Remove-AzureServiceDiagnosticsExtension
    You can use the Set-AzureServiceDiagnosticsExtension method to enable diagnostics extension on a cloud service. One of the parameters on this cmdlet is the XML configuration file. This file is slightly different from the diagnostics.wadcfgx file. You can
    create this file from scratch by either following the article that you are referring to or  you can modify the wadcfgx file and pass in the modified file as a parameter to the powershell cmdlet.
    To modify the wadcfgx file –
    Make a copy the .wadcfgx.
    Remove the following elements from the Copy:
    <DiagnosticsConfiguration xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration">
       <PrivateConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration">
         <StorageAccount name=" " endpoint="https://core.windows.net/" />
       </PrivateConfig>
       <IsEnabled>false</IsEnabled>
    </DiagnosticsConfiguration>
    Make sure the top of the file still has xml version and encoding –
       <?xml version="1.0" encoding="utf-8"?>
    Effectively you are stripping down the Wadcfgx to only contain the <PublicConfig> section and the <?xml> header. You can then call the PowerShell cmdlet along with the appropriate parameters for the staging slots and roles:
    $storage_name = ‘
    <storagename>’
    $key= ‘<key>’
    $service_name = '<servicename>'
    $public_config = '<thepublicconfigfrom_diagnostics.wadcfgx>'
    $storageContext = New-AzureStorageContext –StorageAccountName $storage_name –StorageAccountKey $key
    Set-AzureServiceDiagnosticsExtension -StorageContext $storageContext -DiagnosticsConfigurationPath $public_config –ServiceName $service_name -Slot ‘Staging’ -Role ‘WebRole1’
    Hope this helps !
    Regards,
    Sowmya

  • How do I setup multiple users to use one source but manage playlists separately?  For example, play counts are specific to each user.

    How do I setup multiple users to use one source but manage playlists separately?  For example, play counts are specific to each user.

    Hi,
    Thank you for posting in Windows Server Forum.
    Here adding to the words of “Tim”, a forwarder is a DNS server on a network used to forward DNS queries for external DNS names to DNS servers outside of that network. You can also forward queries according to specific domain names using conditional forwarders.
    A DNS server on a network is designated as a forwarder by having the other DNS servers in the network forward the queries they cannot resolve locally to that DNS server. You can refer information regarding forwarders and how to configure from beneath link.
    Understanding forwarders
    http://technet.microsoft.com/en-us/library/cc782142(v=ws.10).aspx
    Configure a DNS Server to Use Forwarders
    http://technet.microsoft.com/en-us/library/cc754941.aspx
    Hope it helps!
    Regards.

  • Hi! How do I switch existing automation data for one parameter to another?

    Hello! I am new to Logic 8, and actually to using the Logic automation feature. Here is the deal: I have some old-school Logic event data that I recorded as a kind of automation (before the current kind of dedicated automation) for an EXS24 parameter. I need to change this event 'automation' to actual Logic automation, as Logic 8 does not read it (as Logic 7 did before it). So, I highlighted all of the events in my old-school 'automation' region in the event list. I had to change the number value to a common number assignment so that I could Change Selected Event Data into Track Automation. I selected volume (number seven, I believe), and than highlighted all events, clicked on the event region, than went to options and selected my desired automation transformation destination (say that three times fast). Great! I have my region as automation! Logic 8 knows what to do with this! But, it is volume. I need it to control the LFO 2 Rate of the EXS24. Problem is, that when I switch the automation parameter from Volume to LFO 2 Rate, the automation disappears. It seems as though that automation data is only for Volume. So, how might a charming gentleman, such as myself, go about keeping/converting the same automation, but for the parameter I want to change it to?
    Thanks! =)

    Hold "command" when switching between automation parameters.
    Then you will get a dialog offering you the option to convert the existing parameter's data into the destination parameter, or "copy and convert" (which does the same but leaves the existing volume automation data intact.)

Maybe you are looking for

  • Generate PDF file from sql query

    I have a table having some now i want to perform some calculation on tht data and then want to export it into a  pdf file. All this thing thing i want to write in a sql query so tht i can put it in a stored procedure . Kindly suggest.

  • Can I use one Apple TV unit to distribute content to two TV's?

    Can I use one Apple TV unit to distribute content to two TV's? If so, how? I want to be able to purchase movies off of the iTunes store on both TV's and to display photos and listen to music. Thank you for looking.

  • My ipod touch 1st generation doesnt wanna charge and its not recognized on iTunes!

    Ok so when i try to charge my ipod touch 1st gen 8 gigs wit the usb it doesnt want too even wit the outlet charger it doesnt want too. its not even recognized by itunes either. i have a ipod classic 120 gigs that one does charge by the outlet and usb

  • Having a column name as a variable within a query,Problem!!!!

    I have a problem with a simple query shown below SELECT * FROM DisneyWHERE Upper(COLNAME) LIKE UPPER('%' || SEARCHSTRING || '%');SELECT * FROM Disney WHERE Upper(COLNAME) LIKE UPPER('%' || SEARCHSTRING || '%'); My problem, The colname variable is not

  • How to get data from BW Cube to Another File system?

    Hi, I have a requirement to get the sales data from BW system(infoCube) to another file system as XML file using XI as a middleware. Can anybody give the suggestions on this please? Thanks & Regards, Satish