MEASURING WITH HP 54600 SCOPE

Hello
I am new user of labview and I am trying to make a program that measures the output of a simple circuit (ie LP filter) and shows it on the labview display window in time domain and to modify it to show the magnitude as a function of frequency.I managed to have the right sinusoidal curve on the oscilloscope (hp54600B) display but I only get a ramp on my labview window
I am using version 7.1 and using a GPIB and the hp 33120A waveform generator
i appreciate if any one can take a look at the program code and advice how to change it.
Thanks in advance
Teddy555
Attachments:
HP54600B-MEAS.vi ‏159 KB

Don't have to shout.
Did you also download the driver for the 33120A? I would recomend you do that as well. One of the functions is ag33xxx - config Waveform. This has a frequency input control. You can put the function gen and scope VIs inside a loop and increment the frequency with each iteration of the loop. There are several ways to do that. You can create an array of values and with each iteration, you index the next frequency value out of the array. You can also do it with a calculation. You can provide a start value, an increment, and a final value. You would wire your start value to a shift register and with each iteration, add the step size, and terminate the while loop when the value in the shift register is greater or equal to the final value.

Similar Messages

  • NI 5124, max points measured with 256MB memory

    I want to know if there is a simple calculation that i can do to know
    the maximum amount of points that i can measure with the 5124 scope.
    With My application i want to measure a burst of 520 mS of signal at
    ones, the signal is downconverted from 406MHz  with the 5600. So
    the signal is at a minimum frequency of 5MHz. This burst i want t save
    completely and still be able to retrieve offline the following features:
    - Frequency
    - Power
    - IQ
    - Decode the message.
    For this i need a high enough sample rate (minimum niquist). So as
    absolute minimum i need 520mS signal, but it would be nice to have at
    least 1 second of signal..
    is there an ellegant sollution for this?
    Regards
    Joost van Heijenoort

    So I've implemented the change and it seems to work a bit better. I don't know if there's a drastic speed increase, but the intervals between stage motion (the only indication that an acquisition is done in the auto acquire mode) seem more uniform and on the quicker side of where they were last night.
    I opted against using the "Fetch Relative To" property node in your example (which is for continuous data acquisition) in favor of the "Fetch Offset" property node used in the 'Fetch in Chunks' NI-SCOPE example (which is more along the lines of a single triggered acquisition like my situation). I also found that fetching relative to the read position wouldn't start the data stream at the normal pretrigger location, even though the position is supposed to be set to zero after each acquisition start. I would presume I would get the same waveform out using 'fetch relative to:read position' as I would using the scope soft panel, but it doesn't.
    A similar problem I have now that I've messed with the "Fetch Relative To" property node is that now it seems like I have to make an extra acquisition before anything like trigger delay or number of samples to read is registered. In the manual mode, I used the "Fetch Relative To: Pretrigger" in order to get the section of data to be properly oriented in time. But it still exhibits the updating only after an acquisition has been made. Example: I setup a 1Msample acquisition at 50MS/S w/ a trigger delay of zero. I acquire and see my pretty graph. There's an interesting feature at t=8ms, so I set my trigger delay to 8ms, but I have to acquire twice in order to see my graph update with a waveform that starts near the interesting feature. Any idea why this is happening now?
    Also, I've attached a screenshot of the acquisition loop. Let me know if this is approximately what you had in mind for fetching 65000 bytes (or 65000 data points w/ my I8 acquisition) at a time and streaming that to disk. Now that I look at it, it might make more sense to use another queue (like in your example) that passes the waveform ID and the chunk of data to a separate loop that is always waiting on data to write to disk. Also, when I save the files, I'd like to be able to load them and see the original Voltage/timing information on the file and I'm at somewhat of a loss as to how to include information like offset, gain, t0 and dt. Right now, even though I think it creates a weird header at the beginning of the file, I use the waveform attributes to keep this information.  I ended up saving those four items in reasonably related waveform attributes, but if there's a better way you can think of, let me know.
    Attachments:
    FetchinChunks.jpg ‏181 KB

  • "Semi/Non additive measure with" : "LastChild Across All Dimension" OR "LastChild Across Time and MAX across All Other Dimension"

    Hey All
    I am using the SQL SERVER 2008R2 standard edition so please be mindful when posting any suggestion if that will work on this edition or not.
    Here is the situation: I have Cube with one fact table "Positions". Measure I am querying is "AUM" and Dimensions across user can query are combinations of {"Portfolio", "Security" and "Time"}.
    I am receiving the "AUM" values at "portfolio+Security+Time" level. AUM is actually the Portfolio level value. so sample data I receive would like this.
    PORTFOLIO SECURITY TME AUM
    A SEC1 JAN-1 100
    A SEC2 JAN-1 100
    B SEC1 JAN-1 200
    B SEC3 JAN-1 200
    A SEC1 JAN-2 300
    A SEC2 JAN-2 300
    B SEC1 JAN-2 400
    B SEC3 JAN-2 400
    So, as shown in above example , we received same "AUM" value for a given portfolio and time. (regardless of what security value is).  
    And as AUM is the snapshot values, we can not sum across time as well, so if user is looking across time we need to pick Last values.
    Now when user wants to view this data from cube, two different ways s/he wants to look at it.
    option one: AUM at port+sec+time level. that is same as shown above in example.
    option two: AUM at portfolio level only,  (notice that this is across time so results are picking values for time=JAN-2)
    PORTFOLIO AUM
    A 300
    B 400
    GRAND TOTAL 700
    I have applied the "LastChild" aggregation property to measure to achieve "Do no Sum across Time "behavior. but with this I also get "SUM Across Except time" and hence my AUM values get summed across Portfolio and security as
    well. I have also enable Time Intelligence to enable MTD, YTD calculations.
    I tried to add SCOPE statement for Security before and after the "time calculations" yet I am not able to achieve expected behavior.
    here is the calculated member definition, and scope statement i have tried.
    CREATE MEMBER
    CURRENTCUBE.[MEASURES].[AUMNetBaseCurrency] AS
    [Measures].[AUMNetBaseCurrency_Raw_Max],
    FORMAT_STRING = "#,##0.00 ;( #,##0.00 )",
    VISIBLE = 1 , DISPLAY_FOLDER = 'Position';
    After this I have MTD and YTD time calculation for few measures including AUM.
    I tried placing the following SCOPE for AUM before and after the Time calculations but it no difference in result, in fact it seems like it has no affect of scope at all due to the "LastChild" aggregate property applied on measure.
    SCOPE([MEASURES].[AUMNetBaseCurrency]);
    SCOPE([Security].[Symbol].[All]);
    THIS = MAX([Measures].[AUMNetBaseCurrency_Raw_Max]);
    END SCOPE;
    END SCOPE;
    so options I am thinking of but do not have clear idea how to implement those are:
    0) why my Security Scope Statement are not showing any effect, and can we do something about it?
    1) Can be change "SUM" behavior to "MAX" / "MIN" across all Dimension except time.
    2) Is there a way to implement "LastChild" across all dimension instead only across Time.
    3) Should I need to make any changes at Fact table level to achieve this ? I refer one Q&A in this forum that suggested to create such measure in different measure group but it doesn't have any clarification.
    Any help would be appreciated.
    Thanks
    Anup Shah

    hi,
    you will have to enable cache in your nqsconfig.ini file under the section
    # Query Result Cache Section
    [ CACHE ]
    ENABLE     = YES;
    cache seeding is also a mechanism through which the data would be updated in the cache.

  • How to use external timebase for Two Edge Separation measurement with PCIe 6351

    Hi
    I am working with PCIe 6351 x series DAQ card.
    counter measurements.
    Here i need to measure the time/no edges between two edges. Falling to Raising of two signals.
    I got the measurement with internal time base.
    here i need to synchronize the measurements with external clock from the external setup.
    so i need to use that external clock,timebase for the counters.
    Any solution please..
    Regards,
    Hari

    First, the VI you mention is NOT for STC timer/counters. Therefore, it won't work.
    If you want to get the time interval between two edges, you want to use the Functions -> Data Acquisition -> Counter -> Count Events or Time VI. This will do the function you are looking for.
    Mark

  • Unit of measure with language Key

    Hello Guruz,
    WHile creationd of purchase order, there is a check kept which compares the vendor communication language which gets updated in the Po header - Communucation tab with Db table T006A which holds the Unit of measurements with Language Key and pops up an error message.
    It cheks the order unit in the PO, and the T006A whether the unit is maintained in the Language key as in the PO header - Communiction tab.
    But while creating a unit of measurement in CUNI, we dont have the option to maintain the language key for the UOM.
    In CUNI, it defaults the language key as EN in table T006A.
    I need a prompt support from you guys on this issue. Please suggest how to go about with this?
    Best Regards
    Greg.

    Hello,
    When i am trying to go for the language key in CUNI, it says that the said language is 'not available in server'. I guess it is checking whether the language patch is updated in the server.
    But it doesn't sound logical to maintain all the unit of measurements in all the vendor languages. Suppose if we are having vendors with 75 communication languages and we are having 50 units os measurements, so does that mean that we have to maintain units of measurements 75 X 50  times ?
    Kindly provide your inputs on the issue. We are facing lots of issues with this control kept in PO header checking for the T006A with the vendor communication language.
    Regards
    Greg

  • Samples Not Acquired on NI 660x or NI 6624 Devices When using pulse width, semi-period or two-edge separation buffered measurements with the NI 660x and NI 6624 on counters 0 and 1, some samples may not be acquired under certain circumstances.

    In the DAQmx 9.4 Readme the following is stated:
    Samples Not Acquired on NI 660x or NI 6624 Devices When using pulse width, semi-period or two-edge separation buffered measurements with the NI 660x and NI 6624 on counters 0 and 1, some samples may not be acquired under certain circumstances.
    Could anyone elaborate on this? I am particularly interested in knowing under what circumstances this problem arises for the 6602 PCI devices.
    How should I understand 'NI 660x and NI 6624 on counters 0 and 1'. Is it only a problem on counter 0 and 1 for the 660x or does this only apply for the 6624 devices.
    /mola 

    Are there any NI Product Support Engineers out there who has an answer to this question and who are willing to share it with me?

  • Personal recruitment How to do new measure with function call Update IT4000

    Hi out there and happy new year to forum,
    i want to update to implement a recruitment measure (Transaction PB40). I did it already with batch input, but i guess it's much more comfortable to do with funtion module, bapi or calsses.
    After choosing the measure, the infogroup will be processed. So i guess, i just have to use the corresponding function modules to update the infotypes in a given sequence. is this suggestion right?
    What are these function modules? How to update f.e. IT4000? Does anyone has a coding snippet?
    Is it HR_INFOTYPE_OPERATION? 
    In our Infogroup there are also customer specific infotypes f.e. 9***. Can I use the sap standard function modules for "own implemented" infotypes?
    I also appreciate an abap oo pattern or any similar....
    Best regards,
    T

    Andy Goris wrote:
    But I don't think that it's possible to hire an application completly with function modules (there is no PERNR in the beginning so I would keep the BIM).
    The applicants already exists in system, i just want to implement a measure with exact one click!
    Otherwise there is a BAPI for applicant_create.
    I'm not very expierienced focussing the implementation in pa-recruiting. I'm really looking for a good documentation or anything like that f.e. sap press.
    Andy Goris wrote:
    ... but why would you use the old recruiting database? This is hardly used anymore.
    What are the alternatives to it. (I'm not looking forward to upgrade releases or using wd4a or webgui). we are on ecc 6.0 There is already a lot of z-coding in recruitment. We are also using an self implemented e-recruiting approach
    Edited by: Timo Ehl on Jan 3, 2012 2:30 PM

  • Exporting set of measures with aliases into essbase cube

    We were wandering if there is an easy way to import bunch of measures with aliases into Essbase cube rather then entering them one by one?
    thank you

    Simplest way, would be to create a spreadsheet with the Measures and alias listed in column format.
    You can organize them in Generation, Level or Parent-Child format.
    Create a dimension load rule that maps the columns to properties in Essbase.
    When you want to modify the Measures dimension layout, change the spreadsheet and rerun the load rule.
    Brian Chow

  • Error "CONVT_NO_NUMBER" while performing user measurement with USMM

    Hi,
    While running user measurement with USMM I am getting following run
    time error.
    To be specific - After preselections in USMM->system measurement->while
    checking Green Check mark this error is screaming up.
    Ae per Note - 1230640 :- You call transaction USMM and want to cancel
    the user measurement using the cancel button (red 'X') on the dialog
    box of the user measurement. However, the measurement continues and is
    completed as if you had chosen the ok button (green check mark).
    Above note is applied if the problem comes if we check red button
    rather than green check mark.
    Please suggest us the best way to procees.
    ST22 details:-
    Runtime Errors CONVT_NO_NUMBER
    Exception CX_SY_CONVERSION_NO_NUMBER
    Date and Time 09/16/2008 07:42:32
    Short text
    Unable to interpret " 1,410 " as a number.
    What happened?
    Error in the ABAP Application Program
    The current ABAP program "RSUVM001" had to be terminated because
    it has
    come across a statement that unfortunately cannot be executed.
    What can you do?
    Note down which actions and inputs caused the error.
    To process the problem further, contact you SAP system
    administrator.
    Using Transaction ST22 for ABAP Dump Analysis, you can look
    at and manage termination messages, and you can also
    keep them for a long time.
    Error analysis
    An exception occurred that is explained in detail below.
    The exception, which is assigned to
    class 'CX_SY_CONVERSION_NO_NUMBER', was not
    caught and
    therefore caused a runtime error.
    How to correct the error
    Whole numbers are represented in ABAP as a sequence of numbers,
    possibly
    with an algebraic sign.
    The following are the possibilities for the representation of
    floating
    point numbers:
    [mantissa]E[algebraic sign][exponent]
    [whole number part].[fraction part]
    For example, -12E+34, +12E-34, 12E34, 12.34
    If the error occurred in your own ABAP program or in an SAP
    program you modified, try to remove the error.
    If the error occures in a non-modified SAP program, you may be able
    to
    find an interim solution in an SAP Note.
    If you have access to SAP Notes, carry out a search with the
    following
    keywords:
    "CONVT_NO_NUMBER" "CX_SY_CONVERSION_NO_NUMBER"
    "RSUVM001" or "RSUVM001"
    "START-OF-SELECTION"
    If you cannot solve the problem yourself and want to send an error
    notification to SAP, include the following information:
    1. The description of the current problem (short dump)
    Information on where terminated
    Termination occurred in the ABAP program "RSUVM001" - in "START-OF-
    SELECTION".
    The main program was "RSUVM001 ".
    In the source code you have the termination point in line 96
    of the (Include) program "RSUVM001".
    85 * Analyze the user records.
    86 CALL FUNCTION 'SAPGUI_PROGRESS_INDICATOR'
    87 EXPORTING
    88 text = 'Die Benutzerstammsätze werden analysiert.'(002).
    89 * PERFORM analyze_user_records.
    90
    91 CALL FUNCTION 'SLIM_USER_MEASUREMENET'
    92 CHANGING
    93 SLIM_TUREP = LT_TUREP.
    94 .
    95 LOOP AT LT_TUREP INTO LS_TUREP .
    >>>>> MOVE-CORRESPONDING LS_TUREP TO ITUREP.
    97 APPEND ITUREP.
    98 ENDLOOP.
    99
    100 * Submit Application measurements as batch jobs.
    101 * Save additional data about this analysis in the DB.
    102 * we need meas-no ,v-date at the batch
    any ideas?

    Hi Srikanth,
    The below Notes helps you to solve the issue.
    Please review the below Notes.
    Note 1117010 - Runtime error CONVT_NO_NUMBER in RSUVM001
    Note 1115924 - USMM composite SAP Note: Problems after importimg new SPs
    Thanks & Regards,
    Pradeep Srigiri

  • Please let me know the TCODE for configuring Unit of Measure with settings

    Dear All,
    Please let me know the TCODE for configuring Unit of Measure with settings.
    Thanks

    Dear All,
    Please let me know that can we get PO amount with Taxes included in report ME80FN.
    Is there ant other standard report to get the same?
    Please help.
    Thanks.
    Shreya

  • Build the Bar chart from Cross Tab which have 3 measures with cell refr.cal

    Hello Friends. I want to Build the Bar chart from Cross Tab which have 3 measures with Actual , Plan and Forecast
    But Issue here is each cell is defferent like i have to do different calculation for each cell for actual, plan, forecast
    like A = Jan act/Budget actual, Where as B = Feb Act/ Jan Budg+Febbdgt_March Budget
    So now when I right click to say turn to Chart it is not generating the chart
    Please help how to build bar chart since i do not have 1 key figure which i can drag and drop in view structure of chart so how to handle this
                             Jan,  Feb, March, Aprl, May, June
    Actual                  A       B       C     D
    Plan                     N       M       O     P
    Forecast
    SONIYA

    Hi Soniya,
    It will be better if you can get these keyfigures calculated from the BEx with the help of structures in the BEx.
    At the report level it will be complex to create these variables and also not sure whether going to work:
    But you can try something like this:
    A= If (L01 Moths="JAN") then Sum (Actual)/Sum (Bud) Else 0 similarly for other months.
    Then you need to create one more variable (measure) which should be plotted on the chart as:
    Final Act = If([L01 Month]="JAN") Then A ElseIf([L01 Month]="FEB") Then B ElseIf([L01 Month]="MAR") Then C
    Final Plan= If([L01 Month]="JAN") Then N ElseIf([L01 Month]="FEB") Then M ElseIf([L01 Month]="MAR") Then O
    similarly you have to create variables for Forecast, then finally you can plot final 3 variables in the graph.
    Hope it helps you.
    Regards,
    Rohit

  • Configuring measures with custom aggregation for hierarchies

    Can I Configuring different measures with different aggregation for hierarchies . in AWM the aggregation for measures are disabled, am I missing something.
    For example
    Dimension : supplier
    Measure : risk score
    Aggregation : maximum
    Dimension : supplier
    Measure : spent
    Aggregation : sum

    Yes it is possible. But make sure that you have turned the compression off in the cube

  • AC resistivity measurement with MyDAQ on TEC

    Hello,
    I would like to ask for advice: we would like to measure resisitivity of Peltier-elements. If we use DC method, because of the Seebeck-voltage of the element, we do not get the real resistivity value. The common method to use AC measurement in case of such devices. I would like to know, if it is possible to use a MyDAQ device for this task? As a first approach we do not need the best accuracy, this would be more as a "broken-element" test (there is usually a significant shift away from the nominal AC resistivity value stated by the manufacturer).
    What would be the easiest way to carry out a simple measurement? What about if I use one analogue output of the MyDAQ lets say with 1kHz, and I drive the Peltier-element with this voltage AC source? And I connect a known resistor in series, and I measure the AC voltage drop on this. From this data, I could calculate the current, and the AC resistance of the Peltier-element? Of course I have to choose a resistor so the MyDAQ can drive the network.
    what is your opinion, where to start?
    Thanks!
    Solved!
    Go to Solution.

    Hello, thanks for the reply!
    Actually I am just going to replace 60 Peltier elements in a "home-made" 15 years old heat exchanger (I expect to find some mice too ). This assembly has dropped performance (heat transfer power) and also I know that, a few years ago after a malfunction with the hot side cooling fans, this device was running for a certain time without forced cooling. What is quite bad in case of Peltier-elements. I have already bought 65 brand new Peltier-modules and after I get some heat conducting paste, I will replace them. And I will have 60 pieces of old Peltier modules, to play with them
    Since I have this NI MyDAQ, I thought it would be fun (and would give also valuable information) to try to measure the old elements' AC resistivity. Since this task is not like I should do often, so for first, I would not like to go for more pro solutions, I will try the MyDAQ first.
     I expect that some of the modules got structural degradation during the heat shock, and this may be measurable in a shifted AC resistance.
    I would do the measurement with zero applied DC voltage. These modules are Melcor PT3-12-30L (discontinued product), the nominal module resistance is about 5 Ohm.
    EDIT: and would you recommend a certain circuit to try? Shell I use a known resistor for current measurement via measuring the voltage on it? Or I could use a known AC current to drive the Peltier, and I would measure the resulted AC Voltage on it?

  • Create new object (measure) with IIF function on BW Olap universe:mdx error

    hello
    i have a universe based on a BEX query
    i want to create a new object (a measure) with the IIF(,,) function
    i created the following select:
    <EXPRESSION>IIF(@Select(Currency Type\CurrencyTypeCode)="10",@Select(Key Figures\RevenueGlobalCurrency),@Select(Key Figures\RevenueLocalCurrency))</EXPRESSION>
    defined as number
    where
    CurrencyTypeCode is a dimension (character format)
    RevenueGlobalCurrency and RevenueLocalCurrency are measure (number format)
    i always have a parse which is OK but
    when i want to  use the object on a webi report i always have an MDX error : ...invalid MDX with .... (WIS 10901)
    and advise ?
    Thanks

    Hi Didier,
    Thanks a lot for your help.
    Now it works : i do not understand why
    i explain:
    the correct syntaxe of my new measure is :
    <EXPRESSION>([Measures].[4FVLHG7OAJMNY9WC06X6JLOZQ], [0CURTYPE].[10])</EXPRESSION>
    it does not works with my "L01 currency type" BO dimension
    ([0CURTYPE].[LEVEL01])
    when i had :
    -my universe is based on a BW BEX query
    -my universe has been automaticaly generated with the assistant
    -the Currency Type Class was define as below:
           - 1 BO dimension object : L01 Currency Type =>
    [0CURTYPE].[LEVEL01]
           - 3 BO detail objects : L01 Currency Type techName =>
    [0CURTYPE].[LEVEL01].[TECH_NAME]
                                              L01 Currency Type Key =>
    [0CURTYPE].[LEVEL01].[[20CURTYPE]].[Value]
                                             L01 Currency Type Long Name =>
    [0CURTYPE].[LEVEL01].[[40CURTYPE]].[Value]
    i always had a wrong mdx syntax error on the webi report
    i updated the class with the follow
         - 1 BO dimension object : L01 Currency Type =>
    [0CURTYPE].[LEVEL01] (idem)
           - 3 BO detail objects : L01 Currency Type techName =>
    [0CURTYPE].[LEVEL01].[TECH_NAME]  (idem)
                                              L01 Currency Type Key =>
    [0CURTYPE].[LEVEL01].[NAME] (updated)
                                             L01 Currency Type Long Name =>
    [0CURTYPE].[LEVEL01].[DESCRIPTION] (updated)
    then it was working
    Then i tried to reproduce from scratch and now it also works with the previous version of the class.
    I do not understand anything
    Is there something like a cache file of the MDX query (in RSRT, cache mode is set to inactive for the bex query) ?
    ps: this is the same if a use webi under infoview (java) or webi rich client...
    Edited by: Frederic Nez on Jan 19, 2010 6:35 PM
    (edit : there is mistake in the post : objects are always between [] even if it is not dispay => change objects to code mode)

  • Alternate Unit of Measure with several Qty KF's

    Case is as follows:
    3 different plain KFs related to qty. Sales volume, net weight anddelivery qty.
    All need to be dispalyed in the same report. All have different base unit of measure.
    I have only one variable [YTAUNIT) for entering required unit.
    Now if I enter all KFs using this same YTAUNIT variable for them each separately I get error message on the query that this is not permitted.
    How should I change the structure ?
    - add new variable for each KF ? (YTAUNIT1, YTAUNIT2, .....) -> do I need to change any of the ABAP codes? is there a link only picking YTAUNIT ?
    Please can You give me an example (detailed) how to proceed ?
    Thanks in advance
    Petri

    Hi,
    I have worked on single variable But from the understanding of working I feel you need to create three diff variable (input var) YTAUNIT1, YTAUNIT2 ect Because they are on same base unit of measure with different units of measures. And you will have to take corresponding variable for every key figure in the calculated KF.
    Just try with different variables.
    Assign points if it helps.
    Regards,
    Wrushali

Maybe you are looking for

  • Pop Up Error using MVC method after Validation in  DO_HANDLE_EVENT Method

    Hi Experts,. I am rather new to BSP applications and I am working using MVC at the moment. I have a requirement which apparently is simple, but am yet to get that done successfully. I have a couple of buttons on my BSP View screen, on click of these

  • REGIONAL OPTIONS with WEBUTIL

    When trying to run an application with Webutil while having regional settings other than English, the application fails to start. Webutil seems to be unable to create the webutil.properties file in the Jinitiator directory on the client with regional

  • Windows 7 loading hangs - Bootcamp

    Hi guys, i just got a i5 13" MBA 256Gb, LG dvdrw drive and windows 7 pro. problem is that after setting bootcamp parition, it restarts to setup mode, and it hangs at the loading bar, about 3 fingers long, it hangs forever. did anybody encounter this?

  • ILOM0001 - functional location description  (pltxt)

    Hi, problem: the (same) description (PLTXT) of functional location has to be saved in TWO different languages every time any change is done on a functional location AND  save is (pressed) processed. User exit ILOM0001 is already implemented and can b

  • EXCH 2003 2010 Transition, /PrepareLegacyExchangePermissions failure.

    I've searched and searched and cannot find the same scenario with a resolution to this problem. We have planned the transition of Exch 2003 to 2010 but have not accounted for a scenario during test, now we are supposed to be beginning the role out an