Issue when programmin​g a Diolan DLN-2

Greetings Ni community,
   I am trying to program a Diolan DLN-2 that works with a magnetometer, simliar to what I've done with a USB-8451. However, I receive an error coe of 1172 - "Error calling method Dln.I2cMaster.Port.Write, (System.Reflection.TargetInvocationException: Exception has been thrown by the target of an invocation.
    Inner Exception: Dln.Exceptions.DlnError: DLN adapter returns error: InvalidAddress.) <append><b>Dln.Exceptions.DlnError</b> in i2c-master-configure.vi"
  Attached is the zip file. It contains the DLL, the file I am trying to code with the Diolan (i2c-master-configure.vi), the data sheet for the magnetometer I am working with, and the other VI that I wrote to work with the magnetometer via the NI USB-8451 (Working Magnetometer.vi).
    Apologies if my code is not right on the Diolan - again I'm trying to program it similar to the one I have working for the 8451,
  I am by no means an expert at LabVIEW, but I am getting better day by day.
   My code is in 2012. Let me know if you need a previous version.
  Thanks!
   TheLT
Attachments:
Diolan Zipped.zip ‏371 KB

Hi there, check this article. It explains that error 1172 is basically a warning coming from the .NET operation performed. I would suggest to check the manual of the device in case you are not invoking references appropriately. There are some VIs missing so I can't see all the code but I assume that somewhere on the "i2c-master-configure.vi" there is a Constructor Node. Since you are new to LabVIEW I would recommend to look at the examples on the DLN-2 manual and then try to reproduce them in LabVIEW. If you get a little lost, open the examples that uses the Constructor node in order to make yourself more familiar with it. I hope this helps.
Alejandro | Academic Program Engineer | National Instruments

Similar Messages

  • Issue when uploading Sales data from DSO to Cube.

    Dear All,
    I have an issue when I am uploading Sales data from DSO to Cube.  I am using BI 7.0 and I have uploaded all sales document level data to my DSO.  Then I use a transformation rule to calculate the Sales Value when I am doing DTP to cube.  Cube has customer wise aggregation data.
    In DSO I have NetPrice(KF) and Delivered_QTY(KF).  I do a simple multiplication routine in the transformation from DSO to Cube.
    RESULT =   SOURCE_FIELDS-NET_PRICE * SOURCE_FIELDS-DLV_QTY .
    At the moment I use Active Table(With out Archive) on the DSO to get the data since this is my first load.
    Issue is that the figure (Sales Value) in the cube is incorrect. I am getting a very large values which is impossible.
    Can someone please help me.
    Shanka

    Hi,
    are you sure that cube has customer wise aggregated data..since it will always aggregate the values of the key figures for the same set of characters...
    did you checked the values for the key figures as well and if even they are inflated or is the problem with this key figure only.
    During the data load the records may aggregates first for the same values of characters and then multiplication will be happening....if that is the case then you may have to multiply the values before storing it in the data package and then let it aggregate....can be achived through start routine...
    but first verify if other key figures are also having the same issue.
    Thanks
    Ajeet

  • File formatting issues when saving PDF as PPTX in Acrobat XI Pro

    Hi guys,
    I have some issues when I'm saving my PDF (initially comming from InDesign) as PPTX. The format is suddenly changing, leaving a blank zone, and some of the pages are rotated.
    Is it a problem comming from the Indesign or from the Acrobat options itself? I have already tried changing the chekboxes in the option box when I'm saving, but nothing worked.
    If someone has an idea how to fix that, i would be very grateful! Thanks in advance

    Which dot version of Acrobat XI are you currently using in your export?  Is the PDF file tagged?  If not, then you may want to tag the PDF file before exporting it.

  • Mid 2009 17" MacBook Pro battery shutdown issue when waking from sleep.

    My unibody 17" MacBook Pro occasionally shuts down on waking from sleep when it isn't plugged in.
    When I open the lid to wake it I can hear the hard drive and fans starting to spool, and then the next second it's off and silent. The laptop doesn't mind being turned on again after it does this and might go a week without doing it again, equally it might do it next time I try and wake it from sleep. It does this regardless of battery charge level.
    coconutBattery says the battery still has 86% of its designed capacity, and I haven't noticed anything else untoward regarding battery charging or performance when in use.
    So my question is two fold:
    1) Does it sound like the battery's internal processor is malfunctioning, or could it be something else?
    2) If it is the battery's internal processor, will replacing the battery replace the processor, or are they fitted separately (in defference to the 'internal' part of its name)?
    Cheers,
    James
    nb. just in case it makes any difference, the laptop is a 17" Mid-2009 MacBook Pro, 2.8GHz Core 2 Duo with 4Gb Ram running OSX 10.7.3

    James
    I have had issues when i press the space bar from sleep, sometimes it wakes but others i'm forced to power it on with the power button.
    Is this similar to anything you've experianced?
    Its so inconsistantly these days, but i'm also connected to an external screen.

  • I reinstall the Yosemite due to the wifi issue, when finished, the iMovie was not installed, I try to download it from app store, but it's not free. How can I get iMovie back?

    I'm using a mid 2014 13" rmbp. I reinstall the Yosemite due to the wifi issue, when finished, the iMovie was not installed, I try to download it from app store, but it's not free. How can I get iMovie back?

    It's something to do with the main user account (not iTunes account) on my Mac because I logged in with another account and it downloaded fine, I installed it on the HDD and even though the date in App Store was 16th October 2014, it is the full installer of 10.10.2.
    To try to resolve this problem, I have copied the installer app elsewhere and deleted it but App Store still will not allow me to download it with my main user account yet just allowed me to update iTunes.
    I had a brainwave last night that it might have been related to me changing the default downloads folder to one not on my boot drive (to save disk space on my main boot drive) but having changed it back, tried again, rebooted and tried yet again, deleted my App Store preferences. nothing has worked.

  • OBIEE 11g - issue when export hierarchy to a csv

    Hello,
    I have an issue when I try to export to a csv file a report containing a Logical Dimension field, when I export just the first level of hierarchy is exported and not all the records. When I export to another format (such as excel) all the records are exported correctly. Do you know what can I do in order to have all the records exported to a csv?
    Thanks!
    Solange.

    This is a bug see below.
    Bug 13083479 : EXPORT DATA TO CSV FORMAT DOES NOT DOWNLOAD EXPANDED HIERARCHICAL COLUMNS
    Proposed workaround : OBIEE 11.1.1.5: Pivot Table Report Export To Csv Does Not Include Hierarchical Data [ID 1382184.1]
    Mark if answered.!
    Thanks,
    SVS

  • Performance issues when creating a Report / Query in Discoverer

    Hi forum,
    Hope you are can help, it involves a performance issues when creating a Report / Query.
    I have a Discoverer Report that currently takes less than 5 seconds to run. After I add a condition to bring back Batch Status that = ‘Posted’ we cancelled the query after reaching 20 minutes as this is way too long. If I remove the condition the query time goes back to less than 5 seconds.
    Please see attached the SQL Inspector Plan:
    Before Condition
    SELECT STATEMENT
    SORT GROUP BY
    VIEW SYS
    SORT GROUP BY
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    TABLE ACCESS BY INDEX ROWID GL.GL_CODE_COMBINATIONS
    AND-EQUAL
    INDEX RANGE SCAN GL.GL_CODE_COMBINATIONS_N2
    INDEX RANGE SCAN GL.GL_CODE_COMBINATIONS_N1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUES_N1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUE_SETS
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUE_SETS_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES_TL
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUES_TL_U1
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUE_NORM_HIER_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_LINES
    INDEX RANGE SCAN GL.GL_JE_LINES_N1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_HEADERS
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_DAILY_CONVERSION_TYPES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_SOURCES_TL
    INDEX UNIQUE SCAN GL.GL_JE_SOURCES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_CATEGORIES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    INDEX UNIQUE SCAN GL.GL_BUDGET_VERSIONS_U1
    INDEX UNIQUE SCAN GL.GL_ENCUMBRANCE_TYPES_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_BATCHES
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_PERIODS
    INDEX RANGE SCAN GL.GL_PERIODS_U1
    After Condition
    SELECT STATEMENT
    SORT GROUP BY
    VIEW SYS
    SORT GROUP BY
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    TABLE ACCESS FULL GL.GL_JE_BATCHES
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_HEADERS
    INDEX RANGE SCAN GL.GL_JE_HEADERS_N1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    INDEX UNIQUE SCAN GL.GL_ENCUMBRANCE_TYPES_U1
    INDEX UNIQUE SCAN GL.GL_DAILY_CONVERSION_TYPES_U1
    INDEX UNIQUE SCAN GL.GL_BUDGET_VERSIONS_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_SOURCES_TL
    INDEX UNIQUE SCAN GL.GL_JE_SOURCES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_CATEGORIES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_LINES
    INDEX RANGE SCAN GL.GL_JE_LINES_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    TABLE ACCESS BY INDEX ROWID GL.GL_CODE_COMBINATIONS
    INDEX UNIQUE SCAN GL.GL_CODE_COMBINATIONS_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_PERIODS
    INDEX RANGE SCAN GL.GL_PERIODS_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUES_N1
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUE_NORM_HIER_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES_TL
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUES_TL_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUE_SETS
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUE_SETS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    Is there anything i can do in Discoverer Desktop / Administration to avoid this problem.
    Many thanks,
    Lance

    Hi Rod,
    I've tried the condition (Batch Status||'' = 'Posted') as you suggested, but the qeury time is still over 20 mins. To test i changed it to (Batch Status||'' = 'Unposted') and the query was returned within seconds again.
    I’ve been doing some more digging and have found the database view that is linked to the Journal Batches folder. See below.
    I think the problem is with the column using DECODE. When querying the column in TOAD the value of ‘P’ is returned. But in discoverer the condition is done on the value ‘Posted’. I’m not too sure how DECODE works, but think this could be the causing some sort of issue with Full Table Scans. How do we get around this?
    Lance
    DECODE( JOURNAL_BATCH1.STATUS,
    '+', 'Unable to validate or create CTA',
    '+*', 'Was unable to validate or create CTA',
    '-','Invalid or inactive rounding differences account in journal entry',
    '-*', 'Modified invalid or inactive rounding differences account in journal entry',
    '<', 'Showing sequence assignment failure',
    '<*', 'Was showing sequence assignment failure',
    '>', 'Showing cutoff rule violation',
    '>*', 'Was showing cutoff rule violation',
    'A', 'Journal batch failed funds reservation',
    'A*', 'Journal batch previously failed funds reservation',
    'AU', 'Showing batch with unopened period',
    'B', 'Showing batch control total violation',
    'B*', 'Was showing batch control total violation',
    'BF', 'Showing batch with frozen or inactive budget',
    'BU', 'Showing batch with unopened budget year',
    'C', 'Showing unopened reporting period',
    'C*', 'Was showing unopened reporting period',
    'D', 'Selected for posting to an unopened period',
    'D*', 'Was selected for posting to an unopened period',
    'E', 'Showing no journal entries for this batch',
    'E*', 'Was showing no journal entries for this batch',
    'EU', 'Showing batch with unopened encumbrance year',
    'F', 'Showing unopened reporting encumbrance year',
    'F*', 'Was showing unopened reporting encumbrance year',
    'G', 'Showing journal entry with invalid or inactive suspense account',
    'G*', 'Was showing journal entry with invalid or inactive suspense account',
    'H', 'Showing encumbrance journal entry with invalid or inactive reserve account',
    'H*', 'Was showing encumbrance journal entry with invalid or inactive reserve account',
    'I', 'In the process of being posted',
    'J', 'Showing journal control total violation',
    'J*', 'Was showing journal control total violation',
    'K', 'Showing unbalanced intercompany journal entry',
    'K*', 'Was showing unbalanced intercompany journal entry',
    'L', 'Showing unbalanced journal entry by account category',
    'L*', 'Was showing unbalanced journal entry by account category',
    'M', 'Showing multiple problems preventing posting of batch',
    'M*', 'Was showing multiple problems preventing posting of batch',
    'N', 'Journal produced error during intercompany balance processing',
    'N*', 'Journal produced error during intercompany balance processing',
    'O', 'Unable to convert amounts into reporting currency',
    'O*', 'Was unable to convert amounts into reporting currency',
    'P', 'Posted',
    'Q', 'Showing untaxed journal entry',
    'Q*', 'Was showing untaxed journal entry',
    'R', 'Showing unbalanced encumbrance entry without reserve account',
    'R*', 'Was showing unbalanced encumbrance entry without reserve account',
    'S', 'Already selected for posting',
    'T', 'Showing invalid period and conversion information for this batch',
    'T*', 'Was showing invalid period and conversion information for this batch',
    'U', 'Unposted',
    'V', 'Journal batch is unapproved',
    'V*', 'Journal batch was unapproved',
    'W', 'Showing an encumbrance journal entry with no encumbrance type',
    'W*', 'Was showing an encumbrance journal entry with no encumbrance type',
    'X', 'Showing an unbalanced journal entry but suspense not allowed',
    'X*', 'Was showing an unbalanced journal entry but suspense not allowed',
    'Z', 'Showing invalid journal entry lines or no journal entry lines',
    'Z*', 'Was showing invalid journal entry lines or no journal entry lines', NULL ),

  • Urgent help - PO issue when receiving

    Dear gurus
    i have an urgent production issue, i have done a mass conversion of all the master data from one plant to another, the scenario is that they are closing down and stopping operations at old plant and moving the entire operations to new plant under the same company code, so for this i did the entire MM master data conversions
    i also changed the delivering plant and address using MASS T-code for all the open PO's line items from old plant to new plant (some of line items had partial open qty and some of them entire open qty)
    <b><u>1st issue:</u></b>
    But when i am doing VL31N inbound delivery, the PO qty is NOT showing up automatically in the delivery qty field, so it is causing issues when trying the do a GR
    <b><u>2nd issue</u></b>
    also another issue is that in the Goods Movement Data tab while using VL31N for GR, it is by default showing the old plant still and is grayed out even though the PO line item contains the new plant code and delivery address, so how can we change the receiving plant to new plant and why it is not picking the new plant still
    so can you please help me on this issue and let me know if there is any settings or resolution
    thanks a lot in advance and appreciate your expert advices
    chakri

    dear chakri,
    you have written that you have converted the material master data, but what about the existing puchase orders, you have to copy all the orders in the new plant .
    or if possible try to chane the delivery address in in the item details of the P.O.
    rewards if useful.
    ashish

  • Iteration Speed issue when Indexing 3D array wried to a while-loop

    Brief Description of my Program:
    I am working on the real-time signal generation by using LabView and DAQmx (PCI-6110 card). My vi reads the big data file (typically 8MB txt file containing about 400,000 samples (complex double precision). Then, the signal is pre-processed and I end up with a huge 3D array to feed while-loop (typically 3D array dimension is N x 7 x M where N & M >> 7). Inside the loop, that 3D array is indexed and processed before the results are written to the DAQmx outputs. I have a speed issue when indexing the large 3D array (i.e, 3D array having a large sub-array size). My while-loop could not run fast enough to top-up the DAQmx AO buffer (at the output rate of 96kHz). It could run faster only if I use smaller 3D array (i.e, smaller-sized sub-arrays). I do not quite understand why the size of 3D sub-array affects the rate of looping although I am indexing same sub-array size at each iteration. I really appreciate your comments, advices and helps.
    I include my 3D array format as a picture below.
    Question on LabView:
    How does indexing an 3D array which wires to the while-loop affect the speed of the loop iteration? I found that large dimension of sub-arrays in the 3D array slows down the iteration speed by comparing to indexing the same size of sub-array from smaller-sized sub-arrays of the 3D array to perform signal processing inside the while-loop. Why? Is there any other way of designing LabView Program to improve speed of iteration?
    attachment:

    Thank you all for your prompt replies and your interests. I am sorry about my attachment. But, I have now attached a jpg format image file as you suggested.
    I had read the few papers on large data handling such as "LabVIEW Performance and Memory Management". Thus, I had already tried to avoid making unnecessary copies of data and growing arrays in my while-loop. I am not an expert on LabView, so I am not sure if the issues I have are just LabView fundamental limitations or there are any other ways to improve the iteration speed without reducing the input file size and DAQ output rate.
    As you request, I also attach my top-level vi showing essential sections such as while-loop and its indexing. The attached file is as an image jpg format because the actual vi including Sub-VIs are as big as 3MB in total. I hope my attachment would be useful for anyone who would like to reply my question. If anyone would like to see my whole vi & llb files, I would be interesting to send it to you by an e-mail privately and thus please provide your e-mail address.
    The dimension of my 3D array is N x 7 x M (Page x Row x Column), where N represents number of pages in 3D array, and M represents the size of 1D array.  The file I am currently using forms 3D array of N = 28, & M = 10,731.  Refering to the top-level vi picture I attached, my while-loop indexes each page per iteration and wrap-around.  The sub-VI called "channel" inside the while-loop will further index its input (2D array) into seven of 1D arrays for other signal processsing.  The output from that "channel" sub-VI is the superposition of those seven arrays.  I hope my explaination is clear. 
    Attachement: 3Darray.jpg and MyVi.jpg
    Kind Regards,
    Shein
    Attachments:
    3Darray.jpg ‏30 KB
    MyVI.jpg ‏87 KB

  • OBIEE 11g - issue when export to a csv file

    Hello,
    I have an issue when I try to export to a csv file a report having numeric feild, when I export and open the file in notepad the trailing zeros to the right of decimal points do not show up. When export to another formats (such as excel, pdf) the trailing zeros appears correctly. e..g, 100.00 appears as 100, 100.50 appears as 100.5.
    Do you know what can be done in order to have trailing zeros to the right of decimal points to appear in a csv?
    Thanks

    This is a bug see below.
    Bug 13083479 : EXPORT DATA TO CSV FORMAT DOES NOT DOWNLOAD EXPANDED HIERARCHICAL COLUMNS
    Proposed workaround : OBIEE 11.1.1.5: Pivot Table Report Export To Csv Does Not Include Hierarchical Data [ID 1382184.1]
    Mark if answered.!
    Thanks,
    SVS

  • Issue when scrolling a pdf - There was a problem reading this document (135)

    This is when the PDF is created and rendered/previewed from Acrobat Reader and the issue is there for latest acrobat version too.
    When you open the PDF in Acrobat Reader and then when you scroll it and if you have more that one page in the PDF, it gives an error saying "There was a problem reading this document (135)".
    Is this a known issue in Acobat? Any resolution? Reinstalling acrobat version didn't help.
    Thank you in advance...!
    /Manu

    It's hard to tell for me as they are files such bank statement or other document from websites...how can i know?
    Thanks,
    Date: Wed, 9 May 2012 01:29:45 -0600
    From: [email protected]
    To: [email protected]
    Subject: Issue when scrolling a pdf - There was a problem reading this document (135)
        Re: Issue when scrolling a pdf - There was a problem reading this document (135)
        created by Nikhil.Gupta in Adobe Reader - View the full discussion
    Hi Manulak_Tech,
    As you are able to view PDF I had provided, the issue seems to be PDF specific.
    So when you scroll the PDF and reach to a page where Adobe Reader cannot read it, the error comes. I am sure the problem lies the way PDF was created. Can you describe how the PDF was created?
         Replies to this message go to everyone subscribed to this thread, not directly to the person who posted the message. To post a reply, either reply to this email or visit the message page: http://forums.adobe.com/message/4390058#4390058
         To unsubscribe from this thread, please visit the message page at http://forums.adobe.com/message/4390058#4390058. In the Actions box on the right, click the Stop Email Notifications link.
         Start a new discussion in Adobe Reader by email or at Adobe Forums
      For more information about maintaining your forum email notifications please go to http://forums.adobe.com/message/2936746#2936746.

  • Printing issues when printing to Server 2008 R2 Print Server

    We have a problem with netwok printing that encompasses many different areas and versions of Windows OSes so I will be posting this in a few different places.
    Let me lay this out for you. We have a single Server 2008 R2 print server with 300 printers of various models. This print server has both 32 and 64-bit drivers installed for every printer. For the HP printers we are using the HP Universal Printing PCL 6
    (version 61.94.3.37) and for the Ricoh printers we are using their PCL6 driver for universal printing version 3.5. All drivers run in Isolation mode and no printer gets published to AD.
    We have a mix of Windows clients printing to these printers, both XP Pro and Windows 7 32 and 64-bit.
    Now for the problems: The other day one Isolation host on the server went crazy, processing was hitting 100% and memory was fluctuating and growing a lot. After investigating a little that isolation host was having issues accessing the Ricoh Driver (which
    was 3.4 at the time). In an attempt to resolve the issue we updated the driver to V. 3.5. This seemed to fix the isolation issues and things calmed down however this is the beginning of many other issues for the clients. Both Windows OS versions (XP and 7)
    don't seem to be either downloading and installing the updated driver and/or doing so correctly. Clients are now experiencing extremely slow printing, applications such as Word, Outlook, Excel and others open slowly and it can take 2-3 minutes for print dialog
    boxes to open. In all cases it seems that we have to have a tech delete the printer(s), delete the driver and then re-install the printer to make things better but for some reason, which is really unknown, this process doesn't work for every machine and on
    some we can't even remove the old drivers at all without booting into SafeMode and doing it that way.
    With all of that said, are there any suggestions as to why the clients aren't downloading and installing the drivers from the server correctly? How to avoid these issues when we need to upgrade the drivers on the server in the future? Most importantly right
    now, are there any recommendations or scripts on how to force the deletions of the printer, driver, reset the print spoolers and then re-add the printers in a more automated fashion?
    I have played with both wmic, pnputil and printuidll and am trying to put something together but without a set solution that will work consistently it becomes a lot more difficult.

    We are experiencing a very similar issue.
    Windows Server 2008 R2. Printers are deployed over group policy.
    Clients range from Windows 7 to Windows 8. Citrix XenDesktop and XenApp environment also connects to these endpoints.
    We have A wide assortment of HP printers, and Ricoh Printers.
    For the ricoh's we have been historically using the PCL6 drivers for the individual device type (not universal print). Up till recently, it all was working fine, until about some time in the middle of june, when everything just seemed to go crazy.
    The issue was resolved by moving off from the PCL 6 drivers and using the PCL5e driver counterparts on all of our ricoh printers.
    It would crash XenDesktop applications, it would crash office applications.
    The print spooler would lock up, even if the ricoh printers were not actually used, just as long as they were loaded.
    Even after loading PCL5 on the servers, some of the clients would simply not download the PCL5 version of the driver from the print server. In order to resolve this, I have created the steps below. Be advised, this may possibly cause damage on your system.
    We just found it worked great for us. Local printers will likely have to be manually reinstalled after this.
    Directions
    Stop the print spooler
    Open C:\Windows\System32\Spool and delete the following
    All files in SERVERS
    All files in PRINTERS
    All files in drivers
    All files in prtprocs
    Start the print spooler
    Open a command prompt as administrator, and run the following command
    Cscript C:\windows\system32\Printing_Admin_Scripts\en-US\prndrvr.vbs -x
    This should further delete all print drivers.
    Open a command prompt as a standard user (not administrator) and run the following
    Gpupdate
    Now wait. It takes a while, but the printers will slowly start to repopulate.
    I am still not sure what is going on with the PCL6 drivers for ricoh, and why this issue is happening. I'm guessing it has something to do with the Ricoh auto update that is included in their pcl6 drivers. Maybe some sort of auto update happened recently.

  • HT4913 How do I resolve this issue when I try to download iTunes. I receive this error message " iTunes64Setup.exe is corrupt or invalid"

    How do I resolve this issue when I try to download iTunes on my laptop?  It fails because of this error message " iTunes64Setup.exe is corrupt or invalid"

    That suggests that the installer is getting damaged during the download.
    I'd first try downloading an installer from the Apple website using a different web browser:
    http://www.apple.com/itunes/download/
    If you use Firefox instead of IE for the download (or vice versa), do you get a working installer?

  • Issue when running acutal Off-cycle payroll in India!

    Dear experts!
    I have an urgent issue when running ACTUAL off-cycle payroll in India as below: "Payroll already performed once in future".
    In testing mode, everything is fine. I have tried to check payroll result in pc_payresult and pu01, and Infotype 3. All of them show the last payroll in 31.07.2011. But now I run off-cycle payroll on 20.08.2011, this error is displayed. So I cannot run actual payroll for them.
    Please help me! I am really appreciated.
    Regards!
    Woody.

    Hello
    Normally this error message reflects an inconsistency in the payroll
    directory cluster, that you can view via RPUDIR00. What you can do to
    restore the consistency is to run the report RPUDIR00 for this pernr
    with the following parameters on the selection screen:
    Payroll for country                           
    Personnel number             
    Compare directory            
    Payroll results structure   
    detailed log                
    Save new directory             X          <  this will write the new directory.  If you want to test first, do not select this parameter.
    Afterwards it should be possible to run payroll for this employee.
    Thanks and Kind Regards
    Ramana

  • JMS issues when migration from weblogic 9.2 to 10.3.5

    We are facing some issues when migration from weblogic 9.2 to 10.3.5
    In  weblogic 9.2 :_
    BMP Entity EJBs used in our project are read-only in nature using entity cache, below is the configuration details
    <!DOCTYPE weblogic-ejb-jar PUBLIC "-//BEA Systems, Inc.//DTD WebLogic 6.0.0 EJB//EN" "http://www.bea.com/servers/wls600/dtd/weblogic-ejb-jar.dtd">
    <weblogic-ejb-jar>
    <weblogic-enterprise-bean>
    <ejb-name>
    Company
    </ejb-name>
    <entity-descriptor>
    <pool>
    <max-beans-in-free-pool>300</max-beans-in-free-pool>
    <initial-beans-in-free-pool>150</initial-beans-in-free-pool>
    </pool>
    <entity-cache>
    <max-beans-in-cache>3500</max-beans-in-cache>
    <idle-timeout-seconds>100000</idle-timeout-seconds>
    <read-timeout-seconds>0</read-timeout-seconds>
    <concurrency-strategy>ReadOnly</concurrency-strategy>
    </entity-cache>
    Entity beans will get refreshed using the JMS messges. with in the MDB descriptor files(weblogic-ejb-jar.xml) we are using the provider URL directly and XA enabled connection factory is set to false.
    migration to Weblogic 10.3.5_
    With the same configurations MDB are not not getting deployed in weblogic 10 with some exception, so we removed the provider URL from weblogic-ejb-jar.xml and changed the JMS configuration to use foreign JMS and XA enable connection factory is set to true. Now when ever the JMS message is triggered Entity bean is not getting refreshed with the updated values. i.e values are stale.
    Can some one look into this and provide your inputs to resolve this issue.

    I think the Entity bean refresh problem appears to be unrelated to MDBs. The MDB is only responsible for getting the message to your application (which in turn interacts with Entity beans). You might want to try posting your question to an EJB newsgroup.
    Tom

Maybe you are looking for