Niggling SAN issues

Hardware:
2 x M5000 (both have exact same issue as described below)
Hitachi VSP Storage
Brocade Switches in between.
Storage has configured so we see 8 paths to each SAN Disk (which solved the PLOGI errors)
28 SAN Disks shared to each of the M5000's (to be able to move zones between each M5000 with minimal effort)
A SAN disk in use on one, is not in use on the other, obviously .
Software:
Solaris 11.2 (sparc)
Multipathing enabled to each disk
ZFS File systems.
5 zones/containers per
Issue:
I'm running Zabbix to graph the many kstat output.  The one I'm having issues with is the "ssderr::::/Illegal Request/"
In zabbix I graph this data on this particular error:
  kstat -c device_error -p "ssderr:::/Illegal Request/"
Each ssd has 8 paths.
Each ssd error "Illegal Request" count goes up hourly, about ~224 "Illegal Requests" every 60 minutes (all at once, over a minute or two time-frame), like an hourly cron job, but there's no correlating cron/smoking gun, on either server.
Is there some default background daemon that deals with SAN triggered events I'm overlooking?  There is no cron in root or other that matches these errors.
dmesg is void of error messages since the PLOGI issue was resolved.
Has anyone come across this issue.
Rick
Edit: more info
28 San Disks, each with 8 paths (= 224)
On the 30 minute mark (224 "Illegal Request"), so something is checking each path, and I'm trying to determine what it is.  I'm running all root crons while monitoring kstat in real time, all errors happen over 10 to 15 seconds time period.  No cron correlates.

Hardware:
2 x M5000 (both have exact same issue as described below)
Hitachi VSP Storage
Brocade Switches in between.
Storage has configured so we see 8 paths to each SAN Disk (which solved the PLOGI errors)
28 SAN Disks shared to each of the M5000's (to be able to move zones between each M5000 with minimal effort)
A SAN disk in use on one, is not in use on the other, obviously .
Software:
Solaris 11.2 (sparc)
Multipathing enabled to each disk
ZFS File systems.
5 zones/containers per
Issue:
I'm running Zabbix to graph the many kstat output.  The one I'm having issues with is the "ssderr::::/Illegal Request/"
In zabbix I graph this data on this particular error:
  kstat -c device_error -p "ssderr:::/Illegal Request/"
Each ssd has 8 paths.
Each ssd error "Illegal Request" count goes up hourly, about ~224 "Illegal Requests" every 60 minutes (all at once, over a minute or two time-frame), like an hourly cron job, but there's no correlating cron/smoking gun, on either server.
Is there some default background daemon that deals with SAN triggered events I'm overlooking?  There is no cron in root or other that matches these errors.
dmesg is void of error messages since the PLOGI issue was resolved.
Has anyone come across this issue.
Rick
Edit: more info
28 San Disks, each with 8 paths (= 224)
On the 30 minute mark (224 "Illegal Request"), so something is checking each path, and I'm trying to determine what it is.  I'm running all root crons while monitoring kstat in real time, all errors happen over 10 to 15 seconds time period.  No cron correlates.

Similar Messages

  • CUCM 10.5 - CSR/SAN issue

    Just upgraded this past weekend to CUCM 10.5.  Today, I'm trying to generate updated certs for the servers per our security policy.  What i'm finding, though, is that the old way of adding additional SANs doesn't seem to work anymore.  We use Active Directory's CA, and usually I would just populate the SANs in the additional attributes field.  Doing that now, though, prevents the cert from being accepted at all.  Using the multi-server option, it will only take SANs that have the FQDN, which is great, but we also use just the hostname to access servers as well...
    Has anyone come across a good way to address this kind of issue?

    Thanks Chris.  I've tried that, and it's partially working.  Where I'm stuck is having just hostnames as well as FQDNs.  The TAC engineer I'm working with mentioned attaching a list of separate hostnames to the CSR request (doing a multi-server CSR).  I'm wondering, though, do I just attach the .txt file and generate the CSR or do I need to do anything else to make sure that it populates the hostnames in the text file as well as the automatically discovered ones?

  • SAN issue

    All,
    Our shop has recently installed a HP EVA 5000 and made available a LUN of 250gb for use on our 3 node RAC.
    The purpose is to setup a single point repository of files that are shared by all (e.g. .sql files). The LUN appears
    to all 3 nodes on the RAC; however, if I make a change on one node (e.g add a file) the change fails to
    appear on the LUN from the other 2 nodes.
    I have looked around the web and have found postings regarding the masking of the SAN, but nothing specific
    to this problem.
    Was wondering if anyone has had/seen similar problems and or have an idea of a solution work-around.
    V/R,
    Gerry

    Hi,
    Once contact HP for support they will help you in this type of problems. Since i was also facing the same type of problem with HP. They have installed a rpm package in the linux box which has synchronized with LUN.
    other problem is everything in RAC setup correctly once check since one small error will give this error also.
    Thanks

  • InDesign font issue - displaying spaces as "D" using Fontin Sans

    I've been having a strange font problem in InDesign CS3 for a couple days. When opening docs using ex libris' font "Fontin Sans" (in various weights) ID is messing with the formatting and instead of spaces is displaying a capital "D". I figured it was a corrupt font, but I deleted and reinstalled the font and it works fine in other programs. The issue isn't 100%, either. Yesterday I switched to another font and then switched back to Fontin Sans, and it worked fine, for a time. I've been using this font for a while in a variety of projects without issue.
    Besides that it works in other programs, another reason I suspect this isn't a specific issue with this font is that when I opened a couple web pages yesterday (NY Times and my company's own home page) much of the plain text was completely garbled (symbols and such, no recognizable words), and that certainly wasn't a Fontin Sans issue. The two things seem unrelated, but as they both started happening yesterday I wonder if they're connected. The web text-display issue seems to have gone away today, however, while the Fontin/ID issue remains.
    I don't even know where to start on this one. I'm using CS3, Mac Pro Intel, OS 10.4. No updates in the past couple days on font managers or CS3.

    Delete your font cache. Download a free font cache cleaner utility. I have font finagler which may or may not be free. Can't recall.

  • N97 mini - niggling issues I want fixed but can't ...

    Is there some kind of nokia hotline I can call that isn't going to cost me?
    I am facing a number of niggling little issues (Some i have posted on here with no reply). Biggest ones:
    - Music purchased from Ovi music. If i leave the song on the music widget for a while then come back later and play it, it doesn't play. I have to select a different song then go back and select the song i want to listen to.
    -The phone beeps every now and then. Not often enough to be able to time it but often enough to know it's happening. Could this be the exchange server updating maybe? I don't know cause there's noone to ask. This is particularly noticable when listening to podcasts as the sound on the podcast goes down, phone beeps then the sound comes back up.
    These are the two main issues I am having at the moment and would appreciate any advice anyone has.
    My other problems are Ovi Music software related (the pc one) so I will post those elsewhere. It seems hard to justify to everyone around me why I have a nokia and not an iphone or something. I really don't like the whole iphone thing but all these little issues are things that you don't get on other phones. Seems that nokia's have great hardware but really **bleep** software. Or is that just me?
    Thanks,
    J

    * beep * I have the same problem, already mailed Nokia support team and have no problem with ovi things because I don't use it. I also have no new applications installed or open programs, I rarely use the music features of the phone and only have Sygic Mobile maps 10 installed, but the beep does not come from that program since it is only in use while navigating and switched off to save battery life, no Bluetooth enabled, no gps in use, latest firmware according the SW-upgrade service. The problem came "out of the blue" and no other software is installed. Tried 2 different providers because I also own a HTC Touch Pro 2 with Hi and Telfort so swapped the sim cards but the problem still occurs. The beep also occurs when another profile like "silent" is in use. I am quite frustrated about it since it interrupts me my sleeping hours constantly. When nokia answers my question, I will translate it and post it here since I am from the Netherlands and probably will get a response in Dutch.

  • Write Performanc Issues.

    Hi
    Oracle : 11.2.0.2.0 EE.
    Linux : Rhel 5.6.
    Dell R720 :
    EMC VNX SAN Storage :
    Here is the scenario:
    The dell server is connected to the SAN . We have 2 instances on this same server. When we create a tablespace on one instance it takes twice as long as creating the tablespace on the 2nd instance. The tablespaces are created on the san.
    At this point we have eliminated any hardware/ SAN issues. Since both databases are on the same server and hardware, connecting to the same lun on the san.
    Any help would be appreciated.
    Below is the relative information of the creation scripts and init.ora parameters.
    Tablesapace creation Script for ORCL1 (Notice the 8k block size corresponding to the db_block_size on the db ).
    CREATE BIGFILE TABLESPACE TEST_IO BLOCKSIZE 8k LOGGING
    DATAFILE '/uP01/oracle/oradata/orcl1/TEST_IO_01.dbf'
    SIZE 256G AUTOEXTEND OFF
    EXTENT MANAGEMENT LOCAL UNIFORM SIZE 256K
    SEGMENT SPACE MANAGEMENT AUTO
    PERMANENT ONLINE;
    14 minutes to create
    Tablespace creation script for orcl 2 (Notice the 32k block size corresponding to the db_block_size on the db ).
    CREATE BIGFILE TABLESPACE TEST_IO BLOCKSIZE 32k LOGGING
    DATAFILE '/uP01/oracle/oradata/orcl2/TEST_IO_01.dbf'
    SIZE 256G AUTOEXTEND OFF
    EXTENT MANAGEMENT LOCAL UNIFORM SIZE 256K
    SEGMENT SPACE MANAGEMENT AUTO
    PERMANENT ONLINE;
    30minutes to create
    Init.ora on orcl1 :
    orcl1.__db_cache_size=494927872
    orcl1.__java_pool_size=4194304
    orcl1.__large_pool_size=4194304
    orcl1.__oracle_base='/opt/app/oracle/ora11g'#ORACLE_BASE set from environment
    orcl1.__pga_aggregate_target=536870912
    orcl1.__sga_target=1073741824
    orcl1.__shared_io_pool_size=0
    orcl1.__shared_pool_size=515899392
    orcl1.__streams_pool_size=25165824
    *._optimizer_extend_jppd_view_types=FALSE
    *._optimizer_group_by_placement=FALSE
    *._replace_virtual_columns=FALSE
    *.audit_file_dest='/opt/app/oracle/ora11g/admin/orcl1/adump'
    *.audit_trail='none'
    *.compatible='11.1.0.0.0'
    *.db_block_size=8192
    *.db_domain=''
    *.db_name='orcl1'
    *.diagnostic_dest='/opt/app/oracle/ora11g'
    *.local_listener=''
    *.open_cursors=1000
    *.pga_aggregate_target=536870912
    *.processes=500
    *.recyclebin='OFF'
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sec_case_sensitive_logon=FALSE
    *.sessions=555
    *.sga_target=1073741824
    *.undo_tablespace='UNDOTBS1'
    orcl2
    *.db_block_size=32768
    *.db_cache_size=23068672000
    *.db_domain=''
    *.db_file_multiblock_read_count=32
    *.db_files=10000
    *.db_flashback_retention_target=0
    *.db_keep_cache_size=134217728
    *.db_name='orcl2'
    *.db_unique_name='ORCL2'
    *.diagnostic_dest='/opt/app/oracle/ora11g'
    *.fast_start_mttr_target=1200
    *.filesystemio_options='SETALL'
    *.java_pool_size=268435456
    *.job_queue_processes=4
    *.large_pool_size=134217728
    *.log_buffer=134217728
    *.log_checkpoint_timeout=0
    *.log_checkpoints_to_alert=TRUE
    *.open_cursors=1000
    *.pga_aggregate_target=31457280000
    *.processes=300
    *.query_rewrite_integrity='STALE_TOLERATED'
    *.recyclebin='OFF'
    *.remote_login_passwordfile='EXCLUSIVE'
    *.resource_limit=TRUE
    *.resumable_timeout=7200
    *.sessions=335
    *.shared_pool_reserved_size=134217728
    *.shared_pool_size=2147483648
    *.star_transformation_enabled='TRUE'
    *.trace_enabled=FALSE
    *.undo_retention=36000
    *.undo_tablespace='UNDOTBS1'

    >
    The dell server is connected to the SAN . We have 2 instances on this same server. When we create a tablespace on one instance it takes twice as long as creating the tablespace on the 2nd instance. The tablespaces are created on the san.
    At this point we have eliminated any hardware/ SAN issues. Since both databases are on the same server and hardware, connecting to the same lun on the san.
    >
    If those are really the config paramters for your instances then you may have some serious configuration issues.
    Instance 1
    1. you have shown SGA_TARGET is set (implying Automatice Shared Memory Management) but then you show the individual components have also been set: SHARED_POOL_SIZE, LARGE_POOL_SIZE, etc.
    Which are trying to use manual or automatic shared memory management?
    2. SGA_TARGET is set to only 1 GB. Why so small?
    Instance 2
    1. You have NOT set DB_nK_CACHE_SIZE but say you want to use 32 KB block size. This parameter MUST BE SET.
    2. You have set DB_CACHE_SIZE to 23 GB. Why? Your block size is 32 KB that you want to use. Only the SYSTEM tablespace will use the standard 8K blocks so why the enormous cache? You only used a cache of .5 GB and total memory target of 1 GB for instance 1.
    3. SGA_TARGET is NOT set - why not? why are you not using the same memory management as what you tried to do in instance 1.
    What is going on with these two instances that their config is so radically different?
    Daniel suggested a possible issue with the 32 KB block size. My hypothesis is that your configuration for both instances is faulty and that for instance 2 in particular your failure to provide any setting for DB_nK_CACHE_SIZE is a likely suspect for causing the issue.
    Review the DBA Guide for how to configure memory and the requirements for using non-standard block sizes.
    I would expect that once you have your instances configured properly you won't have the problem you reported.
    http://docs.oracle.com/cd/E11882_01/server.112/e25494/memory004.htm
    About Automatic Shared Memory Management
    >
    Automatic Shared Memory Management simplifies SGA memory management. You specify the total amount of SGA memory available to an instance using the SGA_TARGET initialization parameter and Oracle Database automatically distributes this memory among the various SGA components to ensure the most effective memory utilization.
    >
    See the example for automatic management
    >
    You can take advantage of automatic shared memory management by setting Total SGA Size to 992M in Oracle Enterprise Manager, or by issuing the following statements:
    ALTER SYSTEM SET SGA_TARGET = 992M;
    ALTER SYSTEM SET SHARED_POOL_SIZE = 0;
    ALTER SYSTEM SET LARGE_POOL_SIZE = 0;
    ALTER SYSTEM SET JAVA_POOL_SIZE = 0;
    ALTER SYSTEM SET DB_CACHE_SIZE = 0;
    ALTER SYSTEM SET STREAMS_POOL_SIZE = 0;
    where 992M = 1200M minus 208M.
    >
    For non-standard block sizes see the section 'Setting the Buffer Cache Initialization Parameters'
    >
    Oracle Database supports multiple block sizes in a database. If you create tablespaces with nonstandard block sizes, you must configure nonstandard block size buffers to accommodate these tablespaces. The standard block size is used for the SYSTEM tablespace. You specify the standard block size by setting the initialization parameter DB_BLOCK_SIZE. Legitimate values are from 2K to 32K.
    If you intend to use multiple block sizes in your database, you must have the DB_CACHE_SIZE and at least one DB_nK_CACHE_SIZE parameter set. Oracle Database assigns an appropriate default value to the DB_CACHE_SIZE parameter, but the DB_nK_CACHE_SIZE parameters default to 0, and no additional block size caches are configured.
    The sizes and numbers of nonstandard block size buffers are specified by the following parameters:
    DB_2K_CACHE_SIZE
    DB_4K_CACHE_SIZE
    DB_8K_CACHE_SIZE
    DB_16K_CACHE_SIZE
    DB_32K_CACHE_SIZE
    Each parameter specifies the size of the cache for the corresponding block size.
    >
    Don't forget - when using non-standard block sizes you MUST set both the standard cache parameter (8k) and the non-standard parameter (32k for your use case).

  • Final Cut X read "The selected SAN location is already in use". Now will not even open. Any advice would be really grateful.

    Initially while working on two projects I began to run out of disk space on the external hard drive. Soon I was given this message:"The selected SAN location is already in use" To try and remedy the situation I decided to move one project to another harddrive that I use for another project. Initially it then showed nothing(projects and events) on either harddrive after that. Now FCX won't even boot up at all. Please help. I'm approching deadlines on all of these projects.
    Thank you for any advise.

    RE : "The selected SAN location is already in use"
    I spent 2 hours with Apple technical support and  even did a remote session as they looked at my desk top live.
    1st the bad news, there is no real solution nor is there an answer as to why this happens. The SAN is supposed to only work through fiber optics and is strictly designed for multiple users accessing the same drives … It should only be activated manually and only through fiber optics . No one at Apple could understand why this is happening, why did my thunderbolt disk appear to be on a SAN and unexplainably unable to be accessed by my computer because it was being accessed by my computer. A conundrum indeed
    The semi good news is that I was able to open the original session eventually.
    Here is the workaround… And it ain't pretty .
    If you can access the disk on the desktop copy the whole content of the disk to another drive , this is because you will have to reformat the drive which is giving you the problems.
    Format the original problematic drive and after you have cleared all of the user preferences for Final Cut X relaunch Final Cut.
    Final Cut should now see the drive without any of the SAN issues.
    Start a new event on the original reformatted drive and give it a somewhat generic name so as not to confuse it with the original event that went bad.
    Reimport into this new fresh event all of the original media ( video/audio/graphic ) files and copy them to the reformatted drive.
    Do not start a new project related to this new event
    Shutdown Final Cut X
    Now go to the drive in which you have copied all of your original files and copy the original problematic Final Cut Projects folder right next to the Final Cut Events folder at the root level of the newly formatted drive.
    The new drive should only have 2 folders in it now, one is the new Final Cut Events folder that you created by copying over all of original media  and the 2nd one should be the Final Cut Projects folder which you just copied back from the backup drive.
    I only had to copy the one project which started giving me problems with the SAN issue, because it was the only FCX on my drive at the time therefore I kept intact the original Final Cut folder hierarchy, theoretically even if you have many different projects inside that Final Cut Projects folder it should work, but if that doesn't work try only copying the problematic Final Cut Project.
    Start Final Cut again and it should read everything except that the media will be off-line ( unlinked ) and you should be able to re-link all of the media if the files had not change names somewhere along the process.
    The one caveat about this is that my edits were off against the music. It almost seems like there was some kind of format issue. Something was lost in translation . My project was only 2 min. long and it had about 50 edits. The 1st three edits were completely off after that it was hard to tell ,some would be on some would be off.
    Now I was going on about 2 hours of sleep so i am a bit blurry but what I am certain is that the 1st few edits were 100% off by at least 60 frames.
    Again this may have just worked in my instance and it may not be applicable every time this problem pops up.
    BTW whenever I connected an older drive everything worked fine there were no SAN issues.
    I am on an iMac 27 inch mid 2011, 3.4 GHz Intel core i7 with 16 GB of RAM
    running Mac OS X version 10.7.4
    and Final Cut X 10.0.4 and
    my drive is a Western Digital my book thunderbolt Duo 4 TB formatted in RAID 0
    as a final note I tried everything I read on every forum before calling Apple.
    Hope this helps someone
    Giorgio
    new root level drive look

  • UNRESOLVED CAPTIVATE ISSUES

    In response to a question by Rod Ward in another thread, herewith the unresolved issues with Captivate:
    - Programme not listed in "uninstall list" - that is in Windows Cotrol Panel
    - The Jigsaw widget just hangs when puzzle is not built in time
    - When using Drop and Drag and or Jigsaw widget in project with multiple attemps the widgets "fall over" the moment you add a quiz somehere in the project
    - The Notes widget does not work when you publish to WIndows exe. (Even Adobe experts online could not solve this - several hours with control of my PC)
    - For some reason in some projects the background audio keeps looping, even when turned of 9Again experts could not solve)
    - Randomly, projects will publish with images, smartshapes etc. all filled with red. Then magicly, if you try enough times, it publishes ok.
    - The software hangs (freezes) OFTEN. This could be while working in a project, when running a preview, anything. This has been an issue since my Captivate 5.5.
    - Background audio volume control is inconsistent
    - Fill and other formatting tools are PATHETIC when compared to Microsoft Office and even smaller software houses like Serif.
    - Captivate is VERY limited as to the format of files that it can use. Compare with Corel.
    - Buttons (and some other tools) do not behave consistently in publishing vs previewing. this EATS UP MY WORK TIME!! Again pathetic when compared to software such as Serif.
    - The package as a whole seems "unstable" (Hanging, inconsistent behaviour) Remainds me of stuff we used in the eighties.
    - Captivate is VERY resource intensive.
    - It is "slow" software. Takes ages to do previews etc. Compare with good o'll powerpoint,
    - Rediculously expensive for what it does.
    - The turorials provided by Adobe is rather weak. it always shows things like widgets in the simplest application imagineable. Then I have to spend HOURS AND DAYS trying to figure out all the pitfalls associated with tools in a slightly more advanced use.
    - I have worked on CAD, Webdesign, graphics, video editing, ERP, MRP, office (I.e, MS Office,Lotus, Multimate) and learning software. I do some coding work in visual basic. I teach software applications. Adobe Captivate is most difficult to get to know. It is riddled with pitfalls. I spend about 30% of my time to figure out the software.
    In case you wonder, I use an i5 computer with four gigs of ram.
    I process MASSIVE videos in Corel which NEVER HANGS.
    And then there are many niggling little issues that I have learnt to work around!!

    Hi THYSG333,
           Sorry for the inconveneince. I will try to respond to some of the them(added in-line). There are couple of issues, I need sample projects. You can upload them to workspaces.acrobat.com and send me a shared link.  You can also contact me at sankaram at adobe dot com for discussing the issues.
    -Programme not listed in "uninstall list" - that is in Windows Control Panel
    Sankaram: It seems to be an issue with cleaning up. Can you try with creative suite cleaner to remove Captivate 7 and install it again? Creative suite cleaner is available at http://www.adobe.com/support/contact/cscleanertool.html
    - The Jigsaw widget just hangs when puzzle is not built in time
    Sankaram: When Jigsaw puzzle is not built in time, we have seen movie pauses (NOT hang). Let us know if you observe the same.
    - When using Drop and Drag and or Jigsaw widget in project with multiple attempts the widgets "fall over" the moment you add a quiz somewhere in the project
    Sankaram: Jigsaw widget : This is a bug in Jigsaw widget.
    Drag and Drop : Are you using Captivate 7 in-built Drag and Drop or any widget? 
    - The Notes widget does not work when you publish to WIndows exe. (Even Adobe experts on-line could not solve this - several hours with control of my PC)
    Sankaram: This is a bug in Notes Widget.
    - For some reason in some projects the background audio keeps looping, even when turned of 9Again experts could not solve)
    Sankaram: This is a bug in Captivate.
    - Randomly, projects will publish with images, smartshapes etc. all filled with red. Then magicly, if you try enough times, it publishes ok.
    Sankaram: Can you try “Force Publish” option? Let us know if you still encountering the issue after this.
    - The software hangs (freezes) OFTEN. This could be while working in a project, when running a preview, anything. This has been an issue since my Captivate 5.5.
    - The package as a whole seems "unstable" (Hanging, inconsistent behaviour) Remainds me of stuff we used in the eighties.
    Sankaram: Can you provide some set of steps or scenarios for us to investigate.
    - Background audio volume control is inconsistent
    Sankaram: It would be great if you could send the sample project for us to investigate.
    - Buttons (and some other tools) do not behave consistently in publishing vs previewing. this EATS UP MY WORK TIME!! Again pathetic when compared to software such as Serif.
    Sankaram: Could you please share the project?
    - Captivate is VERY limited as to the format of files that it can use. Compare with Corel.
    Sankaram : May I know what all additional formats you are looking for and why?
    Thanks,
    Sankaram
    Captivate Engg Team.

  • [scanner] sane: every second scan fails

    Hi everybody,
    I got a Canon MP510. I'm already in the group 'scanner' and using sane to scan images:
    scanimage --format=tiff >image.tiff
    scanimage -L shows:
    device `pixma:05B43737_53450F' is a CANON Canon PIXMA MP510 multi-function peripheral
    Every second time I get an error:
    scanimage: sane_read: Error during device I/O
    So i have to scan everything twice.
    dmesg says:
    usb 1-3: usbfs: interface 1 claimed by usblp while 'scanimage' sets config #1
    'rmmod usblp' has no effect. /etc/sane.d/canon.conf seems to be ok.
    so what is the problem?

    Once I had a issue like yours, so I decided to go merciless against dll.conf. I deleted every single line which wasn't required by my scanner and problem solved.
    p.s. check sane's man page about SANE_DEBUG env variable. It will help a lot debugging sane issues.
    Luck

  • Transaction Log File Drive is missing from SAN

    HI all,, we had some SAN issues and we dont have Transaction Log files for some databases..
    This is SQL Server 2008 R2 Cluster.. The drive which was holding this Tlog files were missing.. PLease let me know, how to bring back databases.. Awaiting for early reply..

    As others have said, the SAN folks needs to get their act together and bring back disk with the log files.
    If the log files are truly lost, you should restore a clean backup. If you don't have a clean backup, well, there are some people in your company that are likely to ask you some questions of what is going on the data centre.
    It was suggested that you should detach the data file and reattach it to have a new log file created. I strongly recommend against this. You will get a database that is likely to have corruption and inconsistency on both SQL Server level and application,
    due to transactions that were in flight when the log files were lost.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • X201 - Random standby, blank/bright screen and left side USB issue

    I have a bunch of issues with my X201 that slowly developed over the past months. The motherboard had been replaced once due to USB/standby issues and worked fine for a few months, but now it seems to be returning again.
    I'm not sure if this is again a motherboard issue or something else altogether. In rough sequence of when the faults started appearing.
    Initial symptoms: the left side yellow USB ports will stop working after returning from standby/hibernation. Windows 7 gives a USB device error and basically everything connected to that same USB controller won't work. Sometimes it recovers by itself after  a few more standby/wake cycles, sometimes if I disable/enable the USB controller in Device Manager but often nothing works except a shutdown and cold boot.
    Then the screen started going extra bright or more accurate high gamma white-out when returning from standby. Everything becomes very whitish/bright but slowly will settle down to normal within 30 seconds or so.
    I could live with all these niggling small issues, but recently my X201 started going crazy with random drops into STANDBY. I know there is a thread about random shutdown but that isn't the issue here.
    Just out of the blue, connected to power with 100% battery, the X201 will go into standby. Sometimes it will wake almost immediately by itself, then go back into standby again. Other times, I have to hit the Fn button to wake it, but frequently it will just go right back to standby again. The number of times this happen before it stabilized again is random. Reboot may or may not fix it the issue. Like a hiccup, it starts without reason and stops without reason, but makes it extremely annoying and impossible to work when it does.
    The worst part is that often when returning from standby, the screen may go blank, even the backlight is off. Sometimes this happen immediately on return, but sometimes it takes a few seconds before the screen dies. During the blank screen, the system is responsive. I could type my login password blind and see hard disk activity and hear the login music.
    According to the Windows event log, this appears to be triggered by some phantom power button or Lid close event.
    Would anybody have a conclusive idea of what might be the root cause of these problems?
    I can't really have a downtime without my laptop for days sending it for repairs and hope to be able to isolate the part so that I can just call in on parts warranty and get it fixed in a few hours.

    I had a similar problem with my X201 after hibernation due to no AC power and battery depletion.
    I was loathe to replace the motherboard.
    Then one day I tried this. Went into BIOS. Disabled, then enabled various  ports under supervisor password control. (I heard don't use extended password which encrypts. use the simple 8 character one).
    I had disabled the usb card reader all usb ports. I had taken off the hard drive password I had put on earlier.
    Then after a power off cycle.  I enabled the usb card reader and all the usb ports again.
    Then like some miracle windows detected the SD card reader and re-installed the usb root hub drivers at the same time. For some reason it seems the SD card reader is tied to the usb root hub. Maybe it's already part of it already.
    Then all the ports were working again. I was so relieved that I didn' have to call Lenovo to replace my MB. I didn't even have to update by flashing the BIOS.
    It's all ok now. I like the X201 and the keyboard. From what i've seen of the X220 they've abandoned the traditional 90's style keyboard this time. I should have bought an extra X201 when Lenovo put it on special pricing bakc in 2011. Now they are completely gone from Lenovo and the prices on the major resellers of Lenovo are way too high for all the Lenovo notebooks too.

  • Restored standby database from primary; now no logs are shipped

    Hi
    We recently had a major network/SAN issue and had to restore our standby database from a backup of the primary. To do this, we restored the database to the standby, created a standby controlfile on the primary, copied this across to the control file locations and started in standby recover and applied the logs manually/registered to get it back up to speed.
    However, no new logs are being shipped across from the primary.
    Have we missed a step somewhere?
    One thing we've noticed is that there is no RFS process running on the standby:
    SQL> SELECT PROCESS, CLIENT_PROCESS, SEQUENCE#, STATUS FROM V$MANAGED_STANDBY;
    PROCESS CLIENT_P SEQUENCE# STATUS
    ARCH ARCH 0 CONNECTED
    ARCH ARCH 0 CONNECTED
    MRP0 N/A 100057 WAIT_FOR_LOG
    How do we start this? Or will it only show if the arc1 process on the primary is sending files?
    The arc1 process is showing at OS level on the primary but I'm wondering if its faulty somehow?
    There are NO errors in the alert logs in the primary or the standby. There's not even the normal FAL gap sequence type error - in the standby it's just saying 'waiting for log' and a number from ages ago. It's like the primary isn't even talking to the standby. The listener is up and running ok though...
    What else can we check/do?
    If we manually copy across files and do an 'alter database register' then they are applied to the standby without issue; there's just no automatic log shipping going on...
    Thanks
    Ross

    Hi all
    Many thanks for all the responses.
    The database is 10.2.0.2.0, on AIX 6.
    I believe the password files are ok; we've had issues previously and this is always flagged in the alert log on the primary - not the case here.
    Not set to DEFER on primary; log_archive_dest_2 is set to service="STBY_PHP" optional delay=720 reopen=30 and log_archive_dest_state_2 is set to ENABLE.
    I ran those troubleshooting scripts, info from standby:
    SQL> @troubleshoot
    NAME DISPLAY_VALUE
    db_file_name_convert
    db_name PHP
    db_unique_name PHP
    dg_broker_config_file1 /oracle/PHP/102_64/dbs/dr1PHP.dat
    dg_broker_config_file2 /oracle/PHP/102_64/dbs/dr2PHP.dat
    dg_broker_start FALSE
    fal_client STBY_PHP
    fal_server PHP
    local_listener
    log_archive_config
    log_archive_dest_2 service=STBY_PHP optional delay=30 reopen=30
    log_archive_dest_state_2 DEFER
    log_archive_max_processes 2
    log_file_name_convert
    remote_login_passwordfile EXCLUSIVE
    standby_archive_dest /oracle/PHP/oraarch/PHParch
    standby_file_management AUTO
    NAME DB_UNIQUE_NAME PROTECTION_MODE DATABASE_R OPEN_MODE
    PHP PHP MAXIMUM PERFORM PHYSICAL S MOUNTED
    ANCE TANDBY
    THREAD# MAX(SEQUENCE#)
    1 100149
    PROCESS STATUS THREAD# SEQUENCE#
    ARCH CONNECTED 0 0
    ARCH CONNECTED 0 0
    MRP0 WAIT_FOR_LOG 1 100150
    NAME VALUE UNIT TIME_COMPUTED
    apply finish time day(2) to second(1) interval
    apply lag day(2) to second(0) interval
    estimated startup time 8 second
    standby has been open N
    transport lag day(2) to second(0) interval
    NAME Size MB Used MB
    0 0
    On the primary, the script has froze!! How long should it take? Got as far as this:
    SQL> @troubleshoot
    NAME DISPLAY_VALUE
    db_file_name_convert
    db_name PHP
    db_unique_name PHP
    dg_broker_config_file1 /oracle/PHP/102_64/dbs/dr1PHP.dat
    dg_broker_config_file2 /oracle/PHP/102_64/dbs/dr2PHP.dat
    dg_broker_start FALSE
    fal_client STBY_R1P
    fal_server R1P
    local_listener
    log_archive_config
    log_archive_dest_2 service="STBY_PHP" optional delay=720 reopen=30
    log_archive_dest_state_2 ENABLE
    log_archive_max_processes 2
    log_file_name_convert
    remote_login_passwordfile EXCLUSIVE
    standby_archive_dest /oracle/PHP/oraarch/PHParch
    standby_file_management AUTO
    NAME DB_UNIQUE_NAME PROTECTION_MODE DATABASE_R OPEN_MODE SWITCHOVER_STATUS
    PHP PHP MAXIMUM PERFORMANCE PRIMARY READ WRITE SESSIONS ACTIVE
    THREAD# MAX(SEQUENCE#)
    1 100206
    NOW - before you say it - :) - yes, I'm aware that fal_client as STBY_R1P and fal_server as R1P are incorrect - should be PHP - but it looks like it's always been this way! Well, as least for the last 4 years where it's worked fine, as I found an old SP file and it still has R1P set in there...?!?
    Any ideas?
    Ross

  • VM server frequently crashes

    We have installed VM server 3.2.4 on a couple of servers and provisioned one RHEL 6 guest per physical. Every now and again the VM server appears to have a problem with multipathing and the attached LUNS go read only, so the guests crash. We did a default VM server install, with multipathing enabled, all our storage is SAN (emc vmax I think). When this last happened some other physical servers in the same enclosure had issues, basically they almost froze, but these servers did not crash or go read-only.  I've had an SR on this, but supports answer was that we had a switch or SAN issue, nothing helpful. Any suggestions as to what I should look at? Thanks.

    Personally, I would never run Oracle VM in production against an EMC array without using Powerpath. EMC arrays have long had issues with various linux distributions native multipathing.... Hence the existence of "EMC Powerpath".
    Do the other servers you mentioned use powerpath.?

  • Backup Failed in SQL server 2008

    Backup is failing for only one database on the server.
    Msg 3201, Level 16, State 1, Line 1
    Cannot open backup device 'F:\databasename_backup_2014_07_29.bak'.
     Operating system error 5(failed to retrieve text for this error. Reason: 15105).
    Msg 3013, Level 16, State 1, Line 1
    BACKUP DATABASE is terminating abnormally.
    ** SAME DATABASE BACKUP IS COMPLETED SUCCESSFULLY ON OTHER DRIVE **
    we have tried the below steps but still it is failed.
    Microsoft SQL Server 2008 R2 (SP2) - 10.50.3720.0 (X64)   Apr 27 2012 13:51:37   Copyright (c) Microsoft Corporation  
     Standard Edition (64-bit) on Windows NT 6.1 <X64> (Build 7601: Service Pack 1) (Hypervisor)  
    1.checked for any spaces is th database name --No space
    2.added service account to administrator .
    3.provided full access to backup folder.
    4.Compatibilty is 80 .
    5.created backup devide to same location and tried still it got failed.
    10 percent processed.
    20 percent processed.
    30 percent processed.
    40 percent processed.
    50 percent processed.
    60 percent processed.
    70 percent processed.
    Msg 3202, Level 16, State 1, Line 1
    Write on "F:\Data\Backup\New_Backup_databasename.bak" failed: 665(failed to retrieve text for this error. Reason: 15105)
    Msg 3013, Level 16, State 1, Line 1
    BACKUP DATABASE is terminating abnormally.
    6.Changed the database ower to service account,NO luck

    Hi,
    How big is database ? Does F drive has enough space on it. Issue does not seems with SQL Server backup it is with F drive or network connecting Server to drives.
    Can you give more information about F drive is it a SAN, Local disk ?
    If SQL server backup would have been issue it would have not processed uptill 70 % ( unles its bug). First error message with OS error 5 means access denied. Ater then you gave necessary rights and now backup issue is not due to SQL Server but while writing
    to F drive some unusual thing is happeing which is causing backup to fail.Can you also involve SAN team as well.
    ** SAME DATABASE BACKUP IS COMPLETED SUCCESSFULLY ON OTHER DRIVE **
    Does this mean backup of same database completes successfully on other drive if so it is a SAN issue or connection issue between SAN and server
    Is SQL server 2008 updated to latest service pack ? Why do you use 80 compatibility mode. Although these questions are not realted to backup
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • SQL 2012 AlwaysOn cluster IP not moving after failover, causing database to be read-only

    SQL Server Cluster Name: SQLDAG01
    SQL Server Cluster IP: 10.0.0.50
    Cluster Listener IP: 10.0.0.60
    Node 1 Name: SQL01
    Node 1 IP: 10.0.0.51
    Node 2 Name: SQL02
    Node 2 IP: 10.0.0.52
    Everything is fine when SQL01 is the primary. When failing over to SQL02, everything looks fine in the dashboard but for some reason the cluster IP, 10.0.0.50, is stuck on node 1. The databases are configured to provide secondary read access. When executing
    a query on SQLDAG01, I get an error that the database is in read-only mode. Connectivity tests verify that SQLDAG01, 10.0.0.50, connects to SQL01 even though SQL02 is now the primary.
    I've been Googling this for the better part of the day with no luck. Any suggestions? Is there a Powershell command force the cluster IP to move to the active node or something? Also I'm performing the failover as recommended, from Management Studio connected
    to the secondary node.

    This was the answer, it had been setup to use the cluster name instead of the application name. Whoever installed Sharepoint connected it to SBTSQLDAG01 instead of SHAREPOINT01. Once we changed Sharepoint to connect to SHAREPOINT01, the failover worked as
    expected. We did have a secondary issue with the ARP cache and had to install the hotfix from http://support.microsoft.com/kb/2582281 to resolve it. One of the Sharepoint app servers was failing to
    ping the SQL node after a failover, the ARP entry was stuck pointing to the previous node. This article actually helped a lot resolving that: http://blog.serverfault.com/2011/05/11/windows-2008-and-broken-arp/
    One thing I did notice is that the SQL failover wizard does not move cluster groups "Available Storage" and "Cluster Group", I had to move those through the command line after using the wizard. I'm going to provide the client with a Powershell script that
    moves all cluster groups when they need to do a manual failover. This also happens to be why the Sharepoint issue started, "Cluster Group" is what responds to the cluster name SBTSQLDAG01. Moving that group over to the node that has the active SQL cluster
    group also made it work properly, but using the application name is the correct method.
    Thanks everyone for all your help. Although the nitpicking about terminology really didn't help, that was a pointless argument and we really could have done without it. Yeah I know 2008 called is "Failover Cluster Manager" and MSCS is the "2003 term" but
    really, they're basically the same thing and we don't really need to derail the conversation because of it. Also, If you look at the screenshot below you can clearly see "AlwaysOn High Availability" in SQL Management Studio. That's what it's called in SQL,
    that's where you do all the work. Trying to tell me it's "not a feature" is wrong, pointless, and asinine, and doesn't get us anywhere.
    Sorry it took so long to get back, I was off the project for a couple weeks while they were resolving some SAN issues that caused the failover to happen in the first place.

Maybe you are looking for