Best way to search audit logs in cDOT

.. Title says it all.. i'm having trouble searching audit logs in CDOT... In 7-mode I used a powershell cmdlet, i'm not finding one in cdot

Hi, Please refer, file access and protocols Management Guide,
https://library.netapp.com/ecm/ecm_download_file/ECMP1196891  Thanks

Similar Messages

  • Best way to move redo log from one disk group to another in ASM?

    Hi All,
    Our db is 10.2.0.3 RAC db. And database servers are window 2003 server.
    We need to move more than 50 redo logs (some are regular and some are standby) which are not redundant from one disk group to another. Say we need to move from disk group 1 to 2. Here are the options we are thinking about but not sure which one is the best from easiness and safety prospective.
    Thank you very much for your help in advance.
    Shirley
    Option 1:
    1)     shutdown immediate
    2)     copy log files from disk group 1 to disk group2 using RMAN (need to research on this)
    3)     startup mount
    4)     alter database rename file ….
    5)     Open database open
    6)     delete the redo files from disk group 1 in ASM (how?)
    Option 2:
    1)     create a set of redo log groups in disk group 2
    2)     drop the redo log groups in disk group 1 when they are inactive and have been archived
    3)     delete the redo files associated with those dropped groups from disk group 1 (how?) (According to Oracle menu: when you drop the redo log group the operating system files are not deleted and you need to manually delete those files)
    Option 3:
    1)     create a set of redo members in disk group 2 for each redo log group in disk group 1
    2)     drop the redo log memebers in disk group 1
    3)     delete the redo files from disk group 1 associated with the dropped members

    Absolutely not, they are not even remotely similar concepts.
    OMF: Oracle Managed Files. It is an RDMBS feature, no matter what your storage technology is, Oracle will take care of file naming and location, you only have to define the size of a file, and in the case of a tablespace on an OMF DB Configuration you only need to issue a command similar to this:
    CREATE TABLESPACE <TSName>; So the OMF environment creates an autoextensible datafile at the predefined location with 100M by default as its initial size.
    On ASM it should only be required to specify '+DGroupName' as the datafile or redo log file argument so it can be fully managed by ASM.
    EMC. http://www.emc.com No further commens on it.
    ~ Madrid
    http://hrivera99.blogspot.com

  • What is the best way to search in a CLOB, in batch operation ?

    Hi All,
    My application runs batch for searching 'text' into clobs and accordingly inserts into another table-
    code snippet is like this -
    FORALL i IN 1 .. l_snme_sn_temp.COUNT
    INSERT INTO snme_rmanotes_out
    (report_request_id, serial_number, scrubbed_sn, product_id, pk1_value, last_update_date, event_date, event)
    SELECT pi_report_request_id
    , l_snme_sn_temp_orig (i)
    , l_snme_sn_temp (i)
    , '' product_id
    , pk1_value
    , TO_CHAR (last_update_date, 'DD-MON-YYYY') last_update_date
    , TO_CHAR (last_update_date, 'DD-MON-YYYY') event_date
    , 'IN RMA NOTES' event
    FROM c3_rma_notes
    WHERE contains (notes, l_snme_sn_temp (i)) > 0; **
    here - l_snme_sn_temp contains the text that the query searches ( also evident from the where clause, marked **), and is a variable of a TYPE.
    I am populating l_snme_sn_temp in a chunk of 500 +( using bulk collect limit 500)+ from a table having at least 1 million records +(and the upper cap will not be more than 1.5 M)+.
    The procedure that does this, takes 24+ hours to complete.
    If you guys have any better way to do this or any idea to tune to reduce time kindly suggest.
    I'd appreciate you kind response.
    regards,
    Raj Pandit

    Hi Herald,
    Thanks for writing.
    Oracle text index is there on 'notes' column. This table is part of petty big application of a huge company,
    so though I dont have access to it data dictionary tables, I can expect that the indexes are analyzed regularly.
    insert with HIT takes 2 mins
    insert with MISS takes less than 1 sec
    select with HIT takes minimum 2 mins and maximum 3 mins
    select with MISS takes less than a minute.
    It is expected that there would be more MISS then HITs ( HITs are not desirable)
    This source table is huge+(count in billions)+ and most of the clobs would be bigger. +(in most cases entire email conversation is stored in it)+.
    However there are many restriction on access, they have given us provision to see currently running query in a session ..
    when checked I found most of the time spent in executing insert+select only. As the entire query involves insert+select so I don't know which part is taking time.
    As much I know select + insert would be faster. The destination table+(having just one index)+ is merely used for this operation+(and have no other use in the application)+ only so there is no way insertion would be taking time. Also this procedure is just for insertion+(into that dest. table)+ no other operation is done
    Storage optimization is beyond my domain of discussion as we are just developers and organization is huge. There is no way DBA would listen to us unless there is an acute urgency hampering business operations. However in dev environment I analyzed the index but it didn't help.

  • What is the best way to search for tags in Finder?

    There are a lot of articles out there written about Tags in Mavs. All of them cover how to apply tags, but there is hardly anything written about searching and finding files.
    Now, does anyone here actually use tags in Finder? How do you search for them? Did anyone ditch folders or stytematic file-naming?

    I am currently using and experimenting with Tags. Searching, finding, and managing your tags is pretty straight forward. On the far left side of any Finder window, you have a selected number of Tags predetermined and an option for "all tags." If you click on this link, it will take you to a menu where ALL tags in use are listed. Change the name and color for them just as you would any other file name (I have a usb mouse, so it's a right click for me).
    If you used the "labels" feature (color-coding documents and folders), you have used tags; its the same idea, but now you define a word as what you are sorting by.
    To search for a tag, I either use the in-Finder search or Spotlight (which may be the same, but I am not sure--I am also a newer Mac convert and am still learning the ropes). I have started to organize my schoolwork solely by tagging it, but I have run into issues where I have "Exam 01" for one class and need to make a second "Exam 01" file for another class (I created different tags for each class).
    I will say that, since I started to only use the minimum amount of folders, I have noticed somewhat significant speed improvements in startup times and normal use. However, this coincided with the release of Mavericks so I do not know truly whether the folders help or not.
    One of the things I both love and hate about tags is that you can search across multiple folders (or drives in my case) for documents matching the tag in question. When I download classwork from our school's website, I tag with the appropriate class name, and then move to external hard drive (yeah, I am paranoid). Unfortunately, if I then don't delete the copy inside the "downloads" folder, the document will show up if I try to list all of my "Class" tags.
    Best of luck,
    ~Zach

  • Best way of deleting archive logs on standby when backup taken on primary

    Hello,
    I have a 11gr2 dataguard configuration without the broker.
    main problem in my particular situation arises from my FRA size on primary much more than on standby (90Gb vs 30Gb) and in some situations I'm constrained with FRA space on standby.... I'm going to align the FRAs but in the mean time....
    can anyone confirm that if I have configuration as in subject with maximum performance mode and I'm not using the broker I should consider a sort of "external" way for archive log deletion on the physical standby?
    In fact I'm setting on primary
    - retention policy = redundancy 1
    - archivelog deletion = applied on standby (without using any "backed up...." option)
    Considering the standby it seems from docs that:
    - retention policy always wins over deletion policy (so no archive eligible for deletion is deleted if retention is not satisfied)
    - I can set on standby archivelog deletion = applied on standby (without using any "backed up...." option)
    But it is not clear to me the effect of this config in case this is a standby itself and without any other cascade standby databases
    The manual says this has to be met:
    1) The archived redo log files have been applied to the required standby databases.
    ---> what does this mean in case the db is a standby.. met or not met as I have no other cascade standby dependant on this one?
    2) The logs are not needed by the BACKED UP ... TIMES TO DEVICE TYPE deletion policy. If the BACKED UP policy is not set, then this condition is always met.
    ---> I should be ok
    So the last question for standby is
    - what to set retention policy to?
    I have not understood if I set it to "NONE", this means no retention at all or retention for ever....
    sometimes I read that there is no policy; in other examples that files will be retained forever (because no file considered as obsolete).... or at least so I understood...
    I would like to have a sort of policy on standby where the archive logs can be deleted as soon as they are applied during the continuous media recovery phase...
    It seems I can't find a self contained set of configurations across the dataguard....
    Thanks in advance.
    Gianluca

    I agree with you.... It is a temporary contingency....
    I already stressed the fact that sizes between sites must be quite identical...
    we are going towards a production dataguard, migrating from a windows config (without it) to a Linux based config....
    but this contingency made me think about these considerations regarding policies and automatic dataguard operations
    I also stressed on equal size grow capability for both sites file systems, also for data tablespaces, so that for example in case of high fragmentation we have to avoid a limit situation where
    1) a file system full for DR datafiles causes abend of automatic media recovery
    2) this generates FRA growth in DR because archive logs are shipped but not applied any more
    3) FRA goes 100% in DR and primary cannot ship any more
    4) archive logs cannot be automatically deleted on primary because not already shipped
    5) FRA full on primary and DB hangs....
    Thanks anyway for input

  • Best way to acquire and log N number of samples at a specific repetition rate and write to file

    Hello,
    I am interested in acquiring N number of samples (e.g., 10000) at a specific acquistion rate (e.g., 1MHz), and writing this data to file - and repeating this acquisition/logging at a specific repetition rate (e.g., 50Hz).  Pretty straightforward I think, but I am unsure of how to go about doing this in order to maximize the repetition rate.     
    Any insight is appreciated!
    Thanks,
    Vic

    First thing's first: get the acquire and log working.
    Look in the example finder for a simple Analog Input (don't remember the exact name right now).  You will want to configure the task to be Finite Samples.  So you acquire N samples at Fs rate, read the data, and log it to a TDMS file.
    Once you have that working, put it inside of a While loop.  Use a wait inside of the loop to regulate how often you run this task.  You can actually setup the task before this close and close it out after the loop.  Same for the TDMS file.

  • Best way to search colors

    Hi there,
    Please some help with my problem.
    What I am doing is saving records in a sqlite db. Each record has some data and a color.
    In that color field i save 6 colors like this: 66ffff^0^66^0^ff00^0
    So on my screen i have 6 color pickers and i save the selected colors in the db with a special character between them. (In my case the ^)
    Now my problem;
    I want to create a search by color.
    But when i give the user a color picker.. and he/she selects red to search for, i have numerous color nuances that could define as red.
    something like this:
    f37575 = red
    f03737 = red
    a31a1a = red
    Is there a nice solution to get all items that have a red like color? (of course not only red, but lets say the main colors red, yellow, green, blue, purple)
    Or perhaps another solution to achieve this?
    please some help,
    Greets, jacob

    Colors are not subjective in a sense, but when you are dealing with numbers that introduce slight variations in colors, you are forced to more deeply understand how those colors are represented using the numbers.
    In this case, you need to determine what range of colors you consider red, blue, green, etc. because one person may say a given number represents red, but another may say at the far end of the spectrum, it is less like red and more like orange.
    Is the last VBox background in the following app red or orange?
    <?xml version="1.0" encoding="utf-8"?>
    <mx:Application xmlns:mx="http://www.adobe.com/2006/mxml">
      <mx:VBox width="100" height="100" backgroundColor="0xf37575"/>
      <mx:VBox width="100" height="100" backgroundColor="0xf03737"/>
      <mx:VBox width="100" height="100" backgroundColor="0xa31a1a"/>
      <mx:VBox width="100" height="100" backgroundColor="0xFF0000"/>
      <mx:VBox width="100" height="100" backgroundColor="#E86C2E"/>
    </mx:Application>

  • Does anyone know the best way to go about logging a formal complaint?

    I was treated deplorably by a customer service rep a week ago while trying to resolve a billing mistake made by Verizon. I was promised a call from a supervisor. I have yet to get that call. I've been w/ Verizon since 2009 & I think it's time to call it quits. This company has been a nightmare since day 1. I should have bailed years ago. I am not respected or valued. I'm sick of the run around & am hoping someone can point me in the right direction. Thanks!

    Michelle,
    Thanks for the response. Other than you, no one has reached out to me. I was given a partial credit. At this point I'm more concerned w/ the treatment I've received. I have shared this & no one is concerned w/ righting the wrong. I am currently looking into Sprint or T Mobile because they will buy out  my contracts.
    I appreciate the fact that you took time to contact me. Have a nice day.
    Monique

  • How to extract audit log data from every document library in site collection using powershell?

    Hi All,
    I have n number of document library in one site collection,
    My query is- How to extract audit log data from every document library in a site collection using powershell?
    Please give solution as soon as possible?

    Hi inguru,
    For SharePoint audit log data, These data combine together in site collection. So there is no easy way to extract audit log data for document library.
    As a workaround, you can export the site collection audit log data to a CSV file using PowerShell Command, then you can filter the document library audit log data in Excel.
    More information:
    SharePoint 2007 \ 2010 – PowerShell script to get SharePoint audit information:
    http://sharepointhivehints.wordpress.com/2014/04/30/sharepoint-2007-2010-powershell-script-to-get-sharepoint-audit-information/
    Best Regards
    Zhengyu Guo
    TechNet Community Support

  • Best Practice for Searching HTML Source

    Hello
    What is the best way to search the content of html files?
    These would be files I create.
    I have tried to use SpotLight but it's not up to it.
    The best thing I have found thus far is Multi-File Search in BBEdit.
    There I find my annotations under '/* NOTES:'
    Thanks

    Thanks for the response, but unfortunately the provided links are not much help:
    - The first link is about an article I don't have access to (i'm not a registered user)
    - The second link is about Integration Services. This is nice for Integration stuff, but I need to have a functionality within a frontend. 
    - The third link is for use in Excel.
    Maybe I'm looking for the wrong thing when wanting to create an extra column with "cleaned" up data. Maybe there's another solution from within my frontend or business layer, but I simply want a textbox on a form where users can type a search-value like
    "BAKKER". The result of the search should return names like "DEBACKER", "DE BEKKER", "BACKER", "BAKRE", ...
    I used to work in a hospital where they wrote their own SQL-function (on an Interbase database) to do this: They had a column with the original name, and a column with a converted name:
    => DEBACKER => Converted = DEBAKKER
    => DE BEKKER => Converted = DEBEKKER
    => BACKER => Converted = BAKKER
    => BAKRE => Converted = BAKKER
    When you searched for "BAKKER", you did a LIKE operation on the converted column ...
    What I am looking for is a good function to convert my data as above.
    Greetz,
    Tim

  • Search-MailboxAuditLog is empty - Mailbox Audit Logging not working in Exchange 2013 CU6 environment

    Hello,
    i activated Mailbox Audit Logging for Admin, delegate and owner with all supported operations (update, delete, etc..)
    like mentioned here:
    http://exchangeserverpro.com/using-exchange-server-2013-mailbox-audit-logging/
    But also two days later (and also one Server reboot later) search-MailboxAuditLog is still empty.
    any ideas how to fix this?
    Best,
    martin

    Hi S.Nithyanandham,
    i looked up the mailboxfolderstatistics. There are items in the folder: 
    [PS] C:\Windows\system32>Get-MailboxFolderStatistics mailboxname |where{$_.Name -like "*audit*"}
    RunspaceId : a95e32b8-93c3-4330-8d42-45cade9d64d4
    Date : 18.09.2014 16:35:20
    Name : Audits
    FolderPath : /Audits
    FolderId : LgAAAADmBpGVdb8iQp3F89WOcmcHAQBpQNFODkTESLeLj74B887wAAAAAAESAAAB
    FolderType : Audits
    ItemsInFolder : 147
    DeletedItemsInFolder : 0
    FolderSize : 434.2 KB (444,649 bytes)
    ItemsInFolderAndSubfolders : 147
    DeletedItemsInFolderAndSubfolders : 0
    FolderAndSubfolderSize : 434.2 KB (444,649 bytes)
    OldestItemReceivedDate :
    NewestItemReceivedDate :
    OldestDeletedItemReceivedDate :
    NewestDeletedItemReceivedDate :
    OldestItemLastModifiedDate :
    NewestItemLastModifiedDate :
    OldestDeletedItemLastModifiedDate :
    NewestDeletedItemLastModifiedDate :
    ManagedFolder :
    DeletePolicy :
    ArchivePolicy :
    TopSubject :
    TopSubjectSize : 0 B (0 bytes)
    TopSubjectCount : 0
    TopSubjectClass :
    TopSubjectPath :
    TopSubjectReceivedTime :
    TopSubjectFrom :
    TopClientInfoForSubject :
    TopClientInfoCountForSubject : 0
    SearchFolders :
    Identity : mailboxname\Audits
    IsValid : True
    ObjectState : New
    What do you think?
    why cant i search and find these entries the auditlog?
    best, 
    martin

  • Exchange 2013 SP1 - The attempt to search the administrator audit log failed.

    During migration process from Exchange 2010 to 2013, after moving Arbitration mailbox from Exchange 2010 database to Exchange 2013 SP1 database, cmdlet Search-AdminAuditLog fails with following error.
    The attempt to search the administrator audit log failed. Please try again later.
    + CategoryInfo : NotSpecified: (:) [Search-AdminAuditLog], AdminAuditLogSearchException
    + FullyQualifiedErrorId : [Server=EX2013,RequestId=517873e3-a623-4363-bfdc-e5aa23595c33,TimeStamp=29. 4. 2014
    8:38:37] [FailureCategory=Cmdlet-AdminAuditLogSearchException] 2774D0CF,Microsoft.Exchange.Management.SystemConfig
    urationTasks.SearchAdminAuditLog
    + PSComputerName : ex2013.domainname.local

    Hi,
    First, please make sure the Microsoft Exchange Search and the Microsoft Exchange Search Host Controller service are running and please run the get-mailbox -arbitration cmdlet to check the result.
    Besides, please check the properties of the DiscoverySearchMailbox and verify that the homeMDB attribute is set to a mounted database.
    If the steps above don't work, please try to re-create a new Discovery System Mailbox to check the result. You can refer to the following article.
    Re-Create the Discovery System Mailbox
    http://technet.microsoft.com/en-gb/library/gg588318(v=exchg.150).aspx
    Best regards,
    Belinda
    Belinda Ma
    TechNet Community Support

  • What is the best way to store and search 2D data

    Hi,
    There is a set of data (~10k records ) in 2D dimension.
    like this :
    Col 1, Col 2, Col3....
    What is the best way to store and search those records ?
    Thanks in advance
    Wilson

    Hi,
    Either userObjet[][] if you know how much data you have, and the data size is fixed, or use a list of lists. E.g. A Vector of Vectors (some will probably say that you should use an ArrayList instead, and that could be the case, but it sounds like you would want to display the data later on, and a DefaultTableModel (for JTables) uses a Vector as data holder).
    Kaj

  • What's the best way to check whether a user is logged in or not?

    I have a question about basic session handling.
    I'm running Tomcat 5.0.30 and have a web application where users can register with a username and password, and then log into a "member site".
    What is the best way of making sure that a user actually has logged in or not?
    What I've done in previous applications I've made is that I've just put a simple variable into the users session after he has successfully entered his password (i.e. Boolean loggedIn=true). Then I just test if this flag is true to grant him access to the member site. Is that a smart way of doing it?
    Are there any libraries I can use which handles sessions for me in a secure way?
    All comments and suggestions are appreciated!

    Maintain a flag using session attributes it to one when the user is logs in set it to value.... and change set it zero or invalidate that session... when the user is logged out...
    use something like this while user had logged in
    session.setAttribute("flag",<unique_number>);
    for checking whether the user had logged in or not...
    if(Integer.parseInt(session.getAttribute("flag").toString())=!<unique_number>){
    out.println("The session had expired");
    out.close();
    }

  • I need to know best way to edit MXF files, with least rendering, is that converting the MXF or Log and Transfer? the end product is dvd, but I have lots and lots of these projects to do.

    I am about to comence my Edit, and was told of Mpeg Streamclip, however this doesn't seem to work with my Canon XF305 Clips, the reason I was told to use it was so I wouldn't have to Render in Final Cut Pro all the time.. I do have the Canon Disc with the Plug in Utility and the update from their site.
    Is it best to Convert these MXF files first and import them into FCP or use 'Log and Transfer', so far I am unable to do either!!
    The projects on the timeline will have introduction .Movs already on it as a Template to re-use again and again with the new Footage ( MXF Files ) to be added and edited with transitions etc..This process is repeated again and again but with different MXF clips each time.
      I am looking for the quickest and best way for workflow as there are 80 or more projects to do, each with 10 or so clips to be added to the timeline and Exported for DVD...Will I have to then Compress to fit onto DVD's??
    I know Streamclip is an Mepeg 2 app but I thought the Canon XF305 Shot as mpeg2....
    Is Log and Transfer my only Option?
    Basically I need to know the quickest for Render purposes, and of course, the best way so the files aren't so huge etc etc.
    Please, any suggestions would be greatly received....

    We only work with PR422 files in FC7.  We use Shane's workflow with all 5D & MXF footage.
    Make your life easier and stick to the ProRes family for your footage and edit sequences. Avoid H264 & convert any you have to ProRes.
    Unless you are considering Blu-Ray the only way to deliver video on a DVD is to convert it to SD.
    Depending on the length of your show you can make a Blu-Ray in Compressor but you have very limited authoring options when compared to DVD Studio Pro.
    Compressor can also convert your HD to SD mpg-2 & Ac3 files. The best results usually come from setting the frame controls in the mpg-2 setting of Compressor to best. Depending on the the length of the show and your Qmaster settings this can take quite a while.
    I personally find Compressor's HD to SD DVD conversion varies. Sometimes it is wonderful and sometimes not so much. I prefer the following as it is usually is faster and very consistent.
    Export a quicktime movie (not QT conversion) of your final sequence
    Import the exported HD file back into FC.
    Create a SD 16x9 sequence and then drop in the recently exported HD seq and render. Rendering the sequence basically has FC doing the HD to SD conversion. Export the SD sequence and then convert to mpg-2 & ac3 in Compressor. While this does take time I still find it faster than Compressor with Qmaster utilizing all cores.
    Another option is using jobs in Compressor
    http://pixelcorps.cachefly.net/macbreak-154-540p-h264.mov
    In this workflow you would have Compressor convert the HD file to SD and then through jobs convert the SD file to mpg2 & Ac3. 
    In theory the jobs version could be faster (based on a number of variables- length, Qmaster, etc) but I haven't had time to compare the two. 

Maybe you are looking for