Event based monitoring for xmonad

I am using a conky + dzen2 status bar and xmonad DynamicLog hook for monitoring.
What I am looking for is something event based not like conky which runs all the time or at a given interval.
As an example, volume of the audio system. I want to display it continuously but I want to change the display only when it changes. Another example would be keyboard layout.
Any idea ?
regards,
Hardik
Last edited by rangalo (2010-05-31 20:37:26)

Hi,
i`ll recommend you post your Question in the Data Services Space.
Data Services and Data Quality
Regards
-Seb.

Similar Messages

  • How to find what are all the rules and events are monitoring for specific server in SCOM 2007

    how to find what are all the rules and events are monitoring for specific server in SCOM 2007.
    I need to know what are all the events, services, and rules are monitored for the specific server.
    Kindly help me friends.

    how to find what are all the rules and events are monitoring for specific server in SCOM 2007.
    I need to know what are all the events, services, and rules are monitored for the specific server.
    Kindly help me friends.
    Thanks for the question Sandoss. This is something that we all come across at sometime.
    Thanks & regards, Sumit Agrawal
    The lack of this feature is an inexcusable oversight for an enterprise management product.  They have some serious lightweights making design decisions on the SCOM team. 5 thumbs down. 
    BTW the answer is probably LOTS of stuff. Literally 100's of workflows are running on any server. 
    The following OpsMgr DB query will list all running monitors on a server. Change the name of @srv. I think this works pretty well.  I'd like some feedback if something is missing.
    DECLARE @srv varchar(30)
    SET @srv = 'your name here'
    SELECT mon.displayName as monitor, bme.FullName,bme.DisplayName as object,
    case
        when s.HealthState = 1 then 'healthy'
        when s.HealthState = 2 then 'Warning'
        when s.HealthState = 3 then 'Critical'
        else 'N/A'
    end as Health
    FROM state AS s WITH (NOLOCK)
    left join BaseManagedEntity as bme WITH (NOLOCK) on s.basemanagedentityid = bme.basemanagedentityid
    left join dbo.MonitorView Mon WITH (NOLOCK) on Mon.ID = s.monitorid
    where
    bme.FullName like '%' + @srv + '%'
    and s.HealthState <> 0
    and mon.IsInternalRollupMonitor = 0
    and mon.IsExternalRollupMonitor = 0
    order by bme.DisplayName, mon.displayName

  • Best Event-Based notifications for Sox Complience?

    We are looking for the best way to send notifications to various people to comply with sox controls.  For instance, we need to send certain data when and employee is terminated, hired, changes jobs, locations or supervisors.  We have considered Alerts, but they are very generic and cannot be formatted, unless someone has additional information.  Is there a way to format an alert message?
    Another option, possibly using Life Event Checklists, not sure if we can provide employee data in the notifications to satisfy the requirement.  Is there a way to pull in a message that contains employee specific data?
    All information is appreciated.
    Thank you,
    Patty

    Hi Patty,
    If you're using 12.1, you can try Checklists.
    Everything explained in detail :
    Understanding and Using Life Event Checklists in Oracle HRMS R12 [ID 740618.1]
    Another option is to use an audit tool( audit-trial or a third-party solution)
    You can audit all the data and then base your reports on the audit data.
    Audit Trial - http://oracle.anilpassi.com/audit-trail-in-oracle-hrms.html
    Third-party tools - http://www.caosys.com/csaudit.php
    Cheers,
    Vignesh

  • How to set a default start and/or end date for New Events based on trigger date.

    I'm using the CalendarActivityListener to get current row when clicking on an existing event. As per previous posts this listener gives you access to event detail including Start Date, End Date, etc.
    However, what I want to do is to default the start (and end) dates for New Events based on the trigger date.
    I've tried the CalendarListener and can grab the Trigger Date from it - however, I can't see a way to pass this directly to the popup/dialog I'm using to create the new event.
    At present I'm putting the TriggerDate into the ADFContext session scope e.g. ADFContext.getCurrent().getSessionScope().put("TriggerDate",calendarEvent.getTriggerDate());
    Then, I've tried multiple approaches to try and "get" the TriggerDate from session scope to drop it into my new Calendar Event basically, I'm trying to default the InputField(s) associated with the Start Date using the value from the session - I've tried
    1. setting the default value for the InputField in the jspx using a binding expression i.e. value="#{sessionScope.TriggerDate}" - this actually sets the value appropriately when the jspx is rendered but, when I go to create I get a NPE and I can't debug. I assumed that it might be a Date type issue - it would appear that CalendarListener provides a date of type java.util.Date and that the StartDate attribute of my VO/EO/table is a DATE and therefore requires oracle.jbo.domain.Date so I tried casting it - to no effect
    2. Using a Groovy expression *(StartDate==null?adf.context.sessionScope.TriggerDate:StartDate)* in my calendar's EventVO to default the Start Date to the same result
    Any thoughts or ideas?

    John,
    Thanks for that suggestion - could not get it to work. However, I did manage a different approach. I finally determined the sequence of events in terms of how the various events and listeners fire (I think).
    Basically, the CalendarActivityListener fires, followed by the listener associated with the Calendar object's Create facet, followed finally by the CalendarEventListener - the final is where the TriggerEvent is available and then finally, control is passed to the popup/dialog in the Create facet. So, my approach of trying to set/get the TriggerDate in the user's HTTP session was doomed to failure because it was being get before it had been set :(
    Anyway, I ended up adding a bit of code to the CalendarEvent listener - it grabs the current BindingContext, navigates through the DCBindingContainer to derive an Iterator for the ViewObject which drives the calendar and then grabs the currently active row. I then do a few tests to make sure we're working with a "new" row because I don't want to alter start & end dates associated with an existing calendar entry and then I define the Start and End dates to be the Trigger Date.
    Works just fine. Snippet from the listener follows
    BindingContext bindingContext = BindingContext.getCurrent();+
    *if ( bindingContext != null )    {*+
    DCBindingContainer dcBindings = (DCBindingContainer) bindingContext.getCurrentBindingsEntry();+
    DCIteratorBinding iterator = dcBindings.findIteratorBinding("EventsView1Iterator");+
    Row currentRow = iterator.getCurrentRow();+
    if ( currentRow.getAttribute("StartDate") == null)+
    currentRow.setAttribute("StartDate", calendarEvent.getTriggerDate());+
    if (currentRow.getAttribute("EndDate")==null)+
    currentRow.setAttribute("EndDate", calendarEvent.getTriggerDate());+
    *}*

  • Span-Based Monitor of IPPA Agents at remote site for UCCX 9X

    I understand to be able to monitor/record agents using IPPA (for UCCX), you need to utilize Span-based monitoring as opposed to Desktop Monitoring. According to the SRND, it also states that there should be no routing devices between agent phone (IPPA) and UCCX server which means no routers in between. Question I have is if I put the failover UCCX at the same location as the IPPA agents, would this work? I understand there might be an issue with redundancy with this scenario. Thoughts?

    It is possible without QM. When using desktop monitoring, the phone is configured to Span to PC port. The agent must be running CAD and the PC running it must be attached to the PC port on the phone. CAD sees the spanned packets from the phone and then forwards them to the recording service on the CCX server. The big gotcha is that the NIC in the PC must not discard 802.1q-tagged packets. If the NIC discards them instead of forwarding them into the NDIS stack, CAD will never see them. Sometimes this takes a Google search and a registry setting change to fix if it doesn't work at first.

  • Error handler for event based messaging framework

    I've been very interested in using the event based messaging framework (described here http://forums.ni.com/t5/LabVIEW/Community-Nugget-2009-03-13-An-Event-based-messageing-framework/td-p...) for my next large application.
    My main concern is the fact that it seems like typos would be very difficult to debug since you need to ignore unknown commands to make this system work.
    To solve this problem I've been considering the idea of having a single message error handler VI which will store all valid commands and check all sent commands to see if they are valid.  Each VI would send out a register message on startup with their name and all messages they can send and receive.  The message error handler would store these and then check all future messages to be sure it is a valid message, throwing an error if it is not.
    My basic problem is this: for this to work the message error handler VI would have to be started before any messages are sent so that it can capture all the register events.  If this is a VI that will be continuously running the entire application how can I ensure it starts first since I cannot wait for it to complete? (I.e. the usual method of running an error out wire or using a sequence structure will not work since everything will then wait for it to complete which will not happen until the program is ready to shut down)
    I'm assuming the answer might be to use an asynchronous call but I'm not very familiar with this method.  
    Any help is appreciated.  Thanks. 

    Could you just use the error handler as a subVI inside a case structure that is only called when you have new message to be checked? I'm not sure I understood the exact functionality you are looking for, so sorry if this does not apply.
    Zach P.
    Product Support Engineer | LabVIEW R&D | National Instruments

  • Rule based monitors (RZ20) - Show system availability for selected systems

    Hi experts,
    I'm new with CCMS monitoring and need an advice for rule based monitors.
    I want to create three alert monitors (RZ20). One for our developing systems, one for our quality assurance systems and one to monitor our production systems.
    For the start, I want to show data in each of this monitor for free space of the filesytems (MTE class: "FilesystemFreeSpace"), for users which are locked out (MTE class "R3SyslogSecurity") and for the availabilty of the ABAP systems (MTE class "Availability_SysPercent_ABAP").
    I started to create the monitor for our developing system and got a problem. I started with the rule "CCMS_DEFINE_R3_SYSTEMS" to select our developing systems "HRT", "CRT" and so on. Under this nodes, I defined the mte classes ("CCMS_GET_MTE_BY_CLASS") which needs to be shown: "FilesystemFreeSpace", "R3SyslogSecurity" and "Availability_SysPercent_ABAP".
    When I checked the results, I was able to see the free space of the filesystems and whether a user has been locked out. Unfortunately, I recognized that the availability is not shown. Instead there is the message "No MTEs currently available".
    I'm aware, the availability cannot be shown on this way, since the agents are part of the central monitoring system and hence i need to choose the central system for "CCMS_DEFINE_R3_SYSTEMS". However I'd like to show the availability separated for each system under "HRT", "CRT" and so on. That means for example, there should me a node "HRT", which contains the free space of the filesystems, whether users are locked out, and the availability for the system.
    It's not a problem to get a solution with a static monitor. I simply create a virtual node "HRT" and tick the checkbox under "<central system>, ..., <availability>, <hrt> ..."
    However, I'd like to use rule based monitor to keep the effort at a minimum by not ticking thousand of check-boxes when I expand the monitored objects.
    I also tried to use the rule "CCMS_GET_AVAILABILITY_FOR_SYSTEM", but it doesn't work and SAP documentations say "This rule is reserved by SAP for later use".
    Does anybody know a solution for my problem?
    In addition, I've got one more question concerning rule based monitors: If I want to create a monitor for my developing systems, I'll not be able to choose "<all>" for "CCMS_DEFINE_R3_SYSTEMS". Do I need to create "CCMS_DEFINE_R3_SYSTEMS" for each system separately? Or is there another possible way?
    Thanks for your help in advance.
    Bob
    Edited by: Bob Norwich on Oct 17, 2010 12:04 PM
    I read something about logical MTE classes: http://help.sap.com/saphelp_nw04s/helpdata/en/31/5462405c910272e10000000a155106/frameset.htm --> Does I need to use this?

    Sorry Sean for my very late anwser.
    I was very busy in my company.
    Yes, I had tried you suggestion but it didn't work for me.
    The problem is, that the Agent "CCMSPING" is part of the central monitoring system by definition.
    Furthermore I tried to copy the default availabiliy monitor to look how it is definied. It didn't help.
    I made the decision to define my monitor differently (an extra monitor for the availability for all system) and replace this temporalily solution when ""CCMS_GET_AVAILABILITY_FOR_SYSTEM" will be available.
    The situation is not so bad after all.
    Thx for your help. I appreciate it!!
    Bob

  • IMac as an extended monitor for a windows based laptop

    Can I use my iMac as an extended monitor for a windows based Laptop while also still using the iMac as an iMac?

    Hi tpskoog,
    Certain iMacs support a mode known as Target Display Mode which allows them to function as an external display, but at the moment only other Macs with certain video out ports are supported as the primary (source) machine. You may find the following article helpful:
    Target Display Mode: Frequently Asked Questions (FAQ)
    http://support.apple.com/kb/ht3924
    Regards,
    - Brenden

  • Documentation for tab.user_data with event based jobs

    Hi,
    Are all the atributes for tab.user_data when used as part of an event based job documented somewhere?
    I have been looking but all I can find are examples rather than explicit documentation.
    Thanks
    Pete

    Hi Pete,
    The type of the object tab.user_data is the type used to create the AQ queue with. So it varies depending on the type you used to create the queue with.
    If you are consuming messages from the scheduler event queue, the type is sys.scheduler$_event_info and to figure out what fields it has you can use
    desc sys.scheduler$_event_info
    Hope this helps,
    Ravi.

  • Raising event when value for a parameter in alert monitor changes

    Hi All,
    We are using Redwood Cronacle version 7 and have connected it to SAP solution Manager System to get the central monitoring data. Now we want to raise a particular event whenever value for a specific parameter in alert monitor gets changed.
    I had a look of already present functions in JCS_MON package, but was not able to find any pre-existing function.
    Any pointers on how to achieve this...
    Thanks a lot.
    Best regards,
    Vithalprasad

    Hi Anton,
    Thanks for your reply. We want the event to be raised on one of the parameters which is coming from SAP solution Manager. This event is required to be raised every time the value of that parameter gets changed from its current value. So it is not possible to define some severity where value can be defined.
    One way can be to store the previous value in one custom table in repository and keep on checking with the present value. Once the difference is found then raise the even & replace the current value with the new value. But just want to know if there is some better solution for this.
    Thanks a lot.
    Best regards,
    Vithalprasad

  • Monitor for instrument

    Hello to all, I have a problel and for the moment I do not succeed to resolve it. Practically I am writing one lybrari to communicate with one instrument on remote and all good end here. Hour I would that the data that I succeed to recover me appear on the monitor. In the sense that if it watches a spectrum you see a trace that modification in continuation based on the frequency span etc . I would to make like a photo of the image and to make it to appear on the screen. To notice that all the regarding data the spectrum I succeed to recover them and to visualize them without problems in JtextField fields. Someone, kindly, could gives some suggestion to me as I can make, does not know lybrari or method, to reproduce a monitor, does not know. Thanks and hello

    Excuse me for the small caos and for my english but I`m new and I try to explain the problem event if difficult for me.
    I have a spectrum analyzer and I communicate without problems using JNI and interface Java.
    Consequently I can obtain vary datas for example centre frequency, span, trace and so on.
    On the spectrum`s monitor is visualized a trace.
    The problem is: I would to visualize the trace on my computer, for example when I pressed a button I obtain one "photo" of the trace.
    Is neccessary for example a Panel?
    Exist a methods that reproduce one type of this monitor?
    I hope that I have been more clear.
    Thanks

  • Adding an override to enable a monitor for a dynamically populated group has no impact?

    Hello all,
    I have a monitor that I only want to target to a single computer group. The group is dynamically populated.
    My basic approach is that the monitor is disabled by default but enabled with an override for that group. However, the monitor does not appear to be running on those servers (or anywhere else). I was able to confirm this based on the results of running the
    "Show Running Rules and Monitors for this Health Service" task--the overridden monitor does not appear at all in the list of monitors running on the servers in that group.
    I have checked the Operations Manager event log on both the SCOM host and the machines in the group, and have not seen any errors. The monitor being overridden is a PowerShell script monitor, where for script logic issues I have seen that the event log will
    contain information on the error. Nothing appears anywhere to show that the monitor has been enabled for these servers.
    Is there a better way to determine why the monitors would not be running on the servers in the group?
    Thanks!

    Hey Gleb,
    I have validated that the group is being populated and that the monitor's target class matches the class of the group members.
    Checking today, it appears that the monitor has begun running on the members of the group... not sure why there was such a large delay before the monitor began executing on the servers. I still have the dreaded green circles in the health explorer, but have
    narrowed that down to a possible condition in my monitor's PowerShell script where the condition properties are not added to the PropertyBag before the script ends.
    Thanks!

  • Guidelines for Health Monitoring for TimesTen

    This document provides some guidance on monitoring the health of a TimesTen
    datastore. Information is provided on monitoring the health of the
    datastore itself, and on monitoring the health of replication.
    There are two basic mechanisms for monitoring TimesTen:
    1. Reactive - monitor for alerts either via SNMP traps (preferred) or
    by scanning the Timesten daemon log (very difficult) and reacting
    to problms as they occur.
    2. Proactive - probe TimesTen periodically and react if problems, or
    potential problems, are detected.
    This document focusses on the second (proactive) approach.
    First, some basic recommendations and guidelines relating to monitoring
    TimesTen:
    1. Monitoring should be implemented as a separate process which maintains
    a persistent connection to TimesTen. Monitoring schemes (typically based
    on scripts) that open a connection each time they check TimesTen impose
    an unnecessary and undesireable loading on the system and are discouraged.
    2. Many aspects of monitoring are 'stateful'. They require periodic
    sampling of some metric maintained by TimesTen and comparing its
    value with the previous sample. This is another reason why a separate
    process with a persistent connection is desireable.
    3. A good monitoring implementation will be configurable since the values
    used for some of the chcks may depend on e.g. the TimesTen configuration
    in use or the workload being handled.
    MONITORING THE HEALTH OF A DATASTORE
    ====================================
    At the simples level, this can be achieved by performing a simple SELECT
    against one of the system tables. The recommended table to use is the
    SYS.MONITOR table. If this SELECT returns within a short time then the
    datastore can be considered basically healthy.
    If the SELECT does not return within a short time then the datastroe is
    stuck in a low level hang situation (incredibly unlikely and very serious).
    More likely, the SELECT may return an error such as 994 or 846 indicating
    that the datastore has crashed (again very unlikely, but possible).
    A slightly more sophisticated version would also include an update to a
    row in a dummy table. This would ensure that the datastore is also capable
    of performing updates. This is important since if the filesystem holding
    the trsnaction logs becomes full the datastore may start to refuse write
    operations while still allowing reads.
    Now, the SYS.MONITOR table contains many useful operational metrics. A more
    sphisticated monitoring scheme could sample some of these metrics and
    compute the delta between subsequent samples, raising an alert if the
    delta exceeds some (configurable) threshold.
    Some examples of metrics that could be handled in this way are:
    PERM_IN_USE_SIZE and PERM_IN_USE_HIGH_WATER compared to PERM_ALLOCATED_SIZE
    (to detect if datastore is in danger of becoming full).
    TEMP_IN_USE_SIZE and TEMP_IN_USE_HIGH_WATER compared to TEMP_ALLOCATED_SIZE
    (ditto for temp area).
    XACT_ROLLBACKS - excessive rollbacks are a sign of excessive database
    contention or application logic problems.
    DEADLOCKS - as for XACT_ROLLBACKS.
    LOCK_TIMEOUTS - excessive lock timeouts usually indicate high levels of
    contention and/or application logic problems.
    CMD_PREPARES & CMD_REPREPARES - it is very important for performance that
    applications use parameterised SQL statements that they prepare just once
    and then execute many times. If these metrics are continuously increasing
    then this points to bad application programming which will be hurting
    performance.
    CMD_TEMP_INDEXES - if this value is increasing then the optimiser is
    comntinually creating temporary indices to process certain queries. This
    is usually a serious performance problem and indicates a missing index.
    LOG_BUFFER_WAITS - of this value is increasing over timne this indicates
    inadequate logging capacity. Yiou may need to increase the size of the
    datastore log buffer (LogBuffSize) and log file size (LogFileSize). If that
    does not alleviate the problem you may need to change your disk layout or
    even obtain a higher performance storage subsystem.
    LOG_FS_READS - this indicates an inefficieny in 'log snoop' processing as
    performed by replication and the XLA/JMS API. To alleviate this you should
    try increasing LogBuffSize and LogFileSize.
    Checking these metrics is of course optional and not necessary for a basic
    healthy/failed decision but if you do check them then you will detect more
    subtle problems in advance and be able to take remedial action.
    MONITORING THE HEALTH OF REPLICATION
    ====================================
    This is a little more complex but is vital to achieve a robust and reliable
    system. ideally, monitorting should be implemented at both datstores, the
    active and the standby. There are many more failure modes possible for
    a replicated system than for a standalone datastore and it is not possible
    to ennumerate them all here. However the information provided here should
    be sufficient to form the basis of a robist monitoring scheme.
    Monitoring replication at the ACTIVE datastore
    1.     CALL ttDataStoreStatus() and check result set;
    If no connections with type 'replication' exists, conclude that
    replication agents are stopped, restart the agents and skip
    next steps.
    It is assumed here that the replication start policy is 'norestart'.
    An alarm about unstable replication agents should be raised
    if this is Nth restart in M seconds (N and M are configuration parameters).
    The alarm can later be cleared when the agents stayed alive K
    seconds (K is configuration parameter).
    2.     CALL ttReplicationStatus() and check result set;
    This returns a row for every replication peer for this datastore.
    If the pState is not 'start' for any peer, raise an alarm about paused or
    stopped replication and skip rest of the steps.
    It is assumed that master cannot help the fact that state is not
    'start'. An operator may have stopped/paused the replication or
    TimesTen stopped the replication because of fail threshold
    strategy. In former case the operator hopefully starts the replication
    sooner or later (of course, after that TimesTen may stop it again
    because of the fail threshold strategy). In latter case the standby
    side monitor process should recognise the fact and duplicate the data
    store with setMasterRepStart-option which sets state back to 'start'.
    If for any peer, lastMsg > MAX (MAX is a configuration parameter), raise
    an alarm for potential communication problems.
    Note that if replication is idle (nothing to replicate), or there is
    very little replication traffic, the value for lastMsg may become as
    high as 60 seconds without indicating any problem. The test logic
    should cater for this (i.e. MAX must be > 60 seconds).
    3.     CALL ttBookmark();
    Compute the holdLSN delta between the values from this call and the
    previous call and if the delta is greater than maximum allowed
    (configuration parameter), raise an alarm about standby
    that is too far behind. Continue to next step.
    Notice that maximum delta should be less than FAILTHRESHOLD * logSize.
    4.     CALL ttRepSyncSubscriberStatus(datastore, host);
    This step is only needed if you are using RETURN RECEIPT or RETURN TWOSAFE
    with the optional DISABLE RETURN feature.
    If disabled is 1, raise an alarm for disabled return service.
    Continue to next step. If RESUME RETURN policy is not enabled we could,
    of course, try to enable return service again (especially when DURABLE
    COMMIT is OFF).
    There should be no reason to reject TimesTen own mechanisms that
    control return service. Thus, no other actions for disabled return
    service.
    Monitoring replication at the STANDBY datastore
    1.     CALL ttDataStoreStatus();
    If no connections with type 'replication' exists, conclude that
    replication agents are stopped, restart the agents and skip
    next steps.
    It is assumed that replication start policy is 'norestart'.
    An alarm about unstable replication agents should be raised
    if this is Nth restart in M seconds (N and M are configuration parameters).
    The alarm can later be cleared when the agents stayed alive K
    seconds (K is configuration parameter).
    2.     Call SQLGetInfo(...,TT_REPLICATION_INVALID,...);
    If the status is 1, this indicates that the active store has marked this store
    as failed due to it being too far out of sync due to log FAILTHRESHOLD.
    Start recovery actions by destroying the datastore and recreating via a
    'duplicate' operation from the active.
    3.     Check 'timerecv' value for relevant row in TTREP.REPPEERS
    If (timerecv - previous timerecv) > MAX (MAX is a configuration parameter),
    raise an alarm for potential communication problems.
    You can determine the correct row in TTREP.REPPEERS by first getting the
    correct TT_STORE_ID value from TTREP.TTSTORES based on the values in
    HOST_NAME and TT_STORE_NAME (you want the id corresponding to the active
    store) and then using that to query TTREP.REPPEERS (you can use a join if
    you like).
    The recovery actions that should be taken in the event of a problem with
    replication depend on several factors:
    1. The application requirements
    2. The type of replication configuration
    3. The replication mode (asynchronous, return receipt or return twosafe)
    that is in use
    Consult the Timesten replication guide for information on detailed recovery
    procedures for each combination.
    ================================ END ==================================

    The information in the forum article is the abridged text of a whitepaper I wrote recommending best practice for building a monitoring infrastructure for TimesTen. i.e. you write an 'application' in C, C++ or Java that performs these monitoring activities and run it continually in production against your datastores. Various aspects of the behaviour of the application could be controlled by configurable parameters; these are not TimesTen parameters but parameters defined and used by the monitoring application.
    In the specific case you mentioned, the 'lastMsg' value returned by ttReplicationStatus is the number of seconds since the last message was received from that peer. The monitoring application would compare this against some meaningful threshold (maybe 30 seconds) and if lastMsg is > that value, raise an alarm. To allow flexibility, the value compared against )MAX) should be configurable.
    Does that make sense?
    Chris

  • Extended Event Filter Record for PArticular or ONE database

    Hi Folks,
    If am working with Extended Event .In that am use Adventureworks database which are long running queries.Finally i get a result from all  Database those queries are long running queries. But i want to get the result
    from only one database, anyother database long running queries wont come in result., If possible tell me the ways.,
    Thanks

    In this case you can filter the events based on the database id.
    Reference
    http://blog.sqlauthority.com/2010/03/29/sql-server-introduction-to-extended-events-finding-long-running-queries/
    Mention the sqlserver.database_id parameter while creating the event session
    In the below example, I've created session to monitor the long running events for the database_id =15
    CREATE EVENT SESSION LongRunningQuery
    ON SERVER
    -- Add event to capture event
    ADD EVENT sqlserver.sql_statement_completed
    -- Add action - event property
    ACTION (sqlserver.sql_text, sqlserver.tsql_stack,sqlserver.database_id)
    -- Predicate - time 1000 milisecond
    WHERE sqlserver.sql_statement_completed.duration > 1000 and sqlserver.database_id=15
    -- Add target for capturing the data - XML File
    ADD TARGET package0.asynchronous_file_target(
    SET filename='c:\LongRunningQuery.xet', metadatafile='c:\LongRunningQuery.xem'),
    -- Add target for capturing the data - Ring Bugger
    ADD TARGET package0.ring_buffer
    (SET max_memory = 1096)
    WITH (max_dispatch_latency = 1 seconds)
    --Prashanth

  • How to create event based process chains

    Hi All,
    I would like to know about event based process chains. In connection to this, could you please answer the following queries,
    1. How to create events
    2. How to link created event to the process chain in the same BI or BW system and as well as from  
        externel BI system.
    3. How link one process chain with other process chain (i.e, After completion of one process chain, it
        should trigger other dependent process chain)
    Thanks and Regards,
    Kotesh.

    1). Doubt regarding first question.
    For example, i would like to create time based event (it should be trigger daily at specified time),
    where we have to maintain scheduling options while creating event.
    When i checked SM62 there i found only two options a). Event name and b). Description.
    Could please send any doucument link if you have.
    Ans : You can use function modules like "BP_EVENT_RAISE" in a program and schedule the program to trigger.
    2). For externel BIW system also same procedure we need to follow or any difference.
    Ans : Externally you need to trigger the same event.
    3). i found dependent process chain also had scheduling options as direct scheduling insted of start using meta chain or API. As you said dependent process chain should be mata chain. it seems dependent process chain may be Meta chain or Direct scheduilg.
    Ans : Its your choice how you want to schedule it.You can either make that dependent chain a metachain or schedule it separately.
    I found at the end of first process chain they kept one process like Raise event and second process chain connected with the help of raise event process event name. If you have any idea about this process could explain a bit more.
    Ans : May be they are raising the event in the main chain and triggering the dependent chain using this event.
    But Metachain is preferred for such thing.Though it does similar thing.
    Hope this helps.

Maybe you are looking for

  • How can I block spam text messages from an email address?

    Twice per week I am getting spam texts in the middle of the night from a random email address, in reference to a free Target gift card. The problem is, the sender's email address is different every time. Any ideas how to make it stop??

  • Mail accounts are not showing up in Mail

    I recently migrated my 2011 MBP to a 2013 MBP.  I had Mavericks installed on both computers.  Everything works fine except that I can't add other mail accounts.  I put in my primary email (a GoDaddy account) and my local cable account (Time Warner). 

  • How to dispaly the ABAP reports in SUS

    Hi Experts, Please tell me the detailed steps to dispaly the ABAP reports developed in SRM in SUS. Please elt me know the entiire steps. Thanks in Advance, Manu

  • Permission problem when create source system between BI and R/3

    Hi All, I want to create a source system between BI7 and R/3 system Both RFC connections have been created and tested succefully on the 2 systems But when i create the source system on BI I have the message : "Changes to Repository or cross-client cu

  • Importing an iphoto 08 Library from an external hard drive

    I'm trying to migrate my old library from my old MBP to my new one. I've backed up a copy to an external hard drive and then i have tried creating a new library on my new MBP and importing from that hard drive. The file is around 50GB and when it get