Coherence for Persisting HttpSession in OAS

We are moving from WebSphere to Oracle environment and planning to use clustered Oracle Application Servers for fail-over and load balancing applications.
I have read documentation about replicating HttpSession accross application server and persisting either in in-memory or database. Is that possible to setup Oracle Application Server to use Coherence for persisting HttpSession and all the servers in the cluster can use it?
Thanks

hi esjp2000
Maybe this is a good question for this forum:
"Coherence Support" Coherence Support
success
Jan Vervecken

Similar Messages

  • What is the recommended way for persisting JMS messages?

    What is the recommended way for persisting JMS messages?. As per the IMQ admin documentation , using the default built-in persistence type which is through unix flat files is much efficient and faster, compared to the database persistence .
    Tried setting up the jdbc stuff for database persistence on iAS 6.5 . I am getting the following
    error .
    [24/Apr/2002:16:09:20 PDT] [B1060]: Loading persistent data...
    [24/Apr/2002:16:09:21 PDT] Using plugged in persistent store: database connection
    url=jdbc:oracle:thin:@dbatool.mygazoo.com:1521:qa1 brokerid=ias01
    [24/Apr/2002:16:09:23 PDT] [B1039]: Broker "jmqbroker" ready.
    [24/Apr/2002:16:11:56 PDT] ERROR [B4012]: Failed to persist interest
    SystemManager%3ASystemManagerEngine%2BiMQ+Destination%0AgetName%28%29%3A%09%09SM_Response%0AClass%3A%09%09%09com.sun.messaging.Topic%0AgetVERSION%28%29%3A%09%092.0%0AisReadonly%28%29%3A%09%09false%0AgetProperties%28%29%3A%09%7BJMQDestinationName%3DSM_Response%2C+JMQDestinationDescription%3DA+Description+for+the+Destination+Object%7D:
    java.sql.SQLException: ORA-01401: inserted value too large for column
    [24/Apr/2002:16:11:56 PDT] WARNING [B2009]: Creation of consumer SM_Response to destination 1
    failed:com.sun.messaging.jmq.jmsserver.util.BrokerException: Failed to persist interest
    SystemManager%3ASystemManagerEngine%2BiMQ+Destination%0AgetName%28%29%3A%09%09SM_Response%0AClass%3A%09%09%09com.sun.messaging.Topic%0AgetVERSION%28%29%3A%09%092.0%0AisReadonly%28%29%3A%09%09false%0AgetProperties%28%29%3A%09%7BJMQDestinationName%3DSM_Response%2C+JMQDestinationDescription%3DA+Description+for+the+Destination+Object%7D:
    java.sql.SQLException: ORA-01401: inserted value too large for column
    Any thoughts?

    From the output, you are using imq 2.0. In that release
    the key used to persist a durable subscriber in the database
    table has a limit of 100 characters. The output shows that
    your value is:
    SystemManager%3ASystemManagerEngine%2BiMQ+Destination%0AgetName%28%29%3A%09%09SM_Res
    ponse%0AClass%3A%09%09%09com.sun.messaging.Topic%0AgetVERSION%28%29%3A%09%092.0%0Ais
    Readonly%28%29%3A%09%09false%0AgetProperties%28%29%3A%09%7BJMQDestinationName%3DSM_R
    esponse%2C+JMQDestinationDescription%3DA+Description+for+the+Destination+Object%7D:
    which is much longer than 100 characters.
    You might want to shorten the string you use for the
    durable name.
    And yes, the default file-based persistence store is
    more efficient when compared to the plugged-in persistence
    through a database.

  • Re: pattern for persistance management: the solution?

    Thanks to Forte France consulting group and Marc Sonnet (Forte
    France), here is one solution for our problem.
    We don't have test it yet but this seems pretty good.
    thanks for all your replies,
    Corinne Barbat
    ____________________________ Séparateur Réponse ________________________________
    Objet : Re: Bloated clients (was: pattern for persistance management
    Auteur : [email protected] à _INTERNET
    Date : 23/10/98 00:41
    Geoff,
    Fromhat I've understood of Forte's partitioning behaviour, the only
    way to get rid of b, c, d and e Projects from the client partition is by
    deploying them as a shared library and having the Client dynamically
    load them at run-time. Any other way, including creating a complex
    structure of interfaces to avoid the client referencing the other
    projects directly, or even deploying the Server projects as another
    Application, will not prevent these projects from being included in the
    Client Partition.
    The problem I guess is that when we think along the lines of components
    we immediately associate them with Service Objects. Clearly, it's the
    Project and not the SO which can be classified as a component in Forte.
    The questions I have is : Now that we know how Forte's Partioning
    mechanism works, would it significantly impact the way we partition our
    Apps ? I'm not so sure.
    Eric
    ----Original Message Follows----
    Date: Thu, 22 Oct 1998 14:52:29 -0400
    From: Geoff Puterbaugh <[email protected]>
    To: [email protected]
    Subject: Bloated clients (was: pattern for persistance management)
    Reply-To: Geoff Puterbaugh <[email protected]>
    This topic is of keen interest to me as well. At my customer's
    urging, I wrote a little test application spread across
    four or five projects, each supplying the next..
    a << b << c << d << e
    The projects b,c,d, and e each contain a class and a service
    object based on that class. I partitioned it (a client
    partition and four server partitions for b c d e) and made
    a distribution. Then I looked at the client partition
    and discovered b c d e were projects for it!
    If you code-generate this and run fcompile, you will see
    that all the code for all the projects in the entire chain
    of supplier projects winds up in the client partition.
    To be clear about this, if project e contains a class X,
    you will see the source code for X.init and all other
    X methods in the client partition.
    This appears to be the default behavior, and the open
    question is how to change the default behavior, if that
    can be done.
    All my best,
    Geoff
    P.S. If e uses GenericDBMS as a supplier, it will show up
    in the client too. Of course!
    This is at least my understanding as of today.
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

    Thanks to Forte France consulting group and Marc Sonnet (Forte
    France), here is one solution for our problem.
    We don't have test it yet but this seems pretty good.
    thanks for all your replies,
    Corinne Barbat
    ____________________________ S&eacute;parateur R&eacute;ponse ________________________________
    Objet : Re: Bloated clients (was: pattern for persistance management
    Auteur : [email protected] &agrave; _INTERNET
    Date : 23/10/98 00:41
    Geoff,
    Fromhat I've understood of Forte's partitioning behaviour, the only
    way to get rid of b, c, d and e Projects from the client partition is by
    deploying them as a shared library and having the Client dynamically
    load them at run-time. Any other way, including creating a complex
    structure of interfaces to avoid the client referencing the other
    projects directly, or even deploying the Server projects as another
    Application, will not prevent these projects from being included in the
    Client Partition.
    The problem I guess is that when we think along the lines of components
    we immediately associate them with Service Objects. Clearly, it's the
    Project and not the SO which can be classified as a component in Forte.
    The questions I have is : Now that we know how Forte's Partioning
    mechanism works, would it significantly impact the way we partition our
    Apps ? I'm not so sure.
    Eric
    ----Original Message Follows----
    Date: Thu, 22 Oct 1998 14:52:29 -0400
    From: Geoff Puterbaugh <[email protected]>
    To: [email protected]
    Subject: Bloated clients (was: pattern for persistance management)
    Reply-To: Geoff Puterbaugh <[email protected]>
    This topic is of keen interest to me as well. At my customer's
    urging, I wrote a little test application spread across
    four or five projects, each supplying the next..
    a << b << c << d << e
    The projects b,c,d, and e each contain a class and a service
    object based on that class. I partitioned it (a client
    partition and four server partitions for b c d e) and made
    a distribution. Then I looked at the client partition
    and discovered b c d e were projects for it!
    If you code-generate this and run fcompile, you will see
    that all the code for all the projects in the entire chain
    of supplier projects winds up in the client partition.
    To be clear about this, if project e contains a class X,
    you will see the source code for X.init and all other
    X methods in the client partition.
    This appears to be the default behavior, and the open
    question is how to change the default behavior, if that
    can be done.
    All my best,
    Geoff
    P.S. If e uses GenericDBMS as a supplier, it will show up
    in the client too. Of course!
    This is at least my understanding as of today.
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

  • Supported HA Architecture for Persistent Chat

    I have a project that has a requirement for Persistent Chat as part of the infrastructure. The design is two well connected data centers in the same region so there will be two EE Lync pools paired for DR. They are using SQL clustering instead of mirroring.
    My thought was to create an instance in each data center for all Lync databases other than Persistent Chat.  Those instances would be clustered.  Can I then create a second instance in each data center that is not clustered for Persistent Chat
    and use log shipping between the two, basically making one side the active and the other the passive?  Is there any reason that would not be supported?  It is my understanding this is still the only choice for disaster recovery for Persistent Chat.
    As for the Persistent Chat servers themselves there would be two and both would be active since connectivity between data centers is very good.
    Any help is appreciated.  Thanks.

    When doing persistent chat DR between two data-centers, using log shipping is the recommended and support solution.  You can find a drawing here on TechNet:
    http://technet.microsoft.com/en-us/library/jj205211.aspx
    So in DataCenter A, you could have SQL Mirroring, log ship to the primary server in DataCenter B and than configure SQL Mirroring there as well.
    Hope that helps.
    Thanks,
    Richard
    Richard Brynteson, Lync MVP | http://masteringlync.com | http://lyncvalidator.com

  • Oracle Coherence for Java

    I am not sure which version of Oracle Coherence I need for the OpenFire clustering plugin. It says I need Oracle Coherence for Java, and that you need a license for it. There appears to be different flavors of it, Standard, Media, Grid, Real-Time, and enterprise. Oracle Media is cheapest license. Can I use this license or do I need a Standard license.

    The tell tale sign that the application is communicating with the running cache server is the membership information that is logged. In your previous running of the example you would have "members=1" embedded in the logging. From the sounds of what you are doing, you'd expect this to be different as the application would be joining a running instance. Check out section 1.3 of this doc on the system you are using.
    http://download.oracle.com/docs/cd/E15357_01/coh.360/e15831/installcoh.htm#BABIHHFJ
    Cheers,
    Jay

  • OHS Routing to JVM responsible for creating HttpSession (Session Affinity)

    Hello,
    I am using OAS 10.1.3.1 and have a plan in place to create an HttpSession for a just authenticated end user in an OC4J instance/JVM of the my application's choosing. What I then would like to happen is for every subsequent HTTP request initiated by that end user, until such a time as they logout or their session times out, that OHS route their requests to that single OC4J instance/JVM that initially created an HttpSession on their behalf. This is simply Session Affinity.
    What I am wondering is do I simply get this behavior by default if I specify no select method (e.g. random, round robin, etc.) in mod_oc4j.conf? Are their combinations of select method (e.g. roundrobin:local) or other mod_oc4j directives that could result in different behavior than this simple Session Affinity I seek?
    Thanks,
    Doug

    Hello
    Within 10.1.3 OC4Js announce their mount-point(s) in the notifications they send out and mod_oc4j dynamically adjusts its routing table using this information. This eliminates the need for static mount point configuration and enables mod_oc4j to update its mount point configuration dynamically. So. routing is handled by opmn to OHS.
    regards
    Michel

  • Database is not created in SQL server for Persistant Chat

    Hi All,
    We were deploying the Persistant Chat Lync 2013 Role. We have created the topology and published it. Then the Lync 2013 deployment is done on the Persistant Chat servers, while upgrading the cumulative servers soon after the deployment we are getting the
    error as :
    Test-CsDatabase : Cannot open database "mgc" requested by the login. The login
    failed.
    Login failed for user 'domain\username'.
    At line:1 char:1
    + Test-CsDatabase -ConfiguredDatabases -SqlServerFqdn
    SQL Server ...
    Get-MirrorState also was not giving the proper result. Then when I checked in the SQL server, we found that both mgc and mgccomp databases itself not created.
    Could someone assist how we can install the database now?? I googled but confused a bit. Servers are already LYNC deployed, so which command and procedure we need to follow to install the DB on SQL servers and also in mirror server.
    Never giveup till you get what to want.

    Please share Test-CsDatabase -verbose 
    And Get-MirrorState
    in details
    PLEASE REMEMBER, if you see a post that helped you please click "Vote As Helpful" and if it answered your question please click "Mark As Answered"

  • Calculate coherence for a series of Impacts

    Calculating coherence has been discussed several times - including this knowledge base article and community example. But I'm still missing something.
    How can I calculate the coherence so it updates after each hit (right now it's always 1)?
    On a related note, I'm assuming the Y signals in the FRF(Mag-Phase) are the response and the X signals are the stimulus...Transfer Function.vi explicitly states Response and Stimulus for its inputs, but FRF(Mag-Phase) doesn't.
    I attached an abridged version of what I've attemped.
    Andrew
    Attachments:
    CalcCoherences.vi ‏18 KB

    Check this example:
    http://zone.ni.com/devzone/cda/epd/p/id/5810
    You will see that to measure coherence you need to input the same size array or waveform. In your VI, you are sending your input through a shift register and augmenting the size of your array.
    National Instruments
    Applications Engineer

  • Planning for Persistent Chat

    Dears,
    I would like to know if persistent chat databases could use SQL 2012 always ON.
    My design has two data centers and a stretched persistent chat pool. Is there any documentation if this scenario can use SQL always ON or not?
    Thank you,
    Regards

    Always On is currently not supported.  Log shipping is the supported method for cross-site persistent chat database resiliency currently, but as you know isn't exactly an HA replacement, you'd still want mirroring in the primary datacenter.
    http://technet.microsoft.com/en-us/library/jj205007.aspx
    Please remember, if you see a post that helped you please click "Vote As Helpful" and if it answered your question please click "Mark As Answer".
    SWC Unified Communications

  • Which version of SQL server should install for persistent chat?

    hi
    after i got error that i put picture below 
    i decided to install sql server again for service persistent chat
    but i don't know which version should i download? and again i don't know how i should delete last sql server?
    which one i should download?
    1-2008 Express
    2-2008 R2 enterprise 
    3-2008 R2 SP2 - Express Edition

    As per the above error you are trying to install 2008 sql server which is not supported.
    you have to install SQL Server 2008R2 or 2012. Check this link
    http://technet.microsoft.com/en-us/library/gg398495.aspx
    Whenever you see a helpful reply, click on Vote As Helpful & click on Mark As Answer if a post answers your question.

  • How automatic dirty checking works for persistance layer?

    Hi,
    I recently started learning about Java Persistance API and also about Hibernate and I came across this example
    private void addPersonToEvent(Long personId, Long eventId) {
        Session session = HibernateUtil.getSessionFactory().getCurrentSession();
        session.beginTransaction();
        Person aPerson = (Person) session.load(Person.class, personId);
        Event anEvent = (Event) session.load(Event.class, eventId);
        aPerson.getEvents().add(anEvent);
        session.getTransaction().commit();
    }and in the explanation they have given like
    After loading a Person and an Event, simply modify the collection using the normal collection methods. As you can see, there is no explicit call to update() or save(), Hibernate automatically detects that the collection has been modified and needs to be updated. This is called automatic dirty checking, and you can also try it by modifying the name or the date property of any of your objects.
    How does it work? I know that something called dynamic proxy that exist in java, is it by that?
    Let us say I want something to do like an interceptor, whenever I call a particular method on an object, that call will first be address by me and then it will be delegated to the actual object.
    Say
    public void handleEvent(SyncCompletedEvent event)
            // We have to handle this one, this is the only one that will trigger the start up sync
            // however, this request doesn't need response!
            if (event.isSenderDeviceCommunicator())
                // set the request to null, so that no intermediate response will be sent out.
                handleRequest((UpmSyncRequest) null);
            if(event.getSubType().equals(SyncCompletedEvent.SUBTYPE.LEGACY_SERVER_STARTUP_COMPLETED))
                ((UpmService) getService()).scheduleAuditBackup(false);
        }

    What kind of problems do you have with the Firefox 28 release?
    Did you check your security software (firewall, anti-virus) to make sure that the new version in allowed?
    It is possible that your security software (firewall, anti-virus) blocks or restricts Firefox or the plugin-container process without informing you, possibly after detecting changes (update) to the Firefox program.
    Remove all rules for Firefox and the plugin-container from the permissions list in the firewall and let your firewall ask again for permission to get full, unrestricted, access to internet for Firefox and the plugin-container process and the updater process.
    See:
    *https://support.mozilla.org/kb/Server+not+found
    *https://support.mozilla.org/kb/Firewalls
    *https://support.mozilla.org/kb/fix-problems-connecting-websites-after-updating
    You can create a new profile as a test to check if your current profile is causing the problem.
    See "Creating a profile":
    *https://support.mozilla.org/kb/profile-manager-create-and-remove-firefox-profiles
    *http://kb.mozillazine.org/Standard_diagnostic_-_Firefox#Profile_issues
    If the new profile works then you can transfer files from a previously used profile to the new profile, but be cautious not to copy corrupted files to avoid carrying over the problem
    *http://kb.mozillazine.org/Transferring_data_to_a_new_profile_-_Firefox

  • [Solved] udev rules: serial for persistent device naming disappeared

    Hello,
    since the latest upgrade of udev and kmod I cant't mount my external e-SATA hard disk with my working udev rules.
    This is my udev rule:
    $ cat /etc/udev/rules.d/66-persistent_esata.rules
    KERNEL=="sd?1", SUBSYSTEMS=="block", ENV{ID_SERIAL_SHORT}=="S1VSJ1LS301247", NAME="aluguard"
    When plugged in, I could decrypt and mount it with "sudo cryptsetup luksOpen /dev/aluguard aluguard && mount /mnt/aluguard".
    Since the update above I get "Device /dev/aluguard doesn't exist or access denied."
    While I re-checked https://wiki.archlinux.org/index.php/Ma … _with_udev, I discovered, that the serial entry in the output of udevadm has vanished:
    $ udevadm info -a -p $(udevadm info -q path -n /dev/sdg)
    Udevadm info starts with the device specified by the devpath and then
    walks up the chain of parent devices. It prints for every device
    found, all possible attributes in the udev rules key format.
    A rule to match, can be composed by the attributes of the device
    and the attributes from one single parent device.
    looking at device '/devices/pci0000:00/0000:00:09.0/host3/target3:0:0/3:0:0:0/block/sdg':
    KERNEL=="sdg"
    SUBSYSTEM=="block"
    DRIVER==""
    ATTR{range}=="16"
    ATTR{ext_range}=="256"
    ATTR{removable}=="0"
    ATTR{ro}=="0"
    ATTR{size}=="1953525168"
    ATTR{alignment_offset}=="0"
    ATTR{discard_alignment}=="0"
    ATTR{capability}=="50"
    ATTR{stat}==" 96 24 762 780 0 0 0 0 0 780 780"
    ATTR{inflight}==" 0 0"
    ATTR{events}==""
    ATTR{events_async}==""
    ATTR{events_poll_msecs}=="-1"
    looking at parent device '/devices/pci0000:00/0000:00:09.0/host3/target3:0:0/3:0:0:0':
    KERNELS=="3:0:0:0"
    SUBSYSTEMS=="scsi"
    DRIVERS=="sd"
    ATTRS{device_blocked}=="0"
    ATTRS{type}=="0"
    ATTRS{scsi_level}=="6"
    ATTRS{vendor}=="ATA "
    ATTRS{model}=="SAMSUNG HD103SI "
    ATTRS{rev}=="1AG0"
    ATTRS{state}=="running"
    ATTRS{timeout}=="30"
    ATTRS{iocounterbits}=="32"
    ATTRS{iorequest_cnt}=="0xa2"
    ATTRS{iodone_cnt}=="0x98"
    ATTRS{ioerr_cnt}=="0xc"
    ATTRS{evt_media_change}=="0"
    ATTRS{queue_depth}=="31"
    ATTRS{queue_ramp_up_period}=="120000"
    ATTRS{queue_type}=="simple"
    looking at parent device '/devices/pci0000:00/0000:00:09.0/host3/target3:0:0':
    KERNELS=="target3:0:0"
    SUBSYSTEMS=="scsi"
    DRIVERS==""
    looking at parent device '/devices/pci0000:00/0000:00:09.0/host3':
    KERNELS=="host3"
    SUBSYSTEMS=="scsi"
    DRIVERS==""
    looking at parent device '/devices/pci0000:00/0000:00:09.0':
    KERNELS=="0000:00:09.0"
    SUBSYSTEMS=="pci"
    DRIVERS=="ahci"
    ATTRS{vendor}=="0x10de"
    ATTRS{device}=="0x0554"
    ATTRS{subsystem_vendor}=="0x1849"
    ATTRS{subsystem_device}=="0x0554"
    ATTRS{class}=="0x010601"
    ATTRS{irq}=="23"
    ATTRS{local_cpus}=="00000000,00000003"
    ATTRS{local_cpulist}=="0-1"
    ATTRS{numa_node}=="0"
    ATTRS{dma_mask_bits}=="64"
    ATTRS{consistent_dma_mask_bits}=="64"
    ATTRS{enable}=="1"
    ATTRS{broken_parity_status}=="0"
    ATTRS{msi_bus}==""
    looking at parent device '/devices/pci0000:00':
    KERNELS=="pci0000:00"
    SUBSYSTEMS==""
    DRIVERS==""
    What is the reason for this strange behaviour, and how can I get back the serial entry?
    Thank in advance!
    Last edited by indianahorst (2012-01-31 13:03:20)

    $ udevadm info -q all -n /dev/sdg1
    P: /devices/pci0000:00/0000:00:09.0/host3/target3:0:0/3:0:0:0/block/sdg/sdg1
    N: sdg1
    S: disk/by-id/ata-SAMSUNG_HD103SI_S1VSJ1LS301247-part1
    S: disk/by-id/scsi-SATA_SAMSUNG_HD103SIS1VSJ1LS301247-part1
    S: disk/by-id/wwn-0x50024e900136a03e-part1
    S: disk/by-path/pci-0000:00:09.0-scsi-3:0:0:0-part1
    S: disk/by-uuid/bb024e1b-78a5-4ed0-9e80-aaa7633aa493
    E: DEVLINKS=/dev/disk/by-id/ata-SAMSUNG_HD103SI_S1VSJ1LS301247-part1 /dev/disk/by-id/scsi-SATA_SAMSUNG_HD103SIS1VSJ1LS301247-part1 /dev/disk/by-id/wwn-0x50024e900136a03e-part1 /dev/disk/by-path/pci-0000:00:09.0-scsi-3:0:0:0-part1 /dev/disk/by-uuid/bb024e1b-78a5-4ed0-9e80-aaa7633aa493
    E: DEVNAME=/dev/sdg1
    E: DEVPATH=/devices/pci0000:00/0000:00:09.0/host3/target3:0:0/3:0:0:0/block/sdg/sdg1
    E: DEVTYPE=partition
    E: ID_ATA=1
    E: ID_ATA_DOWNLOAD_MICROCODE=1
    E: ID_ATA_FEATURE_SET_AAM=1
    E: ID_ATA_FEATURE_SET_AAM_CURRENT_VALUE=254
    E: ID_ATA_FEATURE_SET_AAM_ENABLED=1
    E: ID_ATA_FEATURE_SET_AAM_VENDOR_RECOMMENDED_VALUE=254
    E: ID_ATA_FEATURE_SET_APM=1
    E: ID_ATA_FEATURE_SET_APM_ENABLED=0
    E: ID_ATA_FEATURE_SET_HPA=1
    E: ID_ATA_FEATURE_SET_HPA_ENABLED=1
    E: ID_ATA_FEATURE_SET_PM=1
    E: ID_ATA_FEATURE_SET_PM_ENABLED=1
    E: ID_ATA_FEATURE_SET_PUIS=1
    E: ID_ATA_FEATURE_SET_PUIS_ENABLED=0
    E: ID_ATA_FEATURE_SET_SECURITY=1
    E: ID_ATA_FEATURE_SET_SECURITY_ENABLED=0
    E: ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=194
    E: ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=194
    E: ID_ATA_FEATURE_SET_SMART=1
    E: ID_ATA_FEATURE_SET_SMART_ENABLED=1
    E: ID_ATA_SATA=1
    E: ID_ATA_SATA_SIGNAL_RATE_GEN1=1
    E: ID_ATA_SATA_SIGNAL_RATE_GEN2=1
    E: ID_ATA_WRITE_CACHE=1
    E: ID_ATA_WRITE_CACHE_ENABLED=1
    E: ID_BUS=ata
    E: ID_FS_TYPE=crypto_LUKS
    E: ID_FS_USAGE=crypto
    E: ID_FS_UUID=bb024e1b-78a5-4ed0-9e80-aaa7633aa493
    E: ID_FS_UUID_ENC=bb024e1b-78a5-4ed0-9e80-aaa7633aa493
    E: ID_FS_VERSION=1
    E: ID_MODEL=SAMSUNG_HD103SI
    E: ID_MODEL_ENC=SAMSUNG\x20HD103SI\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20
    E: ID_PART_ENTRY_TYPE=0x83
    E: ID_PART_TABLE_TYPE=dos
    E: ID_PATH=pci-0000:00:09.0-scsi-3:0:0:0
    E: ID_PATH_TAG=pci-0000_00_09_0-scsi-3_0_0_0
    E: ID_REVISION=1AG01131
    E: ID_SCSI_COMPAT=SATA_SAMSUNG_HD103SIS1VSJ1LS301247
    E: ID_SERIAL=SAMSUNG_HD103SI_S1VSJ1LS301247
    E: ID_SERIAL_SHORT=S1VSJ1LS301247
    E: ID_TYPE=disk
    E: ID_WWN=0x50024e900136a03e
    E: ID_WWN_WITH_EXTENSION=0x50024e900136a03e
    E: MAJOR=8
    E: MINOR=97
    E: PART_ENTRY_DISK=8:96
    E: PART_ENTRY_NUMBER=1
    E: PART_ENTRY_OFFSET=63
    E: PART_ENTRY_SCHEME=dos
    E: PART_ENTRY_SIZE=1953520002
    E: SUBSYSTEM=block
    E: UDEV_LOG=3
    E: UDISKS_PARTITION=1
    E: UDISKS_PARTITION_ALIGNMENT_OFFSET=0
    E: UDISKS_PARTITION_NUMBER=1
    E: UDISKS_PARTITION_OFFSET=32256
    E: UDISKS_PARTITION_SCHEME=mbr
    E: UDISKS_PARTITION_SIZE=1000202241024
    E: UDISKS_PARTITION_SLAVE=/sys/devices/pci0000:00/0000:00:09.0/host3/target3:0:0/3:0:0:0/block/sdg
    E: UDISKS_PARTITION_TYPE=0x83
    E: UDISKS_PRESENTATION_NOPOLICY=0
    E: USEC_INITIALIZED=6897248001
    OK, there is the serial.... but the problem persists - udev doesn't execute the rule and doesn't create /dev/aluguard.

  • Servlets or EJB for Persistent external socket connections

    We have a Servlet that establishes a socket connection to another
    external server and sends and receives messages. We have the Servlet defined as pre-loaded when JRun starts.
    When JRun is started, the Servlet is started and connection is
    established with the external server. The problem is when a new message is being sent, the Servlet is instantiate again and the connection is stopped and restarted. We do a forward request from another Servlet to this.
    Am I totally misunderstanding Servlets because reading articles
    and books seems to indicate that I should create an EJB instead of a
    Servlet if you want a persistent connection.
    Thanks in advance for your help.

    I assume then that your connection to the external server is achieved in the "init" method of your servlet. Therefore every time the servlet container is called upon to accesss your servlet, it instantiates it - thus a new connection will be being made each time.
    I would take out the connection properties from the init method. I.e. you only want it to happen once.
    T

  • Is periodic maintenance required for app deployed on OAS on Windows?

    Hi,
    We have an application that uses 7 BPEL processes and one web service deployed on OAS. Application is into production from the last 4 months.
    We are bringing down the App server once in a while due to applying break fixes, etc.
    But, now, considering the App is quite stable, we won't be planning to bring down the app server for few months. My question is, do we need to schedule planned downtime and restart the application server once in a while in view of memory leaks etc as the App is deployed on Windows Operating System?
    It would be a great if I get some info/links that will give me the exact information.
    Thanks,
    Sasi Bhushan.

    Hi Chris,
    Thanks for the reply.
    Yes. Infact we have taken care of log files. But wondering if memory leaks etc could cause total memory consumption and making the app server hung etc.
    So, I am looking for some suggestion on whether we need the app server restart periodically or it's not required at all.
    Thanks,
    Sasi Bhushan.

  • Is there a cure for persistent file corruptions across the suite?

    i am having terrible and consistent problems with file corruptions across the suite. for example…
    1: i collected for output (indesign) – corrupted over 100 small images that were linked.
    2: opened an indesign job and all linked files failed to update and could not be imported (simply would not import those files and were all corrupted).
    3: had a hang in photoshop followed by a 150:30 error (would not open any of my applications). this prompted a clean install of system and adobe cs4 see below
    4: most recently, had a hang in photoshop (photoshop will also no "force quit") whilst indesign was open. two of my image files corrupted and, most surprisingly, all my illustrator files that i had placed also corrupted (will not even open in illustrator any more). illustrator was not even running at the time, so, get this – a hang in photoshop destroys illustrator files simply because they are placed in indesign.
    this last event has made me more certain than ever that there is a fundamental problem at the cs4 core.
    can anyone help?
    please note the following:
    i am running a macintosh imac, 2.66 GHz Intel Core 2 Duo (latest imac), with 4 GB 1067 MHz DDR3 ram.
    before this last disaster, i had reloaded the system software (archive and install), and updated it all the way to version 10.5.8
    i then uninstalled the creative suite, ran the cleanscript from adobe, reloaded the creative suite, downloaded all the cs4 updates
    (so, latest mac, latest system, latest cs4 – all clean installs).
    is there something i can do? or is this just a very unstable and disappointing upgrade – i have had nothing but problems from day one!
    many thanks if anyone has any advice.

    Re: 150:30 error:
    Please try this KB article's solutions, and report back on how it went for you. If you can provide details about any recent OS updates, backups, or migrations, that would help is nail down the root cause of the new rash of this error for Mac users recently:
    http://kb2.adobe.com/cps/512/cpsid_51260.html

Maybe you are looking for

  • Sending mail in a Java Stored Procedure

    Question: What is the best way to have a Java Stored Procedure connect to our mailhost and send an email? I have been trying to use URL and URLConnection, but I found out URL somehow does not support the "mailto" protocol in the Java Stored Procedure

  • Tracing doesn't work

    Hi, when I put tracing enabled it behaves like it wasn't enabled. I'm using Teststand 3.1 with patch 3.1f1. Best regards, Tito Fidalgo

  • Mac App Store Stuck - Cannot Sign In

    Hello, I've recently upgraded my Mac OS X from 10.6.2 to 10.6.6 . Now the App Store is installed and shown in my dock , but I can't use it because it cannot connect to iTunes Store . When I open the Mac App Store , "Connecting to iTunes Store ..." ap

  • Can a Solr collection index CLOB?

    Hello, everyone, Right now we are trying to figure a way to increase capacity for varchar2 beyond 4000 bytes.  One option on the table is to convert the varchar2 to CLOB.  But that makes me wonder if a Solr collection can index CLOB?? V/r, ^_^

  • Print preview problem after SP upgrade in ECC 6.0

    Hi Everyone, We have recently upgraded support packages in our sap system. During tetsing we have come up some issues with print preview in sp01 transaction. If we click on print preview This is missing a lot of the formatting e.g. lines and logos. H