LiveCycle 2.6.1 Data Management with The ColdFusion 8.0 DataManagement Event Gateway Issue

Hello all,
     I've recently been developing a project that involves sending out events from ColdFusion to LiveCycle 2.6.1 using the Data Management event gateway to Flex 4.0 clients (LiveCycle and ColdFusion are on different Instances, but the same server).  To begin with, I used ColdFusion assemblers, DAO's, and models and everything worked fine locally.  After deploying this setup to a beta site, I decided that this setup would be very troublesome in terms of configuring clustered instances across multiple servers.  I then decided to convert my assemblers, DAO's, and models to Java.  The conversion went well and the flex clients see the exact same data as they did with the ColdFusion adapter.
     Once I tried to send an update through from my ColdFusion application to a Flex client, I get an error stating that:
"Unable to find the Flex adapter for destination My_Dest in the RMI registry on localhost:1099.The Flex adapter may not be running or the destination may be incorrect."
After seeing this error, downloaded a Java-based RMI inspector to see what was going on.  To get a good idea of what was happening when the ColdFusion adapter was being used, I switched my data-management-config file back to the CF adapter.  I noticed that the RMI entry was as follows:
localhost:1099/cfdataserviceadapter/My_Dest
localhost:1099/cfassembler/my_cf_instance
Once I gathered this data as the base, I converted back to the Java adapter in my data-management-config file, restarted the servers, and ran the RMI inspector again.  Only the "localhost:1099/cfassembler/my_cf_instance" was showing.  (This one shows because I have "Enable Remote Adobe LiveCycle Data Management Access" checked in my CF instance's CF Admin -> Flex Integration).  Since I don't need this checked anymore, I unchecked it and re-ran the RMI inspector.  As it should, the "localhost:1099/cfdataserviceadapter/My_Dest" went away.  Since no destination shows up, it means that the Flex adapter isn't registering my "my_Dest" destination with RMI.  Since it isn't registered, I can't see it when I try to send a message through the CF Data Management event gateway.
Can anyone help me out here?  I certainly may be missing something when it comes to RMI (I don't work with Java very often).  Any advice would be greatly appreciated!
Thank you,
Dustin Blomquist

Dustin,
Without the ColdFusion based data management destination defined on the LCDS server, the destination will not show up in the RMI registry.  It is only the CF adapter code that does this.  The 'stock' LCDS adapter does not support invoking via RMI the way the CF version does.
I would recommend you run the LCDS MessageBrokerServlet inside the ColdFusion web application.  This will give you two things:
1. You will not have the overhead of RMI between CF and LCDS as they will share the same VM (better performance!).
2. You will be able to use the CF Data Management Gateway to pass messages to Java-based destinations.  The APIs the gateway uses should work fine with either CF or Java based Data Management destinations.
The CF/LCDS integration doesn't support what you are trying to do when you run two seperate instances.

Similar Messages

  • How many domains can Prime Collaboration Advanced manage with the BE6000?

    The BE6000 Administration guide states that "Most BE6K deployments have a single domain as part of a Standard Prime installation. Multiple domains are available with Prime Collaboration Advanced (available for purchase) that can be used for complex Business Edition 6000 deployments."
    How many domains can Prime Collaboration Advanced manage with the BE6000 solution? How do we order and deploy Prime Collaboration Advanced with the BE6000 solution?

    http://docwiki.cisco.com/wiki/System_Capacity_for_Cisco_Prime_Collaboration_10.0

  • Get the month from a date column with the calculated column

    I am trying to get the month from a date field with the help of calculated column. However I get this syntax error whenever I want to submit the formula:
    Error 
    The formula contains a syntax error or is not supported. 
    the default language of our site is German and [datum, von] is a date field.

    Hi,
    I have created two columns
    Current MM-YY
    Calculated (calculation based on other columns)
    Today
    Date and Time
    Current MM-YY is calculated value with formula as
    =TEXT(Today,"mmmm")
    But the output shows as December instead of May.
    I have tried =TEXT([Datum, von];"mmmm") but no help.
    I am trying to populated the column automatically with current month..ex: if its May the field should show May, next month it should show June an so on.
    Any kind help is grateful.
    Regards,
    Pradeep

  • Errors in the high-level relational engine. A connection could not be made to the data source with the DataSourceID of '

    When I deploy the cube which is sitting on my PC (local) the following 4 errors come up:
    Error 1 The datasource , 'AdventureWorksDW', contains an ImpersonationMode that that is not supported for processing operations.  0 0 
    Error 2 Errors in the high-level relational engine. A connection could not be made to the data source with the DataSourceID of 'Adventure Works DW', Name of 'AdventureWorksDW'.  0 0 
    Error 3 Errors in the OLAP storage engine: An error occurred while the dimension, with the ID of 'Customer', Name of 'Customer' was being processed.  0 0 
    Error 4 Errors in the OLAP storage engine: An error occurred while the 'Customer Alternate Key' attribute of the 'Customer' dimension from the 'Analysis Services Tutorial' database was being processed.  0 0 

    Sorry hit the wrong button there. That is not entire solution and setting it to default would work when using a single box and not in a distributed application solution. If you are creating the analysis database manually or using the wizard then you can
    set the impersonation to your heart content as long as the right permission has been set on the analysis server.
    In my case I was using MS Project Server 2010 to create the database in the OLAP configuration section. The situation is that the underlying build script has been configured to use the default setting which is the SQL Service account and this account does
    not have permission in Project Server I believe.
    Changing the account to match the Project service account allowed for a successful build \ creation of the database. My verdict is that this is a bug in Project Server because it needs to include the option to choose impersonation when creating the Database
    this way it will not use the default which led to my error in the first place. I do not think there is a one fix for all in relations to this problem it is an environment by environment issue and should be resolved as such. But the idea around fixing it is
    if you are using the SQL Analysis server service account as the account creating the database and cubes then default or service account is fine. If you are using a custom account then set that custom account in the impersonation details after you have granted
    it SQL analysis administrator role. You can remove that role after the DB is created and harden it by creating a role with administrative permissions.
    Hope this helps.

  • How to keep data integrity with the two business service in OSB 10.3.1.0

    How to keep data integrity with the two business service in OSB 10.3.1.0
    In our customer system, customer want to keep data integerity between two businness service. I thinks this is XA transaction issue.
    Basing customer requirment, I created a testcase but I can't keep data integerity, For detail information, please refer the attached docs.

    Can you please explain what you meant my data integrity in your use case?
    Manoj

  • How can I make the message date appear with the time for unread messages?

    Mail systems I am moving from display the date & time with the messages. I need this displayed to determine aging and importance of received mail. When archiving/saving old mail it helps filing mail off and not duplicate archived messages.

    Thunderbird displays time and date for all messages except for the present day messages. The time only is displayed for today and the date is appended at midnight.

  • The collection you specified does not exists or is not registered with the ColdFusion Search Service.

    While upgrading to CF7 from CF5, I am creating the new collections from Cf Admin. After creating, I indexed it from cf admin and then trying to search for the documents in that collection, and i am getting the following error. This was working yesterday. I have cheched the logs etc and nothing seems helpful in discovering what the issue is.
    Document count is 11,745 and the size is 25,948 kb, which is not too much for a collection. I have tried to restart Cold Fusion Search Services and then ColdFusion Application services, but all in vain.
    Any help on this would be greatly appreciated.
    Detail
    The collection you specified does not exists or is not registered with the ColdFusion Search Service.
    Message
    The collection rc_collectiom does not exist.
    StackTrace
    coldfusion.tagext.search.CollectionDoesNotExistException: The collection rc_collectiom does not exist. at coldfusion.tagext.search.SearchTag.verifyLocale(SearchTag.java:819) at coldfusion.tagext.search.SearchTag.doSearch(SearchTag.java:200) at coldfusion.tagext.search.SearchTag.doStartTag(SearchTag.java:159) at coldfusion.runtime.CfJspPage._emptyTag(CfJspPage.java:1915) at cfeFull2dText2dReDirect2ecfm511389924.runPage(C:\Inetpub\wwwroot\External\FullText\eFull- Text-ReDirect.cfm:43) at coldfusion.runtime.CfJspPage.invoke(CfJspPage.java:152) at coldfusion.tagext.lang.IncludeTag.doStartTag(IncludeTag.java:349) at coldfusion.runtime.CfJspPage._emptyTag(CfJspPage.java:1915) at cfApplication2ecfc179940445$funcONREQUEST.runFunction(C:\Inetpub\wwwroot\External\Applica tion.cfc:114) at coldfusion.runtime.UDFMethod.invoke(UDFMethod.java:344) at coldfusion.runtime.UDFMethod$ReturnTypeFilter.invoke(UDFMethod.java:290) at coldfusion.runtime.UDFMethod$ArgumentCollectionFilter.invoke(UDFMethod.java:254) at coldfusion.filter.FunctionAccessFilter.invoke(FunctionAccessFilter.java:56) at coldfusion.runtime.UDFMethod.runFilterChain(UDFMethod.java:207) at coldfusion.runtime.UDFMethod.invoke(UDFMethod.java:169) at coldfusion.runtime.TemplateProxy.invoke(TemplateProxy.java:194) at coldfusion.runtime.TemplateProxy.invoke(TemplateProxy.java:146) at coldfusion.runtime.AppEventInvoker.invoke(AppEventInvoker.java:72) at coldfusion.runtime.AppEventInvoker.onRequest(AppEventInvoker.java:178) at coldfusion.filter.ApplicationFilter.invoke(ApplicationFilter.java:215) at coldfusion.filter.RequestMonitorFilter.invoke(RequestMonitorFilter.java:51) at coldfusion.filter.PathFilter.invoke(PathFilter.java:86) at coldfusion.filter.LicenseFilter.invoke(LicenseFilter.java:27) at coldfusion.filter.ExceptionFilter.invoke(ExceptionFilter.java:69) at coldfusion.filter.BrowserDebugFilter.invoke(BrowserDebugFilter.java:52) at coldfusion.filter.ClientScopePersistenceFilter.invoke(ClientScopePersistenceFilter.java:2 8) at coldfusion.filter.BrowserFilter.invoke(BrowserFilter.java:38) at coldfusion.filter.GlobalsFilter.invoke(GlobalsFilter.java:38) at coldfusion.filter.DatasourceFilter.invoke(DatasourceFilter.java:22) at coldfusion.filter.RequestThrottleFilter.invoke(RequestThrottleFilter.java:115) at coldfusion.CfmServlet.service(CfmServlet.java:107) at coldfusion.bootstrap.BootstrapServlet.service(BootstrapServlet.java:78) at jrun.servlet.ServletInvoker.invoke(ServletInvoker.java:91) at jrun.servlet.JRunInvokerChain.invokeNext(JRunInvokerChain.java:42) at jrun.servlet.JRunRequestDispatcher.invoke(JRunRequestDispatcher.java:257) at jrun.servlet.ServletEngineService.dispatch(ServletEngineService.java:541) at jrun.servlet.jrpp.JRunProxyService.invokeRunnable(JRunProxyService.java:204) at jrunx.scheduler.ThreadPool$DownstreamMetrics.invokeRunnable(ThreadPool.java:318) at jrunx.scheduler.ThreadPool$ThreadThrottle.invokeRunnable(ThreadPool.java:426) at jrunx.scheduler.ThreadPool$UpstreamMetrics.invokeRunnable(ThreadPool.java:264) at jrunx.scheduler.WorkerThread.run(WorkerThread.java:66)

    So at last I figured this out myself and found a temporary resolution for this kind of problem. Majority of the times try creating the Collection through Cfcollection tag and then index it using cfindex tag, rather than performing these operations from the Cold Fusion 7 administrator.
    The only reason I can think of which might have caused the issue is – I have thousands of documents in the collection which I indexed. So while indexing the page gets expired. So I tried again to index it. This time the collection got indexed fairly quickly and I got the response that the collection has been indexed successfully. So for the first time when I tried to index the collection, there might be some piece of code, which did not register the collection properly and hence giving this error.
    When I did the same function using the code, with exactly same set of documents, it did not throw any error, even when I executed it for the first time. So the collection indexed properly.
    I am not claiming that technically this theory is correct, but this seems to be the problem.
    If someone knows technically what has happened at the back, please share it. For the time being, this is the fix – if you experience any problem in creating/indexing collections through cfadmin in CF7, do it thorough code.

  • CF8 Verity "The collection you specified does not exist or is not registered with the ColdFusion Search Service."

    I'm running ColdFusion 8 Enterprise on linux. I'm able to
    create collections and index them through cfadmin as well as in cfm
    application pages, but when trying to search I get the error
    "The collection you specified does not exist or is not
    registered with the ColdFusion Search Service."
    I'm using the collection name in cfsearch and not the full
    path.

    Would I have been better off posting this in the General
    Discussion section? Could the moderators move it if so,
    please?

  • Master Data Management with SAP

    Hi
    Master data management in SAP platform is in Compilation
    of report structure where you have additional variables such as field documentation code a field processor and
    a system capable to do this compilation./In other words SAP Master Data Management is a code programmer In Java
    . Master data Mangement is a programmer's capacity to do
    his compilation in Java. Vb C++ is the system mangement architecture. Code is fabricated in this environment for Sap Master Data Management
    K Balakrishnan

    Hello Michelle,
    First of all, MDM Admin UI and SRM-MDM Search UI both are in same sca file (SRMMDMCAT20SP03P05.sca) and there is another sca file with API and MDM Connector (MDMJAVAAPI06_2.sca).
    Remember that files name might be change as per the business content version.
    So there are two file which you need to deploy. You can deploy these two file on any Web Application Server (WAS 7.0 and Above) using Software Deployment Manager (SDM) of SAP NW AS J2EE installation.
    For further information about these files, you can refer the SAP Note- 1057316 -- SRM-MDM Catalog 2.0 - Installation & Configuration.
    For checking the successful deployment of these SCA files, you can check with the mentioned below URL:
    http://ServerName:PortNo/webdynpro/dispatcher/sap.com/tcmdmsrmcat~uisearch/MDM_SRM_UI_App
    go to the above link, try to connect with the MDM Server which you have installed. After connecting the MDM Server, Choose your SRM-MDM Catalog repository and try to connect with the user as "Master" with blank pwd. If you are able to login in the repository, it means you have successfully deployed the correct sca files.
    For further details regarding this scenario, you can refer my answer in below thread:
    installion of SRM-MDM Catalog 2.0
    I hope, this will clear your doubt.If you need further help, Kindly update us.
    TNR,
    Saurabh...

  • Syndication process - Tags in Data manager Selecting the map in Syndication

    Hello Experts,
    I am supporting a repository in MDM 7.1, and there is a part of the syndication process that i don´t know how does it work. This is the process what i mean:
    The repository works with 3 different maps, the map is selected in  MDM Data Manager belong a 'syndicatable' tag. Two of the maps are selected inside Qalified tables.
    What i need to know is where could i see the configuration for that automatic mapping (it is performed belong MDSS), could anybody help me?
    Thanks in advance,
    Mariano Pinkava.

    Hi Mariano,
    In addition, please go through page 223 (mds.ini) & 272 of Console Reference Guide.
    Syndication File Location:
    When MDSS completes a syndication to a port, it places the syndication file in the portu2019s Ready folder. The Ready folder is part of the following fixed directory structure, located beneath the MDM Serveru2019s distribution root directory:
    root/DBMSinstance_DBMStype/RepositoryName/Outbound/ RemoteSystem/PortName/Ready
    where:
    u2022 root is the distribution root directory (set in mds.ini).
    u2022DBMSinstance is the network identifier used to specify the DBMS instance name and DBMStype is the four-character identifier for the DBMS type (i.e. MSQL, ORCL, IDB2).
    u2022RepositoryName, RemoteSystem, and PortName are the values entered in the Code property for each item in MDM Console.
    Thanks,
    Priti

  • How is index data managed within the cluster?

    Hello
    We are in the process of evaluating Coherence.
    With regard to indexes, am I correct in thinking that the data for the index is stored in an (internally managed) cache on the same node as the attendant indexed values? I'm not interested in getting at the index data, but rather just understanding how the index data itself is treated/stored within the cluster.
    So in a 2 node cluster where an index is applied to a distributed cache that only has 2 entries, will there be an index cache on each node that has 1 {Binary,key} entry where the key points to the value that is stored on that same node? That is, I trust that the index data is spread out across the cluster in the case of a partitioned topology?
    Regards
    Peter

    user11279467 wrote:
    Hello
    We are in the process of evaluating Coherence.
    With regard to indexes, am I correct in thinking that the data for the index is stored in an (internally managed) cache on the same node as the attendant indexed values? I'm not interested in getting at the index data, but rather just understanding how the index data itself is treated/stored within the cluster.
    So in a 2 node cluster where an index is applied to a distributed cache that only has 2 entries, will there be an index cache on each node that has 1 {Binary,key} entry where the key points to the value that is stored on that same node? That is, I trust that the index data is spread out across the cluster in the case of a partitioned topology?
    Regards
    PeterHi Peter,
    up to 3.4.x indexes were supported only on partitioned topology, and yes, indexes on each node contain only data corresponding to entries which have their primary copy on that node (referred to as local entries from now on).
    The indexes are not stored in a cache but they are managed as part of the information Coherence maintains about the backing map (the map which contains local entries).
    The indexes are made of 2 parts:
    - forward index: this is a mapping from the cache key in an internal representation (backing map keys from now on) to the value extracted with the extractor used to create the index from the entry belonging to the cache key (extracted value from now on)
    - reverse index (aka. reverse map): this is a mapping from the extracted value to a set of backing map keys of entries from which the reverse index key was extracted with the extractor used to create the index
    The index can be sorted which means that the reverse index will be a SortedMap (so sorting order is the ordering of extracted values)
    Let's see an example. Assuming you have the following cache content within a particular storage-enabled node :
    A --> Object (getX=5 ,...)
    B --> Object (getA=3 ,...)
    C --> Object (getA=3 ,...)
    D --> Object (getA=4 ,...)
    The forward index in that node will contain:
    Binary(A) --> 5
    Binary(B) --> 3
    Binary(C) --> 3
    Binary(D) --> 4
    The reverse index will contain:
    3 --> { Binary(B), Binary(C) }
    4 --> { Binary(D) }
    5 --> { Binary(A) }
    Where Binary(...) means the internal (binary) representation of the ... object.
    If the index is not ordered, then the order of iteration on entries in the reverse index are not deterministic.
    If the index is ordered, then the iteration of entries in the reverse index will be as shown above. Also, you can cast the reverse index to SortedMap so you have the very useful headMap and tailMap and firstKey and lastKey methods.
    Hope this helps.
    Best regards,
    Robert

  • I did a software upgrade today and got a failed error Now all I get is picture of a data cable with the red Itunes logo. Restores and updates all fail now

    I tried to update my iPhone 6 to the newest version today. I got a message that the update failed. I now only see a picture of the data cable and the red iTunes picture. All updates and restores fail.

    I had a similar problem with one of my phones yesterday. It went I to recovery mode. I could not get it to complete the restoration process.  I had to restart my computer in order to complete the restoration through iTunes. I don't know if that will help but it is worth a try. Good luck.

  • I am trying to download the new episode of Game of Thrones on my iPad. We are currently travelling in Italy and have an unlimited pre- paid data plan with the Italian carrier Wind.  When I try download the episode through iTunes I get an error message

    Hi there, traveling through Italy with my iPad. Purchased a unlimited prepaid data plan with Italian carrier WIND.  Has been working excellent, but when I try to download a new episode of Game of Thrones, I get an error message which says that I have to connect to a wi fi network or download the episode on my MacBook.  Why can't I download using the microsim and prepaid data plan?  Any help will be great thank you!

    Unfortunately downloads larger than 20mbps require wifi due to most carriers having data limits. This "requirement" keeps people who are not familiar with data and what takes up a lot of data from going over.

  • How to install Oracle BPEL Process Manager with the BEA WebLogic

    Hi ,
    I will install Oracle BPEL Process Manager with BEA WebLogic 9.2(MP2). I have download orabpel_10133_WebLogic.zip ,then Modify the following mandatory installation properties in the orabpel_10133_WebLogic\bpelDomain.properties file:
    # BEA_HOME is the path where Weblogic is Installed
    BEA_HOME=/opt/bea
    # JAVA_HOME is the path of jdk folder inside your weblogic
    JAVA_HOME=/opt/bea/jrockit90_150_10
    # DOMAIN_HOME is the path where you wish to create your domain called BPELDomain
    DOMAIN_HOME=/opt/bea/user_projects/domains
    # APPS_HOME is the path where you wish to copy your applications and adapters that are required for oracleBPELServer
    APPS_HOME=/opt/bea/user_projects/apps
    # BEA_HOME is the path where BPEL PM is Installed
    BPEL_HOME=/home/oracle/bpel/product/10.1.3.1/OraBPEL_1/bpel
    # DRIVER_TYPE is the datasource class that installable use to create a datasources for oracleBPELServer
    DRIVER_TYPE=oracle.jdbc.xa.client.OracleXADataSource
    # DB_URL is the url to connect to orabpel schema
    DB_URL=jdbc:oracle:thin:@16.157.134.17:1521:orcl
    # DB_USER is the user Id for orabpel shema in database
    DB_USER=ORABPEL
    #DB_PASSWORD is the password for orabpel schema in database
    DB_PASSWORD=bpel
    #BPEL_SERVER_NAME is the server i.e. to be created under BPELDomain
    BPEL_SERVER_NAME=oracleBPELServer
    #PROXY_HOST is the Host name of the proxy server
    PROXY_HOST=www-proxy.us.oracle.com
    #PROXY_HOST=
    #PROXY_PORT is the Port where the proxy server is running
    PROXY_PORT=80
    #PROXY_PORT=
    #NON_PROXY_HOST is the list of non proxy hosts that are divided by a | symbol
    #NON_PROXY_HOST=*.oracle.com|*.oraclecorp.com|localhost|127.0.0.1|stbbn10|stbbn10.us.oracle.com
    NON_PROXY_HOST=*.oracle.com|*.oraclecorp.com|localhost|127.0.0.1|stbbn10|stbbn10.us.oracle.com|16.157.134.135
    When I run the setup.sh , it will report
    BUILD FAILED
    /opt/software/WL_Installables/build.xml:131: Traceback (innermost last):
    File "./wl_scripts/bpelDomain.py", line 22, in ?
    File "./wl_scripts/createGroupsAndUsers.py", line 4, in ?
    weblogic.management.utils.AlreadyExistsException: [Security:090267]Group BpelGroup
    Actully ,there is no BpelGroup in Weblogic. Does anybody know how to solve it ?

    MAke sure you have not set ANY environment variable related to Oracle / BEA / Java / LD_library path. Use the following script to unset / set the initial settings:
    #!/bin/sh
    unset ORACLE_BASE ORACLE_HOME ORACLE_SID ORACLE_TERM
    unset LD_LIBRARY_PATH LD_LIBRARY_PATH_64
    unset CLASSPATH JAVA_HOME
    export PATH=.:/usr/sbin:/usr/bin:/usr/local/bin:/opt/VRTS/bin
    export BEA_HOME=/appl/oracle/products/9.2/weblogic
    Marc
    http://orasoa.blogspot.com

  • How to compare the date value with the date datatype column?

    Hi,
    I am executed a query to get output for the particular date in two different database.
    I got output in one db environment, but in the another environment i didnt get the output.
    APP_DECISION_DATE column contains date datatype
    ex:
    SQL> select DECISIONED_STAGE_ID,DECISIONED_STAGE_USER_ID from naap10_application where APP_DECISION_DATE = TO_DATE('23-JAN-07',
    'DD-MON-YY'); 2
    DECISIONED_STAGE_ID DECISIONED_STAG
    2006060106 SYSTEM
    DATABASE - II
    SQL> select DECISIONED_STAGE_ID,DECISIONED_STAGE_USER_ID from naap10_application where APP_DECISION_DATE=
    TO_DATE('31-JAN-2007','DD-MON-RRRR') and application_id=2007010028552;
    no rows selected
    in the same db i got the output while i used the TO_DATE(APP_DECISION_DATE)
    SQL> select DECISIONED_STAGE_ID,DECISIONED_STAGE_USER_ID from naap10_application where TO_DATE(APP_DECISION_DATE)=
    TO_DATE('31-JAN-2007','DD-MON-YYYY') and application_id=2007010028552; 2
    DECISIONED_STAGE_ID DECISIONED_STAG
    2006060113 SYSTEM
    is it necessary to use TO_DATE function while i am using the data datatype column in the WHERE CLAUSE?
    Thanks in advance.
    siva

    Sorry & thanks sathish,
    Now i got the result.
    SQL> select DECISIONED_STAGE_ID,DECISIONED_STAGE_USER_ID from naap10_application where APP_DECISION_DATE BETWEEN TO_DATE ('31.01.2007 00:00:00', 'DD.MM.YYYY HH24:MI:SS') AND TO_DATE ('31.01.2007 23:59:59', 'DD.MM.YYYY HH24:MI:SS')
    2 and application_id=2007010028552;
    DECISIONED_STAGE_ID DECISIONED_STAG
    2006060113 SYSTEM
    ================================
    But one doubt,
    When i created the column with the date datatype how it gets the time value,
    And in one environment db i got the result but in another environment i dont get,
    Shall i want to change any session status of date?

Maybe you are looking for

  • How to get denomination of Notes ? for the amount paid

    Dear Freinds    i have requirement where the total amount obtained by a employee through Cash . They wanted to have denomination for the amount paid ito that employee ex : An X employee got RS.5550 rupees as cash. They wanted to know how much money h

  • Could I use EJB in mission-critical task,like telecom realtime billing?

    We will begin a project about telecom billing.The key point of the project is the realtime billing,which is must be processed with high speed and efficiency.We have a plan to use Tuxedo on which to run some services coded by c lanaguage.Other managem

  • Can't Find Photos

    Hi there, I am having a problem with my new iPhone 3G where one of my friends took a load of pictures of my graduation. I took them from his digital camera onto my laptop in a folder and then just copied that folder onto the iPhone's hard drive. The

  • Curve fitting during daq

    I would like to perform curve fitting during my data acquisition such that, as more data is collected, the curve will be more accurate. So the fit curve should be anchored at some point in time, and then update on my chart as data is collected. I wou

  • Problem installing premiere pro cs6

    I have a 64 bit windows 7 hp pavilion dv7 6135dx and when I try to launch premier after installing it gives me an error saying i need to update my video display drivers.  My drivers are up to date and I've used premiere pro cs5.5 on this computer wit