Different process revisions with inbound Database Adapters

Hi,
we've got a BPEL-process deployed in different revisions (1.0 and 1.1). The process state of Revision 1.0 is still "on", but it is marked as "inactive" because no new instances of this process are needed, but old instances are still running. Revision 1.1 of the process is marked as default revision with state "on/active".
The processes are started by a database adapter, polling a single view that is based on a single table for new records and in the process flow we've got further adapters, waking up the process-instance, while waiting for input from other components of our system. The views are accessed by synonyms.
Correlation of messages to the instances is done by correlation sets, based on data read from the view.
Because of some refactorings in our application, we had to rename the views and simply create different synonyms, pointing on the renamed views, so that both processes can use it. The problem is now that we monitored, old process-instances of revision 1.0 which are not running further. Analyzing the problem, we find all messages for the old revision in the callback section of the administration-tab in the BPEL-Console, pointing to the new revision and the new view name, so this messages could never be delivered to old process-revision!
What I can imagine is that the adapter has problems with accessing the same view with different synonyms. Or is it possible that there are some issues with the adapters activation agents?
We are using SOA Suite version 10.1.3.4. Any help is appreciated.
Thanks in advance
Sven
PS: Here a short example of the scenario regarding the view:
Before: View: test1_v ( id, name ); accessible by synonym test1_v
After: View: test2_v ( id, name ); accessible by synonyms test1_v and test2_v

Hi,
we've got a BPEL-process deployed in different revisions (1.0 and 1.1). The process state of Revision 1.0 is still "on", but it is marked as "inactive" because no new instances of this process are needed, but old instances are still running. Revision 1.1 of the process is marked as default revision with state "on/active".
The processes are started by a database adapter, polling a single view that is based on a single table for new records and in the process flow we've got further adapters, waking up the process-instance, while waiting for input from other components of our system. The views are accessed by synonyms.
Correlation of messages to the instances is done by correlation sets, based on data read from the view.
Because of some refactorings in our application, we had to rename the views and simply create different synonyms, pointing on the renamed views, so that both processes can use it. The problem is now that we monitored, old process-instances of revision 1.0 which are not running further. Analyzing the problem, we find all messages for the old revision in the callback section of the administration-tab in the BPEL-Console, pointing to the new revision and the new view name, so this messages could never be delivered to old process-revision!
What I can imagine is that the adapter has problems with accessing the same view with different synonyms. Or is it possible that there are some issues with the adapters activation agents?
We are using SOA Suite version 10.1.3.4. Any help is appreciated.
Thanks in advance
Sven
PS: Here a short example of the scenario regarding the view:
Before: View: test1_v ( id, name ); accessible by synonym test1_v
After: View: test2_v ( id, name ); accessible by synonyms test1_v and test2_v

Similar Messages

  • Goods receipt against process order with inbound idoc

    Hi Experts,
    We need to post good receipt against process order with movement type 101 and for By-product 531.
    Can we use inbound idoc-WMMBID02,with message type WMMBXY to achieve this functionality?
    Or please advise if we can achieve this  from any other Idoc type.
    We are getting below field values for external system
    1. process order number
    2.plant
    3.material
    4.quantity in unit of entry
    5.unit of measure
    6.storage location
    7. movement type
    8.Batch
    Kindly advise how can we achieve this.
    Best Regards
    CS

    261 and 262 are two different movement types INDEPENDENT of each other. I definitely know for the fact that SAP allows you to create a 262 without a 261. if you were to restrict it... you can disable access through a tcode for that mov. type 262 and ask users to use MBST to cancel any material documents created via 261 movt type.. that would ensure a smooth process

  • Getting error: ORA-04030: out of process memory with EBS database

    Dear experts,
    I am getting this error with the EBS 11i database during the running of the system specially in the rush hour of the work day leads to disconnect between the database server and client:
    ORA-04030: out of process memory when trying to allocate 917536 bytes (joxcx callheap,ioc_allocate ufree)
    OS: Windows Server 2003 32bit
    Memory: 4GB RAM
    Oracle database version: 9.2.0.6.0
    db_cache_size=629145600
    sga_max_size=1400897536
    pga_aggregate_target=1073741824
    sessions=800
    processes=400
    Thanks

    This error is expected, especially if enough memory is not allocated and the system is heavily loaded, as it seems to be in your case. Pl see this MOS Doc for possible resolution
    Master Note for Diagnosing OS Memory Problems and ORA-4030 [ID 1088267.1]     
    HTH
    Srini

  • Starting Process Flows with a database trigger

    Hi all,
    Is it possible to start a process flow using a trigger? I am using the sqlplus_exec_template.sql at the moment.
    Thanks in advance,
    Regards,
    Sebastian

    Hi Javor,
    Thanks for your reply. I am using version 10g . After embedding the process flow execution into a procedure, I am getting the following error.
    Connecting to the database userx.
    ORA-20001: Task not found - Please check the Task Type, Name and Location are correct.
    ORA-06512: at "OWB_REPOS.WB_RT_API_EXEC", line 704
    ORA-06512: at "userx.RUN_PF", line 27
    ORA-06512: at line 2
    Stage 1: Decoding Parameters
    | location_name=MY_WORKFLOW
    | task_type=PROCESSFLOW
    | task_name=MY_PROCESS_FLOW
    Stage 2: Opening Task
    Process exited.
    Disconnecting from the database userx.
    Extract of the procedure (just like the template).
    -- Initialize Return Code
    exec_return_code := wb_rt_api_exec.RESULT_FAILURE;
    -- Run Task
    exec_return_code := wb_rt_api_exec.run_task(LOCATION_NAME,TASK_TYPE,TASK_NAME,CUSTOM_PARAMS,SYSTEM_PARAMS,OEM_FRIENDLY,OWB_BACKGROUND);
    Regards,
    Sebastian

  • Three database or multiple Database adapters in One Composite

    Hi All,
    I have below requirement:
    1) I have two applications A and B.I am integrating them with SOA.I have to recieve data from A and update that data into application B for different Device IDs.
    I am using one database adapter(insert operation) to recieve the data from A and insert that data into a staging table.(I have to hold on to the data into a table for 24 hrs and update the whole data into B at the ene of the day.)
    BPEL-1 : Application A --- > Staging table(insert DB Adapter) (for every half an hour interval in that day)
    That data(two request schema example) would be like below:
    1) <devID> 345 </devID>
    <Con>45</devId>
    2) <devID> 343 </devID>
    <Con>45</devId>
    I will get the above schema for evry interval for dfifferent devIds. I am inserting the data into only one staging table even though the data would be for different device IDs.
    After inserting the data for 24 hrs in the staging table,I have to use Application B WSDL url to update data into B.
    I am using two Database adapters(one for each deviceId) for selecting the data for two device IDS and inserting into Application B.
    BPEL-2:DB adapter(Select operation-- Staging table which has the data for this deviceID 345) ----------> application B WSDL
    BPEL-3 : DB adapter(Select operation-- Staging table which has the data for this deviceIDs 343) ----------> application B WSDL
    As of now after entering the data for couple of intervals for two deviceIDs,I am getting instances for BPEL-1 and after BPEL-2 which is for device ID 345.
    But I am not getting instances for BPEL-3 which is for device ID 343.
    Can I use three DB adapters(one for Insert,two for selecting the data from same table?
    Why I am not able to update the data related to device Id 343 into B.?
    Please guide me on this.

    Hi Arik,
    I am using seperate DB adapters (for each operation of insert(for any device ID),Select(for device ID1),Select(for Device ID 2) etc.
    I have created seperate BPEL processes for above each operation as every device should be in seperate request.
    Where do u want me to commit?
    As of now, after u said I have set the Unit of wrok value as true in all _db.jca files.
    Please find the below jca files for each operation:
    Insert operation:
    <adapter-config name="InsertWinamrData" adapter="Database Adapter" wsdlLocation="InsertWinamrData.wsdl" xmlns="http://platform.integration.oracle/blocks/adapter/fw/metadata">
    <connection-factory location="eis/DB/SOADB" UIConnectionName="SOADB" adapterRef=""/>
    <endpoint-interaction portType="InsertWinamrData_ptt" operation="insert">
    <interaction-spec className="oracle.tip.adapter.db.DBWriteInteractionSpec">
    <property name="DescriptorName" value="InsertWinamrData.Imdreading"/>
    <property name="DmlType" value="insert"/>
    <property name="MappingsMetaDataURL" value="InsertWinamrData-or-mappings.xml"/>
    <property name="DetectOmissions" value="true"/>
    <property name="GetActiveUnitOfWork" value="true"/>
    </interaction-spec>
    </endpoint-interaction>
    </adapter-config>
    Select ope on one device ID:
    <adapter-config name="GetIMDReadingIntoMDM" adapter="Database Adapter" wsdlLocation="GetIMDReadingIntoMDM.wsdl" xmlns="http://platform.integration.oracle/blocks/adapter/fw/metadata">
    <connection-factory location="eis/DB/SOADB" UIConnectionName="SOADB" adapterRef=""/>
    <endpoint-activation portType="GetIMDReadingIntoMDM_ptt" operation="receive">
    <activation-spec className="oracle.tip.adapter.db.DBActivationSpec">
    <property name="DescriptorName" value="GetIMDReadingIntoMDM.Imdreading"/>
    <property name="QueryName" value="GetIMDReadingIntoMDMSelect"/>
    <property name="MappingsMetaDataURL" value="GetIMDReadingIntoMDM-or-mappings.xml"/>
    <property name="PollingStrategy" value="LogicalDeletePollingStrategy"/>
    <property name="MarkReadColumn" value="READSTATUS"/>
    <property name="MarkReadValue" value="YES"/>
    <property name="MarkUnreadValue" value="NO"/>
    <property name="PollingInterval" value="240"/>
    <property name="MaxRaiseSize" value="4"/>
    <property name="MaxTransactionSize" value="8"/>
    <property name="NumberOfThreads" value="1"/>
    <property name="ReturnSingleResultSet" value="false"/>
    <property name="EnableStreaming" value="false"/>
    <property name="GetActiveUnitOfWork" value="false"/>
    </activation-spec>
    </endpoint-activation>
    </adapter-config>
    seldct of other device ID:
    <adapter-config name="GetIDMReadingIntoOne" adapter="Database Adapter" wsdlLocation="GetIDMReadingIntoOne.wsdl" xmlns="http://platform.integration.oracle/blocks/adapter/fw/metadata">
    <connection-factory location="eis/DB/SOADB" UIConnectionName="SOADB" adapterRef=""/>
    <endpoint-activation portType="GetIDMReadingIntoOne_ptt" operation="receive">
    <activation-spec className="oracle.tip.adapter.db.DBActivationSpec">
    <property name="DescriptorName" value="GetIDMReadingIntoOne.Imdreading"/>
    <property name="QueryName" value="GetIDMReadingIntoOneSelect"/>
    <property name="MappingsMetaDataURL" value="GetIDMReadingIntoOne-or-mappings.xml"/>
    <property name="PollingStrategy" value="LogicalDeletePollingStrategy"/>
    <property name="MarkReadColumn" value="READSTATUS"/>
    <property name="MarkReadValue" value="YES"/>
    <property name="MarkUnreadValue" value="NO"/>
    <property name="PollingInterval" value="360"/>
    <property name="MaxRaiseSize" value="4"/>
    <property name="MaxTransactionSize" value="8"/>
    <property name="NumberOfThreads" value="1"/>
    <property name="ReturnSingleResultSet" value="false"/>
    <property name="GetActiveUnitOfWork" value="false"/>
    </activation-spec>
    </endpoint-activation>
    </adapter-config>
    Please guide me on this.
    Edited by: 899283 on Nov 25, 2012 10:27 PM

  • SMON process terminated with error

    Hi,
    Oracle Version:10.2.0.3
    Operating system:Linux
    I alert log file i found that smon process was terminated with error.Below is the contents of the alert log file.
    Fri Dec 31 05:12:32 2010
    sculkget: failed to lock /home/oracle/oracle/oracle/product/10.2.0/db_1/dbs/lkinstqfundeca exclusive
    sculkget: lock held by PID: 9127
    Oracle Instance Shutdown operation failed. Another process may be attempting to startup or shutdown this Instance.
    Failed to acquire instance startup/shutdown serialization primitive
    Fri Dec 31 05:56:23 2010
    sculkget: failed to lock /home/oracle/oracle/oracle/product/10.2.0/db_1/dbs/lkinstqfundeca exclusive
    sculkget: lock held by PID: 9127
    Oracle Instance Shutdown operation failed. Another process may be attempting to startup or shutdown this Instance.
    Failed to acquire instance startup/shutdown serialization primitive
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    WARNING: inbound connection timed out (ORA-3136)
    Fri Dec 31 06:40:53 2010
    MMNL absent for 7904 secs; Foregrounds taking over
    Fri Dec 31 06:40:53 2010
    Starting background process EMN0
    Fri Dec 31 06:40:53 2010
    The value (30) of MAXTRANS parameter ignored.
    Fri Dec 31 06:40:53 2010
    Thread 1 cannot allocate new log, sequence 62004
    Checkpoint not complete
      Current log# 1 seq# 62003 mem# 0: /u02/oradata/qfundeca/redo01a.log
      Current log# 1 seq# 62003 mem# 1: /u03/oradata/qfundeca/redo01b.log
    EMN0 started with pid=29, OS id=7176
    Fri Dec 31 06:40:53 2010
    Shutting down instance: further logons disabled
    Fri Dec 31 06:40:53 2010
    ksvcreate: Process(m001) creation failed
    Fri Dec 31 06:40:53 2010
    kkjcre1p: unable to spawn jobq slave process, error 1089
    Fri Dec 31 06:40:53 2010
    kkjcre1p: unable to spawn jobq slave process, error 1089
    Fri Dec 31 06:40:54 2010
    Stopping background process QMNC
    Fri Dec 31 06:40:54 2010
    Stopping background process CJQ0
    Fri Dec 31 06:40:54 2010
    Thread 1 cannot allocate new log, sequence 62004
    Private strand flush not complete
      Current log# 1 seq# 62003 mem# 0: /u02/oradata/qfundeca/redo01a.log
      Current log# 1 seq# 62003 mem# 1: /u03/oradata/qfundeca/redo01b.log
    Fri Dec 31 06:40:56 2010
    Stopping background process MMNL
    Fri Dec 31 06:40:56 2010
    Thread 1 advanced to log sequence 62004
      Current log# 2 seq# 62004 mem# 0: /u02/oradata/qfundeca/redo02a.log
      Current log# 2 seq# 62004 mem# 1: /u03/oradata/qfundeca/redo02b.log
    Fri Dec 31 06:40:57 2010
    Starting background process QMNC
    Fri Dec 31 06:40:57 2010
    Errors in file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_3065.trc:
    ORA-00443: background process "QMNC" did not start
    Fri Dec 31 06:41:04 2010
    Stopping background process MMON
    Fri Dec 31 06:41:05 2010
    Errors in file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_9127.trc:
    ORA-00604: error occurred at recursive SQL level 1
    ORA-12663: Services required by client not available on the server
    ORA-36961: Oracle OLAP is not available.
    ORA-06512: at "SYS.OLAPIHISTORYRETENTION", line 1
    ORA-06512: at line 15
    Fri Dec 31 06:41:05 2010
    Shutting down instance (immediate)
    License high water mark = 93
    Fri Dec 31 06:41:05 2010
    Stopping Job queue slave processes
    Fri Dec 31 06:41:05 2010
    Job queue slave processes stopped
    Fri Dec 31 06:41:57 2010
    All dispatchers and shared servers shutdown
    Fri Dec 31 06:41:59 2010
    ALTER DATABASE CLOSE NORMAL
    Fri Dec 31 06:41:59 2010
    SMON: disabling tx recovery
    SMON: disabling cache recovery
    Fri Dec 31 06:42:00 2010
    Shutting down archive processes
    Archiving is disabled
    Fri Dec 31 06:42:05 2010
    ARCH shutting down
    ARC1: Archival stopped
    Fri Dec 31 06:42:11 2010
    ARCH shutting down
    ARC0: Archival stopped
    Fri Dec 31 06:42:12 2010
    Thread 1 closed at log sequence 62004
    Successful close of redo thread 1
    Fri Dec 31 06:42:12 2010
    Completed: ALTER DATABASE CLOSE NORMAL
    Fri Dec 31 06:42:12 2010
    ALTER DATABASE DISMOUNT
    Completed: ALTER DATABASE DISMOUNT
    ARCH: Archival disabled due to shutdown: 1089
    Shutting down archive processes
    Archiving is disabled
    Archive process shutdown avoided: 0 active
    ARCH: Archival disabled due to shutdown: 1089
    Shutting down archive processes
    Archiving is disabled
    Archive process shutdown avoided: 0 active
    Fri Dec 31 06:49:52 2010
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =27
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
      processes                = 150
      sessions                 = 250
      __shared_pool_size       = 385875968
      __large_pool_size        = 16777216
      __java_pool_size         = 16777216
      __streams_pool_size      = 33554432
      streams_pool_size        = 33554432
      sga_target               = 1610612736
      control_files            = /u02/oradata/qfundeca/control01.ctl, /u02/oradata/qfundeca/control02.ctl, /u03/oradata/qfundeca/control03.ctl
      db_block_size            = 8192
      __db_cache_size          = 1140850688
      compatible               = 10.2.0.3.0
      log_archive_dest_1       = LOCATION=/u02/oradata/arch/qfundeca
      log_archive_format       = %t_%s_%r.dbf
      archive_lag_target       = 2700
      db_file_multiblock_read_count= 16
      db_recovery_file_dest    = /u02/oradata/flash_recovery_area
      db_recovery_file_dest_size= 21474836480
      standby_file_management  = AUTO
      undo_management          = AUTO
      undo_tablespace          = UNDOTBS1
      remote_login_passwordfile= EXCLUSIVE
      db_domain                =
      dispatchers              = (PROTOCOL=TCP) (SERVICE=qfundecaXDB)
      job_queue_processes      = 17
      cursor_sharing           = EXACT
      background_dump_dest     = /home/oracle/oracle/admin/qfundeca/bdump
      user_dump_dest           = /home/oracle/oracle/admin/qfundeca/udump
      core_dump_dest           = /home/oracle/oracle/admin/qfundeca/cdump
      audit_file_dest          = /home/oracle/oracle/admin/qfundeca/adump
      db_name                  = qfundeca
      open_cursors             = 300
      pga_aggregate_target     = 839909376
    PMON started with pid=2, OS id=9591
    PSP0 started with pid=3, OS id=9593
    MMAN started with pid=4, OS id=9595
    DBW0 started with pid=5, OS id=9597
    LGWR started with pid=6, OS id=9599
    CKPT started with pid=7, OS id=9601
    SMON started with pid=8, OS id=9603
    RECO started with pid=9, OS id=9605
    CJQ0 started with pid=10, OS id=9607
    MMON started with pid=11, OS id=9609
    Fri Dec 31 06:49:54 2010
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=12, OS id=9611
    Fri Dec 31 06:49:54 2010
    starting up 1 shared server(s) ...
    Oracle Data Guard is not available in this edition of Oracle.
    Fri Dec 31 06:49:54 2010
    ALTER DATABASE   MOUNT
    Fri Dec 31 06:49:58 2010
    Setting recovery target incarnation to 1
    Fri Dec 31 06:49:58 2010
    Successful mount of redo thread 1, with mount id 256948178
    Fri Dec 31 06:49:58 2010
    Database mounted in Exclusive Mode
    Completed: ALTER DATABASE   MOUNT
    Fri Dec 31 06:49:58 2010
    ALTER DATABASE OPEN
    Fri Dec 31 06:49:58 2010
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=16, OS id=9619
    Fri Dec 31 06:49:58 2010
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=17, OS id=9621
    Fri Dec 31 06:49:58 2010
    Thread 1 opened at log sequence 62004
      Current log# 2 seq# 62004 mem# 0: /u02/oradata/qfundeca/redo02a.log
      Current log# 2 seq# 62004 mem# 1: /u03/oradata/qfundeca/redo02b.log
    Successful open of redo thread 1
    Fri Dec 31 06:49:58 2010
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    Fri Dec 31 06:49:58 2010
    ARC1: Becoming the heartbeat ARCH
    Fri Dec 31 06:49:58 2010
    SMON: enabling cache recovery
    Fri Dec 31 06:49:58 2010
    Successfully onlined Undo Tablespace 1.
    Fri Dec 31 06:49:58 2010
    SMON: enabling tx recovery
    Fri Dec 31 06:49:58 2010
    Database Characterset is WE8ISO8859P1
    replication_dependency_tracking turned off (no async multimaster replication found)
    Starting background process QMNC
    QMNC started with pid=18, OS id=9623
    Fri Dec 31 06:49:59 2010
    Errors in file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_9617.trc:
    ORA-00604: error occurred at recursive SQL level 1
    ORA-12663: Services required by client not available on the server
    ORA-36961: Oracle OLAP is not available.
    ORA-06512: at "SYS.OLAPIHISTORYRETENTION", line 1
    ORA-06512: at line 15
    Fri Dec 31 06:50:00 2010
    Completed: ALTER DATABASE OPEN
    Fri Dec 31 06:50:00 2010
    db_recovery_file_dest_size of 20480 MB is 0.00% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    Fri Dec 31 06:55:59 2010
    The value (30) of MAXTRANS parameter ignored.
    kupprdp: master process DM00 started with pid=36, OS id=12004
             to execute - SYS.KUPM$MCP.MAIN('QFUNDECA_PROD_BKP', 'QFUNDECAPROD', 'KUPC$C_1_20101231065559', 'KUPC$S_1_20101231065559', 0);
    kupprdp: worker process DW01 started with worker id=1, pid=37, OS id=12009
             to execute - SYS.KUPW$WORKER.MAIN('QFUNDECA_PROD_BKP', 'QFUNDECAPROD');
    Fri Dec 31 07:04:59 2010
    Thread 1 cannot allocate new log, sequence 62005
    Private strand flush not complete
      Current log# 2 seq# 62004 mem# 0: /u02/oradata/qfundeca/redo02a.log
      Current log# 2 seq# 62004 mem# 1: /u03/oradata/qfundeca/redo02b.log
    Thread 1 advanced to log sequence 62005
      Current log# 3 seq# 62005 mem# 0: /u02/oradata/qfundeca/redo03a.log
      Current log# 3 seq# 62005 mem# 1: /u03/oradata/qfundeca/redo03b.log
    Fri Dec 31 07:30:01 2010
    Thread 1 cannot allocate new log, sequence 62006
    Private strand flush not complete
      Current log# 3 seq# 62005 mem# 0: /u02/oradata/qfundeca/redo03a.log
      Current log# 3 seq# 62005 mem# 1: /u03/oradata/qfundeca/redo03b.log
    Thread 1 advanced to log sequence 62006
      Current log# 1 seq# 62006 mem# 0: /u02/oradata/qfundeca/redo01a.log
      Current log# 1 seq# 62006 mem# 1: /u03/oradata/qfundeca/redo01b.log
    Fri Dec 31 07:54:42 2010
    System State dumped to trace file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_24979.trc
    Fri Dec 31 07:55:06 2010
    System State dumped to trace file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_25057.trc
    Fri Dec 31 07:55:37 2010
    System State dumped to trace file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_25209.trc
    Fri Dec 31 07:59:33 2010
    System State dumped to trace file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_26237.trc
    Fri Dec 31 08:00:25 2010
    System State dumped to trace file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_26523.trc
    Fri Dec 31 08:00:40 2010
    System State dumped to trace file /home/oracle/oracle/admin/qfundeca/udump/qfundeca_ora_26909.trc
    Fri Dec 31 08:06:34 2010
    Errors in file /home/oracle/oracle/admin/qfundeca/bdump/qfundeca_pmon_9591.trc:
    *ORA-00474: SMON process terminated with error*
    *Fri Dec 31 08:06:34 2010*
    *PMON: terminating instance due to error 474*
    Instance terminated by PMON, pid = 9591
    Fri Dec 31 08:08:31 2010
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =27
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
      processes                = 150
      sessions                 = 250
      __shared_pool_size       = 234881024
      __large_pool_size        = 16777216
      __java_pool_size         = 16777216
      __streams_pool_size      = 33554432
      streams_pool_size        = 33554432
      sga_target               = 1610612736
      control_files            = /u02/oradata/qfundeca/control01.ctl, /u02/oradata/qfundeca/control02.ctl, /u03/oradata/qfundeca/control03.ctl
      db_block_size            = 8192
      __db_cache_size          = 1291845632
      compatible               = 10.2.0.3.0
      log_archive_dest_1       = LOCATION=/u02/oradata/arch/qfundeca
      log_archive_format       = %t_%s_%r.dbf
      archive_lag_target       = 2700
      db_file_multiblock_read_count= 16
      db_recovery_file_dest    = /u02/oradata/flash_recovery_area
      db_recovery_file_dest_size= 21474836480
      standby_file_management  = AUTO
      undo_management          = AUTO
      undo_tablespace          = UNDOTBS1
      remote_login_passwordfile= EXCLUSIVE
      db_domain                =
      dispatchers              = (PROTOCOL=TCP) (SERVICE=qfundecaXDB)
      job_queue_processes      = 17
      cursor_sharing           = EXACT
      background_dump_dest     = /home/oracle/oracle/admin/qfundeca/bdump
      user_dump_dest           = /home/oracle/oracle/admin/qfundeca/udump
      core_dump_dest           = /home/oracle/oracle/admin/qfundeca/cdump
      audit_file_dest          = /home/oracle/oracle/admin/qfundeca/adump
      db_name                  = qfundeca
      open_cursors             = 300
      pga_aggregate_target     = 839909376
    PMON started with pid=2, OS id=648
    PSP0 started with pid=3, OS id=650
    MMAN started with pid=4, OS id=652
    DBW0 started with pid=5, OS id=654
    LGWR started with pid=6, OS id=656
    CKPT started with pid=7, OS id=658
    SMON started with pid=8, OS id=660
    RECO started with pid=9, OS id=662
    CJQ0 started with pid=10, OS id=664
    MMON started with pid=11, OS id=666
    Fri Dec 31 08:08:32 2010
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=12, OS id=668
    Fri Dec 31 08:08:32 2010
    starting up 1 shared server(s) ...
    Oracle Data Guard is not available in this edition of Oracle.
    Fri Dec 31 08:08:32 2010
    ALTER DATABASE   MOUNT
    Fri Dec 31 08:08:36 2010
    Setting recovery target incarnation to 1
    Fri Dec 31 08:08:36 2010
    Successful mount of redo thread 1, with mount id 257031232
    Fri Dec 31 08:08:36 2010
    Database mounted in Exclusive Mode
    Completed: ALTER DATABASE   MOUNT
    Fri Dec 31 08:08:36 2010
    ALTER DATABASE OPEN
    Fri Dec 31 08:08:37 2010
    Beginning crash recovery of 1 threads
    Fri Dec 31 08:08:37 2010
    Started redo scan
    Fri Dec 31 08:08:37 2010
    Completed redo scan
    31201 redo blocks read, 754 data blocks need recovery
    Fri Dec 31 08:08:38 2010
    Started redo application at
    Thread 1: logseq 62004, block 66343
    Fri Dec 31 08:08:38 2010
    Recovery of Online Redo Log: Thread 1 Group 2 Seq 62004 Reading mem 0
      Mem# 0: /u02/oradata/qfundeca/redo02a.log
      Mem# 1: /u03/oradata/qfundeca/redo02b.log
    Fri Dec 31 08:08:38 2010
    Recovery of Online Redo Log: Thread 1 Group 3 Seq 62005 Reading mem 0
      Mem# 0: /u02/oradata/qfundeca/redo03a.log
      Mem# 1: /u03/oradata/qfundeca/redo03b.logHow to know what is the reason for this one.On that day the database hangs two time as i mentioned in my previous post.
    Thanks & Regards,
    Poorna Prasad.

    Hi Chinar,
    Here is the contents for the file /home/oracle/oracle/admin/qfundeca/bdump/qfundeca_pmon_9591.trc
    /home/oracle/oracle/admin/qfundeca/bdump/qfundeca_pmon_9591.trc
    Oracle Database 10g Release 10.2.0.3.0 - Production
    ORACLE_HOME = /home/oracle/oracle/oracle/product/10.2.0/db_1
    System name:     Linux
    Node name:     virqts2ora001.localdomain
    Release:     2.6.9-5.ELsmp
    Version:     #1 SMP Wed Jan 5 19:30:39 EST 2005
    Machine:     i686
    Instance name: qfundeca
    Redo thread mounted by this instance: 1
    Oracle process number: 2
    Unix process pid: 9591, image: [email protected] (PMON)
    *** 2010-12-31 08:06:34.723
    *** SERVICE NAME:(SYS$BACKGROUND) 2010-12-31 08:06:34.723
    *** SESSION ID:(250.1) 2010-12-31 08:06:34.723
    Background process SMON found dead
    Oracle pid = 8
    OS pid (from detached process) = 9603
    OS pid (from process state) = 9603
    dtp = 0x2000ed88, proc = 0x7f3528a0
    Dump of memory from 0x2000ED88 to 0x2000EDB4
    2000ED80                   0000007C 7F3528A0          [|....(5.]
    2000ED90 00000000 00000000 4E4F4D53 00000200  [........SMON....]
    2000EDA0 00002583 FFFFFFFF 00000001 5634F590  [.%............4V]
    2000EDB0 00040081                             [....]           
    Dump of memory from 0x7F3528A0 to 0x7F352E58
    7F3528A0 00000102 00000000 00000000 00000000  [................]
    7F3528B0 00000000 7F5FDF30 7F95787C 7F4C61E0  [....0._.|x...aL.]
    7F3528C0 7F956AE4 00000000 7F956B48 7F956B48  [.j......Hk..Hk..]
    7F3528D0 7F957870 00001601 7F4AA968 7F4C61E0  [px......h.J..aL.]
    7F3528E0 00000008 7F4C7430 7F4C75B4 00000000  [....0tL..uL.....]
    7F3528F0 7F5FCE18 7F5FDF48 00000000 00000000  [.._.H._.........]
    7F352900 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    7F352930 00000000 00000000 00000000 000C0000  [................]
    7F352940 00000000 00080000 00000000 00050000  [................]
    7F352950 00000000 00040000 00000000 00040000  [................]
    7F352960 00000000 00080000 00000000 000C0000  [................]
    7F352970 00000000 000C0000 00000000 00050000  [................]
    7F352980 00000000 00110000 00000000 000F0000  [................]
    7F352990 00000000 00000000 00000000 00000000  [................]
            Repeat 3 times
    7F3529D0 00000002 7F3529D4 7F3529D4 00000000  [.....)5..)5.....]
    7F3529E0 00000001 00000000 7F3529E8 7F3529E8  [.........)5..)5.]
    7F3529F0 00000000 00000000 00000000 00000000  [................]
            Repeat 1 times
    7F352A10 00000021 00000018 00000095 00000015  [!...............]
    7F352A20 7F366E98 7F351D30 00010000 00000000  [.n6.0.5.........]
    7F352A30 00000000 00000000 00000000 00000000  [................]
            Repeat 1 times
    7F352A50 00000000 00000000 00002583 00000000  [.........%......]
    7F352A60 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    7F352A90 00000000 00000000 7F3528A0 00000000  [.........(5.....]
    7F352AA0 00000000 00000000 00000000 00000000  [................]
            Repeat 6 times
    7F352B10 00000000 7F352B14 7F352B14 00000000  [.....+5..+5.....]
    7F352B20 00000000 00000001 00000000 00000000  [................]
    7F352B30 3461800E 0000000C 00002583 FFFFFFFF  [..a4.....%......]
    7F352B40 B74796C0 00000000 00000000 00000000  [..G.............]
    7F352B50 00000000 00000000 00000000 00000000  [................]
    7F352B60 00000000 00000000 00000008 FFFFFFFF  [................]
    7F352B70 00000000 00000000 00000000 00000000  [................]
            Repeat 7 times
    7F352BF0 00000000 00000000 6361726F 0000656C  [........oracle..]
    7F352C00 00000000 00000000 00000000 00000000  [................]
    7F352C10 00000000 00000000 00000006 71726976  [............virq]
    7F352C20 6F327374 30306172 6F6C2E31 646C6163  [ts2ora001.locald]
    7F352C30 69616D6F 0000006E 00000000 00000000  [omain...........]
    7F352C40 00000000 00000000 00000000 00000000  [................]
    7F352C50 00000000 00000000 00000000 00000019  [................]
    7F352C60 4E4B4E55 004E574F 00000000 00000000  [UNKNOWN.........]
    7F352C70 00000000 00000000 00000000 00000000  [................]
    7F352C80 00000008 33303639 00000000 00000000  [....9603........]
    7F352C90 00000000 00000000 00000000 00000004  [................]
    7F352CA0 6361726F 7640656C 74717269 726F3273  [oracle@virqts2or]
    7F352CB0 31303061 636F6C2E 6F646C61 6E69616D  [a001.localdomain]
    7F352CC0 4D532820 00294E4F 00000000 00000000  [ (SMON).........]
    7F352CD0 00000027 00000002 00000001 7864736B  ['...........ksdx]
    7F352CE0 6B747366 312B2928 6B2D3C39 63786473  [fstk()+19<-ksdxc]
    7F352CF0 2B292862 31323331 73732D3C 65737570  [b()+1321<-sspuse]
    7F352D00 2B292872 3C323031 3930302D 41374131  [r()+102<-0091A7A]
    7F352D10 00000032 00000000 00000000 00000000  [2...............]
    7F352D20 00000000 00000000 00000000 00000000  [................]
            Repeat 4 times
    7F352D70 00000000 00000000 00000000 00060200  [................]
    7F352D80 0000691D FFFFFFFF 00000000 00000000  [.i..............]
    7F352D90 7F38693C 7F35334C 7F3527DC 00000000  [<i8.L35..'5.....]
    7F352DA0 7F5D3CEC 00000000 00000000 00000000  [.<].............]
    7F352DB0 00000000 00000000 00000000 00000000  [................]
    7F352DC0 7F352DC0 7F352DC0 001B0000 00130000  [.-5..-5.........]
    7F352DD0 00029615 0007CBB1 001E9615 0000EB2C  [............,...]
    7F352DE0 00000000 00005F28 00000000 0002C440  [....(_......@...]
    7F352DF0 00000000 00000814 00000000 00000088  [................]
    7F352E00 00000000 00000814 00000000 00000000  [................]
    7F352E10 00000000 00000000 00000000 00000000  [................]
            Repeat 3 times
    7F352E50 00000000 00000003                    [........]       
    error 474 detected in background process
    ORA-00474: SMON process terminated with error

  • How to install Oracle BPEL Process Manager with the BEA WebLogic

    Hi ,
    I will install Oracle BPEL Process Manager with BEA WebLogic 9.2(MP2). I have download orabpel_10133_WebLogic.zip ,then Modify the following mandatory installation properties in the orabpel_10133_WebLogic\bpelDomain.properties file:
    # BEA_HOME is the path where Weblogic is Installed
    BEA_HOME=/opt/bea
    # JAVA_HOME is the path of jdk folder inside your weblogic
    JAVA_HOME=/opt/bea/jrockit90_150_10
    # DOMAIN_HOME is the path where you wish to create your domain called BPELDomain
    DOMAIN_HOME=/opt/bea/user_projects/domains
    # APPS_HOME is the path where you wish to copy your applications and adapters that are required for oracleBPELServer
    APPS_HOME=/opt/bea/user_projects/apps
    # BEA_HOME is the path where BPEL PM is Installed
    BPEL_HOME=/home/oracle/bpel/product/10.1.3.1/OraBPEL_1/bpel
    # DRIVER_TYPE is the datasource class that installable use to create a datasources for oracleBPELServer
    DRIVER_TYPE=oracle.jdbc.xa.client.OracleXADataSource
    # DB_URL is the url to connect to orabpel schema
    DB_URL=jdbc:oracle:thin:@16.157.134.17:1521:orcl
    # DB_USER is the user Id for orabpel shema in database
    DB_USER=ORABPEL
    #DB_PASSWORD is the password for orabpel schema in database
    DB_PASSWORD=bpel
    #BPEL_SERVER_NAME is the server i.e. to be created under BPELDomain
    BPEL_SERVER_NAME=oracleBPELServer
    #PROXY_HOST is the Host name of the proxy server
    PROXY_HOST=www-proxy.us.oracle.com
    #PROXY_HOST=
    #PROXY_PORT is the Port where the proxy server is running
    PROXY_PORT=80
    #PROXY_PORT=
    #NON_PROXY_HOST is the list of non proxy hosts that are divided by a | symbol
    #NON_PROXY_HOST=*.oracle.com|*.oraclecorp.com|localhost|127.0.0.1|stbbn10|stbbn10.us.oracle.com
    NON_PROXY_HOST=*.oracle.com|*.oraclecorp.com|localhost|127.0.0.1|stbbn10|stbbn10.us.oracle.com|16.157.134.135
    When I run the setup.sh , it will report
    BUILD FAILED
    /opt/software/WL_Installables/build.xml:131: Traceback (innermost last):
    File "./wl_scripts/bpelDomain.py", line 22, in ?
    File "./wl_scripts/createGroupsAndUsers.py", line 4, in ?
    weblogic.management.utils.AlreadyExistsException: [Security:090267]Group BpelGroup
    Actully ,there is no BpelGroup in Weblogic. Does anybody know how to solve it ?

    MAke sure you have not set ANY environment variable related to Oracle / BEA / Java / LD_library path. Use the following script to unset / set the initial settings:
    #!/bin/sh
    unset ORACLE_BASE ORACLE_HOME ORACLE_SID ORACLE_TERM
    unset LD_LIBRARY_PATH LD_LIBRARY_PATH_64
    unset CLASSPATH JAVA_HOME
    export PATH=.:/usr/sbin:/usr/bin:/usr/local/bin:/opt/VRTS/bin
    export BEA_HOME=/appl/oracle/products/9.2/weblogic
    Marc
    http://orasoa.blogspot.com

  • Getting Process instance with key '0' does not exist.

    I am doing a request based provisioning task.When I request for a resource the approvers are able to approve the request but the user is not created.
    I am getting the below exception at the backend:-
    [XELLERATE.APIS],Class/Method: tcProvisioningOperationsBean/getProcessDetailData encounter some problems: Process instance with key '0' does not exist.
    [11/2/12 0:30:13:640 EDT] 00000097 SystemOut O ERROR,02 Nov 2012 00:30:13,639,[XELLERATE.WEBAPP],Class/Method: RequestApprovalDetailAction/requestDetail encounter some problems: Process instance with key '0' does not exist.
    Thor.API.Exceptions.tcAPIException: Process instance with key '0' does not exist.
         at com.thortech.xl.ejb.beansimpl.tcProvisioningOperationsBean.getProcessDetailData(Unknown Source)
         at com.thortech.xl.ejb.beansimpl.tcProvisioningOperationsBean.getProcessDetail(Unknown Source)
         at com.thortech.xl.ejb.beans.tcProvisioningOperationsSession.getProcessDetail(Unknown Source)
         at com.thortech.xl.ejb.interfaces.EJSRemoteStatelesstcProvisioningOperations_6b2f800a.getProcessDetail(Unknown Source)
         at com.thortech.xl.ejb.interfaces._tcProvisioningOperations_Stub.getProcessDetail(Unknown Source)
         at Thor.API.Operations.tcProvisioningOperationsClient.getProcessDetail(Unknown Source)
         at sun.reflect.GeneratedMethodAccessor309.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at Thor.API.Base.SecurityInvocationHandler$1.run(Unknown Source)
         at java.security.AccessController.doPrivileged(Native Method)
         at javax.security.auth.Subject.doAs(Subject.java:337)
         at com.ibm.websphere.security.auth.WSSubject.doAs(WSSubject.java:118)
         at Thor.API.Security.LoginHandler.websphereLoginSession.runAs(Unknown Source)
         at Thor.API.Base.SecurityInvocationHandler.invoke(Unknown Source)
         at $Proxy6.getProcessDetail(Unknown Source)
         at com.thortech.xl.webclient.actions.RequestApprovalDetailAction.setStandardApprovalDetail(Unknown Source)
         at com.thortech.xl.webclient.actions.RequestApprovalDetailAction.requestDetail(Unknown Source)
         at sun.reflect.GeneratedMethodAccessor665.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.apache.struts.actions.DispatchAction.dispatchMethod(DispatchAction.java:280)
         at com.thortech.xl.webclient.actions.tcLookupDispatchAction.execute(Unknown Source)
         at com.thortech.xl.webclient.actions.tcActionBase.execute(Unknown Source)
         at com.thortech.xl.webclient.actions.tcAction.execute(Unknown Source)
         at org.apache.struts.action.RequestProcessor.processActionPerform(RequestProcessor.java:484)
         at org.apache.struts.action.RequestProcessor.process(RequestProcessor.java:274)
         at org.apache.struts.action.ActionServlet.process(ActionServlet.java:1482)
         at org.apache.struts.action.ActionServlet.doPost(ActionServlet.java:525)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:763)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:856)
         at com.ibm.ws.webcontainer.servlet.ServletWrapper.service(ServletWrapper.java:1096)
         at com.ibm.ws.webcontainer.servlet.ServletWrapper.service(ServletWrapper.java:1037)
         at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:145)
         at com.thortech.xl.webclient.security.CSRFFilter.doFilter(Unknown Source)
         at com.ibm.ws.webcontainer.filter.FilterInstanceWrapper.doFilter(FilterInstanceWrapper.java:190)
         at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:130)
         at com.thortech.xl.webclient.security.SecurityFilter.doFilter(Unknown Source)
         at com.ibm.ws.webcontainer.filter.FilterInstanceWrapper.doFilter(FilterInstanceWrapper.java:190)
         at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:130)
         at com.ibm.ws.webcontainer.filter.WebAppFilterChain._doFilter(WebAppFilterChain.java:87)
         at com.ibm.ws.webcontainer.filter.WebAppFilterManager.doFilter(WebAppFilterManager.java:832)
         at com.ibm.ws.webcontainer.filter.WebAppFilterManager.doFilter(WebAppFilterManager.java:679)
         at com.ibm.ws.webcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:566)
         at com.ibm.ws.wswebcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:478)
         at com.ibm.ws.webcontainer.servlet.CacheServletWrapper.handleRequest(CacheServletWrapper.java:90)
         at com.ibm.ws.webcontainer.WebContainer.handleRequest(WebContainer.java:748)
         at com.ibm.ws.wswebcontainer.WebContainer.handleRequest(WebContainer.java:1466)
         at com.ibm.ws.webcontainer.channel.WCChannelLink.ready(WCChannelLink.java:119)
         at com.ibm.ws.http.channel.inbound.impl.HttpInboundLink.handleDiscrimination(HttpInboundLink.java:458)
         at com.ibm.ws.http.channel.inbound.impl.HttpInboundLink.handleNewInformation(HttpInboundLink.java:387)
         at com.ibm.ws.http.channel.inbound.impl.HttpICLReadCallback.complete(HttpICLReadCallback.java:102)
         at com.ibm.ws.ssl.channel.impl.SSLReadServiceContext$SSLReadCompletedCallback.complete(SSLReadServiceContext.java:1818)
         at com.ibm.ws.tcp.channel.impl.AioReadCompletionListener.futureCompleted(AioReadCompletionListener.java:165)
         at com.ibm.io.async.AbstractAsyncFuture.invokeCallback(AbstractAsyncFuture.java:217)
         at com.ibm.io.async.AsyncChannelFuture.fireCompletionActions(AsyncChannelFuture.java:161)
         at com.ibm.io.async.AsyncFuture.completed(AsyncFuture.java:136)
         at com.ibm.io.async.ResultHandler.complete(ResultHandler.java:195)
         at com.ibm.io.async.ResultHandler.runEventProcessingLoop(ResultHandler.java:743)
         at com.ibm.io.async.ResultHandler$2.run(ResultHandler.java:873)

    Process instance with key '0' does not exist
    Such issue typically appears when the Provisioning is initiated without "Auto Save" i.e. before ORC_KEY entry goes in the OIU table...
    Second possibility can be that the mandatory attributes in the process form are not populated..
    Third possibility can be that the Status is Waiting... And Resource is dependent on another Resource...
    Are you able to Direct Provision the account? I mean without raising Request Based provisioning?
    Please give us following details:
    (1) Resource Profile:-status of Auto Save
    (2) Approval Process:- status of Auto Save and Auto Pre-Pop
    (3) Prov Process:-status of Auto Save and Auto Pre-Pop.
    (4) Is the object form properly populated?
    (5) Is the Process form properly populated?
    (6) Any Exception occurred during Approval process?
    (6) In the OIM Database, fire the following query:-
    SELECT OIU.ORC_KEY
    FROM OIU, USR, OBI
    WHERE OIU.USR_KEY=USR.USR_KEY
    AND UPPER(USR.USR_LOGIN)=UPPER('ABCDEFGH_USER')
    AND OBI.OBI_KEY=OIU.OBI_KEY
    AND OBI.OBI_KEY IN (SELECT OBI.OBI_KEY FROM OBI,OBJ WHERE OBI.OBJ_KEY=OBJ.OBJ_KEY AND
    UPPER(OBJ.OBJ_NAME)=UPPER('AD USER'))
    This will show whether any proper ORC_KEY was generated for this resource instance..

  • Essbase Studio - Failed to Establish Connection With SQL Database Server

    Hi all,
    I am new to Hyperion and am having trouble deploying what seems a simple cube in Essbase Studio.
    My environment is Windows 2003, EPM 11.1.1.2, SQL Server 2000.
    I have the following two issues which may be related.
    1. The EPM System Diagnostic tool says that Hyperion Foundation cannot connect to the SQL database.
    Error:
    Failed: Connection to database
    Error: java.net.UnknownHostException: <server name: <server name>
    Recommended Action:
    Every other EPM application is able to connect to the database. I have tried re-configuring Foundation Services and checking the config files and nothing looks wrong. I would appreciate advice on how to fix this.
    2 In Essbase Studio, I was able to connect to the database where the source data is, build the minischema and create dimensions and measures. But when I run the cube deployment wizard I get the error:
    Message: Failed to deploy Essbase cube
    Caused By: Failed to build Essbase cube dimension: (Time)
    Caused By: Cannot incremental build. Essbase Error(1021001): Failed to Establish Connection With SQL Database Server. See log for more information
    …ODBC Layer Error: [08001] ==> [[Microsoft][ODBC SQL Server Driver][TCP/IP Sockets]SQL Server does not exist or access denied.]
    I have checked the DSN as well as the connections set up within Studio, and they are all able to connect to the database. I am using the admin user for the Essbase server and have created a different user for the databases where the data resides and the shared services database. I tried using the same user all the way through and this didn’t help.
    Can you please advise what else I can check or change to resolve this issue.
    Also, can you work in EAS with cubes that are created in Studio? I tried looking at it and I got an error that the rule file couldn’t be accessed because it was created in Studio.
    After the above failures, I tried creating a cube in EIS, with mixed success. I have been able to load the members and data without errors. But when I try to view the data in the cube, I am seeing field names for the measures instead of the data itself.
    ...Definitely more questions than answers at this stage of my learning.
    Regards
    Michelle

    Michelle,
    I don't know if you found an answer to you question after so many months but I was hoping I could be helpful.
    The issue you are experience happens often when the dimension that you are getting an error on, in this case TIME, is built from a snowflake lineage and there is a bad foreign key reference. This dimension is most like high-up in your outline build process for Essbase Studio and this prevents the build from happening usually early on.
    Check your logs also. They are in Hyperion > Products > logs > essbase > essbasestudio.
    You can also make your logs more verbose by setting a configuration variable in the essbase studio server.properties file but that should only be used for debugging as it really saps performance.
    And, yes, you can of course edit an Essbase Studio deployed cube in EAS. However, any changes you make to the cube in EAS are subject to being wiped-out upon the next Essbase Studio deployment of that Applicaion/Database combo.
    If you want to provide more detail, screenshots, etc. I would be glad to help where I can.
    Cheers,
    Christian
    http://www.artofbi.com

  • Failure with cloning database structure and then restore from source db

    Hello, dear experts!
    I am trying to duplicate a database.
    First step was to use the DBCA to create a template. Because there is no downtime possible, I only created a template of the structure without user schema and without objects.
    At another destination I used this template to create a new database "CLON".
    This worked successful. The database is up and running. (But quite empty).
    Then I transferred the latest full backup of my source database to the new database. I have the same structure at the destiantion as at the
    source.
    Now I want to use rman restore to build up the clone with this backup. Therefore I start RMAN and connect to the target without catalog.
    "RMAN> restore database;"
    leads to:
    Starten restore um 27.09.07
    Kanal ORA_DISK_1 wird benutzt
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of restore command at 09/27/2007 16:26:16
    RMAN-06026: some targets not found - aborting restore
    RMAN-06023: no backup or copy of datafile 709 found to restore
    RMAN-06023: no backup or copy of datafile 708 found to restore
    RMAN-06023: no backup or copy of datafile 707 found to restore
    "RMAN> list backup;"
    does not show any results.
    "RMAN> crosscheck backup;"
    Zugewiesener Kanal: ORA_DISK_1
    Kanal ORA_DISK_1: SID=17 Gerõtetyp=DISK
    "RMAN> list backup;"
    does not show any results.
    To me this looks as if RMAN does not identify the/any backup.
    What has to be done to solve the problem?
    Thank you in advance,
    Hergen.

    Hi, I assume that by clone you mean a duplicate of the DB including ALL tablespaces and not just a few right? YOu want to also rename the DB according to what I read.
    As a note, your new database does not see the backups because you do not connect to the catalog, and you have no backup info as a result.
    Correct me if I am wrong if I misundestand your situation:
    First if you want to duplicate, you do not need to create an empty DB. Just create an new instance without any controlfiles or datafiles. Then you restore your rman backup
    You could perhaps adapt this procedure I have used recently to clone databases with RMAN backups: You just need RMAN backups, taken as hot backups is fine. You also need the controlfile autobackup from RMAN. If you do not want to use the catalog, you can use the controlfile as catalog, and register the backup info to the controlfile, then restore.
    Here is the procedure:
    RMAN RESTORE DB FROM COMPLETE LOSS (without use of catalog)
    1-     Install oracle db home and patch.
    Note: As much as possible, patch to the same version of the DB to be restored. If you do not, you should patch to a later release (do not do this if at all possible, as problems may occur). After opening with reset logs, you will have to startup with the upgrade option and run the cat upgrade script to update the DB to the current patchset.
    2-     Install or configure an instance on the desired node. It is not necessary to create a database.
    a.     Run oradim to create the instance.
    Oradim –new –SID sid –SYSPWD sysdbapass –STARTMODE manual|automatic –PFILE ‘….path to pfile….\initsid.ora’
    NOTE: The pfile is optional, required if you do not use a spfile
    b.     Create and configure a listener, as rman requires oracle net connectivity.
    i.     Use oracle utilities to create listener.
    ii.     Use oracle utilities to configure listener to register the instance just created. (or manually edit the listener.ora file).
    c.     Configure a tnsnames.ora file so that rman can reach the database through the listener.
    d.     Create all directories for datafiles, admin, redologs, archive, flash_recovery_area etc if required.
    3-     Create a pfile or get one (ideally the backups should include it).
    a.     Change paths in the pfile to reflect new file structure. Also change any settings that are relevant.
    Note: if using an spfile, then the rman controlfile autobackup will have it. In that case startup in nomount, start RMAN after setting ORACLE_SID=sid, with ‘RMAN target / NOCATALOG.
    Run this in rman
    RUN {
    SET CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO ‘………………..\CONTROL_%F’;
    RESTORE SPFILE TO PFILE ‘………\initsid.ora’ FROM AUTOBACKUP;
    SHUTDOWN ABORT;
    NOTE:
    -     If you do not have a specific archive file location (archivelog mode only) the flash recovery area size may need to be increased, depending on the number of archive files to recover.
    -     If restoring to a different instance (want to restore an old backup to a different db name, set the db_name in the pfile the same as the db name in the backups. You can rename the db later.
    4-     Get the backups desired to restore the DB, and make them available to the node where the restore is to be done.
    5-     Startup the database in nomount mode. Startup the service is necessary, then startup the database in nomount mode: specify the pfile location if not in a standard location.
    ‘STARTUP NOMOUNT PFILE=’……\initsid.ora’’;
    6-     Get DB id from controlfile autobackup. Get the most recent controlfile autobackups from the backup files, and find the database id. Below is an example
    # CONTROL_c-XXXXXXXXXX-YYYYMMDD-ZZ
    # XXXXXXXXXX IS THE DBID
    7-     START RMAN after setting ORACLE_SID=sid, with ‘RMAN target / NOCATALOG’.
    8-     SET DBID. Get the dbid from above. Run ‘SET DBID XXXXXXXXXXXX;’
    9-     Restore controlfile
    RUN {
    SET CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '… \CONTROL_%F';
    RESTORE CONTROLFILE FROM AUTOBACKUP;
    ALTER DATABASE MOUNT;
    NOTE: If restoring old files, then you must add MAXDAYS=xxx to the restore controlfile command, so that it will find the desired file. Default is 7 days. (explaining why it can’t find older then 7 days).
    10-     Catalog backup files. It is probable that the backup files restored will be in a different location once restored (applicable to disk backups). You need to catalog them in RMAN so that RMAN knows about them. List all backup files and run these commands for each backup and controlfile autobackup:
    ‘CATALOG BACKUPPIECE ….\CONTROL_C-310452700-20070918-01';’
    ‘CATALOG BACKUPPIECE …\BACKUP_HEIS8FIK_1_1';
    NOTES: If the backups are in the same location as they have always been, then a recatalog is not necessary, as the control file has the right backup files information.
    If the backups were restored to a different location, then a recatalog must be done so that RMAN can find them. If the regular location of backups is not available, then rman needs to be updated to reflect that. In that case run these commands:
         CROSSCHECK BACKUPS;
         DELETE EXPIRED BACKUP;
         CROSSCHECK ARCHIVELOG ALL;
         DELETE EXPIRED ARCHIVELOG ALL;
    If this is test restore (to test disaster recovery for example), likely all the backups are still there, but it is desired to use the restored backup files only. RMAN needs to be updated to reflect that the regular backups do not exist. Make the regular backup files unavailable to RMAN (such are renaming a directory, or preventing access through security), and then run the above commands.
    11-     Find the datafile names and numbers. This is only required if datafiles are in a different location on the database to be restored.
    Using SQLPlus, set the oracle_sid, then run sqlplus / ;
    Run the following command
         Select file# as “file/grp#’, name from v$datafile; (for datafiles).
         Select group#, member from v$logfile. (for redologs)
    12-     Restore the database
    (use only if all datafile locations are the same!)
    RUN
    # allocate a channel to disk
    ALLOCATE CHANNEL c1 DEVICE TYPE DISK MAXPIECESIZE 2 G FORMAT '…path…\BACKUP_%U';
    RESTORE DATABASE;
    RECOVER DATABASE;
    (use if datafiles are not in the same location)
    RUN
    ALLOCATE CHANNEL C1 DEVICE TYPE DISK FORMAT ‘…\BACKUP_%U';
    SET NEWNAME FOR DATAFILE 1 TO ‘….\SYSTEM01.DBF';
    SET NEWNAME FOR DATAFILE 2 TO '…\UNDOTBS01.DBF';
    SET NEWNAME FOR DATAFILE 3 TO '…\SYSAUX01.DBF';
    SET NEWNAME FOR DATAFILE 4 TO '…\USERS01.DBF';
    SET NEWNAME FOR DATAFILE 5 TO '…\TOOLS_01.DBF ';
    # add other tablespace locations here.
    # We need to tell the controlfile the new location for the redologs
    SQL "ALTER DATABASE RENAME FILE ''…\REDO01.LOG'' TO      ''…\REDO01.LOG''";
    SQL "ALTER DATABASE RENAME FILE ''…\REDO02.LOG'' TO      ''…\REDO02.LOG''";
    SQL "ALTER DATABASE RENAME FILE ''…\REDO03.LOG'' TO      ''…\REDO03.LOG''";
    RESTORE DATABASE;
    # we need to tell the controlfile the location of the new datafiles.
    SWITCH DATAFILE ALL;
    RECOVER DATABASE;
    13-     Reset the logs. ALTER DATABASE OPEN RESETLOGS;
    NOTE: If the oracle home is of a newer revision then the database, the open resetlogs will generate an error. In that case open the database with the upgrade option and then run the catalog upgrade as per patch instructions.
    14-     Create a temporary tablespace. The system thinks it has a temp tablespace, but it needs to be recreated
    a.     Create another temp tablespace (modify as needed).
         CREATE TEMPORARY TABLESPACE temp2
              TEMPFILE '......\temp02.dbf' SIZE 128M REUSE
              AUTOEXTEND ON NEXT 16M MAXSIZE unlimited
              EXTENT MANAGEMENT LOCAL UNIFORM SIZE 1M;
    b.     Point the server to the new file
    ALTER DATABASE DEFAULT TEMPORARY TABLESPACE temp2;
    c.     Delete the missing one:
              DROP TABLESPACE temp INCLUDING CONTENTS AND DATAFILES;
    d.     RECREATE THE temp tablespace;
              CREATE TEMPORARY TABLESPACE temp
                   TEMPFILE 'd:\......\temp_01.dbf' SIZE 128M REUSE
              AUTOEXTEND ON NEXT 16M MAXSIZE unlimited
              EXTENT MANAGEMENT LOCAL UNIFORM SIZE 1M;
    e.     Change system default tablespace again to temp:
              ALTER DATABASE DEFAULT TEMPORARY TABLESPACE temp;
    f.     Delete the transition temp tablespace temp2:
    DROP TABLESPACE temp2 INCLUDING CONTENTS AND DATAFILES;
    15-     Rename the database if the restored database name needs to change.
    Use the nid utility to rename the db. IMPORTANT: the temporary tablespace needs to exist for this to work, or the db rename will fail and a restore will have to be done.
    Use :
    -     Shutdown database in immediate mode.
    -     Startup in nomount.
    -     Open command prompt.
    -     set oracle_sid = sid
    -     nid sys/pass dbname=newname logfile=….\logfile.txt
    -     shutdown the db (if required)
    -     Change the dbname in the pfile or spfile to the new dbname
    -     Create a new password file:
    o     Orapwd file=….. password=…
    IMPORTANT NOTE: if you rename a DB, your existing backups will not be valid, do backups of the newly renamed db!!!
    16-     Open the database and test.
    NOTE: Whenever you open with resetlogs, YOU MUST BACKUP YOUR DATABASE. Your old backups can only be used to restore the old DB.
    I hope this helps.

  • Delivery date from PO not updated with inbound delivery (EDI) date

    Dear all,
    The delivery date in the purchase order (as per my understanding) is the purchase order creation date + the planned delivery time.
    We are using EDI and an inbound delivery gets created when the shipping notification is received from the vendor.
    The real inbound delivery (EDI) date is usually different from the planned delivery date of the purchase order. But the problem is that the delivery date from the PO is not getting updated with the EDI delivery date.
    Is it possible to have the delivery date from the PO updated with the real EDI delivery date?
    I need this because I want to calculate planned delivery costs based on the EDI delivery date, and so far when I post the goods receipt (Price Date Category = "Delivery date" in the PO), the system is taking the price valid at the delivery date from the PO instead of taking the price valid at the delivery date of the EDI message.
    Thanks a lot,
    Annabelle R.

    We have applied the note 793478 "Change of the pricing date in EXIT SAPLEINR 003".
    In our process flow, an inbound delivery is created with reference to the PO, then the goods receipt is done with reference to this inbound delivery.
    When the option 5 "Goods receipt date" is selected in the Conditions Control tab of the purchase order, now when the goods receipt is posted, the price is calculated based on the inbound delivery date, which is what we wanted.

  • ORA-00480: LCK* process terminated with error

    RA-00480: LCK* process terminated with error very frequently getting this said error in our oracle 10g RAC database.
    We have Node1 / Node2,
    Node1 : we have scheduled jobs running & in addition to that end user access the database.
    Node2: 24*7 application process the jobs runnning.
    your suggestion / comments will be highly appreciated.
    Regards,
    Mansoor Ali

    System Information:
    CPU info:
    4 Intel(R) Itanium 2 9000 series processors (1.59 GHz, 24 MB)
    532 MT/s bus, CPU version C2
    8 logical processors (2 per socket)
    Memory: 32736 MB (31.97 GB)
    Firmware info:
    Firmware revision: 02.03
    FP SWA driver revision: 1.18
    IPMI is supported on this system.
    BMC firmware revision: 5.14
    Platform info:
    Model: "ia64 hp server rx6600"
    Machine ID number: 6737643a-4da1-11dc-ab2c-d7eddf21d43f
    Machine serial number: DEH472820A
    OS info:
    Nodename: Node2
    Release: HP-UX B.11.31
    Version: U (unlimited-user license)
    Machine: ia64
    ID Number: 1731683386
    vmunix releaseversion:
    @(#) $Revision: vmunix: B.11.31_LR FLAVOR=perf
    Alert log details
    System State dumped to trace file /oracle/product/admin/tap3plus/bdump/tap3plus2_diag_27593.trc
    Mon Jul 5 20:47:47 2010
    Errors in file /oracle/product/admin/tap3plus/bdump/tap3plus2_lmd0_27603.trc:
    ORA-00480: LCK* process terminated with error
    Mon Jul 5 20:47:47 2010
    Errors in file /oracle/product/admin/tap3plus/bdump/tap3plus2_lms0_27616.trc:
    ORA-00480: LCK* process terminated with error
    Mon Jul 5 20:47:47 2010
    Errors in file /oracle/product/admin/tap3plus/bdump/tap3plus2_lmon_27601.trc:
    ORA-00480: LCK* process terminated with error
    Mon Jul 5 20:47:47 2010
    Errors in file /oracle/product/admin/tap3plus/bdump/tap3plus2_lms1_27634.trc:
    ORA-00480: LCK* process terminated with error
    Mon Jul 5 20:47:49 2010

  • Problems polling two mailboxes using JCA 1.5 - inbound resource adapters.

    Hello,
    I have a requirement to poll two or more mailboxes from my application for unread messages using inbound resource adapters.
    Whilst I'm successful in implementing the above for a single mailbox, I 'm facing problems when tried for 2.
    Here is a snippet of code that worked for single mailbox.
    ejb-jar.xml :
    <ejb-jar version="2.1" xmlns="http://java.sun.com/xml/ns/j2ee" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://java.sun.com/xml/ns/j2ee http://java.sun.com/xml/ns/j2ee/ejb-jar_2_1.xsd">
    <display-name>JavaMailMDB</display-name>
    <enterprise-beans>
    <message-driven>
    <display-name>JavaMailMDB</display-name>
    <ejb-name>JavaMailMDB</ejb-name>
    <ejb-class>samples.connectors.mailconnector.ejb.mdb.JavaMailMessageBean</ejb-class>
    <messaging-type>samples.connectors.mailconnector.api.JavaMailMessageListener</messaging-type>
    <transaction-type>Container</transaction-type>
    <activation-config>
    <activation-config-property>
    <activation-config-property-name>serverName</activation-config-property-name>
    <activation-config-property-value>xxx</activation-config-property-value>
    </activation-config-property>
    <activation-config-property>
    <activation-config-property-name>userName</activation-config-property-name>
    <activation-config-property-value>xxx</activation-config-property-value>
    </activation-config-property>
    <activation-config-property>
    <activation-config-property-name>password</activation-config-property-name>
    <activation-config-property-value>xxx</activation-config-property-value>
    </activation-config-property>
    <activation-config-property>
    <activation-config-property-name>protocol</activation-config-property-name>
    <activation-config-property-value>imap</activation-config-property-value>
    </activation-config-property>
    <activation-config-property>
    <activation-config-property-name>folderName</activation-config-property-name>
    <activation-config-property-value>Inbox</activation-config-property-value>
    </activation-config-property>
    <activation-config-property>
    <activation-config-property-name>interval</activation-config-property-name>
    <activation-config-property-value>1</activation-config-property-value>
    </activation-config-property>
    </activation-config>
    <resource-ref>
    <res-ref-name>TheMailSession</res-ref-name>
    <res-type>javax.mail.Session</res-type>
    <res-auth>Container</res-auth>
    <res-sharing-scope>Shareable</res-sharing-scope>
    </resource-ref>
    </message-driven>
    </enterprise-beans>
    <assembly-descriptor>
    <container-transaction>
    <method>
    <ejb-name>JavaMailMDB</ejb-name>
    <method-name>onMessage</method-name>
    <method-params>
    <method-param>javax.mail.Message</method-param>
    </method-params>
    </method>
    <trans-attribute>Required</trans-attribute>
    </container-transaction>
    </assembly-descriptor>
    </ejb-jar>
    I 'm using four Java classes - ActivationSpecImpl.java,DeliveryThread.java,EndpointConsumer.java,MailServerFolder.java,PollingThread.java.
    I'm newbie to resouce adapters & JCA. Can someone help me how to replicate the functionality for two mailboxes or more.
    Many thanks

    Hema wrote:
    How does it decide which files to pickup ? It will pick up files with modification dates more recent than the last processed file.
    sometimes it picks up all old files from FTP directoryIf it is picking up the same files, probably it is not being able to calculate the modification dates.
    That can be a configuration problem, take a look at this:
    http://docs.oracle.com/cd/E17904_01/integration.1111/e10231/adptr_file.htm#BABECEGD
    Note: We are not deleting already read files from the folderProbably, not a good thing to do. May you can archive them instead.
    Cheers,
    Vlad
    Give points - it is good etiquette to reward an answerer points (5 - helpful; 10 - correct) for their post if they answer your question. If you think this is helpful, please consider giving points

  • I am in the process of expanding a database of chemistry journal articles.  These materials are ideally acquired in two formats when both are available-- PDF and HTML.  To oversimplify, PDFs are for the user to read, and derivatives of the HTML versions a

    I am in the process of expanding a database of chemistry journal articles.  These materials are ideally acquired in two formats when both are available-- PDF and HTML.  To oversimplify, PDFs are for the user to read, and derivatives of the HTML versions are for the computer to read.  Both formats are, of course, readily recognized and indexed by Spotlight.  Journal articles have two essential components with regards to a database:  the topical content of the article itself, and the cited references to other scientific literature.  While a PDF merely lists these references, the HTML version has, in addition, links to the cited items.  Each link URL contains the digital object identifier (doi) for the item it points to. A doi is a unique string that points to one and only one object, and can be quite useful if rendered in a manner that enables indexing by Spotlight.  Embedded URL's are, of course, ignored by Spotlight.  As a result, HTML-formatted articles must be processed so that URL's are openly displayed as readable text before Spotlight will recognize them.  Conversion to DOC format using MS Word, followed by conversion to RTF using Text Edit accomplishes this, but is quite labor intensive.
      In the last few months, I have added about 3,500 articles to this collection, which means that any procedure for rendering URL's must be automated and able to process large batches of documents with minimal user oversight.  This procedure needs to generate a separate file for each HTML document processed. Trials using Automator's "Get Specified Finder Items" and "Get Selected Finder Items", as well as "Ask For Finder Items"  (along with "Get URLs From Web Pages") give unsatisfactory results.  When provided with multiple input documents, these three commands generate output in which the URLs from multiple input items are merged into a single block, which yields a single file using "Create New Word Document" as the subsequent step.  A one-to-one, input file to output file result can be obtained by processing one file at a time, but this requires manual selection of each item and one-at-a-time processing. What I need is a command that accepts multiple input documents, but processes them one at a time, generating a separate output for each file processed.  Is there a way for Automator to do this?

    Hi,
    With the project all done, i'm preparing for the presentation. Managed to get my hands on a HD beamer for the night (Epason TW2000) and planning to do the presentation in HD.
    That of course managed to bring up some problems. I posted a thread which i'll repost here . Sorry for the repost, i normally do not intend to do this, but since this thread is actually about the same thing, i'd like to ask the same question to you. The end version is in AfterEffects, but that actually doesn't alter the question. It's about export:
    "I want to export my AE project of approx 30 min containing several HD files to a Blu Ray disc. The end goal is to project the video in HD quality using the Epson  EMP-TW2000 projector. This projector is HD compatible.
    To project the video I need to connect the beamer to a computer capable of playing a heavy HD file (1), OR burn the project to a BRD (2) and play it using a BRplayer.
    I prefer option 2, so my question is: which would be the preferred export preset?
    Project specs:
                        - 1920x1080 sq pix  (16:9)
                        - 25 fps
                        - my imported video files (Prem.Pro sequences) are also 25 fps and are Progressive (!)
    To export to a BRD compatible format, do i not encounter a big problem: my projectfiles are 25 fps and progressive, and I believe that the only Bluray preset dispaying 1920x1080 with 25 fps requests an INTERLACED video  (I viewed the presets found on this forum, this thread)... There is also a Progr. format, BUT then you need 30 fps (29,...).
    So, is there one dimension that can be changed without changing the content of the video, and if yes which one (either the interlacing or the fps).
    I'm not very familiar with the whole Blu-ray thing, I hope that someone can help me out."
    Please give it a look.
    Thanks,
    Jef

  • ICR process 03 with GL open items and

    Hi gurus
    do you have any idea about ICR with process 03 with GL open items and GL account reconciliation.
    thanks in advance
    best regards

    Hello Lucio,
    I am afraid you will have to be a little more specific in your question
    It is possible to reconcile GL open items using process 003. In FBIC032 you need to specify data source Documents of Other Process. If you want to only read GL open items for a specific company then you create only 1 entry in FBIC032 for this company and specify data source Documents of Other Process. If you want to read both AR/AP open items and GL open items you have to create 2 entries for this company with different SEQNRs. For one you have to specify data source Documents of Current Process, for the other you have to specify data source Documents of Other Process.
    I hope this helps,
    Ralph

Maybe you are looking for

  • Remote Desktop Client 8.1.3 Display Issue

    We're running Microsoft Remote Desktop 2012R2 services including a Gateway/Connection Broker/Licensing server and  Remote Desktop Session Hosts servers.  Our clients include Windows systems, iPads and iPhones. Since the upgrade to Microsoft Remote De

  • Net.Msmq Binding Not Picking Up Messages in Queue as soon as the application pool gets recycled

    Hi, We are trying to implement a wcf service that exposes a net.msmq endpoint.  We are running windows server 2008 R2/IIS 7.5 running on a DOMAIN mode. 1. Our service is called service.svc and it is deployed under an application called Test in the de

  • White Screen, Grey Folder with Question Mark and no USB

    Hello I have a iMac, that is in a classroom of similar machines, they are running os v 10.4.6 (ish). The mac in question had been reported with a fault so i went along to check it and this was a new fault for me. When the mac boots, it bongs the hdd

  • Mass Orders in SAPAP CRM

    Hi friends, I wanted to create mass orders in SAP CRM.Here R/3 is not Back end in my scenario.Any path or Tcode is available to craete mass orders in SAP CRM Thanks&Regards kishore kumar

  • Domains stats?

    Does anyone know how to gather per-service invocation counts and response times from a domains gateway (e.g. gwtdomain & gwidomain)? We're upgrading from Connect TCP v2.1 to eLink TCP v3.1. Since Connect's TPSR process was a Tuxedo server, usage of t