Using Asynch comminication
It is a good thing. But I cannot it use for receiver in real projects. I am always need to make sure are data stored correctly or not.
Asynch and Transport acknowledgement doesn't help since acknowledgements are sent only in case of success.
What do you think about this issue? is there another way to design with asych interfaces
Hi Sergey,
everything is described there. You must only read it.
Send an asynch message to BPM. BPM tries to get the receiving system and sends in case of success a success acknowledgement back, in case of error an error acknowledgement.
Regards,
Udo
Similar Messages
-
Asynch Proxy - XI - Synch Http - XI - Asynch File
Hi,
Currently I have Scenario:
Asynch Proxy (SAP Delivered BADI) ->XI ->Asynch HTTP
Now the Requirement is that from HTTP I need to get the response back So I need:
Asynch Proxy ->XI ->Synch Http ->XI ->Asynch File Scenario.
The catch is the proxy is SAP delivered BADI and makes Asynch call. Also all the Design time objects in XI on Sender side (Proxy) are predelivered by SAP and so Interfaces are Asynchronous.
Is there any way I can achieve the required Asynch Proxy ->XI ->Synch Http ->XI ->Asynch File scenario? If yes How?
I have gone through weblog /people/arpit.seth/blog/2005/06/27/rfc-scenario-using-bpm--starter-kit
but it is not working for me as the proxy I am using is SAP Delivered Asynch and also the Business content in XI is for Asynch comminication.
Thank you in advance.
-Adrean.
Message was edited by:
Adrean HartHi Adrean,
<i>In the First Recieve step, Do I still have the Original Message that I got from the Source SAP or it is after the Mapping?</i>
>>>In the first Receiver Step you have the original Message you got from SAP e.g ReqMessageProxy.
In the BPM you have a sync send Say SYNCMessage.(Sync, Abstract Interface) To create this you need to have following 2 Interface
1) Req 2) Response.
i.e your Req = ReqMessageProxy - Abstract Async
Resp = RespMessageHttp (this response is used to send to a file) (Abstract Async)
Now for the Receiver Http Side, you have a sync Interface. e.g ReqHttp, RespHttp
Mapping Outside the bpm: In the blog Diagram, 11, 12,13,14
ReqMessageProxy(Abstract) -->ReqHttp
RespHttp-->RespMesageHttp.
Integration Directory Config:
1) Receiver Determination: from SAP to BPM (No mapping, no receiver agreement)
2) Receiver Determination: from BPM to Http (Use the above mapping and Reciever agreement)
3) Receiver Determination: from BPM to File ( No mapping , Receiver agreement)
Hope this helps,
Rgdss,
Moorthy -
Copy Attachments of sharepoint list item to a document library using Javascript on click of a button
The requirement is as follows:
1. We have a SP list. All items in the list have mulitple attachments.
2. On click of a button, All the attachments of all items should be copied to a document library
This is to be done from client side only. No server code is acceptable. We have tried following things:
1. USing Asynch call
context = new SP.ClientContext("my site name");
this.oWebsite = context.get_web();
var lists = oWebsite.get_lists();
var list = lists.getByTitle('my list name');
context.load(oWebsite);
var folderPath = 'Lists/<my list name>/Attachments/' + folderId;
var Folder = oWebsite.getFolderByServerRelativeUrl(folderPath);
context.load(Folder);
Files = Folder.get_files();
context.load(Files);
context.executeQueryAsync(Function.createDelegate(this, this.ExecuteCopyOnSuccess), Function.createDelegate(this, this.ExecuteCopyOnFailure));
function ExecuteCopyOnSuccess(sender, args) {
for(var p=0;p<this.Files.get_count();p++)
var file = Files.itemAt(p);
var filename = file.get_name();
if (filename != null) {
var newUrl = 'document library url';
file.copyTo(newUrl, true);
context.executeQueryAsync(null,null);
In this case, Files.get_count() throws error - The collection has not been initialized. It has not been requested or the request has not been executed. It may need to be explicitly requested.
2. Using copy.asmx service
It copies files to document library but some files are blank and are having size 0.
Thanks!Hi,
I think you can use Document Content types or document sets -
Here are links that you can refer to
http://office.microsoft.com/en-us/windows-sharepoint-services-help/introduction-to-content-types-HA010121570.aspx
http://msdn.microsoft.com/en-us/library/office/ms472236(v=office.14).aspx
http://office.microsoft.com/en-us/sharepoint-server-help/create-and-configure-a-new-document-set-content-type-HA101782461.aspx
http://office.microsoft.com/en-us/sharepoint-server-help/introduction-to-document-sets-HA101782466.aspx?CTT=5&origin=HA101782461
Hope this helps!
MCITP: SharePoint 2010 Administrator
MCTS - MOSS 2007 Configuring, .NET 2.0
| SharePoint Architect | Evangelist |
http://www.sharepointdeveloper.in/
http://ramakrishnaraja.blogspot.com/ -
Asynch web services having serialization problem wint 7.0SP1
Hi,
I've tried creating a simple async web service that sends a Java bean to a web
service that's implemented
as an MDB. i.e. The java bean is of type malcolm.AimosMatchedRequest and is therefore
delivered as
an ObjectMessage.
I keep getting a deserialising error listed below in the Weblogic log.
Now if I do a serialver on my Javabean class I get the value 6661344068394585445
in agreement with the log for the stream classdesc. I therefore presume that
the
local class, is the server view which I guess is the SOAP unmarshaled/generated
class that
WebLogic is expecting???
How do I deal with this problem? Note that explitiely forcing the serialuid in
my Javabean class
does not help, which to me confirms that WebLogic is geenrating some SOAP equiv
class.
regards
--malcolm
weblogic.jms.common.JMSException: Error deserializing object
at weblogic.jms.common.ObjectMessageImpl.getObject(ObjectMessageImpl.jav
a:140)
at malcolm.AimosMatchedBean.onMessage(AimosMatchedBean.java:68)
at weblogic.ejb20.internal.MDListener.execute(MDListener.java:356)
at weblogic.ejb20.internal.MDListener.transactionalOnMessage(MDListener.
java:290)
at weblogic.ejb20.internal.MDListener.onMessage(MDListener.java:271)
at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2303)
at weblogic.jms.client.JMSSession.execute(JMSSession.java:2226)
at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
----------- Linked Exception -----------
java.io.InvalidClassException: malcolm.AimosMatchedRequest; Local class not comp
atible: stream classdesc serialVersionUID=6661344068394585445 local class serial
VersionUID=7727103220180333880
at java.io.ObjectStreamClass.validateLocalClass(ObjectStreamClass.java:5
18)
at java.io.ObjectStreamClass.setClass(ObjectStreamClass.java:562)
at java.io.ObjectInputStream.inputClassDescriptor(ObjectInputStream.java
:931)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:361)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:231)
at java.io.ObjectInputStream.inputObject(ObjectInputStream.java:1181)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:381)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:231)
at weblogic.jms.common.ObjectMessageImpl.getObject(ObjectMessageImpl.jav
a:128)
at malcolm.AimosMatchedBean.onMessage(AimosMatchedBean.java:68)
at weblogic.ejb20.internal.MDListener.execute(MDListener.java:356)
at weblogic.ejb20.internal.MDListener.transactionalOnMessage(MDListener.
java:290)
at weblogic.ejb20.internal.MDListener.onMessage(MDListener.java:271)
at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2303)
at weblogic.jms.client.JMSSession.execute(JMSSession.java:2226)
at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
weblogic.jms.common.JMSException: Error deserializing object
at weblogic.jms.common.ObjectMessageImpl.getObject(ObjectMessageImpl.jav
a:140)
at malcolm.AimosMatchedBean.onMessage(AimosMatchedBean.java:68)
at weblogic.ejb20.internal.MDListener.execute(MDListener.java:356)
at weblogic.ejb20.internal.MDListener.transactionalOnMessage(MDListener.
java:290)
at weblogic.ejb20.internal.MDListener.onMessage(MDListener.java:271)
at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2303)
at weblogic.jms.client.JMSSession.execute(JMSSession.java:2226)
at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
----------- Linked Exception -----------
java.io.InvalidClassException: malcolm.AimosMatchedRequest; Local class not comp
atible: stream classdesc serialVersionUID=6661344068394585445 local class serial
VersionUID=7727103220180333880
at java.io.ObjectStreamClass.validateLocalClass(ObjectStreamClass.java:5
18)
at java.io.ObjectStreamClass.setClass(ObjectStreamClass.java:562)
at java.io.ObjectInputStream.inputClassDescriptor(ObjectInputStream.java
:931)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:361)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:231)
at java.io.ObjectInputStream.inputObject(ObjectInputStream.java:1181)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:381)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:231)
at weblogic.jms.common.ObjectMessageImpl.getObject(ObjectMessageImpl.jav
a:128)
at malcolm.AimosMatchedBean.onMessage(AimosMatchedBean.java:68)
at weblogic.ejb20.internal.MDListener.execute(MDListener.java:356)
at weblogic.ejb20.internal.MDListener.transactionalOnMessage(MDListener.
java:290)
at weblogic.ejb20.internal.MDListener.onMessage(MDListener.java:271)
at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2303)
at weblogic.jms.client.JMSSession.execute(JMSSession.java:2226)
at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)Bruce Stephens <[email protected]> wrote:
<!doctype html public "-//w3c//dtd html 4.0 transitional//en">
<html>
Hi Malcolm,
<p>There's been a significant amount of work done on async services in
conjunction with reliable ws messaging that will be available in the
next
release.
<p>Can you/your customer be part of the upcoming beta?Certainly keen to participate. I presume from what you've said that SP1 is still
problematic for async. I've found that SP1 seems pretty good for synch web services
(addressed issues I had with 7.0).
regards
--malcolm
<p>Thanks,
<br>Bruce
<p>Malcolm Robbins wrote:
<blockquote TYPE=CITE>Wow,
<br> based on no replies it appears that no one is actually using
asynch web services.
<p>It looks like I better recommend to my (large bea customer) that WebLogic
async
<br>web services are not ready for prime time...
<p>regards
<p>--malcolm</blockquote>
</html> -
Hi all,
I am building an interface from SAP to a third party system which understands SOAP. So I will be using XI SOAP adapter. I am planning to use asynch mode. In Asynch mode if the third party system is down, the interface should retry after some time.
How can I achieve this retry?
How can it retry at some interval?
How can I define after how many retries the interface should stop sending data?
thanks in advance,
LSPHi,
You can make use of BPM to achieve this.
Use Send Step with Acknowledgement(System Acknowledgement) , and after that you can check for this using Switch and make call again.
<i>How can it retry at some interval?</i>
>>>You can set time by using wait step
<i>How can I define after how many retries the interface should stop sending data?</i>
>> You can put a loop for a specific amount of time. And then you can throw exception once the time out is over.
Hope it gives some idea.
Thanks,
Moorthy -
Hi,
I have a scenario where i pick up a request from JMS queue and send to WEBService. The response i get from WebService is put into another queue.
I understand that we have to use Asynch Synch bridge for this.
But can anyone specify what is the actual need for the Asynch Synch Brdige.
Thanks
VargheseHey
async-sycn bridge is used in BPM when the sender system can only send/receive async calls but the receiver system can send/receive synchronous calls.
there is a blog on SDN,File to RFC with BPM.
in that blog you have a async File system(sender) sending a request to RFC and getting back the response and posting it back in the File system
Thanx
Aamir
Edited by: Aamir Suhail on Feb 19, 2008 6:15 PM -
Database Performace Is Very Poor On IBM AIX Compared To Windows NT
Hi,
Recently we have migrated Our Oracle 10g DataBase from Windows NT to IBM AIX Box. Unfortunately, the Database Performance is gone down when compared to Windows NT environment. Since been a week we are working to pick the problem. We have altered the init.ora parameters to see the database behaviour., But there no Improvement is been observerd.
Below are the Init.Ora Parameters ,
Name Value Description
tracefile_identifier null trace file custom identifier
lock_name_space null lock name space used for generating lock names for standby/clone database
processes 395 user processes
sessions 439 user and system sessions
timed_statistics TRUE maintain internal timing statistics
timed_os_statistics 0 internal os statistic gathering interval in seconds
resource_limit TRUE master switch for resource limit
license_max_sessions 0 maximum number of non-system user sessions allowed
license_sessions_warning 0 warning level for number of non-system user sessions
cpu_count 16 number of CPUs for this instance
instance_groups null list of instance group names
event null debug event control - default null string
sga_max_size 15032385536 max total SGA size
pre_page_sga FALSE pre-page sga for process
shared_memory_address 0 SGA starting address (low order 32-bits on 64-bit platforms)
hi_shared_memory_address 0 SGA starting address (high order 32-bits on 64-bit platforms)
use_indirect_data_buffers FALSE Enable indirect data buffers (very large SGA on 32-bit platforms)
lock_sga TRUE Lock entire SGA in physical memory
shared_pool_size 0 size in bytes of shared pool
large_pool_size 0 size in bytes of large pool
java_pool_size 0 size in bytes of java pool
streams_pool_size 50331648 size in bytes of the streams pool
shared_pool_reserved_size 84724940 size in bytes of reserved area of shared pool
java_soft_sessionspace_limit 0 warning limit on size in bytes of a Java sessionspace
java_max_sessionspace_size 0 max allowed size in bytes of a Java sessionspace
spfile /oracle/app/product/10.2.0.3.0/dbs/spfileCALMDB.ora server parameter file
instance_type RDBMS type of instance to be executed
trace_enabled FALSE enable KST tracing
nls_language AMERICAN NLS language name
nls_territory AMERICA NLS territory name
nls_sort null NLS linguistic definition name
nls_date_language null NLS date language name
nls_date_format null NLS Oracle date format
nls_currency null NLS local currency symbol
nls_numeric_characters null NLS numeric characters
nls_iso_currency null NLS ISO currency territory name
nls_calendar null NLS calendar system name
nls_time_format null time format
nls_timestamp_format null time stamp format
nls_time_tz_format null time with timezone format
nls_timestamp_tz_format null timestampe with timezone format
nls_dual_currency null Dual currency symbol
nls_comp null NLS comparison
nls_length_semantics BYTE create columns using byte or char semantics by default
nls_nchar_conv_excp FALSE NLS raise an exception instead of allowing implicit conversion
fileio_network_adapters null Network Adapters for File I/O
filesystemio_options asynch IO operations on filesystem files
disk_asynch_io FALSE Use asynch I/O for random access devices
tape_asynch_io TRUE Use asynch I/O requests for tape devices
dbwr_io_slaves 0 DBWR I/O slaves
backup_tape_io_slaves FALSE BACKUP Tape I/O slaves
resource_manager_plan null resource mgr top plan
cluster_interconnects null interconnects for RAC use
file_mapping FALSE enable file mapping
gcs_server_processes 0 number of background gcs server processes to start
active_instance_count null number of active instances in the cluster database
sga_target 15032385536 Target size of SGA
control_files /oradata10/oradata/CALMDB/control/CONTROL02.CTL control file names list
db_file_name_convert null datafile name convert patterns and strings for standby/clone db
log_file_name_convert null logfile name convert patterns and strings for standby/clone db
control_file_record_keep_time 0 control file record keep time in days
db_block_buffers 0 Number of database blocks cached in memory
db_block_checksum TRUE store checksum in db blocks and check during reads
db_block_size 8192 Size of database block in bytes
db_cache_size 2147483648 Size of DEFAULT buffer pool for standard block size buffers
db_2k_cache_size 0 Size of cache for 2K buffers
db_4k_cache_size 0 Size of cache for 4K buffers
db_8k_cache_size 0 Size of cache for 8K buffers
db_16k_cache_size 0 Size of cache for 16K buffers
db_32k_cache_size 0 Size of cache for 32K buffers
db_keep_cache_size 0 Size of KEEP buffer pool for standard block size buffers
db_recycle_cache_size 0 Size of RECYCLE buffer pool for standard block size buffers
db_writer_processes 6 number of background database writer processes to start
buffer_pool_keep null Number of database blocks/latches in keep buffer pool
buffer_pool_recycle null Number of database blocks/latches in recycle buffer pool
db_cache_advice ON Buffer cache sizing advisory
max_commit_propagation_delay 0 Max age of new snapshot in .01 seconds
compatible 10.2.0.3.0 Database will be completely compatible with this software version
remote_archive_enable TRUE remote archival enable setting
log_archive_config null log archive config parameter
log_archive_start FALSE start archival process on SGA initialization
log_archive_dest null archival destination text string
log_archive_duplex_dest null duplex archival destination text string
log_archive_dest_1 null archival destination #1 text string
log_archive_dest_2 null archival destination #2 text string
log_archive_dest_3 null archival destination #3 text string
log_archive_dest_4 null archival destination #4 text string
log_archive_dest_5 null archival destination #5 text string
log_archive_dest_6 null archival destination #6 text string
log_archive_dest_7 null archival destination #7 text string
log_archive_dest_8 null archival destination #8 text string
log_archive_dest_9 null archival destination #9 text string
log_archive_dest_10 null archival destination #10 text string
log_archive_dest_state_1 enable archival destination #1 state text string
log_archive_dest_state_2 enable archival destination #2 state text string
log_archive_dest_state_3 enable archival destination #3 state text string
log_archive_dest_state_4 enable archival destination #4 state text string
log_archive_dest_state_5 enable archival destination #5 state text string
log_archive_dest_state_6 enable archival destination #6 state text string
log_archive_dest_state_7 enable archival destination #7 state text string
log_archive_dest_state_8 enable archival destination #8 state text string
log_archive_dest_state_9 enable archival destination #9 state text string
log_archive_dest_state_10 enable archival destination #10 state text string
log_archive_max_processes 2 maximum number of active ARCH processes
log_archive_min_succeed_dest 1 minimum number of archive destinations that must succeed
standby_archive_dest ?/dbs/arch standby database archivelog destination text string
log_archive_trace 0 Establish archivelog operation tracing level
log_archive_local_first TRUE Establish EXPEDITE attribute default value
log_archive_format %t_%s_%r.dbf archival destination format
fal_client null FAL client
fal_server null FAL server list
log_buffer 176918528 redo circular buffer size
log_checkpoint_interval 0 # redo blocks checkpoint threshold
log_checkpoint_timeout 0 Maximum time interval between checkpoints in seconds
archive_lag_target 0 Maximum number of seconds of redos the standby could lose
db_files 200 max allowable # db files
db_file_multiblock_read_count 128 db block to be read each IO
read_only_open_delayed FALSE if TRUE delay opening of read only files until first access
cluster_database FALSE if TRUE startup in cluster database mode
parallel_server FALSE if TRUE startup in parallel server mode
parallel_server_instances 1 number of instances to use for sizing OPS SGA structures
cluster_database_instances 1 number of instances to use for sizing cluster db SGA structures
db_create_file_dest null default database location
db_create_online_log_dest_1 null online log/controlfile destination #1
db_create_online_log_dest_2 null online log/controlfile destination #2
db_create_online_log_dest_3 null online log/controlfile destination #3
db_create_online_log_dest_4 null online log/controlfile destination #4
db_create_online_log_dest_5 null online log/controlfile destination #5
db_recovery_file_dest null default database recovery file location
db_recovery_file_dest_size 0 database recovery files size limit
standby_file_management MANUAL if auto then files are created/dropped automatically on standby
gc_files_to_locks null mapping between file numbers and global cache locks
thread 0 Redo thread to mount
fast_start_io_target 0 Upper bound on recovery reads
fast_start_mttr_target 0 MTTR target of forward crash recovery in seconds
log_checkpoints_to_alert FALSE log checkpoint begin/end to alert file
recovery_parallelism 0 number of server processes to use for parallel recovery
logmnr_max_persistent_sessions 1 maximum number of threads to mine
db_flashback_retention_target 1440 Maximum Flashback Database log retention time in minutes.
dml_locks 1000 dml locks - one for each table modified in a transaction
ddl_wait_for_locks FALSE Disable NOWAIT DML lock acquisitions
replication_dependency_tracking TRUE tracking dependency for Replication parallel propagation
instance_number 0 instance number
transactions 482 max. number of concurrent active transactions
transactions_per_rollback_segment 5 number of active transactions per rollback segment
rollback_segments null undo segment list
undo_management AUTO instance runs in SMU mode if TRUE, else in RBU mode
undo_tablespace UNDOTBS1 use/switch undo tablespace
undo_retention 10800 undo retention in seconds
fast_start_parallel_rollback LOW max number of parallel recovery slaves that may be used
resumable_timeout 0 set resumable_timeout
db_block_checking FALSE header checking and data and index block checking
recyclebin off recyclebin processing
create_stored_outlines null create stored outlines for DML statements
serial_reuse disable reuse the frame segments
ldap_directory_access NONE RDBMS's LDAP access option
os_roles FALSE retrieve roles from the operating system
rdbms_server_dn null RDBMS's Distinguished Name
max_enabled_roles 150 max number of roles a user can have enabled
remote_os_authent FALSE allow non-secure remote clients to use auto-logon accounts
remote_os_roles FALSE allow non-secure remote clients to use os roles
O7_DICTIONARY_ACCESSIBILITY FALSE Version 7 Dictionary Accessibility Support
remote_login_passwordfile NONE password file usage parameter
license_max_users 0 maximum number of named users that can be created in the database
audit_sys_operations TRUE enable sys auditing
global_context_pool_size null Global Application Context Pool Size in Bytes
db_domain null directory part of global database name stored with CREATE DATABASE
global_names TRUE enforce that database links have same name as remote database
distributed_lock_timeout 60 number of seconds a distributed transaction waits for a lock
commit_point_strength 1 Bias this node has toward not preparing in a two-phase commit
instance_name CALMDB instance name supported by the instance
service_names CALMDB service names supported by the instance
dispatchers (PROTOCOL=TCP) (SERVICE=CALMDB) specifications of dispatchers
shared_servers 1 number of shared servers to start up
max_shared_servers null max number of shared servers
max_dispatchers null max number of dispatchers
circuits null max number of circuits
shared_server_sessions null max number of shared server sessions
local_listener null local listener
remote_listener null remote listener
cursor_space_for_time FALSE use more memory in order to get faster execution
session_cached_cursors 200 Number of cursors to cache in a session.
remote_dependencies_mode TIMESTAMP remote-procedure-call dependencies mode parameter
utl_file_dir null utl_file accessible directories list
smtp_out_server null utl_smtp server and port configuration parameter
plsql_v2_compatibility FALSE PL/SQL version 2.x compatibility flag
plsql_compiler_flags INTERPRETED, NON_DEBUG PL/SQL compiler flags
plsql_native_library_dir null plsql native library dir
plsql_native_library_subdir_count 0 plsql native library number of subdirectories
plsql_warnings DISABLE:ALL PL/SQL compiler warnings settings
plsql_code_type INTERPRETED PL/SQL code-type
plsql_debug FALSE PL/SQL debug
plsql_optimize_level 2 PL/SQL optimize level
plsql_ccflags null PL/SQL ccflags
job_queue_processes 10 number of job queue slave processes
parallel_min_percent 0 minimum percent of threads required for parallel query
create_bitmap_area_size 8388608 size of create bitmap buffer for bitmap index
bitmap_merge_area_size 1048576 maximum memory allow for BITMAP MERGE
cursor_sharing FORCE cursor sharing mode
parallel_min_servers 10 minimum parallel query servers per instance
parallel_max_servers 320 maximum parallel query servers per instance
parallel_instance_group null instance group to use for all parallel operations
parallel_execution_message_size 4096 message buffer size for parallel execution
hash_area_size 62914560 size of in-memory hash work area
shadow_core_dump partial Core Size for Shadow Processes
background_core_dump partial Core Size for Background Processes
background_dump_dest /oradata28/oradata/CALMDB/bdump Detached process dump directory
user_dump_dest /oradata28/oradata/CALMDB/udump User process dump directory
max_dump_file_size 10M Maximum size (blocks) of dump file
core_dump_dest /oradata28/oradata/CALMDB/cdump Core dump directory
use_sigio TRUE Use SIGIO signal
audit_file_dest /oracle/app/product/10.2.0.3.0/rdbms/audit Directory in which auditing files are to reside
audit_syslog_level null Syslog facility and level
object_cache_optimal_size 102400 optimal size of the user session's object cache in bytes
object_cache_max_size_percent 10 percentage of maximum size over optimal of the user session's object cache
session_max_open_files 20 maximum number of open files allowed per session
open_links 4 max # open links per session
open_links_per_instance 4 max # open links per instance
commit_write null transaction commit log write behaviour
optimizer_features_enable 10.2.0.3 optimizer plan compatibility parameter
fixed_date null fixed SYSDATE value
audit_trail DB enable system auditing
sort_area_size 31457280 size of in-memory sort work area
sort_area_retained_size 3145728 size of in-memory sort work area retained between fetch calls
db_name TESTDB database name specified in CREATE DATABASE
db_unique_name TESTDB Database Unique Name
open_cursors 2000 max # cursors per session
ifile null include file in init.ora
sql_trace FALSE enable SQL trace
os_authent_prefix ops$ prefix for auto-logon accounts
optimizer_mode ALL_ROWS optimizer mode
sql92_security FALSE require select privilege for searched update/delete
blank_trimming FALSE blank trimming semantics parameter
star_transformation_enabled FALSE enable the use of star transformation
parallel_adaptive_multi_user TRUE enable adaptive setting of degree for multiple user streams
parallel_threads_per_cpu 2 number of parallel execution threads per CPU
parallel_automatic_tuning TRUE enable intelligent defaults for parallel execution parameters
optimizer_index_cost_adj 250 optimizer index cost adjustment
optimizer_index_caching 0 optimizer percent index caching
query_rewrite_enabled TRUE allow rewrite of queries using materialized views if enabled
query_rewrite_integrity enforced perform rewrite using materialized views with desired integrity
sql_version NATIVE sql language version parameter for compatibility issues
pga_aggregate_target 3221225472 Target size for the aggregate PGA memory consumed by the instance
workarea_size_policy AUTO policy used to size SQL working areas (MANUAL/AUTO)
optimizer_dynamic_sampling 2 optimizer dynamic sampling
statistics_level TYPICAL statistics level
skip_unusable_indexes TRUE skip unusable indexes if set to TRUE
optimizer_secure_view_merging TRUE optimizer secure view merging and predicate pushdown/movearound
aq_tm_processes 1 number of AQ Time Managers to start
hs_autoregister TRUE enable automatic server DD updates in HS agent self-registration
dg_broker_start FALSE start Data Guard broker framework (DMON process)
drs_start FALSE start DG Broker monitor (DMON process)
dg_broker_config_file1 /oracle/app/product/10.2.0.3.0/dbs/dr1CALMDB.dat data guard broker configuration file #1
dg_broker_config_file2 /oracle/app/product/10.2.0.3.0/dbs/dr2CALMDB.dat data guard broker configuration file #2
olap_page_pool_size 0 size of the olap page pool in bytes
asm_diskstring null disk set locations for discovery
asm_diskgroups null disk groups to mount automatically
asm_power_limit 1 number of processes for disk rebalancing
sqltune_category DEFAULT Category qualifier for applying hintsets pls suggest
Thanks
KrWe have examined the AWR Reports, That shows ,
Snap Id Snap Time Sessions Cursors/Session
Begin Snap: 1074 27-Jul-09 13:00:03 147 16.7
End Snap: 1075 27-Jul-09 14:01:00 150 22.3
Elapsed: 60.96 (mins)
DB Time: 9.63 (mins)
Report Summary
Cache Sizes
Begin End
Buffer Cache: 12,368M 12,368M Std Block Size: 8K
Shared Pool Size: 1,696M 1,696M Log Buffer: 178,172K
Load Profile
Per Second Per Transaction
Redo size: 12,787.87 24,786.41
Logical reads: 7,409.85 14,362.33
Block changes: 61.17 118.57
Physical reads: 0.51 0.98
Physical writes: 4.08 7.90
User calls: 60.11 116.50
Parses: 19.38 37.56
Hard parses: 0.36 0.69
Sorts: 7.87 15.25
Logons: 0.07 0.14
Executes: 50.34 97.57
Transactions: 0.52
% Blocks changed per Read: 0.83 Recursive Call %: 74.53
Rollback per transaction %: 3.29 Rows per Sort: 292.67
Instance Efficiency Percentages (Target 100%)
Buffer Nowait %: 100.00 Redo NoWait %: 100.00
Buffer Hit %: 99.99 In-memory Sort %: 100.00
Library Hit %: 98.40 Soft Parse %: 98.15
Execute to Parse %: 61.51 Latch Hit %: 99.96
Parse CPU to Parse Elapsd %: 24.44 % Non-Parse CPU: 98.99
Shared Pool Statistics
Begin End
Memory Usage %: 72.35 72.86
% SQL with executions>1: 98.69 96.86
% Memory for SQL w/exec>1: 96.72 87.64
Top 5 Timed Events
Event Waits Time(s) Avg Wait(ms) % Total Call Time Wait Class
CPU time 535 92.5
db file parallel write 596 106 177 18.3 System I/O
log file parallel write 3,844 40 10 6.9 System I/O
control file parallel write 1,689 29 17 5.0 System I/O
log file sync 2,357 29 12 5.0 Commit
Time Model Statistics
Total time in database user-calls (DB Time): 578s
Statistics including the word "background" measure background process time, and so do not contribute to the DB time statistic
Ordered by % or DB time desc, Statistic name
Statistic Name Time (s) % of DB Time
sql execute elapsed time 560.61 96.99
DB CPU 534.91 92.55
parse time elapsed 24.16 4.18
hard parse elapsed time 17.90 3.10
PL/SQL execution elapsed time 7.65 1.32
connection management call elapsed time 0.89 0.15
repeated bind elapsed time 0.49 0.08
hard parse (sharing criteria) elapsed time 0.28 0.05
sequence load elapsed time 0.05 0.01
PL/SQL compilation elapsed time 0.03 0.00
failed parse elapsed time 0.02 0.00
hard parse (bind mismatch) elapsed time 0.00 0.00
DB time 577.98
background elapsed time 190.39
background cpu time 15.49
Wait Class
s - second
cs - centisecond - 100th of a second
ms - millisecond - 1000th of a second
us - microsecond - 1000000th of a second
ordered by wait time desc, waits desc
Wait Class Waits %Time -outs Total Wait Time (s) Avg wait (ms) Waits /txn
System I/O 8,117 0.00 175 22 4.30
Commit 2,357 0.00 29 12 1.25
Network 226,127 0.00 7 0 119.83
User I/O 1,004 0.00 4 4 0.53
Application 91 0.00 2 27 0.05
Other 269 0.00 1 4 0.14
Concurrency 32 0.00 0 7 0.02
Configuration 59 0.00 0 3 0.03
Wait Events
s - second
cs - centisecond - 100th of a second
ms - millisecond - 1000th of a second
us - microsecond - 1000000th of a second
ordered by wait time desc, waits desc (idle events last)
Event Waits %Time -outs Total Wait Time (s) Avg wait (ms) Waits /txn
db file parallel write 596 0.00 106 177 0.32
log file parallel write 3,844 0.00 40 10 2.04
control file parallel write 1,689 0.00 29 17 0.90
log file sync 2,357 0.00 29 12 1.25
SQL*Net more data from client 4,197 0.00 7 2 2.22
db file sequential read 689 0.00 4 5 0.37
enq: RO - fast object reuse 32 0.00 2 50 0.02
rdbms ipc reply 32 0.00 1 34 0.02
db file scattered read 289 0.00 1 2 0.15
enq: KO - fast object checkpoint 47 0.00 1 14 0.02
control file sequential read 1,988 0.00 0 0 1.05
SQL*Net message to client 218,154 0.00 0 0 115.61
os thread startup 6 0.00 0 34 0.00
SQL*Net break/reset to client 12 0.00 0 15 0.01
log buffer space 59 0.00 0 3 0.03
latch free 10 0.00 0 8 0.01
SQL*Net more data to client 3,776 0.00 0 0 2.00
latch: shared pool 5 0.00 0 5 0.00
reliable message 79 0.00 0 0 0.04
LGWR wait for redo copy 148 0.00 0 0 0.08
buffer busy waits 19 0.00 0 0 0.01
direct path write temp 24 0.00 0 0 0.01
latch: cache buffers chains 2 0.00 0 0 0.00
direct path write 2 0.00 0 0 0.00
SQL*Net message from client 218,149 0.00 136,803 627 115.61
PX Idle Wait 18,013 100.06 35,184 1953 9.55
virtual circuit status 67,690 0.01 3,825 57 35.87
Streams AQ: qmn slave idle wait 130 0.00 3,563 27404 0.07
Streams AQ: qmn coordinator idle wait 264 50.76 3,563 13494 0.14
class slave wait 3 0.00 0 0 0.00
Back to Wait Events Statistics
Back to Top
Background Wait Events
ordered by wait time desc, waits desc (idle events last)
Event Waits %Time -outs Total Wait Time (s) Avg wait (ms) Waits /txn
db file parallel write 596 0.00 106 177 0.32
log file parallel write 3,843 0.00 40 10 2.04
control file parallel write 1,689 0.00 29 17 0.90
os thread startup 6 0.00 0 34 0.00
log buffer space 59 0.00 0 3 0.03
control file sequential read 474 0.00 0 0 0.25
log file sync 1 0.00 0 11 0.00
events in waitclass Other 148 0.00 0 0 0.08
rdbms ipc message 32,384 54.67 49,367 1524 17.16
pmon timer 1,265 100.00 3,568 2821 0.67
Streams AQ: qmn slave idle wait 130 0.00 3,563 27404 0.07
Streams AQ: qmn coordinator idle wait 264 50.76 3,563 13494 0.14
smon timer 63 11.11 3,493 55447 0.03
SQL ordered by Gets
Resources reported for PL/SQL code includes the resources used by all SQL statements called by the code.
Total Buffer Gets: 27,101,711
Captured SQL account for 81.1% of Total
Buffer Gets Executions Gets per Exec %Total CPU Time (s) Elapsed Time (s) SQL Id SQL Module SQL Text
11,889,257 3 3,963,085.67 43.87 145.36 149.62 8hr7mrcqpvw7n Begin Pkg_Pg_consolidation.Pro...
5,877,417 17,784 330.49 21.69 59.94 62.30 3mw7tf64wzgv4 SELECT TOTALVOL.PERIOD_NUMBER ...
5,877,303 17,784 330.48 21.69 62.01 63.54 g3vhvg8cz6yu3 SELECT TOTALVOL.PERIOD_NUMBER ...
3,423,336 0 12.63 200.67 200.67 6jrnq2ua8cjnq SELECT ROWNUM , first , sec...
2,810,100 2,465 1,140.00 10.37 19.29 19.29 7f4y1a3k1tzjn SELECT /*+CLUSTER(VA_STATIC_CC...
1,529,253 230 6,648.93 5.64 15.92 16.97 6trp3txn7rh1q SELECT /*+ index(va_gap_irlc_P...
1,523,043 230 6,621.93 5.62 16.22 17.18 3fu81ar131nj9 SELECT /*+ index(va_gap_irla_P...
855,620 358 2,390.00 3.16 11.49 13.31 a3g12c11x7yd0 SELECT FX_DATE, FX_RATE, CCY...
689,979 708 974.55 2.55 4.37 4.43 b7znr5szwjrtx SELECT /*+RULE*/ YIELD_CURVE_C...
603,631 2,110 286.08 2.23 11.03 13.40 3c2gyz9fhswxx SELECT ASSET_LIABILITY_GAP, AL...
554,080 5 110,816.00 2.04 2.37 2.44 9w1b11p6baqat SELECT DISTINCT consolidation_...
318,378 624 510.22 1.17 3.20 3.45 1auhbw1rd5yn2 SELECT /*+ index(va_gap_irla_P...
318,378 624 510.22 1.17 3.19 3.42 6gq9rj96p9aq0 SELECT /*+ index(va_gap_irlc_P...
313,923 3 104,641.00 1.16 2.38 2.38 7vsznt4tvh1b5 ...
SQL ordered by Reads
Total Disk Reads: 1,857
Captured SQL account for 2.1% of Total
Physical Reads Executions Reads per Exec %Total CPU Time (s) Elapsed Time (s) SQL Id SQL Module SQL Text
57 36 1.58 3.07 3.55 5.81 c6vdhsbw1t03d BEGIN citidba.proc_analyze_tab...
32 507 0.06 1.72 0.22 0.40 c49tbx3qqrtm4 insert into dependency$(d_obj#...
28 8 3.50 1.51 0.76 3.02 4crh3z5ya2r27 BEGIN PROC_DELETE_PACK_TABLES(...
20 3 6.67 1.08 145.36 149.62 8hr7mrcqpvw7n Begin Pkg_Pg_consolidation.Pro...
10 1 10.00 0.54 6.21 18.11 4m9ts1b1b27sv BEGIN domain.create_tables(:1,...
7 23 0.30 0.38 1.56 2.22 4vw03w673b9k7 BEGIN PROC_CREATE_PACK_TABLES(...
4 4 1.00 0.22 0.29 1.06 1vw6carbvp4z0 BEGIN Proc_ReCreate_Gap_temp_t...
2 182 0.01 0.11 0.06 0.08 2h0gb24h6zpnu insert into access$(d_obj#, or...
2 596 0.00 0.11 0.26 0.29 5fbmafvm27kfm insert into obj$(owner#, name,...
1 1 1.00 0.05 0.01 0.02 7jsrvff8hnqft UPDATE VA_PRR_IRUT_POL_IBCB_R...
SQL ordered by Executions
Total Executions: 184,109
Captured SQL account for 71.6% of Total
Executions Rows Processed Rows per Exec CPU per Exec (s) Elap per Exec (s) SQL Id SQL Module SQL Text
43,255 43,255 1.00 0.00 0.00 4m94ckmu16f9k JDBC Thin Client select count(*) from dual
25,964 24,769 0.95 0.00 0.00 2kxdq3m953pst SELECT SURROGATE_KEY FROM TB_P...
17,784 54,585 3.07 0.00 0.00 3mw7tf64wzgv4 SELECT TOTALVOL.PERIOD_NUMBER ...
17,784 54,585 3.07 0.00 0.00 g3vhvg8cz6yu3 SELECT TOTALVOL.PERIOD_NUMBER ...
2,631 2,631 1.00 0.00 0.00 60uw2vh6q9vn2 insert into col$(obj#, name, i...
2,465 924,375 375.00 0.01 0.01 7f4y1a3k1tzjn SELECT /*+CLUSTER(VA_STATIC_CC...
2,202 36 0.02 0.00 0.00 96g93hntrzjtr select /*+ rule */ bucket_cnt,...
2,110 206,464 97.85 0.01 0.01 3c2gyz9fhswxx SELECT ASSET_LIABILITY_GAP, AL...
2,043 2,043 1.00 0.00 0.00 28dvpph9k610y SELECT COUNT(*) FROM TB_TECH_S...
842 35 0.04 0.00 0.00 04xtrk7uyhknh select obj#, type#, ctime, mti...
SQL ordered by Parse Calls
Total Parse Calls: 70,872
Captured SQL account for 69.7% of Total
Parse Calls Executions % Total Parses SQL Id SQL Module SQL Text
17,784 17,784 25.09 3mw7tf64wzgv4 SELECT TOTALVOL.PERIOD_NUMBER ...
17,784 17,784 25.09 g3vhvg8cz6yu3 SELECT TOTALVOL.PERIOD_NUMBER ...
2,110 2,110 2.98 3c2gyz9fhswxx SELECT ASSET_LIABILITY_GAP, AL...
786 786 1.11 2s6amyv4qz2h2 exp@PSLDB03 (TNS V1-V3) SELECT INIEXT, SEXT, MINEXT,...
596 596 0.84 5fbmafvm27kfm insert into obj$(owner#, name,...
590 590 0.83 2ym6hhaq30r73 select type#, blocks, extents,...
550 550 0.78 7gtztzv329wg0 select c.name, u.name from co...
512 512 0.72 9qgtwh66xg6nz update seg$ set type#=:4, bloc...
480 480 0.68 6x2cz59yrxz3a exp@PSLDB03 (TNS V1-V3) SELECT NAME, OBJID, OWNER, ...
457 457 0.64 bsa0wjtftg3uw select file# from file$ where ...
Instance Activity Stats
Statistic Total per Second per Trans
CPU used by this session 54,051 14.78 28.64
CPU used when call started 53,326 14.58 28.26
CR blocks created 1,114 0.30 0.59
Cached Commit SCN referenced 755,322 206.51 400.28
Commit SCN cached 29 0.01 0.02
DB time 62,190 17.00 32.96
DBWR checkpoint buffers written 3,247 0.89 1.72
DBWR checkpoints 79 0.02 0.04
DBWR object drop buffers written 118 0.03 0.06
DBWR parallel query checkpoint buffers written 0 0.00 0.00
DBWR revisited being-written buffer 0 0.00 0.00
DBWR tablespace checkpoint buffers written 169 0.05 0.09
DBWR thread checkpoint buffers written 3,078 0.84 1.63
DBWR transaction table writes 0 0.00 0.00
DBWR undo block writes 11,245 3.07 5.96
DFO trees parallelized 0 0.00 0.00
DML statements parallelized 0 0.00 0.00
IMU CR rollbacks 29 0.01 0.02
IMU Flushes 982 0.27 0.52
IMU Redo allocation size 1,593,112 435.57 844.26
IMU commits 991 0.27 0.53
IMU contention 3 0.00 0.00
IMU ktichg flush 3 0.00 0.00
IMU pool not allocated 0 0.00 0.00
IMU recursive-transaction flush 1 0.00 0.00
IMU undo allocation size 3,280,968 897.05 1,738.72
IMU- failed to get a private strand 0 0.00 0.00
Misses for writing mapping 0 0.00 0.00
OS Integral shared text size 0 0.00 0.00
OS Integral unshared data size 0 0.00 0.00
OS Involuntary context switches 0 0.00 0.00
OS Maximum resident set size 0 0.00 0.00
OS Page faults 0 0.00 0.00
OS Page reclaims 0 0.00 0.00
OS System time used 0 0.00 0.00
OS User time used 0 0.00 0.00
OS Voluntary context switches 0 0.00 0.00
PX local messages recv'd 0 0.00 0.00
PX local messages sent 0 0.00 0.00
Parallel operations downgraded to serial 0 0.00 0.00
Parallel operations not downgraded 0 0.00 0.00
SMON posted for dropping temp segment 0 0.00 0.00
SMON posted for undo segment shrink 0 0.00 0.00
SQL*Net roundtrips to/from client 266,339 72.82 141.14
active txn count during cleanout 677 0.19 0.36
application wait time 243 0.07 0.13
background checkpoints completed 0 0.00 0.00
background checkpoints started 0 0.00 0.00
background timeouts 17,769 4.86 9.42
branch node splits 0 0.00 0.00
buffer is not pinned count 11,606,002 3,173.19 6,150.50
buffer is pinned count 65,043,685 17,783.53 34,469.36
bytes received via SQL*Net from client 27,009,252 7,384.57 14,313.33
bytes sent via SQL*Net to client ############### 69,310,703.02 134,343,168.92
calls to get snapshot scn: kcmgss 382,084 104.47 202.48
calls to kcmgas 15,558 4.25 8.24
calls to kcmgcs 1,886 0.52 1.00
change write time 488 0.13 0.26
cleanout - number of ktugct calls 628 0.17 0.33
cleanouts and rollbacks - consistent read gets 3 0.00 0.00
cleanouts only - consistent read gets 53 0.01 0.03
cluster key scan block gets 77,478 21.18 41.06
cluster key scans 41,479 11.34 21.98
commit batch/immediate performed 550 0.15 0.29
commit batch/immediate requested 550 0.15 0.29
commit cleanout failures: block lost 0 0.00 0.00
commit cleanout failures: buffer being written 0 0.00 0.00
commit cleanout failures: callback failure 29 0.01 0.02
commit cleanout failures: cannot pin 0 0.00 0.00
commit cleanouts 19,562 5.35 10.37
commit cleanouts successfully completed 19,533 5.34 10.35
commit immediate performed 550 0.15 0.29
commit immediate requested 550 0.15 0.29
commit txn count during cleanout 396 0.11 0.21
concurrency wait time 23 0.01 0.01
consistent changes 1,803 0.49 0.96
consistent gets 26,887,134 7,351.18 14,248.61
consistent gets - examination 1,524,222 416.74 807.75
consistent gets direct 0 0.00 0.00
consistent gets from cache 26,887,134 7,351.18 14,248.61
cursor authentications 773 0.21 0.41
data blocks consistent reads - undo records applied 1,682 0.46 0.89
db block changes 223,743 61.17 118.57
db block gets 214,573 58.67 113.71
db block gets direct 74 0.02 0.04
db block gets from cache 214,499 58.65 113.67
deferred (CURRENT) block cleanout applications 9,723 2.66 5.15
dirty buffers inspected 5,106 1.40 2.71
enqueue conversions 1,130 0.31 0.60
enqueue releases 49,151 13.44 26.05
enqueue requests 49,151 13.44 26.05
enqueue timeouts 0 0.00 0.00
enqueue waits 79 0.02 0.04
exchange deadlocks 0 0.00 0.00
execute count 184,109 50.34 97.57
failed probes on index block reclamation 1 0.00 0.00
free buffer inspected 6,521 1.78 3.46
free buffer requested 8,656 2.37 4.59
global undo segment hints helped 0 0.00 0.00
global undo segment hints were stale 0 0.00 0.00
heap block compress 457 0.12 0.24
hot buffers moved to head of LRU 5,016 1.37 2.66
immediate (CR) block cleanout applications 56 0.02 0.03
immediate (CURRENT) block cleanout applications 4,230 1.16 2.24
index crx upgrade (found) 0 0.00 0.00
index crx upgrade (positioned) 8,362 2.29 4.43
index fast full scans (full) 3,845 1.05 2.04
index fast full scans (rowid ranges) 0 0.00 0.00
index fetch by key 842,761 230.42 446.61
index scans kdiixs1 376,413 102.91 199.48
leaf node 90-10 splits 42 0.01 0.02
leaf node splits 89 0.02 0.05
lob reads 6,759,932 1,848.23 3,582.37
lob writes 11,788 3.22 6.25
lob writes unaligned 11,788 3.22 6.25
logons cumulative 272 0.07 0.14
messages received 133,602 36.53 70.80
messages sent 133,602 36.53 70.80
no buffer to keep pinned count 219 0.06 0.12
no work - consistent read gets 18,462,318 5,047.76 9,783.95
opened cursors cumulative 77,042 21.06 40.83
parse count (failures) 57 0.02 0.03
parse count (hard) 1,311 0.36 0.69
parse count (total) 70,872 19.38 37.56
parse time cpu 542 0.15 0.29
parse time elapsed 2,218 0.61 1.18
physical read IO requests 821 0.22 0.44
physical read bytes 15,212,544 4,159.25 8,061.76
physical read total IO requests 2,953 0.81 1.56
physical read total bytes 48,963,584 13,387.08 25,947.85
physical read total multi block requests 289 0.08 0.15
physical reads 1,857 0.51 0.98
physical reads cache 1,857 0.51 0.98
physical reads cache prefetch 1,036 0.28 0.55
physical reads direct 0 0.00 0.00
physical reads direct (lob) 0 0.00 0.00
physical reads direct temporary tablespace 0 0.00 0.00
physical reads prefetch warmup 0 0.00 0.00
physical write IO requests 6,054 1.66 3.21
physical write bytes 122,142,720 33,394.92 64,728.52
physical write total IO requests 11,533 3.15 6.11
physical write total bytes 199,223,808 54,469.58 105,577.00
physical write total multi block requests 5,894 1.61 3.12
physical writes 14,910 4.08 7.90
physical writes direct 74 0.02 0.04
physical writes direct (lob) 0 0.00 0.00
physical writes direct temporary tablespace 72 0.02 0.04
physical writes from cache 14,836 4.06 7.86
physical writes non checkpoint 14,691 4.02 7.79
pinned buffers inspected 4 0.00 0.00
prefetch clients - default 0 0.00 0.00
prefetch warmup blocks aged out before use 0 0.00 0.00
prefetch warmup blocks flushed out before use 0 0.00 0.00
prefetched blocks aged out before use 0 0.00 0.00
process last non-idle time 2,370 0.65 1.26
queries parallelized 0 0.00 0.00
recovery blocks read 0 0.00 0.00
recursive aborts on index block reclamation 0 0.00 0.00
recursive calls 643,220 175.86 340.87
recursive cpu usage 15,900 4.35 8.43
redo blocks read for recovery 0 0.00 0.00
redo blocks written 96,501 26.38 51.14
redo buffer allocation retries 0 0.00 0.00
redo entries 115,246 31.51 61.07
redo log space requests 0 0.00 0.00
redo log space wait time 0 0.00 0.00
redo ordering marks 3,605 0.99 1.91 -
Hi,
We are sending a file using sender file adapter through PI into SAP system BAPI with receiver RFC adapter.
The input file is received in XI and we don't see any mapping done, but message mapping displaying as succefully delivered to SAP system.
We are using asynch call, no BPM involved. Its just File to RFC.
Any help appreciated.
Thanks.Hi,
Do you make a commit after BAPI call.
There is a parameter in adapter which has to be set.
From documentation:
<i>"Commit Control for Individual BAPI Calls
If you want to use this communication channel to call BAPIs as remote-enabled function modules that change data in the database, set the indicator.
If executed successfully, the transaction is written to the database by calling the function module BAPI_TRANSACTION_COMMIT explicitly. If an error occurs, the transaction is rolled back by BAPI_TRANSACTION_ROLLBACK.
The result is determined by the value of the field TYPE in parameter RETURN. If successful, the tables are empty and the values , S, I, and W are displayed. All other values are regarded as errors.
To change this setting, set the indicator BAPI Advanced Mode.
In the Successful RETURN-TYPE Values table, enter the values that should lead to a successful execution."</i>
Regards,
Wojciech
Message was edited by: Wojciech Gasiorowski -
CO_TXT_ROUTING_BACK_ERROR
Hi Gurus,
IDOC to Webservice
I used BPM (used Asynch-synch bridge )
Orders.Orders05 is my IDOC
I am getting following error.
<SAP:Category>XIServer</SAP:Category>
<SAP:Code area="OUTBINDING">CO_TXT_ROUTING_BACK_ERROR</SAP:Code>
<SAP:P1>,BS_ECC</SAP:P1>
<SAP:P2>,ECC_LS,http://sap.com/xi/XI/Message/30,Acknowledgment</SAP:P2>
<SAP:P3 />
<SAP:P4 />
<SAP:AdditionalText />
<SAP:ApplicationFaultMessage namespace="" />
<SAP:Stack>Error occurred during back-routing Error in communication channel</SAP:Stack>
<SAP:Retry>M</SAP:Retry>
</SAP:Error>
Regards,
Kalpanahi
i think the error is due to the returm message from WS.
you cant handle the reply message for IDOC as u rv doing for other scenario.
to handle the reply message for IDOC we need to configure ALEAUD.
please follow thw link
https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/f6d2d790-0201-0010-9382-b50b499b3fbe
Configuration Tips for a Business Service/Integration Process to send back ALE audit IDOC
regards,
navneet -
IDOC to Webservice (Async/Sync scenario)
Hi Experts,
Can you please help me out need to design an SEnario
IDOc to webservices useing SOAP. to External WEbservice,
I need some step to design.and Need to validate VEndor Number in that. in External webservice.
Reg's
Rajhi raj,
u can use asynch/synch bridge in bpm for synch idoc-xi-soap.
for bpm refer:
/people/arpit.seth/blog/2005/06/27/rfc-scenario-using-bpm--starter-kit
for soap adapter refer;https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/d23cbe11-0d01-0010-5287-873a22024f79
for idoc refer:
/people/prateek.shah/blog/2005/06/08/introduction-to-idoc-xi-file-scenario-and-complete-walk-through-for-starters - IDoc to File
/people/swaroopa.vishwanath/blog/2007/01/22/ale-configuration-for-pushing-idocs-from-sap-to-xi - ALE settings for R3-XI
u can refer all these links and complete ur scenario.
[reward if helpful]
regards,
latika. -
Automatic popup for reason of rejection after async method
Hello,
How to get a <b>reason of rejection</b> popup after an asynchronous method?
Currently I am using the standard <b>singlerelease</b> method of object type BUS2012 for the release of purchase order.
Now once this order gets rejected the user wants an automatic popup for entering the reason for rejection and NOT as another workitem in his inbox for execution.
Please let me know how can I provide this functionality in the current scenario. Also since the Previous method(singlerelease) is asynchronous Advance with Dialog does not work.
Thanks and Regards,
AnujHi Mikko/Anuj,
Added to the FAQ, including techo details.
Cheers,
Mike
3.16. How can I 'advance in dialog' with asynchronous tasks?
This is unfortunately not possible as the very nature of <i>asynchronous</i> tasks contradicts the idea of 'advance in dialog', also known as <i>synchronous</i> dialog chains. The two main reasons for using asynch tasks are either that the termination can happen outside of WF, or to ensure that all database updates are complete. Both of these imply that there is an action happening outside of workflow (database updates can happen outside the user session after the transaction is completed). Since these happen outside the current dialog session, one would have to 'continue' another user's session in the current dialog, which is not possible.
How it works on a more technical level:
Asynchronous tasks use terminating events which live in the instance linkage tables (tx SWEINST). You don't normally have to touch instance linkages, the WF system creates these as necessary, and deletes it when a task completes.
The main difference between instance and type linkages is that instance linkages have a key and are specific to one instance (hence the name), they are only triggered for the object with the same key. Like any other event they are passed on to the event manager who triggers the receiver (the task completion in this case) via RFC (user WF-BATCH). This is why it is theoretically impossible to advance in dialog - because the task completion happens in a separate dialog under a different user, quite possibly on a different server.
Workarounds can include secondary methods (though they have their own restrictions), or redesigning your flow so that it uses a synchronous task instead e.g. by using a wait for event in a separate branch or creating a synchronous task instead. -
Good practice for async databinding, regarding entity-framework.
I have some issue to find a good pattern to keep my UI responsive using asynch databinding and entity framework which is not thread-safe. It would be a shame if the UI came to totaly freeze with no feedback for a connection lost.
So, entity framework is not thread-safe. Which mean I won't be able to rely on lazy loading here and involve than if I do async databinding to a collection of object, each object will have to load their navigations properties with their own context to display
data. Ok, I can handle this. But what if I need more than displaying data and I want to modify a navigation property ?
I need when I set the navigation property in my viewmodel to attach an entity from the same context than the parent entity is from. So here I have one context for multiple thread. What do I do ? =(
Does anyone have some experience in it and can help me to clear this mess from my mind ? =D
All the sample I can find about async databinding don't go any further than make a Thread.Sleep before returning a string, it's not realy helpfull.There are reasons people post simple explanations.
One of those is because they post blog entries when they're learning something rather than when they are expert in it. The blog post is part of their learning process.
I sometimes share context.
What I do is put the context in Application.Current.Resources and then go get it from there when I need to re-use that particular one.
xxxDomainContext DBContext = new xxxDomainContext();
Application.Current.Resources.Add("DBContext", DBContext);
and using that:
context = (xxxxDomainContext)Application.Current.Resources["DBContext"];
You can alternatively share collections in a similar way. Depends on how many related tables you have, but reading a whole tree of static data is sometimes a good idea rather than just one table at a time.
One technique I use is to cache slowly changing data on the client and then go get changes since the last run off the server. For a fair few applications, yesterdays data is good enough for most data sets. It's just the odd one or two which you really
need up to date numbers or whatever.
You can see a collection shared in this sample ( but the sharing is for different reasons ) :
https://gallery.technet.microsoft.com/Dynamic-XAML-Composed-View-e087f3c1
Please don't forget to upvote posts which you like and mark those which answer your question.
My latest Technet article - Dynamic XAML -
Synchronous BPM starting with HTTP Request
Hello there,
I have a requirement to call 3 web services, consolidate their responses, and sending back the results to the caller (a web page in this case). The request is coming over HTTP (synchronously...the file sender scenario that is on SDN does not apply here).
I have tried the sync/async bridge to prototype this synchronous scenario, but I get the following error.
"The process does not support the given synchronous interface or does not support synchronous messages"
Since the Receive step only allows Async and Sync/Async Bridge options (which doesn't really apply here because all of my web service calls are synchronous), and off course the HTTP as well. What options do I have to build this scenario via BPM (or without BPM)?
Looking forward to the replies.
ThanksHi,
I'm trying to build similar ccBPM process which processes a synchronous message sent from XI SOAP sender to a Web Service. Both sender and receiver systems are synchronous.
Instead of going directly to the Web Service receiver, I would like the message to go through ccBPM.
Can you give a sample for such process ?
What receive step mode should be used - "Asynch" or S\A "bridge" ?
I tried to receive the message using "Asynch" mode and than send it with "Synch" mode.
A runtime error is generated:
<SAP:Error xmlns:SAP="http://sap.com/xi/XI/Message/30" xmlns:SOAP="http://schemas.xmlsoap.org/soap/envelope/" SOAP:mustUnderstand="1">
<SAP:Category>XIAdapter</SAP:Category>
<SAP:Code area="BPE_ADAPTER">SYNCHONOUS_MESSAGE</SAP:Code>
<SAP:P1 />
<SAP:P2 />
<SAP:P3 />
<SAP:P4 />
<SAP:AdditionalText />
<SAP:ApplicationFaultMessage namespace="" />
<SAP:Stack>The process does not support the given synchronous interface or does not support synchronous messages Exception CX_WAPI_DEF_PROPERTY_NOT_FOUND occurred (program: SAPMHTTP, include: , line: 0).</SAP:Stack>
<SAP:Retry>N</SAP:Retry>
</SAP:Error>
Please help !
Thanks
Orit -
We have just deployed a 4-node RAC cluster on 10GR2. We force a log switch every 5 minutes to ensure our Dataguard standby site is relatively up to date, we use the ARCH to ship logs. We are running to a very fast HP XP 12000 with massive amounts of write cache, so we never actually write straight to disk. However everytime we do a log switch and archive the log, we see a massive spike in the log file sync event. This is a real-time billing system so we monitor transaction response times in ms. Our response time for a transaction can go from 8ms to around 500ms.
I can't understand why this is happening, not only are our disks fast but we are also using asynch I/O and ASM. Surely with asynch I/O you should never wait for a write to complete.Log file sync event happens when client wait for LGWR finishes write to the log file after client said 'commit'. The way to reduce the number of the 'Log file sync' events is to increase the speed of LGWR process or not to commit that often.
You've described your disk system as very fast - what is the amount of data you write on every log switch? How does the performance of this write relates to your disk system tests? what block size did you use when testing the disk system? as far as I remember the LGWR uses OS block size and not the DB block size to write data to the disk. Try to experiment on your test system - put your log files on the virtual disk created in RAM and run the test case - do you see the delays?
With such restrictions for the transaction time you may want to look at Oracle Times-Ten database (http://www.oracle.com/database/timesten.html)
Since you've mentioned the 10gR2 you could probably use the new feature - asynchronous commit - in this case your transaction will not wait for the LGWR process. Be aware that using the NOWAIT commit opens a small possibility of data loss - the doc describes it quite clear.
http://download-east.oracle.com/docs/cd/B19306_01/appdev.102/b14251/adfns_sqlproc.htm#CIHEDGBF
Mike -
How to achieve failover when the master definition site is down ?
I have two Oracle 8i instances in a multi-master async replication environment. Two J2EE application servers connect to each of the Oracle instance respectively. When one of the oracle instance is down, I'd like to have my J2EE application failover to the second Oracle instance automatically without human intervention. I understand I can use OCI driver to allow my J2EE app to connect to the other Oracle instance when the first is down. But here are some questions:
1. If I understand it correctly, there's only one master definition site. Let's say my master definition site sits with the first Oracle instance. If this server is down, the DBA has to go in to config the second Oracle instance to be the master definition site. Can this be done automatically without the DBA ?
2. If the master definition site is down and the DBA is yet to come in, will the second Oracle instance still work ? I do understand the replication won't work, but will normal database read/write work ? What happens between the master definition site down and the master definition site reconfigured ?
3. Are there any low cost third party tools can handle what I wanted ?
Thanks in advance,To answer your questions.
1. Yes there is only one master def site, it is not important to relocate the master def site, in the event of its failure. The def site is used as a the source for replication definition generation.
2. It depends upon what type of multi-master replication you have implemented. If you are using asynch replication, the site that is up will continue to allow updates and will queue the transactions. When the other site(s) are active again, it will forward the transactions. Note, this can cause untold problems if you have a heavy used system, and narrow bandwidth. If on the other hand you are using synchanous replication, when one site goes down, the other site(s) lock their tables to prevent dml/ddl changes. Then you have to decide whether or not you can wait for the down site to be restored, or break replication to continue operations.
3. Yes, there are a few, one that comes to mind, I believe is Shareplex.
Maybe you are looking for
-
Is there a function to predict the next point accoding to the known point?
For example ten points arranged in a parabola, is there a function to predict the coordinates of the six point behind according to the four points in front of?
-
We are trying to share PDFs over a company network, but are having problems with people not being able to save their changes. Some of our employees work at a site that is out of state, but they are still connected to the company network. We have abou
-
I have restored my IPad to factory settings however cannot now get in as password is incorrect
I have restored IPad to factory settings however can not get now as password is not working. I have tried the initial password and also using 0000.
-
Hi, The SAP report RFUMSV00 dump when i execute it with any selection criteria. Error analysis The following syntax error was found in the program CL_BADI_BASE==================CP : "Method "TYPE_IS_BADI" does not have parameters . . ."
-
Transactional Metrics for Status Changes
I'm using the Plant Maintenance module in PR4, with SAP GUI 720 Final Release, if that helps. Does someone know where I can start to build metrics regarding the quantity per time period of items that went from one user and/or system status to another