URGENT:SMON: Parallel transaction recovery tried

Hi,
I got these following messages in the BDUMP file. Around these time my process was always getting blocked while updating a table. Does this messages mean anything critical:
Dump file d:\oracle\admin\usmdb\bdump\usmdb_smon_1564.trc
Sun Mar 07 03:17:27 2004
ORACLE V9.2.0.1.0 - Production vsnsta=0
vsnsql=12 vsnxtr=3
Windows 2000 Version 5.0 Service Pack 3, CPU type 586
Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
With the Partitioning, OLAP and Oracle Data Mining options
JServer Release 9.2.0.1.0 - Production
Windows 2000 Version 5.0 Service Pack 3, CPU type 586
Instance name: usmdb
Redo thread mounted by this instance: 1
Oracle process number: 6
Windows thread id: 1564, image: ORACLE.EXE
*** 2004-03-07 03:17:27.000
*** SESSION ID:(5.1) 2004-03-07 03:17:27.000
*** 2004-03-07 03:17:27.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 03:58:28.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 04:38:38.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 04:39:23.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 05:20:16.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 06:01:25.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 06:42:27.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 07:23:28.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 08:04:33.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 08:45:31.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 09:26:36.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 10:07:42.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 10:48:46.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 11:29:41.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 12:10:56.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 12:51:52.000
SMON: Parallel transaction recovery tried
*** 2004-03-07 13:32:55.000
SMON: Parallel transaction recovery tried
Thanks,
Tuhin

Hello
No there is no problem. SMON probe a recovery in parallel but failed, and done in serial mode.
lajos

Similar Messages

  • SMON: Parallel transaction recovery tried

    Hi,
    I got these following messages in the BDUMP file. Around these time my process was always getting blocked while updating a table. Does this messages mean anything critical:
    Dump file d:\oracle\admin\usmdb\bdump\usmdb_smon_1564.trc
    Sun Mar 07 03:17:27 2004
    ORACLE V9.2.0.1.0 - Production vsnsta=0
    vsnsql=12 vsnxtr=3
    Windows 2000 Version 5.0 Service Pack 3, CPU type 586
    Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    Windows 2000 Version 5.0 Service Pack 3, CPU type 586
    Instance name: usmdb
    Redo thread mounted by this instance: 1
    Oracle process number: 6
    Windows thread id: 1564, image: ORACLE.EXE
    *** 2004-03-07 03:17:27.000
    *** SESSION ID:(5.1) 2004-03-07 03:17:27.000
    *** 2004-03-07 03:17:27.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 03:58:28.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 04:38:38.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 04:39:23.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 05:20:16.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 06:01:25.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 06:42:27.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 07:23:28.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 08:04:33.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 08:45:31.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 09:26:36.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 10:07:42.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 10:48:46.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 11:29:41.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 12:10:56.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 12:51:52.000
    SMON: Parallel transaction recovery tried
    *** 2004-03-07 13:32:55.000
    SMON: Parallel transaction recovery tried
    Thanks,
    Tuhin

    That would occur because a large transaction (or transactions) had been killed / interrupted while the instance was running (or when the instance was shutdown abort). SMON takes over the job of "cleanup" and may use Parallel Recovery . You should be able to monitor the recovery in the V$FAST_START_TRANSACTIONS view.

  • Parallel Transaction recovery caught exception 30319

    Hi,
    Database - Oracle 11.1.0.7
    Server - RHEL 5.2
    A load process which was to load 7 milllion records was killed after 4 hours because it had several errors in the alert log -
    22238:ORA-04030: out of process memory when trying to allocate 10504 bytes (pga heap,kgh stack)
    After the process was killed, even after 3 hours, there were several trc files being generated with the error -
    Dead transaction 0x0001.04a.00005b78 recovered by 32 server(s)
    SMON: Parallel transaction recovery tried
    Parallel Transaction recovery caught exception 30319
    and
    Parallel Transaction recovery server caught exception 10388
    *** 2010-10-07 20:40:01.124
    *** SESSION ID:(2993.32762) 2010-10-07 20:40:01.124
    *** SERVICE NAME:(SYS$BACKGROUND) 2010-10-07 20:40:01.124
    Does this means that the smon recovery failed? If so, what can I do for this? I need to kick off the load process again.
    these are some other queries I ran -
    select l.message,l.sql_id,l.totalwork,l.sofar,l.*
    from v$session_longops l
    where time_remaining is not null and time_remaining > 0
    Row#     MESSAGE     SQL_ID     TOTALWORK     SOFAR     SID     TIME_REMAINING     ELAPSED_SECONDS
    1     Table Scan: prdb.prpeet: 156648 out of 843440 Blocks done     3w1cxu6vbz4jn     843440     156648     2956     8707     1986
    2     Index Fast Full Scan: prdb.prpeet: 15 out of 216076 Blocks done     7cacc3d07d041     216076     15     2953     3442572     239
    3     Index Fast Full Scan: prdb.prpeet: 192281 out of 206419 Blocks done     d2cu3p7tmuz0z     206419     192281     2953     37     509
    4     Index Fast Full Scan: prdb.prpeet: 198899 out of 216076 Blocks done     f7whmwp813kf8     216076     198899     2953     46     531
    5     Index Fast Full Scan: prdb.prpeet: 5 out of 216076 Blocks done     dwf4gghk4mq0z     216076     5     2953     604999     14
    6     Index Fast Full Scan: prdb.prpeet: 7 out of 216076 Blocks done     dwf4gghk4mq0z     216076     7     2953     864276     28
    7     Index Fast Full Scan: prdb.prpeet: 6 out of 216076 Blocks done     7cacc3d07d041     216076     6     2953     504163     14
    8     Index Fast Full Scan: prdb.prpeet: 9 out of 216076 Blocks done     7cacc3d07d041     216076     9     2953     1128350     47
    9     Index Fast Full Scan: prdb.prpeet: 13 out of 216076 Blocks done     7cacc3d07d041     216076     13     2953     1961187     118
    Can someone please tell me what I need to do so that I can restart the load?
    Thanks.

    I do -
    OPEN cur_n1_C1 ;
    LOOP
    FETCH cur_n1_C1 BULK COLLECT INTO vt_C1 LIMIT 1000000;
    FOR i IN vt_C1.FIRST .. vt_C1.LAST LOOP
    INSERT /*+ APPEND */ INTO prpeet
    VALUES
    END LOOP ;
    COMMIT;
    END IF ;
    END LOOP;
    CLOSE cur_n1_C1 ;
    EXCEPTION
    So, will this commit be a problem? I am loading around 7 million records.
    Also,
    I ran the v$session_longops query again now and this is what it says -
    select l.message,l.sql_id,l.totalwork,l.sofar,l.sid, l.time_remaining, elapsed_seconds
    from v$session_longops l
    where time_remaining is not null and time_remaining > 0
    Row#     MESSAGE     SQL_ID     TOTALWORK     SOFAR     SID     TIME_REMAINING     ELAPSED_SECONDS
    1     Sort Output: : 55613 out of 122712 Blocks done          122712     55613     2987     490     406
    3     Table Scan: prdb.prpeet: 160239 out of 843440 Blocks done     3w1cxu6vbz4jn     843440     160239     2956     21071     4942
    Does this mean it will finish in 21071 seconds?
    Edited by: user12158503 on Oct 7, 2010 9:41 PM

  • Transaction recovery caught exception

    i see my .trc file
    *** 2009-05-17 15:57:03.492
    Parallel Transaction recovery caught exception 30319
    Parallel Transaction recovery caught error 30319
    *** 2009-05-17 16:01:04.040
    Parallel Transaction recovery caught exception 30319
    Parallel Transaction recovery caught error 30319
    what it is and what must do i for this?

    have a look at this article and see if it helps.
    http://el-caro.blogspot.com/2008/11/parallel-rollback.html

  • Oracle 9 and XA transaction recovery

    Just thought I would pass on a bit of advice if you are attempting to enable XA transaction recovery from either Application Server 7.0 or 8.1 with Oracle version 9.2.0.6 and possibly above.
    In 9.2.0.6 Oracle changed the way they responded to XA recovery queries from transaction managers and so to now allow a list of XAID's to be returned without an exception being raised you will need to have the following privileges on the Application Server Oracle user attempting to do the recovery:
    Select on DBA_PENDING_TRANSACTIONS, PENDING_TRANS$, DBA_2PC_PENDING and DBA_2PC_NEIGHBORS
    and the new one execute on DBMS_SYSTEM
    Also I discovered that although a db ping is possible in Application server 8.1 with lower case properties on the user, password and url (JDBC Pool properties) in order to get XA recovery to work for some reason these had to be changed to be case sensitive. See below:
    Not working:
    <jdbc-connection-pool connection-validation-method="table"
    datasource-classname="oracle.jdbc.xa.client.OracleXADataSou
    rce" fail-all-connections="false" idle-timeout-in-seconds="300"
    is-connection-validation-required="true" is-isolation-leve
    l-guaranteed="false" max-pool-size="32"
    max-wait-time-in-millis="60000" name="OraclePool"
    pool-resize-quantity="2" res-typ
    e="javax.sql.XADataSource" steady-pool-size="8"
    validation-table-name="DUAL">
    <description>Oracle Pool</description>
    <property name="user" value="sunone"/>
    <property name="url" value="jdbc:oracle:oci:@SS"/>
    <property name="password" value="sunone"/>
    </jdbc-connection-pool>
    Working!!
    <jdbc-connection-pool connection-validation-method="table"
    datasource-classname="oracle.jdbc.xa.client.OracleXADataSou
    rce" fail-all-connections="false" idle-timeout-in-seconds="300"
    is-connection-validation-required="true" is-isolation-leve
    l-guaranteed="false" max-pool-size="32"
    max-wait-time-in-millis="60000" name="OraclePool"
    pool-resize-quantity="2" res-typ
    e="javax.sql.XADataSource" steady-pool-size="8"
    validation-table-name="DUAL">
    <description>Oracle Pool</description>
    <property name="User" value="sunone"/>
    <property name="URL" value="jdbc:oracle:oci:@SS"/>
    <property name="Password" value="sunone"/>
    </jdbc-connection-pool>
    Just thought I would share that with you!

    I am able to connect and message to Oracle AQ using an XA Datasource using the getVendorConnection method on the Weblogic connection wrapper (WLConnection).
    This, however, seems to be memory leaking quite badly. I've followed the Weblogic documentation (which says that you should not call close() on the Vendor connection).
    I've also tried closing the vendorConnection object and nulling out the reference that the WLConnection holds. This slows the leak, but it appears that the WLConnections are never being released.
    Has anyone else had any success doing this?
    Dom.

  • Transaction Recovery Service failover

    Can anyone explain what the suggested configuration is for the default persistence store? In particular, this is to ensure the proper failover / migration of the Transaction Recovery Service which is required to use the Default Persistence Store which is file based. Based on the following statement from the docs:
    Preparing to Migrate the Transaction Recovery Service
    To migrate the Transaction Recovery Service from a failed server in a cluster to another server (backup server) in the same cluster, the backup server must have access to the transaction log records from the failed server. Therefore, you must store default persistent store data files on persistent storage available to all potential backup servers in the cluster. Oracle recommends that you store transaction log records on a Storage Area Network (SAN) device or a dual-ported disk. Do not use an NFS file system to store transaction log records. Because of the caching scheme in NFS, files on disk may not always be current. Using transaction log records stored on an NFS device for recovery may cause data corruption.
    A SAN storage device is recommended for this but my understanding is that the SAN can only be mounted by one machine at a time. Does this imply then that our failover process needs to include mounting the SAN before starting the failover server (as part of the whole server migration)? The docs here (http://download.oracle.com/docs/cd/E15523_01/core.1111/e12036/net.htm#CIHBDDAA) indicate that a NAS can be used and even give examples of configuring it using NFS mount points:
    The following commands show how to share the SOA TX logs location across different nodes:
    SOAHOST1> mount nasfiler:/vol/vol1/u01/app/oracle/stores/soadomain/soa_cluster/tlogs
    +/u01/app/oracle/stores/soadomain/soa_cluster/tlogs -t nfs+
    SOAHOST2> nasfiler:/vol/vol1/u01/app/oracle/stores/soadomain/soa_cluster/tlogs
    +/u01/app/oracle/stores/soadomain/soa_cluster/tlogs -t nfs+
    Can anyone describe a best-practices approach for how to configure the expected persistent storage solution that will work with proper failover of the transaction recovery service?
    Thanks!
    Gary

    have a look at this article and see if it helps.
    http://el-caro.blogspot.com/2008/11/parallel-rollback.html

  • HT4009 In-App Purchases denied "Please contact iTunes Support to complete transaction" when trying to buy Gems in a game.

    In-App Purchases denied "Please contact iTunes Support to complete transaction" when trying to buy Gems in a game.
    I want to gem In-app very much. I need to do.

    Now. I contract itune support. But I have not received the email from Itune support.
    about 2 days.

  • Transaction recovery: caught and ignored

    Since a week ago I had noticed in my alert log the following message repeated many many times:
    Transaction recovery: caught and ignored
    I suspect this is related with a big datapump import operation that I did and do not end due to I cancelled it because of errors. After that the frecuency of the log switch began to occur too fast(every 20 seconds, redo size 50MB).
    Actually I have not idea about what can be happening so I aprecciate any help you can give me.
    Thanks very much.

    http://serdarturgut.blogspot.com/2010/05/transaction-recovery-lock-conflict.html

  • Urgent Please - Receiving Transaction Processor Issue

    Hi,
    We are on 11.5.10. I am trying to create PO receipts using the interface program "Receiving Transaction Processor". I inserted data into the rcv_headers_interface and rcv_transactions_interface tables. As this dta is for contract labor, i am assuming that there is no need to insert data into shipment_related interface table. Following shows the insert statement I made sure that I am running the process from the correct org (the org that i used in the filed destination_oprganization field). But still the process completes normally but po_interface_errors has a record with error message as :
    "RVTSH-189: Subroutine rvtshiline() returned error
    Cause: Subroutine rvtshiline() returned an internal error.
    Action: Note this error number and the actions you are trying to perform.
    Contact your system administrator for"
    The receipt gets created succesfully from the form. Anyone knows the issue? Please help me...thanks
    Here are the insert statements...
    --Insert data into Interface tables
    INSERT INTO rcv_headers_interface
    (header_interface_id,
    GROUP_ID,
    processing_status_code,
    receipt_source_code,
    transaction_type,
    auto_transact_code,
    last_update_date,
    last_updated_by,
    creation_date,
    created_by,
    expected_receipt_date,
    comments,
    validation_flag,
    packing_slip,
    vendor_id,
    vendor_site_id,
    ship_to_organization_id
    VALUES (rcv_headers_interface_s.NEXTVAL,
    rcv_interface_groups_s.NEXTVAL,
    'PENDING',
    'VENDOR',
    'NEW',
    'RECEIVE',
    SYSDATE,
    fnd_global.user_id,
    SYSDATE,
    fnd_global.user_id,
    SYSDATE,
    'test',
    'Y',
    'ABC',
    po_headers.vendor_id,
    po_headers.vendor_site_id
    po_distributions.destination_organization_id
    --Insert record in transactions_interface
    INSERT INTO rcv_transactions_interface
    (interface_transaction_id,
    header_interface_id,
    processing_status_code,
    receipt_source_code,
    transaction_date,
    GROUP_ID,
    last_update_date,
    last_updated_by,
    creation_date,
    created_by,
    transaction_type,
    processing_mode_code,
    transaction_status_code,
    quantity,
    unit_of_measure,
    source_document_code,
    po_header_id,
    po_line_id,
    po_line_location_id,
    po_distribution_id,
    to_organization_id,
    comments,
    interface_source_code,
    employee_id,
    auto_transact_code,
    destination_type_code,
    attribute_category,
    category_id,
    item_description
    VALUES (rcv_transactions_interface_s.NEXTVAL,
    rcv_headers_interface_s.CURRVAL,
    'PENDING',
    'VENDOR',
    SYSDATE,
    rcv_interface_groups_s.CURRVAL,
    SYSDATE,
    fnd_global.user_id,
    SYSDATE,
    fnd_global.user_id,
    'RECEIVE', --'NEW',
    'BATCH',
    'PENDING',
    2,
    po_lines.unit_meas_lookup_code,
    'PO',
    po_headers..po_header_id,
    po_lines.po_line_id,
    po_line_locations.po_line_location_id,
    po_distributions.po_distribution_id,
    po_distributions.destination_organization_id,
    'TEST',
    'RCV',
    1234, --employee_id
    'RECEIVE',
    'RECEIVING',
    po_distributions.destination_organization_id,
    45074,
    'Standard'
    );

    Pl see you duplicate post here - PO Receipts Interface Error!!
    For urgent issues, pl open an SR with Support
    Srini

  • Urgent Please - Receiving Transaction Processor

    Hi,
    We are on 11.5.10. I am trying to create PO receipts using the interface program "Receiving Transaction Processor". I inserted data into the rcv_headers_interface and rcv_transactions_interface tables. As this dta is for contract labor, i am assuming that there is no need to insert data into shipment_related interface table. Following shows the insert statement I made sure that I am running the process from the correct org (the org that i used in the filed destination_oprganization field). But still the process completes normally but po_interface_errors has a record with error message as :
    "RVTSH-189: Subroutine rvtshiline() returned error
    Cause: Subroutine rvtshiline() returned an internal error.
    Action: Note this error number and the actions you are trying to perform.
    Contact your system administrator for"
    The receipt gets created succesfully from the form. Anyone knows the issue? Please help me...thanks
    Here are the insert statements...
    --Insert data into Interface tables
    INSERT INTO rcv_headers_interface
    (header_interface_id,
    GROUP_ID,
    processing_status_code,
    receipt_source_code,
    transaction_type,
    auto_transact_code,
    last_update_date,
    last_updated_by,
    creation_date,
    created_by,
    expected_receipt_date,
    comments,
    validation_flag,
    packing_slip,
    vendor_id,
    vendor_site_id,
    ship_to_organization_id
    VALUES (rcv_headers_interface_s.NEXTVAL,
    rcv_interface_groups_s.NEXTVAL,
    'PENDING',
    'VENDOR',
    'NEW',
    'RECEIVE',
    SYSDATE,
    fnd_global.user_id,
    SYSDATE,
    fnd_global.user_id,
    SYSDATE,
    'test',
    'Y',
    'ABC',
    po_headers.vendor_id,
    po_headers.vendor_site_id
    po_distributions.destination_organization_id
    --Insert record in transactions_interface
    INSERT INTO rcv_transactions_interface
    (interface_transaction_id,
    header_interface_id,
    processing_status_code,
    receipt_source_code,
    transaction_date,
    GROUP_ID,
    last_update_date,
    last_updated_by,
    creation_date,
    created_by,
    transaction_type,
    processing_mode_code,
    transaction_status_code,
    quantity,
    unit_of_measure,
    source_document_code,
    po_header_id,
    po_line_id,
    po_line_location_id,
    po_distribution_id,
    to_organization_id,
    comments,
    interface_source_code,
    employee_id,
    auto_transact_code,
    destination_type_code,
    attribute_category,
    category_id,
    item_description
    VALUES (rcv_transactions_interface_s.NEXTVAL,
    rcv_headers_interface_s.CURRVAL,
    'PENDING',
    'VENDOR',
    SYSDATE,
    rcv_interface_groups_s.CURRVAL,
    SYSDATE,
    fnd_global.user_id,
    SYSDATE,
    fnd_global.user_id,
    'RECEIVE', --'NEW',
    'BATCH',
    'PENDING',
    2,
    po_lines.unit_meas_lookup_code,
    'PO',
    po_headers..po_header_id,
    po_lines.po_line_id,
    po_line_locations.po_line_location_id,
    po_distributions.po_distribution_id,
    po_distributions.destination_organization_id,
    'TEST',
    'RCV',
    1234, --employee_id
    'RECEIVE',
    'RECEIVING',
    po_distributions.destination_organization_id,
    45074,
    'Standard'
    );

    Pl see your triplicate post here - PO Receipts Interface Error!!
    For urgent issues, pl open an SR with Support.
    Srini

  • URGENT:Error in Transaction HRUSER

    Hi,
    I have installed and configured ESS 1.0 WD version. Now when I go to ECC 6.0 to assign employees to the user ids through the transaction HRUSER. I get an error like error occured cannot assign employee 123456 to user ID ABCD.
    I have also tried entering user id in PA30 infotype 105 subtyp 0001.Thereit gets saved but is not reflected in the EP ESS. The ESS R/3 however works.
    I tried debugging HRUSER and found that the subtyp 001 is passed as 10. I corrected that in debugger so the employee got assigned.
    I want to ask am I doing some step wrong or some configurations needs to be done before using HRUSER transaction.
    or is it more like a bug?
    All suggestions are welcome and will be rewarded accordingly.
    regards
    Justin

    what are the steps you follow that you followed when you run the transaction.

  • Transaction Recovery within an Oracle RAC environment

    Good evening everyone.
    I need some help with Oracle 11gR1 RAC transaction-level recovery issues. Here's the scenario.
    We have a three(3) node RAC Cluster running Oracle 11g R1. The Web UI portion of the application connects through WLS 9.2.3 with connection pooling set. We also have a command-line/SQL*Developer component that uses a TNSNAMES file that allows for both failover and load balancing. Within either the UI or the command line portion of the application, a user can run a process by which invokes one or more PL/SQL Packages to be invoked. The exact location of the physical to the database is dependent on which server is chosen from either the connection pooling or the TNSNAMES.ORA Load Balancing option.
    In the normal world, the process executes and all is good. The status of the execution of this process is updated by the Packages once completed. The problem we are encountering is when an Oracle Instance fails. Here's where I need some help. For Application-level (Transaction Level) recovery, the database instances are first recovered by the database background proccesses and then Users must determine which processes were in flight and either re-execute them (if restart processing is part of the process) or remove any changes and restart from scratch. Given that the database instance does not record which processes are "in flight" it is the responsibility of the application to perform its own recovery processing. Is this still true?
    If an instance fails, are "in flight" transactions/connections moved to other instances in the Grid/RAC environment? I don't think this is possible but I don't remember if this was accomplished through a combination of Application and Database Server features that provide feedback between each other. How is the underlying application notified of the change if such an issue occurs? I remember something similar to this in older versions of Oracle but I cannot remember what it was callled.
    Any help or guidance would be great as our client is being extremely difficult in pressing this issue.
    Thanks in advance
    Stephen Karniotis
    Project Architect - Compuware
    [email protected]
    (248) 408-2918

    You have not indicated whether you are using TAF or FCF ... that would be the first place to start.
    My recommendation would be to let Oracle roll back the database changes and have the application resubmit the most recent work.
    If the application knows what it did since the last "COMMIT" then you should be fine with the possible exception of variables stored
    in packages. Depending on packages retaining values is an issue best solved with PRAGMA SERIALLY_REUSABLE ... in other words
    not using the retention feature.

  • Urgent: Triggering a Transaction BMV0..

    Hi Friends,
    I am having a Zprogram .. i need to trigger the BMV0 .from the Zprogram..it is not a call transaction .. i was asked to create a Event and to trigeer the Program Associate to that Event ... I tried to create a Event from SM62 there i dont have an option for entering the program name... how do i know that this event is specific to this program ... please help me with a code if possible with the steps to associate a program to that Event...
    i tried to crate background job with that Zprogram and the program corresponding to the Transaction BMV0...it says job called be scheduled to the program associate to this transaction...it was a standard programm.....
    please help me in solving this issue...

    hi,
    define a job from sm36 or bmv0
    with event instead of date/time
    A.

  • Help needed urgent please ( parallel port access in linux)

    hi all, i am trying to read and write data to /from the parallel port , i am running IBM j9 vm on linux.under /dev there are lp0..lp3 parport0..parport3, and port devices , i dont know which one to use , i have tried all , they all return with an ioexception saying "cannot find the file" except for port device which does not generate any error but it does not return what i have written on it ,
    any help would be much appreciated.
    i am using fileoutputstream and printwriter calsses .
    i dont want to use javax.comm api cause i think its incompatible with ibms visual age micro edition ( the one i am using ) so i would prefer something that works with my IDE .
    thanks
    yahya
    [email protected]

    However, there is an on-going OpenSource version called "RXTX" (http://jarvi.ezlink.com/rxtx/) that works quite well. Be sure to read all the documentation that comes with it, as you'll need to set some permissions on the ports you want to use.

  • Transaction Recovery

    A user is connect from past 5 hours, and is inserting continuesly, without any idle time, till now user has inserted 2500 rows in a table, but does not commits. suddenly user expreinces client procees failure (i.e. client machine from where user was connected has rebooted). Now PMON process will rollback all the transactions. is there a way to recover those 2500 transtions?
    Regards
    Neo

    you mean redo log file only contains dump of transaction without any ownership & session informationNo, the O.S. infos, Oracle Username, etc are stored in redo logs. If a rollback is performed, it's too late.
    now how to recover those transactions,There is no recovery involved in what you describe. This is normal transaction processing. If you fail a transaction, it is rolled back. Normal functionning.
    If you like to play with data integrity, try to use the Log Miner on your achived redo logs, but I warn you... this is dangerous and will sooner or later make your data inconcistent. One and only warning.
    I suppose DBA's primary motive is to protect data in almost every case of disasterWell... yes?
    not to train the users/developers.There I disagree. Who better than a DBA can teach devs/users how to use a database well?
    becuase in almost every situation there is space for humans to commit a mistake.Yes, and in your case, the mistake has been made by the user (no commit) and the dba (not explaining what could happen in case of problem to the user).
    Regards,
    Yoann.

Maybe you are looking for

  • Jabber for Windows 10.6.1 Problem: Chat History sporadically empty

    Hi, we have Jabber for Windows 10.6.1 running on Win 7 64bit Machines. We have the Problem that sporadically the Chat History windows is empty. If this happens, I cannot see my incoming chats (but I can hear Audio that a IM just arrived) and I also c

  • Taking time to fetch records from BSEG

    Hi all, When iam fetching wrbtr,augdt,zterm fields  from BSEG table by passing VBELN and BUKRS it is taking lot of time to get.My query goes like this   select vbeln wrbtr augdt zterm from bseg into table t_bsegtab for all             entries in t_vt

  • Information about badis

    hai, i want complete detailes about badi  and where they are used and how to use in standard applications,how it is created. plz give me the complete detailes badis. Moderator message : Not enough re-search before posting, read forum rules before pos

  • Create local view with Bex Analyzer 7.0

    Hi, with the Bex Analyzer 3.5 it was possible to create and use a local views based on a query in a workbook. With this "workaround" it was possible to create just one Query for different views and charts. Is there something similar in BI 7.0??? I kn

  • What Firewire Do I Need?

    Just purchased a 1360t desktop.  I need a firewire port to download video from my camcorder.  Anyone know what I need for this?  Thanks in advance! This question was solved. View Solution.