Error while taking GC heap Dump using Microsoft PerfView

Hello,
When I try to take heap dump for an application process using Microsoft
PerfView , error observed
and the error log given below. Can you please let me know the root cause for his issue ?
Steps followed
From the PerfView UI, choose “Take Heap Snapshot,” located on the Memory menu.
And choose the process to capture
Click the “Dump GC Heap” button or simply double click on the process name.
Error Log
Completed: Dumping GC Heap to C:\Install\TM\PerfView\TestProcess.gcDump   (Elapsed Time: 1.156 sec)
Error: HeapDump failed with exit code 1
Directory TestProcess.gcdump does not exist
Started: Dumping GC Heap to C:\Install\TM\PerfView\TestProcess.1.gcDump
Collecting a GC Heap SnapShot for process 1704
[Taking heap snapshot of process '1704' ID 1704 to TestProcess.1.gcdump.  This can take 10s of seconds to minutes.]
During the dump the process will be frozen.   If the dump is aborted, the process being dumped will need to be killed.
Starting dump at 8/08/2014 3:55:15 PM
Starting Heap dump on Process 1704 running architecture AMD64.
set _NT_SYMBOL_PATH=SRV*C:\Users\UserId\AppData\Local\Temp\3\symbols*http://msdl.microsoft.com/download/symbols
Exec: "C:\Users\UserId\AppData\Roaming\PerfView\VER.2014-08-08.13.49.17.346\AMD64\HeapDump.exe"  /MaxDumpCountK=250 "1704" "TestProcess.1.gcdump"
Looking for C:\Users\UserId\AppData\Roaming\PerfView\VER.2014-08-08.13.49.17.346\Microsoft.Diagnostics.FastSerialization.dll
Dumping process 1704 with id 1704.
Process Has DotNet: False Has JScript: False Has ClrDll: False
HeapDump Error: Could not dump either a .NET or JavaScript Heap.  See log file for details
Completed: Dumping GC Heap to C:\Install\TM\PerfView\TestProcess.1.gcDump   (Elapsed Time: 1.172 sec)
Error: HeapDump failed with exit code 1
Directory TestProcess.1.gcdump does not exist

Hi,
Below is the DDL taken from different database. Will this be enough ? One more thing please, what shall be the password should it be DMSYS.....since this will not be used by me but system.
CREATE USER "DMSYS" PROFILE "DEFAULT" IDENTIFIED BY "*******" PASSWORD EXPIRE DEFAULT TABLESPACE "SYSAUX" TEMPORARY TABLESPACE "TEMP" QUOTA 204800 K ON "SYSAUX" ACCOUNT LOCK
GRANT ALTER SESSION TO "DMSYS"
GRANT ALTER SYSTEM TO "DMSYS"
GRANT CREATE JOB TO "DMSYS"
GRANT CREATE LIBRARY TO "DMSYS"
GRANT CREATE PROCEDURE TO "DMSYS"
GRANT CREATE PUBLIC SYNONYM TO "DMSYS"
GRANT CREATE SEQUENCE TO "DMSYS"
GRANT CREATE SESSION TO "DMSYS"
GRANT CREATE SYNONYM TO "DMSYS"
GRANT CREATE TABLE TO "DMSYS"
GRANT CREATE TRIGGER TO "DMSYS"
GRANT CREATE TYPE TO "DMSYS"
GRANT CREATE VIEW TO "DMSYS"
GRANT DROP PUBLIC SYNONYM TO "DMSYS"
GRANT QUERY REWRITE TO "DMSYS"
GRANT SELECT ON "SYS"."DBA_JOBS_RUNNING" TO "DMSYS"
GRANT SELECT ON "SYS"."DBA_REGISTRY" TO "DMSYS"
GRANT SELECT ON "SYS"."DBA_SYS_PRIVS" TO "DMSYS"
GRANT SELECT ON "SYS"."DBA_TAB_PRIVS" TO "DMSYS"
GRANT SELECT ON "SYS"."DBA_TEMP_FILES" TO "DMSYS"
GRANT EXECUTE ON "SYS"."DBMS_LOCK" TO "DMSYS"
GRANT EXECUTE ON "SYS"."DBMS_REGISTRY" TO "DMSYS"
GRANT EXECUTE ON "SYS"."DBMS_SYSTEM" TO "DMSYS"
GRANT EXECUTE ON "SYS"."DBMS_SYS_ERROR" TO "DMSYS"
GRANT DELETE ON "SYS"."EXPDEPACT$" TO "DMSYS"
GRANT INSERT ON "SYS"."EXPDEPACT$" TO "DMSYS"
GRANT SELECT ON "SYS"."EXPDEPACT$" TO "DMSYS"
GRANT UPDATE ON "SYS"."EXPDEPACT$" TO "DMSYS"
GRANT SELECT ON "SYS"."V_$PARAMETER" TO "DMSYS"
GRANT SELECT ON "SYS"."V_$SESSION" TO "DMSYS"
The other database has the DMSYS and the status is EXPIRED & LOCKED but I'm still able to take the dump using datapump??

Similar Messages

  • Error while taking Backup using hpux tape drive

    I got the following error while taking a backup.
    ERROR:   System Error, /opt/ignite/bin/make_medialif failed creating boot LIF 
    ERROR:   Failed to generate LIF on tape.
    ERROR:   System Error, /opt/ignite/bin/make_medialif failed creating boot LIF
    ERROR:   Failed to generate LIF on tape.
    make_medialif: ERROR: Cannot tar and gzip script files; possibly insufficient di
    [24;1H[0K[7mscbackup.log (13%)[m[24;1H[0K/11[24;1H[24;1H[0Ksk space in /var/tmp/make_medialif8916.
    =======  11/10/10 12:17:49 IST  make_tape_recovery completed unsuccessfully
    ERROR:     Ignite tape writing phase had errors
               Exit status: 1
    How can this problem be resolved? Kindly drop in your comments and solutions. 

    Hello Kulvinderbuttar,
    I understand that you are not able to make the recovery media.
    Are you using recovery media creation?
    Here is a link to walk you through the steps to create the recovery media.
    The flash drive that you are using needs to be empty and may need to be reformatted before you can make the recovery media.
    Let me know how everything goes.

  • Ora-00604 error while taking tkprof of a trace file

    Sorry i am giving the full erro but omitting exact table names
    Hi ,
    I have an error while taking tkprof of a trace file.
    I gave the following command ---
    tkprof <source.trc> <file.prc> sys=no sort=exeela,fchela,prsela explain= /
    error is --
    Error in create table of EXPLAIN PLAN table : unix_session_user.prof$paln_table
    ORA-00604: error occurred at recursive SQL level 1
    ORA-20001: Step-6:DDL
    Event Security. You are not permitted to perform the requested structural
    changes to PROF (TABLE)
    Event triggered : CREATE
    ora_login_user
    (session_user) : unix_session_user(dummy)
    Search : select count(*) from
    tabl(dummy table name) where obj_name like '%\%%' escape '\' and obj_type =
    'TABLE' and obj_type = 'USER' and ( event_CREATE = 'Y' or status =
    'Override')
    ORA-06512: at line 162
    ORA-06510: PL/SQL: unhandled
    user-defined exception
    EXPLAIN PLAN option disabled.
    i searched for the error and in oracle forum i found a solution .. http://forums.oracle.com/forums/thread.jspa?threadID=844287&tstart=0
    but after giving the table option it is giving the same error
    tkprof <source.trc> <file.prc> sys=no sort=exeela,fchela,prsela table=old_schema.plan_table explain= /
    it again gave the same error.
    In both two cases it gives elapsed time results,library cache missing etc but before giving this it throws ORA-00604 error as stated above
    then i again correct the tkprof statement ..
    tkprof <source.trc> <file.prc> sys=no sort=exeela,fchela,prsela table=new_schema.plan_table explain= /
    say this schema name here i used is dummy schema name.
    My question is did this error came as we had not sufficient previlages in the old_schema but that previleges we have in new_schema?
    My databse version is 9.2.0.4.0
    Thanks in advance
    Edited by: bp on Feb 3, 2009 11:36 PM
    Edited by: bp on Feb 3, 2009 11:40 PM

    Please post here full error message, there should be lines with ORA-00604 and then some other ORA as well.
    And are there any trace files generated during this error?
    And as You can see from error description, probably You will have to contact with Oracle support in order to solve this case:
    oerr ora 00604
    00604, 00000, "error occurred at recursive SQL level %s"
    // *Cause:  An error occurred while processing a recursive SQL statement
    // (a statement applying to internal dictionary tables).
    // *Action: If the situation described in the next error on the stack
    // can be corrected, do so; otherwise contact Oracle Support.

  • Getting error while taking MAX DB trans log backup.

    Hi,
    I am getting error while taking trans log backup of Maxdb database for archived log through data protector as below,
    [Critical] From: OB2BAR_SAPDBBAR@ttcmaxdb "MAX" Time: 08/19/10 02:10:41
    Unable to back up archive logs: no autolog medium found in media list
    But i am able to take complete data and incremental backup through data protector.
    I have already enabled the autolog for MAX DB database and it is writing that log file directly to HP-UX file system. Now i want to take backup of this archived log backup through data protector i.e. through trans log backup. So that the archived log which is on the file system after trans log backup completed will delete the archived logs in filesystem.  So that i don;t have to manually delete the archived logs from file system.
    Thanks,
    Subba

    Hi Lars,
    Thanks for the reply...
    Now i am able to take archive log backup but the problem is i can take only one archive file backup. Not multiple arhive log files generated by autolog at filesystem i.e /sapdb/MAX/saparch.
    I have enabled autolog and it is putting auto log file at unix directory i.e. /sapdb/MAX/saparch
    And then i am using the DataProtector 6.11 with trans log backup to backup the archived files in /sapdb/MAX/saparch. When i start the trans backup session through data protector it uses the archive stage command as "archive_stage BACKDP-Archive LOGBackup NOVERIFY REMOVE" If /sapdb/MAX/saparch has only one archive file it will backup and remove the file successfully. But if /sapdb/MAX/saparch has multiple archive files it gives an error as below,
      Preparing backup.
                Setting environment variable 'BI_CALLER' to value 'DBMSRV'.
                Setting environment variable 'BI_REQUEST' to value 'OLD'.
                Setting environment variable 'BI_BACKUP' to value 'ARCHIVE'.
                Constructed Backint for MaxDB call '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.
    bsi_in -c'.
                Created temporary file '/var/opt/omni/tmp/MAX.bsi_out' as output for Backint for MaxDB.
                Created temporary file '/var/opt/omni/tmp/MAX.bsi_err' as error output for Backint for MaxDB.
                Writing '/sapdb/data/wrk/MAX/dbm.ebf' to the input file.
                Writing '/sapdb/data/wrk/MAX/dbm.knl' to the input file.
            Prepare passed successfully.
            Starting Backint for MaxDB.
                Starting Backint for MaxDB process '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.
    bsi_in -c >>/var/opt/omni/tmp/MAX.bsi_out 2>>/var/opt/omni/tmp/MAX.bsi_err'.
                Process was started successfully.
            Backint for MaxDB has been started successfully.
            Waiting for the end of Backint for MaxDB.
                2010-09-06 03:15:21 The backup tool is running.
                2010-09-06 03:15:24 The backup tool process has finished work with return code 0.
            Ended the waiting.
            Checking output of Backint for MaxDB.
            Have found all BID's as expected.
        Have saved the Backup History files successfully.
        Cleaning up.
            Removing data transfer pipes.
                Removing data transfer pipe /var/opt/omni/tmp/MAX.BACKDP-Archive.1 ... Done.
            Removed data transfer pipes successfully.
            Copying output of Backint for MaxDB to this file.
    Begin of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
                #SAVED 1009067:1 /sapdb/data/wrk/MAX/dbm.ebf
                #SAVED 1009067:1 /sapdb/data/wrk/MAX/dbm.knl
    End of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
            Removed Backint for MaxDB's temporary output file '/var/opt/omni/tmp/MAX.bsi_out'.
            Copying error output of Backint for MaxDB to this file.
    Begin of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
    End of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
            Removed Backint for MaxDB's temporary error output file '/var/opt/omni/tmp/MAX.bsi_err'.
            Removed the Backint for MaxDB input file '/var/opt/omni/tmp/MAX.bsi_in'.
        Have finished clean up successfully.
    The backup of stage file '/export/sapdb/arch/MAX_LOG.040' was successful.
    2010-09-06 03:15:24
    Backing up stage file '/export/sapdb/arch/MAX_LOG.041'.
        Creating pipes for data transfer.
            Creating pipe '/var/opt/omni/tmp/MAX.BACKDP-Archive.1' ... Done.
        All data transfer pipes have been created.
        Preparing backup tool.
            Setting environment variable 'BI_CALLER' to value 'DBMSRV'.
            Setting environment variable 'BI_REQUEST' to value 'OLD'.
            Setting environment variable 'BI_BACKUP' to value 'ARCHIVE'.
            Constructed Backint for MaxDB call '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.bsi_
    in -c'.
            Created temporary file '/var/opt/omni/tmp/MAX.bsi_out' as output for Backint for MaxDB.
            Created temporary file '/var/opt/omni/tmp/MAX.bsi_err' as error output for Backint for MaxDB.
            Writing '/var/opt/omni/tmp/MAX.BACKDP-Archive.1 #PIPE' to the input file.
        Prepare passed successfully.
        Constructed pipe2file call 'pipe2file -d file2pipe -f /export/sapdb/arch/MAX_LOG.041 -p /var/opt/omni/tmp/MAX.BACKDP-Archive.1 -nowait'.
        Starting pipe2file for stage file '/export/sapdb/arch/MAX_LOG.041'.
            Starting pipe2file process 'pipe2file -d file2pipe -f /export/sapdb/arch/MAX_LOG.041 -p /var/opt/omni/tmp/MAX.BACKDP-Archive.1 -nowait >>/var/tmp/tem
    p1283767880-0 2>>/var/tmp/temp1283767880-1'.
            Process was started successfully.
        Pipe2file has been started successfully.
        Starting Backint for MaxDB.
            Starting Backint for MaxDB process '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.bsi_
    in -c >>/var/opt/omni/tmp/MAX.bsi_out 2>>/var/opt/omni/tmp/MAX.bsi_err'.
            Process was started successfully.
        Backint for MaxDB has been started successfully.
        Waiting for end of the backup operation.
            2010-09-06 03:15:25 The backup tool process has finished work with return code 2.
            2010-09-06 03:15:25 The backup tool is not running.
            2010-09-06 03:15:25 Pipe2file is running.
            2010-09-06 03:15:25 Pipe2file is running.
            2010-09-06 03:15:30 Pipe2file is running.
            2010-09-06 03:15:40 Pipe2file is running.
            2010-09-06 03:15:55 Pipe2file is running.
            2010-09-06 03:16:15 Pipe2file is running.
            Killing not reacting pipe2file process.
            Pipe2file killed successfully.
            2010-09-06 03:16:26 The pipe2file process has finished work with return code -1.
        The backup operation has ended.
        Filling reply buffer.
            Have encountered error -24920:
                The backup tool failed with 2 as sum of exit codes and pipe2file was killed.
            Constructed the following reply:
                ERR
                -24920,ERR_BACKUPOP: backup operation was unsuccessful
                The backup tool failed with 2 as sum of exit codes and pipe2file was killed.
        Reply buffer filled.
        Cleaning up.
            Removing data transfer pipes.
                Removing data transfer pipe /var/opt/omni/tmp/MAX.BACKDP-Archive.1 ... Done.
            Removed data transfer pipes successfully.
            Copying output of Backint for MaxDB to this file.
    Begin of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
    End of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
            Removed Backint for MaxDB's temporary output file '/var/opt/omni/tmp/MAX.bsi_out'.
            Copying error output of Backint for MaxDB to this file.
    Begin of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
    End of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
            Removed Backint for MaxDB's temporary error output file '/var/opt/omni/tmp/MAX.bsi_err'.
            Removed the Backint for MaxDB input file '/var/opt/omni/tmp/MAX.bsi_in'.
            Copying pipe2file output to this file.
    Begin of pipe2file output (/var/tmp/temp1283767880-0)----
    End of pipe2file output (/var/tmp/temp1283767880-0)----
            Removed pipe2file output '/var/tmp/temp1283767880-0'.
            Copying pipe2file error output to this file.
    Begin of pipe2file error output (/var/tmp/temp1283767880-1)----
    End of pipe2file error output (/var/tmp/temp1283767880-1)----
            Removed pipe2file error output '/var/tmp/temp1283767880-1'.
        Have finished clean up successfully.
    The backup of stage file '/export/sapdb/arch/MAX_LOG.041' was unsuccessful.
    2010-09-06 03:16:26
    Cleaning up.
        Have encountered error -24919:
            Can not remove file '/var/tmp/temp1283767880-0'.
            (System error 2; No such file or directory)
        Could not remove temporary output file of pipe2file ('/var/tmp/temp1283767880-0' ).
        Have encountered error -24919:
            Can not remove file '/var/tmp/temp1283767880-1'.
            (System error 2; No such file or directory)
        Could not remove temporary output file of pipe2file ('/var/tmp/temp1283767880-1' ).
    Have finished clean up successfully.
    Thanks,
    Subba

  • Error while taking archive log backup

    Dear all,
    We are getting the below mentioned error while taking the archive log backup
    ============================================================================
    BR0208I Volume with name RRPA02 required in device /dev/rmt0.1
    BR0210I Please mount BRARCHIVE volume, if you have not already done so
    BR0280I BRARCHIVE time stamp: 2010-05-27 16.43.41
    BR0256I Enter 'c[ont]' to continue, 's[top]' to cancel BRARCHIVE:
    c
    BR0280I BRARCHIVE time stamp: 2010-05-27 16.43.46
    BR0257I Your reply: 'c'
    BR0259I Program execution will be continued...
    BR0280I BRARCHIVE time stamp: 2010-05-27 16.43.46
    BR0226I Rewinding tape volume in device /dev/rmt0 ...
    BR0351I Restoring /oracle/RRP/sapreorg/.tape.hdr0
    BR0355I from /dev/rmt0.1 ...
    BR0278W Command output of 'LANG=C cd /oracle/RRP/sapreorg && LANG=C cpio -iuvB .tape.hdr0 < /dev/rmt0.1':
    Can't read input
    ===========================================================================
    We are able to take offline, online backups but we are facing the above mentioned problem while taking archive log backup
    We are on ECC 6 / Oracle / AIX
    The kernel is latest
    The drive is working fine and there is no problem with the tapes as we have tried using diffrent tapes
    can this be a permissions issue?
    I ran saproot.sh but somehow it is setting owner as sidadm and group as sapsys to some of the br* files
    I tried by changing the permissions to oraSID : dba but still the error is the same
    Any suggestions?

    Means you have not initialized the medias but trying to take backups.
    First check how many medias you have entered in your tape count parameter for archive log backups (just go to initSID.sap and check)
    Then increase/reduce them to according to your archive backup plan >> Initialize all the tapes according to their name (same as you have initialized in initSID.sap) >> stick physical label to all the medias according to name >> Schedule archive backups
    It will not ask you for initialization as already you have initialized in second step.
    Suggestion: Use 7 medias per week (one tape per day)
    Regards,
    Nick Loy

  • Error while creating the DWH tables using DAC

    Hi,
    I am getting error while creating the DWH tables using DAC. I have created a ODBC DSN using merant driver with DAC repository DB credentials and the test connection is successful. And while creating the tables i gave the olap dw credentials and the DSN name which i created earlier. But it throws the error as below:
    Please find the below mentioned error message
    =====================================
    STD OUTPUT
    =====================================
    CREATING SIEBEL DATABASE OBJECTS
    F:\DAC\bifoundation\dac\UTILITIES\BIN\DDLIMP /I N /s N /u infdomain /p ******* /c DB_DAC /G "SSE_ROLE" /f F:\DAC\bifoundation\dac/conf/sqlgen/ctl-file/oracle_bi_dw.ctl /b "" /K "" /X "" /W N
    Error while importing Siebel database schema.
    =====================================
    ERROR OUTPUT
    =====================================
    Siebel Enterprise Applications ODBC DDL Import Utility, Version 7.7 [18030] ENU
    Copyright (c) 2001 Siebel Systems, Inc. All rights reserved.
    This software is the property of Siebel Systems, Inc., 2207 Bridgepointe Parkway,
    San Mateo, CA 94404.
    User agrees that any use of this software is governed by: (1) the applicable
    user limitations and other terms and conditions of the license agreement which
    has been entered into with Siebel Systems or its authorized distributors; and
    (2) the proprietary and restricted rights notices included in this software.
    WARNING: THIS COMPUTER PROGRAM IS PROTECTED BY U.S. AND INTERNATIONAL LAW.
    UNAUTHORIZED REPRODUCTION, DISTRIBUTION OR USE OF THIS PROGRAM, OR ANY PORTION
    OF IT, MAY RESULT IN SEVERE CIVIL AND CRIMINAL PENALTIES, AND WILL BE
    PROSECUTED TO THE MAXIMUM EXTENT POSSIBLE UNDER THE LAW.
    If you have received this software in error, please notify Siebel Systems
    immediately at (650) 295-5000.
    F:\DAC\bifoundation\dac\UTILITIES\BIN\DDLIMP /I N /s N /u infdomain /p ***** /c DB_DAC /G SSE_ROLE /f F:\DAC\bifoundation\dac/conf/sqlgen/ctl-file/oracle_bi_dw.ctl /b /K /X /W N
    Connecting to the database...
    28000: [DataDirect][ODBC Oracle driver][Oracle]ORA-01017: invalid username/password; logon denied
    Unable to connect to the database...
    any help is appreciated.
    Thanks,
    RM

    The fact that you are getting an "ORA-01017: invalid username/password; logon denied" message indicates that you are at least talking to the database.
    The log shows that username "infdomain" is being used. Can you double check the username and password you have in DAC in a SQL*Plus/SQL Developer session?
    Please mark if useful/helpful,
    Andy.

  • RequestTimeoutException error while invoking a BPEL process using RMI

    Hi,
    I am getting RequestTimeoutException error while invoking a BPEL process using this code:
    Locator locator = LocatorFactory.createLocator(jndiProps);
    String compositeDN = "default/"+processName+"!1.0";
    Composite composite = locator.lookupComposite(compositeDN);
    String serviceName = "client";
    Service deliveryService = composite.getService(serviceName);
    NormalizedMessage nm = new NormalizedMessageImpl();
    nm.getPayload().put("payload", requestXml);
    NormalizedMessage res = deliveryService.request("process", nm);
    responseMap = res.getPayload();
    The error stack trace is
    weblogic.rmi.extensions.RequestTimeoutException: RJVM response from 'weblogic.rjvm.RJVMImpl@604f2d14 - id: '-361032376059206
    2776S:10.67.232.164:[8001,-1,-1,-1,-1,-1,-1]:emaar_domain:soa_server1' connect time: 'Mon Jan 18 11:34:41 GST 2010'' for 'executeServiceMethod
    (Loracle.soa.management.CompositeDN;Ljava.lang.String;Ljava.lang.String;[Ljava.lang.Object;) 'timed out after: 60000ms.
    oracle.fabric.common.FabricInvocationException: weblogic.rmi.extensions.RequestTimeoutException: RJVM response from 'weblogic.rjvm.RJVMImpl@60
    4f2d14 - id: '-3610323760592062776S:10.67.232.164:[8001,-1,-1,-1,-1,-1,-1]:emaar_domain:soa_server1' connect time: 'Mon Jan 18 11:34:41 GST 20
    10'' for 'executeServiceMethod(Loracle.soa.management.CompositeDN;Ljava.lang.String;Ljava.lang.String;[Ljava.lang.Object;) 'timed out after: 6
    0000ms.
            at oracle.soa.management.internal.facade.ServiceImpl.request(ServiceImpl.java:135)
            at com.gss.common.bo.BpelUtil.invokeBPELProcess(BpelUtil.java:81)
    To add to it the BPEL process is executing successfuly and RMI call timeout is happening.
    Can I know how to increase the related timeout value?                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

    Have got the same problem. Scenario at my end is little different though.
    I am trying to invoke a BPEL process from an ESB Service.
    I am trying to look into it..
    However, would be grateful, if someone can give some insight into this since many are running into this issue without being able to fix.
    Ashish.

  • XML Parser Error while creating Web service Client using JAX RPC

    hello evryone,
    Im facing XML Parser Error while creating web service client using JAX RPC. Im using Net Beans IDE for development purpose. I have wrote configuration file for client. Now i want to create Client stub. However i dont know how to do this in Net Beans. So i tried to do it from Command promt using command :
    wscompile -gen:client -d build -classpath build config-wsdl.xml
    here im getting Error:
    error parsing configuration file: XML parsing error: com.sun.xml.rpc.sp.ParseException:10: XML declaration may only begin entities
    Please help me out.
    Many thanks in advance,
    Kacee

    Can i use the client generated using jdeveloper 11g to import into the oracle forms 10g, i.e., form builder 10g. Currently this is the version we have in our office.

  • Error While Deploying the BPEL Process using obant script

    Hi All,
    I am getting the following error while deploying the BPEL Process using obant script. we are using the BPEL Version 10.1.2.0.2.Any information in this regard will be really helpful.
    Buildfile: build.xml
    main:
    [bpelc] file:/home5102/dibyap/saravana/Test/CreditRatingService.wsdl
    [bpelc] validating "/home5102/dibyap/saravana/Test/CreditRatingService.bpel" ...
    BUILD FAILED
    /home5102/dibyap/saravana/Test/build.xml:15: ORABPEL-01002
    Domain directory not found.
    The process cannot be deployed to domain "default" because the domain directory "/opt02/app/ESIT/oracle/esit10gR2iAS/BPEL10gR2/iAS/integration/orabpel/domains/default/deploy" cannot be found or cannot b
    e written to.
    Please check your -deploy option value; "default" must refer to a domain that has been installed locally on your machine.
    Total time: 23 seconds
    dibyap@ios5102_ESIBT:/home5102/dibyap/saravana/Test>
    Thanks,
    Saravana

    In 10.1.2.0.2 you need to create your own build.xml
    I have found an example, it may be of some help. This does call a property file
    cheers
    James
    <?xml version="1.0" ?>
    <!-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Run cxant on this file to build, package and deploy the
    ASB_EFT BPEL process
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -->
    <project name="ASB_EFT" default="main" basedir=".">
    <!-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Name of the domain the generated BPEL suitcase will be deployed to
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -->
    <property name="deploy" value="default" />
    <!-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    What version number should be used to tag the generated BPEL archive?
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -->
    <property name="rev" value="1.0" />
    <!-- BPEL Best Practices Properties -->
    <!-- Defaults Properties for TARGET environments
    # CHANGE THIS FILE TO REFLECT THE TARGET ENVIRONEMNT
    # either dev, test, or prod.properties
    -->
    <property file="ebusd.properties"/>
    <property name="env" value="${env.name}"/>
    <property name="current.project.name" value="${project.name}"/>
    <property name="target.project.name" value="${project.name}_${env}"/>
    <property name="deployment.profile" value ="${env}.properties"/>
    <property name="source.development.directory" location="${basedir}"/>
    <property name="target.env.directory" location="${basedir}/deploy/${project.name}_${env}"/>
    <property file="${deployment.profile}"/>
    <property name="build.fileencoding" value="UTF-8"/>
    <!-- Prints Environment
    -->
    <target name="print.env" description="Display environment settings">
    <echo message="Base Directory: ${basedir}"/>
    <echo message="Deployment Profile: ${deployment.profile}"/>
    <echo message="target.env.directory: ${target.env.directory}"/>
    <echo message="Deploy to Domain: ${deployToDomain}"/>
    <echo/>
    <echo message="os.name: ${os.name}"/>
    <echo message="os.version: ${os.version}"/>
    <echo message="os.arch: ${os.arch}"/>
    <echo/>
    <echo message="java.home: ${java.home}"/>
    <echo message="java.vm.name: ${java.vm.name}"/>
    <echo message="java.vm.vendor: ${java.vm.vendor}"/>
    <echo message="java.vm.version: ${java.vm.version}"/>
    <echo message="java.class.path: ${java.class.path}"/>
    <echo/>
    <echo message="env: ${env}"/>
    <echo message="current.project.name: ${current.project.name}"/>
    <echo message="target.project.name: ${target.project.name}"/>
    <echo message="server.name: ${server.name}"/>
    </target>
    <!--
    Copies the current directory structure along with
    all the file into the target.env.directory and
    change the name of the project
    -->
    <target name="create.environment">
    <copy todir="${target.env.directory}">
    <fileset dir="${basedir}"/>
    <filterset begintoken="@" endtoken="@">
    <filtersfile file="${deployment.profile}"/>
    </filterset>
    </copy>
    <move file="${target.env.directory}/${current.project.name}.jpr" tofile="${target.env.directory}/${target.project.name}.jpr"/>
    </target>
    <target name="main">
    <!-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    the bpelc task compiles and package BPEL processes into versioned BPEL
    archives (bpel_...jar). See the "Programming BPEL" guide for more
    information on the options of this task.
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -->
    <bpelc input="${basedir}/bpel.xml" rev="${rev}" deploy="${deploy}" />
    </target>
    </project>
    here is a property file
    project.name=ASB_EFT
    env.name=ebusd
    deployToDomain=default
    server.name=[server]
    server.port=7788
    ebusd\:7788=http://[server]:7788/
    IntegrationMailAccount=OracleBPELTest
    IntegrationMailAddress=[email]
    IntegrationMailPassword=[password]
    archivedir=[directory]
    inbounddir=/[directory]
    errordir=[directory]
    outbounddir=[directory]
    bpelpw=bpel
    dbhost1=[dbserver]
    dbhost2=[dbserver]
    dbport=1523
    dbservice=bpel
    dbconnstr=jdbc:oracle:thin:@(DESCRIPTION=(LOAD_BALANCE=YES)(FAILOVER=YES)(ADDRESS_LIST=(ADDRESS=(PROTOCOL=tcp)(HOST=[server])(PORT=1523))(ADDRESS=(PROTOCOL=tcp)(HOST=[server])(PORT=1523)))(CONNECT_DATA=(SERVICE_NAME=ebusd)))

  • I am getting error while connecting to sqlplus by using apps account in 11i

    i am getting error while connecting to sqlplus by using apps account in 11i. Does anyone has similar error.
    [applmgr@appsc05t01 admin]$ sqlplus
    SQL*Plus: Release 8.0.6.0.0 - Production on Mon Sep 13 14:34:52 2010
    (c) Copyright 1999 Oracle Corporation. All rights reserved.
    Enter user-name: apps
    ERROR:
    ORA-12514: Message 12514 not found; product=RDBMS; facility=ORA
    Enter user-name:
    Thanks
    Rao

    yes.
    login as: gorantla
    [email protected]'s password:
    Last login: Mon Sep 13 14:39:23 2010 from glng3wy9g1.actuant.pri
    [gorantla@appsc05t01 ~]$ sudo su - applmgr
    [applmgr@appsc05t01 ~]$ ora_setup
    Available databases:
    x = ATUXX
    y = ATUYY
    z = ATUZZ
    Select database: \c
    y
    ORACLE_SID=ATUYY
    IAS_ORACLE_HOME=/apps/ATUYY/atuyyora/iAS
    ORACLE_HOME=/apps/ATUYY/atuyyora/8.0.6
    APPL_TOP=/apps/ATUYY/atuyyappl
    [applmgr@appsc05t01 ~]$
    [applmgr@appsc05t01 ~]$
    [applmgr@appsc05t01 ~]$
    [applmgr@appsc05t01 ~]$ tnsping ATUYY
    TNS Ping Utility for Linux: Version 8.0.6.3.0 - Production on 13-SEP-2010 15:42:32
    (c) Copyright 1997 Oracle Corporation. All rights reserved.
    Attempting to contact (ADDRESS=(PROTOCOL=tcp)(HOST=dbrac05t01-vip.actuant.pri)(PORT=1611))
    OK (0 msec)
    [applmgr@appsc05t01 ~]$ sqlplus
    SQL*Plus: Release 8.0.6.0.0 - Production on Mon Sep 13 15:42:40 2010
    (c) Copyright 1999 Oracle Corporation. All rights reserved.
    Enter user-name: apps
    ERROR:
    ORA-12514: Message 12514 not found; product=RDBMS; facility=ORA
    Enter user-name:
    Thanks
    Rao

  • I am getting below error while taking RMAN backup.

    Friends,
    I am getting below error while taking RMAN backup.
    The issue is related with a lock while trying to backup the controlfile.
    The lock is more likely caused by the autobackup controlfile process started just before RMAN process was also trying the backup controlfile as shown in the following alert.log excerpt.
    Starting Control File and SPFILE Autobackup at 2009-04-07 09:24:58
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of Control File and SPFILE Autobackup command on ORA_DISK_1 channel at 04/07/2009 09:25:03
    ORA-00230: operation disallowed: snapshot control file enqueue unavailable
    Is there anyway to solve this issue?Thanks in advance.

    Hi,
    SQL> SELECT s.sid, username AS "User", program, module, action, logon_time "Logon", l.*
    FROM v$session s, v$enqueue_lock l
    WHERE l.sid = s.sid and l.type = 'CF' AND l.id1 = 0 and l.id2 = 2; 2 3
    no rows selected
    Oracle version : 10.2.0.3
    I am getting this error once in a month not everyday.
    Starting Control File and SPFILE Autobackup at 2009-04-07 09:24:58
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of Control File and SPFILE Autobackup command on ORA_DISK_1 channel at 04/07/2009 09:25:03
    ORA-00230: operation disallowed: snapshot control file enqueue unavailable
    Please suggest me based on this..Thanks in advance..

  • Getting error while taking preview of Web Dynpro iView

    Hi,
    I am getting this error while taking a preview of Web Dynpro iView :
    500   Internal Server Error
          SAP J2EE Engine/6.40
    Application error occurred during request processing.
    Details:   com.sap.tc.webdynpro.services.exceptions.WDRuntimeException: Wrong WebDynpro-URL: no application name specified
    Exception id: [0011D83FB2E7004A00000CB800001E0400045F7E8F181DF7]
    This error comes if the Web Dynpro iView don't have the Web Dynpro application name specified.
    Regards
    Neha Singh

    Hi,
    After selecting the Webdynpro java iView enter the following parameter correctly
    1. System -  Sap-LocalSystem
    2.Name Space  - local/<your webdynpro project name>
    3.Application name - <your webdynpro application name>
    4. choose plain HTML option
    Regards,
    P.Manivannan

  • Error while applying LM-tools patch using JSPM

    hi all,
    I got error while applying LM-Tool patch using JSPM. following is the error log. How can i rectify it. please help.
    Mar 29, 2010 12:49:57... Info:
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: DROP TABLE "BC_MID_TREE"
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: DROP TABLE "BC_MID_TREE"
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE TABLE "BC_MID_TREE"("NODE_ID" RAW(16)  NOT NULL, "PARENT_I
    D" RAW(16)  , "NODE_TYPE" NUMBER(10) DEFAULT 0 NOT NULL, "NAME" NVARCHAR2(30) DE
    FAULT ' ' NOT NULL, "TS" TIMESTAMP  NOT NULL, "USER_ID" NVARCHAR2(30)  )
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE TABLE "BC_MID_TREE"("NODE_ID" RAW(16)  NOT NULL, "PARENT_I
    D" RAW(16)  , "NODE_TYPE" NUMBER(10) DEFAULT 0 NOT NULL, "NAME" NVARCHAR2(30) DE
    FAULT ' ' NOT NULL, "TS" TIMESTAMP  NOT NULL, "USER_ID" NVARCHAR2(30)  )
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE INDEX "BC_MID_TREE_R" ON "BC_MID_TREE"("PARENT_ID")
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE INDEX "BC_MID_TREE_R" ON "BC_MID_TREE"("PARENT_ID")
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE UNIQUE INDEX "BC_MID_TREE_U" ON "BC_MID_TREE"("PARENT_ID",
    "NAME")
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE UNIQUE INDEX "BC_MID_TREE_U" ON "BC_MID_TREE"("PARENT_ID",
    "NAME")
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: ALTER TABLE "BC_MID_TREE" ADD PRIMARY KEY ("NODE_ID") NOT DEFERRA
    BLE INITIALLY IMMEDIATE ENABLE VALIDATE
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Statement successf
    ully executed: ALTER TABLE "BC_MID_TREE" ADD PRIMARY KEY ("NODE_ID") NOT DEFERRA
    BLE INITIALLY IMMEDIATE ENABLE VALIDATE
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Next archive objec
    t : dbtables/BC_MID_TREE_ITEM.gdbtable
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Next archive objec
    t : dbtables/BC_MID_TREE_ITEM.gdbtable
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  <<< Analyze table
    BC_MID_TREE_ITEM >>>
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  <<< Analyze table
    BC_MID_TREE_ITEM >>>
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  predefined action
    is: >>>DROP_CREATE<<<
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  predefined action
    is: >>>DROP_CREATE<<<
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Action: DROP_CREAT
    E
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:  Action: DROP_CREAT
    E
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:    DROP TABLE "BC_M
    ID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info:  CREATE TABLE "BC_MID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info: (
    Mar 29, 2010 12:49:57... Info: "NODE_ID" RAW(16)  NOT NULL,
    Mar 29, 2010 12:49:57... Info: "ITEM_ID" RAW(16)
    Mar 29, 2010 12:49:57... Info: )
    Mar 29, 2010 12:49:57... Info:  CREATE INDEX "BC_MID_TREE_ITEM_R" ON "BC_MID_TRE
    E_ITEM"
    Mar 29, 2010 12:49:57... Info: (
    Mar 29, 2010 12:49:57... Info: "ITEM_ID"
    Mar 29, 2010 12:49:57... Info: )
    Mar 29, 2010 12:49:57... Info:  ALTER TABLE "BC_MID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info: ADD PRIMARY KEY ("NODE_ID")
    Mar 29, 2010 12:49:57... Info: NOT DEFERRABLE INITIALLY IMMEDIATE ENABLE VALIDAT
    E
    Mar 29, 2010 12:49:57... Info:
    Mar 29, 2010 12:49:57... Info: 12:49:56 2010-03-29 dbs-Info:    DROP TABLE "BC_M
    ID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info:  CREATE TABLE "BC_MID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info: (
    Mar 29, 2010 12:49:57... Info: "NODE_ID" RAW(16)  NOT NULL,
    Mar 29, 2010 12:49:57... Info: "ITEM_ID" RAW(16)
    Mar 29, 2010 12:49:57... Info: )
    Mar 29, 2010 12:49:57... Info:  CREATE INDEX "BC_MID_TREE_ITEM_R" ON "BC_MID_TRE
    E_ITEM"
    Mar 29, 2010 12:49:57... Info: (
    Mar 29, 2010 12:49:57... Info: "ITEM_ID"
    Mar 29, 2010 12:49:57... Info: )
    Mar 29, 2010 12:49:57... Info:  ALTER TABLE "BC_MID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info: ADD PRIMARY KEY ("NODE_ID")
    Mar 29, 2010 12:49:57... Info: NOT DEFERRABLE INITIALLY IMMEDIATE ENABLE VALIDAT
    E
    Mar 29, 2010 12:49:57... Info:
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: DROP TABLE "BC_MID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: DROP TABLE "BC_MID_TREE_ITEM"
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE TABLE "BC_MID_TREE_ITEM"("NODE_ID" RAW(16)  NOT NULL, "ITE
    M_ID" RAW(16)  )
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE TABLE "BC_MID_TREE_ITEM"("NODE_ID" RAW(16)  NOT NULL, "ITE
    M_ID" RAW(16)  )
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE INDEX "BC_MID_TREE_ITEM_R" ON "BC_MID_TREE_ITEM"("ITEM_ID"
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: CREATE INDEX "BC_MID_TREE_ITEM_R" ON "BC_MID_TREE_ITEM"("ITEM_ID"
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: ALTER TABLE "BC_MID_TREE_ITEM" ADD PRIMARY KEY ("NODE_ID") NOT DE
    FERRABLE INITIALLY IMMEDIATE ENABLE VALIDATE
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Statement successf
    ully executed: ALTER TABLE "BC_MID_TREE_ITEM" ADD PRIMARY KEY ("NODE_ID") NOT DE
    FERRABLE INITIALLY IMMEDIATE ENABLE VALIDATE
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE_ITEM successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE_ITEM successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE_ITEM successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Runtime XML of BC_
    MID_TREE_ITEM successfully written
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Analysis and adjus
    tment of tables successfully executed
    Mar 29, 2010 12:49:57... Info: 12:49:57 2010-03-29 dbs-Info:  Analysis and adjus
    tment of tables successfully executed
    Mar 29, 2010 12:49:57... Info: Finished successfully: development component 'tc/
    slm/mid/dd'/'sap.com'/'MAIN_APL70P18_C'/'2712380'/'0', grouped by
    Mar 29, 2010 12:49:57... Info: Starting to save the repository
    Mar 29, 2010 12:49:57... Info: Finished saving the repository
    Mar 29, 2010 12:49:57... Info: Starting: Update: Selected development component
    'tc/archtech/browser/viewer/locator'/'sap.com'/'MAIN_APL70P18_C'/'2712383'/'0' u
    pdates currently deployed development component 'tc/archtech/browser/viewer/loca
    tor'/'sap.com'/'MAIN_APL70VAL_C'/'1496536'/'0'.
    Mar 29, 2010 12:49:57... Info: SDA to be deployed: /usr/sap/IRX/DVEBMGS02/SDM/ro
    ot/origin/sap.com/tc/archtech/browser/viewer/locator/MAIN_APL70P18_C/0/2712383/t
    carchtechbrowserviewerlocator.sda
    Mar 29, 2010 12:49:57... Info: Software type of SDA: J2EE
    Mar 29, 2010 12:49:57... Info: ***** Begin of SAP J2EE Engine Deployment (J2EE A
    pplication) *****
    Mar 29, 2010 12:49:58... Info: Starting cluster instance processes.
    Mar 29, 2010 12:50:03... Info: Starting the instance JC_PRDCIXI_IRX_02 running o
    n host PRDCIXI_svc
    Mar 29, 2010 12:50:03... Info: Starting the instance JC_PRDCIXI_IRX_02 processes
    . The instance is running on host PRDCIXI_svc
    Mar 29, 2010 12:50:03... Info: Starting the process dispatcher
    Mar 29, 2010 12:50:03... Info: Starting the process server0
    Mar 29, 2010 6:49:58 PM  Info: ***** End of SAP J2EE Engine Deployment (J2EE App
    lication) *****
    Mar 29, 2010 6:49:58 PM  Error: Aborted: development component 'tc/archtech/brow
    ser/viewer/locator'/'sap.com'/'MAIN_APL70P18_C'/'2712383'/'0', grouped by softwa
    re component 'LM-TOOLS'/'sap.com'/'MAIN_APL70P19_C'/'1000.7.00.19.0.200904271432
    29''/'0':
    SDM could not start the J2EE cluster on the host PRDCIXI! The online deployment
    is terminated.
    A timeout occured during the cluster running verification.
    (message ID: com.sap.sdm.serverext.servertype.inqmy.extern.EngineApplOnlineDepl
    oyerImpl.performAction(DeploymentActionTypes).STARTUP_CLUSTER)
    Mar 29, 2010 6:49:58 PM  Info: Starting to save the repository
    Mar 29, 2010 6:49:59 PM  Info: Finished saving the repository
    Mar 29, 2010 6:49:59 PM  Info: Starting: Update: Selected software component 'LM
    -TOOLS'/'sap.com'/'MAIN_APL70P19_C'/'1000.7.00.19.0.20090427143229''/'0' updates
    currently deployed software component 'LM-TOOLS'/'sap.com'/'MAIN_APL70P14_C'/'1
    000.7.00.14.1.20080124101556''/'0'.
    Mar 29, 2010 6:49:59 PM  Error: Aborted: software component 'LM-TOOLS'/'sap.com'
    /'MAIN_APL70P19_C'/'1000.7.00.19.0.20090427143229''/'0':
    Failed deployment of SDAs:
    development component 'tc/archtech/browser/viewer/locator'/'sap.com'/'MAIN_APL70
    P18_C'/'2712383'/'0' : aborted
    Please, look at error logs above for more information!
    Mar 29, 2010 6:49:59 PM  Info: Starting to save the repository
    Mar 29, 2010 6:49:59 PM  Info: Finished saving the repository
    Mar 29, 2010 6:49:59 PM  Info: Restoring the state of the instance (JC_PRDCIXI_I
    RX_02) process server0 from Stopped to Running
    Mar 30, 2010 12:50:00... Error: Received exception when restoring Engine state:
    A timeout occured during the cluster state restore verification.
    Mar 30, 2010 12:50:00... Error: -
    At least one of the Deployments
    failed -
    Regards,
    Priya

    Hi Priya,
    As per your error message, there is problem with the SDM is not starting after finishing deployment on SAP J2EE Engine Deployment. Development component are not deployed properly due to error message
    > SDM could not start the J2EE cluster on the host PRDCIXI! The online deployment is terminated. A timeout occured during the cluster running verification. P
    Please check the default timeout and increased as requred.
    Please check the below thread to rectify this issue.
    SDM Troubleshooting  : http://help.sap.com/saphelp_nw04/helpdata/de/76/fb72ec091f4bf8a2d8ba321bb7e8d9/frameset.htm
    SDM Troubleshooting  : https://cw.sdn.sap.com/cw/docs/DOC-29627:
    Hope it helps
    Regards
    Arun

  • Error while load data into Essbase using ODI

    Hi ,
    I'm getting the following error while loading measures into Essbase using ODI, I used the same LOG nd Error file and file path for all my Dimensions , this worked well but not sure why this is not working for measures....need help.
    File "<string>", line 79, in ?
    com.hyperion.odi.common.ODIHAppException: c:/temp/Log1.log (No such file or directory)
    Thanks
    Venu

    Are you definitely running it against an agent where that path exists.
    Have you tried using a different location and filename, have you restarted the agent to make sure there is not a lock on the file.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Error while taking dump using datapump

    getting following error -
    Export: Release 10.2.0.1.0 - Production on Friday, 15 September, 2006 10:31:41
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Starting "XX"."SYS_EXPORT_SCHEMA_02": XX/********@XXX directory=dpdump dumpfile=XXX150906.dmp logfile=XXX150906.log
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    ORA-39125: Worker unexpected fatal error in KUPW$WORKER.GET_TABLE_DATA_OBJECTS while calling DBMS_METADATA.FETCH_XML_CLOB []
    ORA-31642: the following SQL statement fails:
    BEGIN "DMSYS"."DBMS_DM_MODEL_EXP".SCHEMA_CALLOUT(:1,0,0,'10.02.00.01.00'); END;
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.DBMS_METADATA", line 907
    ORA-06550: line 1, column 7:
    PLS-00201: identifier 'DMSYS.DBMS_DM_MODEL_EXP' must be declared
    ORA-06550: line 1, column 7:
    PL/SQL: Statement ignored
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPW$WORKER", line 6235
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    2A68E610 14916 package body SYS.KUPW$WORKER
    2A68E610 6300 package body SYS.KUPW$WORKER
    2A68E610 9120 package body SYS.KUPW$WORKER
    2A68E610 1880 package body SYS.KUPW$WORKER
    2A68E610 6861 package body SYS.KUPW$WORKER
    2A68E610 1262 package body SYS.KUPW$WORKER
    255541A8 2 anonymous block
    Job "XX"."SYS_EXPORT_SCHEMA_02" stopped due to fatal error at 10:33:12
    Action required is contact customer support. And on metalink found a link that states it a bug in 10g release 1 that was suppose to be fixed in 10g release 1 version 4.
    some of the default schemas were purposely dropped from the database. The only default schema available now are -
    DBSNMP, DIP, OUTLN, PUBLIC, SCOTT, SYS, SYSMAN, SYSTEM, TSMSYS.
    DIP, OUTLN, TSMSYS were created again.
    Could this be a cause of problem??
    Thanks in adv.

    Hi,
    Below is the DDL taken from different database. Will this be enough ? One more thing please, what shall be the password should it be DMSYS.....since this will not be used by me but system.
    CREATE USER "DMSYS" PROFILE "DEFAULT" IDENTIFIED BY "*******" PASSWORD EXPIRE DEFAULT TABLESPACE "SYSAUX" TEMPORARY TABLESPACE "TEMP" QUOTA 204800 K ON "SYSAUX" ACCOUNT LOCK
    GRANT ALTER SESSION TO "DMSYS"
    GRANT ALTER SYSTEM TO "DMSYS"
    GRANT CREATE JOB TO "DMSYS"
    GRANT CREATE LIBRARY TO "DMSYS"
    GRANT CREATE PROCEDURE TO "DMSYS"
    GRANT CREATE PUBLIC SYNONYM TO "DMSYS"
    GRANT CREATE SEQUENCE TO "DMSYS"
    GRANT CREATE SESSION TO "DMSYS"
    GRANT CREATE SYNONYM TO "DMSYS"
    GRANT CREATE TABLE TO "DMSYS"
    GRANT CREATE TRIGGER TO "DMSYS"
    GRANT CREATE TYPE TO "DMSYS"
    GRANT CREATE VIEW TO "DMSYS"
    GRANT DROP PUBLIC SYNONYM TO "DMSYS"
    GRANT QUERY REWRITE TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_JOBS_RUNNING" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_REGISTRY" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_SYS_PRIVS" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_TAB_PRIVS" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_TEMP_FILES" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_LOCK" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_REGISTRY" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_SYSTEM" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_SYS_ERROR" TO "DMSYS"
    GRANT DELETE ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT INSERT ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT SELECT ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT UPDATE ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT SELECT ON "SYS"."V_$PARAMETER" TO "DMSYS"
    GRANT SELECT ON "SYS"."V_$SESSION" TO "DMSYS"
    The other database has the DMSYS and the status is EXPIRED & LOCKED but I'm still able to take the dump using datapump??

Maybe you are looking for

  • Pedido de Venda - Guia Imposto não é atualizado o peso líquido

    Prezados, Ao entrar com um pedido de venda no SAP Business One e selecionar um item que possui no Cadastro do item: ficha Dados de venda  as seguintes configurações abaixo: Unidade de medida para vendas: UN Itens por unidade de venda: 1 Unidade de me

  • Automatic assignment responsible group partner function of Activities

    Hi, I need help on below requirement>> Automatic assignment responsible group partner function of Activities when status is set to u201CIncorrectly Assignedu201D. Thanks

  • Can we Restore a Dropped Package in oracle 10g

    Hello, I have dropped a pl/sql package accidantly.The database is Oracle 10g R2. Is there any way to recover it.? Thanx.

  • Im getting a folder with a "!" any success stories??

    I have recieved both an unhappy icon and a this folder, what does it mean as apple give no direct answer. Alsohow do you phone up because they say you have to pay?? How do you organise to send back??

  • URI syntax issue.

    Gday everyone, The following snippet throws a URISyntaxException:Illegal character in query at index... I can understand that it doesnt like he pipes in the url. Any ideas on how to overcome this syntactical problem besides removing the pipes? try {