Transportable Tablespace Importing - Failed to read stderr from process

Hello,
I'm using the normal EM console packaged with Oracle (10gR2) and am trying to import a tablespace on RedHat rel4.
I follow the Transport Tablespaces link on the maintenance tab and get the option of export or import and also the place to put the host credentials in. I select import, enter the credentials and click continue and am given this error: Failed to read stderr from process.
I've tried logging into EM as SYSTEM, SYS, and a DBA account I use and have used both root and the oracle account with no luck.
Searching hasn't gotten me any useful results here or in metalink so I'm stuck. What am I missing?
Thanks

Thank you thank you!
The short version is I traced it back to an error with /bin/nmo which traced back to a failure to run root.sh when upgrading to 10.2.0.3.
Thanks again!

Similar Messages

  • Failed to read data from report file Reason: The table could not be found.

    BO Enterprise XI R2, cannot publish crystal reports using the publishing wizard.
    Failed to read data from report file Reason: The table could not be found.
    Any ideas to get around this would really help out.
    Regards

    Connection used Views, ODBC System DSN is setup properly.
    Approach for import from business view manager and import wizard.  both methods failed to import the Business View and underlying reports.
    I figure I may have imported the Business View wrong? From Business View Manager I exported from my dev server then imported to prod server.
    Apparently I learned exporting my business view also includes the Data Connections that the Business Views are dependent upon.
    So which ever folder you specify it copies it there. Originally the all Data Connections Resides on the root folder. To return it to the original location. I deleted what I had exported. Exported this time to the root folder, then only deleted the business views, foundation, elements. Then exported again to the folder where I intended then only deleted the Data Connection.
    Makes any sense? So I then had to re point the business views and all the dependent objects to the data connection that resides in the root folder.
    I tested the connection, it works fine. I properly updated my crystal reports to the business view in production. Did a sample extract it works as expected.
    However when i try to publish, either from Crystal or Publish wizard i get the same error?
    As a work around i am thinking, after updating the business view in the crystal reports, shall i re map the fields?? or reexport the business views again?
    Any help will be surely appreciated.

  • Failed to read data from report file : Reason: Crystal Reports: Print Engin

    Hi,
    When we try to migrate the crystal reports from BO R2 to BO R3, some of the reports are faling with the error:
    Failed to create a new Report.Reason: Failed to read data from report file C:\DOCUME~1\xxxxx.rpt. Reason: Crystal Reports: Print Engine Error
    Would please someone help me in fixing the issue.
    Thanks and Regards,

    Few quick checks to identify the cause -
    1. Are you able to run report in your R2 system?
    2. In XI3.1, check SIA running on which account.
    3. Check the acount have sufficient rights on file system and registry.
    4. What's your web server?
    Give proper rights and import again. Hopefully it will resolve.
    Edited by: Das on Dec 6, 2010 9:47 PM

  • Failed to read PID from file /run/nginx.pid: Invalid argument

    Hi,
    tried to get an nginx server running to set up an owncloud environment.
    When starting the nginx server
      $  systemctl start nginx.service
    I get the message: "Failed to read PID from file /run/nginx.pid: Invalid argument"
    [root@klaus /etc/nginx]# systemctl status nginx
    nginx.service - A high performance web server and a reverse proxy server
    Loaded: loaded (/usr/lib/systemd/system/nginx.service; disabled)
    Active: active (running) since Sat 2013-10-12 17:50:46 CEST; 8min ago
    Process: 1823 ExecStart=/usr/bin/nginx -g pid /run/nginx.pid; daemon on; master_process on; (code=exited, status=0/SUCCESS)
    Process: 1821 ExecStartPre=/usr/bin/nginx -t -q -g pid /run/nginx.pid; daemon on; master_process on; (code=exited, status=0/SUCCESS)
    Main PID: 1825 (nginx)
    CGroup: /system.slice/nginx.service
    ├─1825 nginx: master process /usr/bin/nginx -g pid /run/nginx.pid; daemon on; master_process on;
    └─1826 nginx: worker process
    Oct 12 17:50:46 klaus systemd[1]: Failed to read PID from file /run/nginx.pid: Invalid argument
    Oct 12 17:50:46 klaus systemd[1]: Started A high performance web server and a reverse proxy server.
    but /run/nginx.pid is readable:
    # cat /run/nginx.pid
    2058
    # ll /run/nginx.pid
    -rw-r--r-- 1 root root 4 Oct 12 17:39 /run/nginx.pid
    It seems nginx is runnung thought but in my browser I only get a blank page.
    Any help appreciated.
    Last edited by wombalton (2013-10-13 14:17:46)

    Actually it's the first time I tried to set up one.
    I Basically followed this guide [1] adopting it to arch.
    That's where i got my nginx.conf from. Taking the standard nginx.conf that comes by install and replacing the server part.
    After some searching I found the owncloud manual[2] with an example nginx.conf. Comparing that with my one I found some differences. Applying them works out fine.
    I get now the owncloud page.
    The systemd error  still ocurs, but I think it does not matter to run the server.
    Thanks for your help.
    [1] https://docs.google.com/file/d/0B0ZsTQd … ring&pli=1
    [2] http://doc.owncloud.org/server/5.0/admi … figuration
    EDIT:
    Just in case someone stumbles over this and tries to use the config:
    This one [3] really works
    [3] http://doc.owncloud.org/server/5.0/admi … thers.html
    Last edited by wombalton (2013-10-14 13:31:22)

  • Failed to read data from report file, failed to read parameter object

    Error message when doing a u201CSchedule List of Valuesu201D
    failed to read data from report file,
    c:\docume~1\user........
    failed to read parameter object
    Earlier in the day it worked just fine nothing had been changed.
    I have admin right on the windows server and in central management console
    Can someone please help with this issue.

    Hi Glenn
    Please let us know the following:
    1.What is the exact version of Crystal Reports Designer?
    2.What is the version of Business objects Enterprise installed on the machine?
    3.Are reports based on Business views,Deski,Webi?
    4.Are the reports migrated from older version of Crystal Reports to new Version?
    Thanks
    Shraddha

  • DCA-40002: Failed to read WSDL from xsd

    Hi,
    I am a newbie to JDeveloper, we are using Oracle JDeveloper 11g to consume a web service from PeopleSoft 9.0.
    We have a WSDL based on a PeopleSoft Component Interface. We can view the wsdl document through the PeopleSoftServiceListeningConnector URL and it appears ok.
    However, when we create a new Web Service Data Control in JDeveloper, it presents us with the following error:
    DCA-40002: The WSDL document is invalid due to the following reason: java.io.IOException: Failed to read WSDL from
    http://machine:port/PSIGW/PeopleSoftServiceListeningConnector/M999999.V1.xsd:
    HTTP connection error code is 500.
    Any advice would be appreciated.
    Thanks in advance.
    Regards,
    Dan

    Hi, after 1 1/2 years you can assume the thread to be closed. So please open your own thread and make sure the WSDL file is accessible from a browser URL.
    Frank

  • Failed to read data from Lock tables

    Hi,
    When i trying to update data using T_CODE j1i8. I am getting message (Failed to read data from Lock tables ) but where as my database tables are not been locked. can anyone plz help me out.
    Regards,
    Anaveer

    Hi
    Probably some other transation is modifying the Database table you are using & it may have locke the table using FOR UPDATE stmt.
    Using Deque_all FM module you can remove the locks
    Thanks
    Sandeep
    Reward if helpful

  • Failed to read data from report file

    Hello ,
    When iam trying to publish crystal report to Bo enterprise using Publish wizard, i get below error
    *"Failed to read data from report file C:\Users\prakash\Desktop\new_obj.rpt. Reason: File I/O error. File ~cidbc64f360135e0.rpt."*
    can you please help me in resolving it
    Regards,
    KumarTP

    Please post this to the [Business Objects Enterprise Administration|BI Platform; forum.
    - Ludek

  • TT16060: Failed to read data from the network. select() timed out

    hi!
    i am working on active standby pair.....i created
    [activedsn]
    Driver=/d01/oracle/tt70/TimesTen/tt70/lib/libtten.so
    DataStore=/d01/oracle/tt70/TimesTen/tt70/info/activedsn
    DatabaseCharacterSet=WE8MSWIN1252
    PermSize=10
    [standbydsn]
    Driver=/d01/oracle/tt70/TimesTen/tt70/lib/libtten.so
    DataStore=/d01/oracle/tt70/TimesTen/tt70/info/standbydsn
    DatabaseCharacterSet=WE8MSWIN1252
    PermSize=10
    [sub3]
    Driver=/d01/oracle/tt70/TimesTen/tt70/lib/libtten.so
    DataStore=/d01/oracle/tt70/TimesTen/tt70/info/sub3
    DatabaseCharacterSet=WE8MSWIN1252
    PermSize=10
    replication schemes; .............all are on same hosts..........
    activedsn datastore.............
    Command> create table readtab(a number not null primary key,b varchar2(31));
    Command> insert into readtab values(101,'aaaa');
    1 row inserted.
    Command> commit;
    command>create active standby pair activedsn on "tap2.test3.com",standbydsn on "tap2.test3.com" return receipt subscriber sub3 on "tap2.test3.com";
    standbydsn datastore..........
    command>ttrepadmin -duplicate -from activedsn -host "tap2.test3.com" -uid adm -pwd adm "dsn=standbydsn";
    sub3 datastore..........
    command>ttrepadmin -duplicate -from standbydsn -host "tap2.test3.com" -uid adm -pwd adm "dsn=sub3";
    now they are working fine when i insert something from activedsn it is replicated to standbydsn and from standbydsn to sub3 ......
    problem:
    but when i test "Recovering from a failure of the standby master data store"
    i ttdestroy standbydsn..............and on activedsn i executed
    command>call ttrepstatesave('failed','standbydsn','tap2.test3.com');
    after that all updates from activedsn were replicated to sub3....in the meanwhile i again duplicated standbydsn from activedsn
    command>ttrepadmin -duplicate -from activedsn -host "tap2.test3.com" -uid adm -pwd adm "dsn=standbydsn";
    now what happens updates from activedsn are replicated to standbydsn but no updates are replicated to sub3 it is giving error
    15:24:29.29 Warn: REP: 7077: SUB3:receiver.c(1931): TT16060: Failed to read data from the network. select() timed out
    15:29:36.09 Warn: REP: 7008: STANDBYDSN:receiver.c(1931): TT16060: Failed to read data from the network. TimesTen replication agent is stopping
    replication agents for all are running.............
    please helppppp..........
    Edited by: Muhammad.Usman on Oct 13, 2009 2:57 AM
    Edited by: Muhammad.Usman on Oct 13, 2009 3:08 AM

    [timesten@tap2 bin]$ ttrepadmin -showconfig activedsn
    Self host "TAP2.TEST3.COM", port auto, name "ACTIVEDSN", LSN 0/916688, timeout 120, threshold 0
    List of subscribers
    Peer name Host name Port State Proto
    SUB3 TAP2.TEST3.COM Auto Start 24
    Last Msg Sent Last Msg Recv Latency TPS RecordsPS
    00:00:03 - -1.00 -1 -1
    Peer name Host name Port State Proto
    STANDBYDSN TAP2.TEST3.COM Auto Start 24
    Last Msg Sent Last Msg Recv Latency TPS RecordsPS
    00:00:03 00:00:04 -1.00 -1 -1
    List of objects and subscriptions
    Table details
    Table : ADM.READTAB Timestamp updates : -
    Master Name Subscriber name
    STANDBYDSN SUB3
    STANDBYDSN ACTIVEDSN
    Table details
    Table : ADM.READTAB Timestamp updates : -
    Master Name Subscriber name
    ACTIVEDSN SUB3
    ACTIVEDSN STANDBYDSN
    Datastore details
    Master Name Subscriber name
    STANDBYDSN SUB3
    STANDBYDSN ACTIVEDSN
    Datastore details
    Master Name Subscriber name
    ACTIVEDSN SUB3
    ACTIVEDSN STANDBYDSN
    [timesten@tap2 bin]$ ttrepadmin -showstatus activedsn
    Replication Agent Status as of: 2009-10-13 20:42:02
    DSN : activedsn
    Process ID : 19000 (Started)
    Replication Agent Policy : manual
    Host : TAP2.TEST3.COM
    RepListener Port : 58698 (AUTO)
    Last write LSN : 0.973840
    Last LSN forced to disk : 0.973840
    Replication hold LSN : 0.968456
    Replication Peers:
    Name : SUB3
    Host : TAP2.TEST3.COM
    Port : 58371 (AUTO) (Connected)
    Replication State : STARTED
    Communication Protocol : 24
    Name : STANDBYDSN
    Host : TAP2.TEST3.COM
    Port : 59000 (AUTO) (Connected)
    Replication State : STARTED
    Communication Protocol : 24
    TRANSMITTER thread(s):
    For : SUB3
    Start/Restart count : 6
    Send LSN : 0.971432
    Transactions sent : 2
    Total packets sent : 158
    Tick packets sent : 112
    MIN sent packet size : 48
    MAX sent packet size : 568
    AVG sent packet size : 59
    Last packet sent at : 20:42:00
    Total Packets received: 158
    MIN rcvd packet size : 48
    MAX rcvd packet size : 96
    AVG rcvd packet size : 64
    Last packet rcvd'd at : 20:42:00
    TRANSMITTER thread(s):
    For : STANDBYDSN
    Start/Restart count : 4
    Send LSN : 0.971432
    Transactions sent : 2
    Total packets sent : 106
    Tick packets sent : 84
    MIN sent packet size : 48
    MAX sent packet size : 560
    AVG sent packet size : 63
    Last packet sent at : 20:42:00
    Total Packets received: 104
    MIN rcvd packet size : 48
    MAX rcvd packet size : 96
    AVG rcvd packet size : 66
    Last packet rcvd'd at : 20:42:00
    Most recent errors (max 5):
    TT16122 in transmitter.c (line 3313) at 20:28:28 on 10-13-2009
    TT16121 in transmitter.c (line 3048) at 20:28:28 on 10-13-2009
    TT16060 in transmitter.c (line 5028) at 20:33:59 on 10-13-2009
    TT16122 in transmitter.c (line 3313) at 20:33:59 on 10-13-2009
    TT16121 in transmitter.c (line 3048) at 20:33:59 on 10-13-2009
    RECEIVER thread(s):
    For : STANDBYDSN
    Start/Restart count : 1
    Transactions received : 0
    Total packets sent : 33
    Tick packets sent : 0
    MIN sent packet size : 48
    MAX sent packet size : 68
    AVG sent packet size : 67
    Last packet sent at : 20:42:00
    Total Packets received: 33
    MIN rcvd packet size : 48
    MAX rcvd packet size : 135
    AVG rcvd packet size : 51
    Last packet rcvd'd at : 20:42:00
    [timesten@tap2 bin]$
    [timesten@tap2 bin]$ ttrepadmin -showstatus standbydsn
    Replication Agent Status as of: 2009-10-13 20:42:35
    DSN : standbydsn
    Process ID : 19102 (Started)
    Replication Agent Policy : manual
    Host : TAP2.TEST3.COM
    RepListener Port : 59000 (AUTO)
    Last write LSN : 0.1007904
    Last LSN forced to disk : 0.1007904
    Replication hold LSN : 0.1002472
    Replication Peers:
    Name : SUB3
    Host : TAP2.TEST3.COM
    Port : 58371 (AUTO) (Connected)
    Replication State : STARTED
    Communication Protocol : 24
    Name : ACTIVEDSN
    Host : TAP2.TEST3.COM
    Port : 58698 (AUTO) (Connected)
    Replication State : STARTED
    Communication Protocol : 24
    TRANSMITTER thread(s):
    For : SUB3
    Start/Restart count : 2
    Send LSN : 0.1005496
    Transactions sent : 1
    Total packets sent : 48
    Tick packets sent : 33
    MIN sent packet size : 48
    MAX sent packet size : 568
    AVG sent packet size : 65
    Last packet sent at : 20:42:30
    Total Packets received: 48
    MIN rcvd packet size : 48
    MAX rcvd packet size : 96
    AVG rcvd packet size : 64
    Last packet rcvd'd at : 20:42:30
    Most recent errors (max 5):
    TT16229 in transmitter.c (line 6244) at 20:38:01 on 10-13-2009
    TRANSMITTER thread(s):
    For : ACTIVEDSN
    Start/Restart count : 1
    Send LSN : 0.1005496
    Transactions sent : 0
    Total packets sent : 36
    Tick packets sent : 34
    MIN sent packet size : 48
    MAX sent packet size : 135
    AVG sent packet size : 50
    Last packet sent at : 20:42:30
    Total Packets received: 36
    MIN rcvd packet size : 48
    MAX rcvd packet size : 68
    AVG rcvd packet size : 67
    Last packet rcvd'd at : 20:42:30
    RECEIVER thread(s):
    For : ACTIVEDSN
    Start/Restart count : 1
    Transactions received : 1
    Total packets sent : 42
    Tick packets sent : 0
    MIN sent packet size : 48
    MAX sent packet size : 96
    AVG sent packet size : 66
    Last packet sent at : 20:42:30
    Total Packets received: 47
    MIN rcvd packet size : 48
    MAX rcvd packet size : 190
    AVG rcvd packet size : 58
    Last packet rcvd'd at : 20:42:30
    [timesten@tap2 bin]$
    [timesten@tap2 bin]$ ttrepadmin -showstatus sub3
    Replication Agent Status as of: 2009-10-13 20:43:05
    DSN : sub3
    Process ID : 18898 (Started)
    Replication Agent Policy : manual
    Host : TAP2.TEST3.COM
    RepListener Port : 58371 (AUTO)
    Last write LSN : 0.707088
    Last LSN forced to disk : 0.707088
    Replication hold LSN : -1.-1
    Replication Peers:
    Name : ACTIVEDSN
    Host : TAP2.TEST3.COM
    Port : 0 (AUTO)
    Replication State : STARTED
    Communication Protocol : 24
    Name : STANDBYDSN
    Host : TAP2.TEST3.COM
    Port : 0 (AUTO)
    Replication State : STARTED
    Communication Protocol : 24
    RECEIVER thread(s):
    For : ACTIVEDSN
    Start/Restart count : 1
    Transactions received : 0
    Total packets sent : 46
    Tick packets sent : 0
    MIN sent packet size : 48
    MAX sent packet size : 96
    AVG sent packet size : 65
    Last packet sent at : 20:43:00
    Total Packets received: 46
    MIN rcvd packet size : 48
    MAX rcvd packet size : 134
    AVG rcvd packet size : 51
    Last packet rcvd'd at : 20:43:00
    RECEIVER thread(s):
    For : STANDBYDSN
    Start/Restart count : 1
    Transactions received : 1
    Total packets sent : 45
    Tick packets sent : 0
    MIN sent packet size : 48
    MAX sent packet size : 96
    AVG sent packet size : 66
    Last packet sent at : 20:43:00
    Total Packets received: 50
    MIN rcvd packet size : 48
    MAX rcvd packet size : 190
    AVG rcvd packet size : 56
    Last packet rcvd'd at : 20:43:00
    [timesten@tap2 bin]$
    [timesten@tap2 bin]$
    Edited by: Muhammad.Usman on Oct 13, 2009 9:57 PM

  • Oracle 10gR2 - Any way to speed up Transportable Tablespace Import?

    I have a nightly process that uses expdp/impdp with Transportable Tablespace option to copy a full schema from one database to another. This process was initially done with just expdp/impdp, however the creation of the indexes started to slow down the process at which time I switched to using Transportable Tablespace.
    In the last 3 months, the number of objects in the schema has not changes however the amount of data has increased 15.73%. At the same time, the Transportable Tablespace impdp has increased 30.77% in time to process.
    Is there any way to speed up a Transportable Tablespace impdp in Oracle 10gR2? The docs show that I cannot use parallel with TTS. I have excluded STATISTICS in the export.
    Eventually I will be moving away from doing this TTS expdp/impdp, however for now I need to do what I can to contain the amount of time it takes to process.
    Environment: Oracle 10.2.0.4 (source and destination) on Sun Solaris SPARC v10. We are not using ASM currently. Tablespace is 6 datafiles for a total of 87.5GB, 791 tables, 1924 indexes, 1773 triggers, and 10 LOBs.
    Any help would be greatly appreciated.
    Thanks!
    -Vorpel

    My reading of this thread was that there was a need to be able to do offline setups when the client had lost oracle lite (we get this on PDAs all the time when they go totally dead), so the CD install is something that can be sent out.
    Any install (CD or otherwise) will be empty in terms of applications and data, not much you can do about that on a CD, but depending on the type of device and the storage possibilities you have you can secure the database and application files seperately. for example on a PDA using the external SD or flash for the data and file storage will keep them when the orace directory is lost from main storage and therefore all thet is needed after the oracle lite recovery is to recreate the odbc file (and possibly polite if you have special parameters). The setting of the crr flags for 'real' data on the server is a trick to bring the database up to date without a total rebuild.
    NOTE if you use SD cards etc. on PDAs/small devices, beware of windows ce/mobile power up differential between main memory and external storage. If you power down in the middle of an update or query, you can get device read/write errors when the device powers up due to processing continuing before the storage media is active

  • Transportable tablespace (metadata export with READ WRITE tablespace set)

    Is it possible to use transportable tablespace backup for import if data files for tablespace set has been copied when tablespaces were READ ONLY but the metadata export (or a part of it) is taken after tablespaces has been set to READ WRITE?

    Actually the size of datafile(s) are too huge so unable to test the same on test box with TTS export backup.
    But my question is if there were no structural changes (DML might be there on tablespace objects) even after making tablespaces to READ WRITE till the time metadata export is completed(data files are copied to staging location already before tablespaces has been set READ WRITE), Shall I use these backup for import?

  • WSDLException: faultCode=OTHER_ERROR: Failed to read WSDL from

    Hi all,
    I enabled inbound security for a process using SSL for Certificate-Based Authentication.Security is successfully enabled but I am getting warning "WSDLException: faultCode=OTHER_ERROR: Failed to read WSDL" when initiating test using HTML form as well as XML form.

    I have done the config based on bpel admin guide -> bpel_admin_10.1.3.1.0.pdf. and one more thing is after enabling SSL , I was not able to deploy from JDeveloper. however I deployed the application from console(bpel_*.jar).

  • Failed to read bytes from InputStream

    Hi,
    I get this excpetion log when I try to read a stream from weblogic. Can anyone help me out why this happens and what
    could be done to overcome this
    ####<Aug 8, 2007 4:48:37 PM EDT> <Error> <HTTP> <sundev01> <fhsserver> <[ACTIVE] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1186606117277> <BEA-101019> <[weblogic.servlet.internal.WebAppServletContext@111f9b - name: 'FFMS', context-path: '/FFMS'] Servlet failed with IOException
    java.io.IOException: failed to read '2678' bytes from InputStream; clen: 58512 remaining: 44902 count: 1394
    at weblogic.servlet.internal.ChunkOutput.writeStream(ChunkOutput.java:411)
    at weblogic.servlet.internal.ChunkOutputWrapper.writeStream(ChunkOutputWrapper.java:168)
    at weblogic.servlet.internal.ServletOutputStreamImpl.writeStream(ServletOutputStreamImpl.java:496)
    at weblogic.servlet.internal.ServletOutputStreamImpl.writeStream(ServletOutputStreamImpl.java:484)
    at weblogic.servlet.FileServlet.sendFile(FileServlet.java:400)
    at weblogic.servlet.FileServlet.doGetHeadPost(FileServlet.java:224)
    at weblogic.servlet.FileServlet.service(FileServlet.java:166)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:856)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:225)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:127)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:272)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:165)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3150)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:121)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:1973)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:1880)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1310)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:207)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:179)
    >

    Hi Aravind,
    Which version of weblogic are u using?
    Cheers,
    Lakshmi

  • DSO activation step has been failing if i activate from process chain

    Hello Frnds,
    one of the DSO activation step is failing from process chain.
    Problem : if i do the activation from PC then only it is failing
    If i do the DSO activation manually then it is fine .
    Kindly provide your suggestions on this issue.
    Thanks & regards
    Ravi

    Hello Lavanya,
    below are the logs for ur reference & process monitor is not genereated
    log from Process Tab:
    Cannot activate request 0000015877(REQU_4E3YPXSZ5F73H6ZNFZJYWMVJH) of DataStore object ZXXXX
    Activation of M records from DataStore object ZOGSDRCA terminated
    Job log:
    Job started
    Step 001 started (program RSPROCESS, variant &0000000015318, user ID ALEREMOTE)
    Cannot activate request 0000015877(REQU_4E3YPXSZ5F73H6ZNFZJYWMVJH) of DataStore object ZOGSDRCA
    Activation of M records from DataStore object ZOGSDRCA terminated
    Cannot activate request 0000015877(REQU_4E3YPXSZ5F73H6ZNFZJYWMVJH) of DataStore object ZOGSDRCA
    Activation of M records from DataStore object ZOGSDRCA terminated
    Entire chain now has status 'A'
    Process Activate DataStore Object Data, variant Activate Data ZOGSDREC has status Undefined (instance )
    Process Generate Index, variant Generated from DROPINDEX ZDELETEREC_CAUSE_INDEX has status Undefined (instance )
    Process Generate Index, variant Generated from DROPINDEX DELETE_INDEX_ZCGSDREC has status Undefined (instance )
    Process Data Transfer Process, variant ZOGSDREC -> ZCGSDREC has status Undefined (instance )
    Process Data Transfer Process, variant ZOGSDRCA -> ZCGSDRCA has status Undefined (instance )
    Process Delete Index, variant Delete Reclamation causes' index has status Undefined (instance )
    Process Delete Index, variant Delete index of ZCGSDREC Reclamations has status Undefined (instance )
    Process Start Process, variant Start Reclamation delta load has status Completed (instance 4EQ978RV7JE1BU3MT2YB20A3X)
    Process Execute InfoPackage, variant 2LIS_05_Q0NOTIF Delta has status Successfully completed (instance REQU_4EQTTCCMY2XEEPL99XGE9EECT)
    Process Execute InfoPackage, variant 2LIS_05_Q0CAUSE Delta has status Successfully completed (instance REQU_4EQG4DHZ6JD3K5UZCR5WP16P9)
    Process Data Transfer Process, variant 2LIS_05_Q0NOTIF / EQ2CLNT210 -> ZOGSDREC has status Undefined (instance DTPR_4EQ97CM4OU8SN3TPQ44G2ZMZX)
    Process Generate Index, variant Generated from LOADING ZPAK_4BFDG83ZVZM4MDUC6L71H1 has status Successfully completed (instance INDX_4EQ97IDIWSIXM0EU3NVNMGOBX)
    Process Activate DataStore Object Data, variant Generated from LOADING ZPAK_4BFDG83ZVZM4MDUC6L71H1 has status Ended with errors (instance ODSA_4EQG4KJGK99HMTOSPBB5WU0BX)
    Job finished
    Here i am giving process types sequence :
    Infopackage -> Create index-> DSO activation.
    does this create index is generating any issues ?
    Regards
    Ravi
    Edited by: BIuser on Aug 10, 2009 1:13 PM
    Edited by: BIuser on Aug 10, 2009 1:13 PM

  • Using import script to read data from another location

    Hi,
    I need to find a way to import, in one location, the data of all other locations (with some additional transformations) using an integration script.
    I think I should be able to read from the recordset returned by farsTable, but I need to know how to find the table name that contains the data from my other locations.
    Thanks in advance,
    Julien

    First you will need to read the tpovlocation table to get the location key and segment key for each location to be read. You can then either read the relevent tdataseg(index) table where the index is the segment key from the location table, or alternatively you could try to access the data from the View (which i think is Vdata) in which case you you will not need the segment key.

Maybe you are looking for