Due to huge data TIME_OUT Error

Hi All
Can anybody help me out from TIME_OUT error
I have declared types as per table FAGLFLEXA field sequence
Before checking the for all entries with table FAGLFLEXA, i was deleted duplicate entries from table BSAS
after deleted the duplicate records in BSAS table contains 3500000 records, because of this reason it goes to TIME_OUT error
Can you please suggest the solution to resolve this TIME_OUT error
Thanks in Advance
Sekhar
Moderator message: one thread only per issue, please.
Edited by: Thomas Zloch on Feb 1, 2012

Hi,
1. Run your program in background and check it in sm37, whether it processed or not.
2. Do performance tuning on select statements , nested loops, non-primary keys,......
3. go to RZ11  -
provide  ' rdisp/max_wprun_time '   -
display and then check the process time, increase the process time by clicking on change button   like( ex: 600  to  1200)
and also check with  ' abap/timeout '
Edited by: Parupelly on Feb 1, 2012 3:30 PM
Edited by: Parupelly on Feb 1, 2012 3:30 PM

Similar Messages

  • Deferred transaction fails due to a "ORA-01403: no data found" error

    I recently acticated replication between two Oracle instances.
    (one being the master definition site and the other one being
    a second master site).
    Everything was going well until two weeks ago, when i realized
    that some datas were different between the two Oracle instances.
    I tried to figure out why and found that some deferred transation
    failed to propagate due to a "ORA-01403: no data found" error.
    Not all transaction fail, only a few.
    All the errors are stored in the DEFERROR table of the second
    master site, the DEFERROR table is empty in the master definition
    site.
    For every error, the source is the master definition site and the
    destination is the second master site, which is OK because the
    second master site is supposed to be "read-only". But why are the
    errors stored in the DEFERROR table of the second master site
    (and not in the master definition site (the source)).
    Most errors occur on UPDATE statements but a few of them occur
    on DELETE statements. Given the fact that the record actually is
    in the second master site, how is it possible to get a
    "ORA-01403: no data found" error on a delete statement ???????
    I can't figure out what data cannot be found,
    Thanks for your help,
    Didier.

    What if i tell you the first transaction in error is an UPDATE
    statement on a record that actually is in the destination table
    in the second master site. So the data is there, it should be
    found ...
    I've tried an EXECUTE_ERROR on that very first transaction
    in error but it doesn't work. Here's the call (as repadmin) :
    begin
    DBMS_DEFER_SYS.EXECUTE_ERROR (
    DEFERRED_TRAN_ID => '6.42.271',
    DESTINATION => 'AMELIPI.SENAT.FR'
    end ;
    And here's the full message :
    ORA-01403: no data found
    ORA-06512: at "SYS.DBMS_DEFER_SYS_PART1" line 430
    ORA-06512: at "SYS.DBMS_DEFER_SYS" line 1632
    ORA-06512: at "SYS.DBMS_DEFER_SYS" line 1678
    ORA-06512: at line 2
    I thought it could be because the primary key constraint on the
    destination table wasn't activated but it was. Could it
    be a problem with the corresponding index ?
    Thanks for your help,
    Didier.

  • Strange error in requests returning huge data

    Hi
    Whenever a request returns huge data, we get the following error, which is annoying because there is no online source or documentation about this error:
    Odbc driver returned an error (SQLFetchScroll).
    Error Details
    Error Codes: OPR4ONWY:U9IM8TAC
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 46073] Operation 'stat()' on file '/orabi/OracleBIData/tmp/nQS_3898_715_34357490.TMP' failed with error: (75) ¾§ . (HY000)
    Any idea what might be wrong?
    Thanks

    The TMP folder is also "working" directory for the cache management (not exclusively). OBIEE tries to check if the report can be run against a cache entry using the info in this file.
    Check if MAX_ROWS_PER_CACHE_ENTRY , MAX_ROWS_PER_CACHE_ENTRY and MAX_CACHE_ENTRY_SIZE are set correctly.
    Regards
    John
    http://obiee101.blogspot.com

  • Due to huge file message failed in Moni

    Hi All
    Scenario is SQL to BI
    My sender Adapter successfully picked up data from SQL. Both Select query and Update querry have done there operations. All the records were selected, there flags are updated successfully.
    Now problem is when it came to Integration Server, message failed saying error "CLIENT_SEND_FAILED" this is due to huge file.
    Now how can i do the roll back because now i cannot resend this particular message because of huge file and all the records flags are updated, so i cannot run the query again.
    Please help
    Regards
    Dheeraj Kumar

    Hi Dheeraj,
    >>Can i see what are the records that got updated by this traction?
    Check the inbound payload in MONI/Communication channel and check which all records were ther. alternatively check the modified record in the specific time at database level
    >>when i am checking payload it is not showing whole file, at the end it is written "Solve the error and then refresh the page" "unable to write onto server".
    Where are you seeing this? Right click and try to view the source of the page
    >>can you suggest me now what workaround i can do to resolve this problem?
    As suggested read fewer records (your select query needs to be altered for this). Or use a sender stored procedure which will read fewer records and send to PI
    Regards
    Suraj

  • OSB: Cannot acquire data source error while using JCA DBAdapter in OSB

    Hi All,
    I've entered 'Cannot acquire data source' error while using JCA DBAdapter in OSB.
    Error infor are as follows:
    The invocation resulted in an error: Invoke JCA outbound service failed with application error, exception: com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/DBAdapter1/RetrievePersonService [ RetrievePersonService_ptt::RetrievePersonServiceSelect(RetrievePersonServiceSelect_inputParameters,PersonTCollection) ] - WSIF JCA Execute of operation 'RetrievePersonServiceSelect' failed due to: Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/soademoDatabase].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.soademoDatabase'. Resolved 'jdbc'; remaining name 'soademoDatabase'.
    ; nested exception is:
    BINDING.JCA-11622
    Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    JNDI Name for the Database pool: eis/DB/soademoDatabase
    JNDI Name for the Data source: jdbc/soademoDatabase
    I created a basic DBAdapter in JDeveloper, got the xsd file, wsdl file, .jca file and the topLink mapping file imported them into OSB project.
    Then I used the .jca file to generate a business service, and tested, then the error occurs as described above.
    Login info in RetrievePersonService-or-mappings.xml
    <login xsi:type="database-login">
    <platform-class>org.eclipse.persistence.platform.database.oracle.Oracle9Platform</platform-class>
    <user-name></user-name>
    <connection-url></connection-url>
    </login>
    jca file content are as follows:
    <adapter-config name="RetrievePersonService" adapter="Database Adapter" wsdlLocation="RetrievePersonService.wsdl" xmlns="http://platform.integration.oracle/blocks/adapter/fw/metadata">
    <connection-factory location="eis/DB/soademoDatabase" UIConnectionName="Connection1" adapterRef=""/>
    <endpoint-interaction portType="RetrievePersonService_ptt" operation="RetrievePersonServiceSelect">
    <interaction-spec className="oracle.tip.adapter.db.DBReadInteractionSpec">
    <property name="DescriptorName" value="RetrievePersonService.PersonT"/>
    <property name="QueryName" value="RetrievePersonServiceSelect"/>
    <property name="MappingsMetaDataURL" value="RetrievePersonService-or-mappings.xml"/>
    <property name="ReturnSingleResultSet" value="false"/>
    <property name="GetActiveUnitOfWork" value="false"/>
    </interaction-spec>
    </endpoint-interaction>
    </adapter-config>
    RetrievePersonService_db.wsdl are as follows:
    <?xml version="1.0" encoding="UTF-8"?>
    <WL5G3N0:definitions name="RetrievePersonService-concrete" targetNamespace="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService" xmlns:WL5G3N0="http://schemas.xmlsoap.org/wsdl/" xmlns:WL5G3N1="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService" xmlns:WL5G3N2="http://schemas.xmlsoap.org/wsdl/soap/">
    <WL5G3N0:import location="RetrievePersonService.wsdl" namespace="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService"/>
    <WL5G3N0:binding name="RetrievePersonService_ptt-binding" type="WL5G3N1:RetrievePersonService_ptt">
    <WL5G3N2:binding style="document" transport="http://www.bea.com/transport/2007/05/jca"/>
    <WL5G3N0:operation name="RetrievePersonServiceSelect">
    <WL5G3N2:operation soapAction="RetrievePersonServiceSelect"/>
    <WL5G3N0:input>
    <WL5G3N2:body use="literal"/>
    </WL5G3N0:input>
    <WL5G3N0:output>
    <WL5G3N2:body use="literal"/>
    </WL5G3N0:output>
    </WL5G3N0:operation>
    </WL5G3N0:binding>
    <WL5G3N0:service name="RetrievePersonService_ptt-bindingQSService">
    <WL5G3N0:port binding="WL5G3N1:RetrievePersonService_ptt-binding" name="RetrievePersonService_ptt-bindingQSPort">
    <WL5G3N2:address location="jca://eis/DB/soademoDatabase"/>
    </WL5G3N0:port>
    </WL5G3N0:service>
    </WL5G3N0:definitions>
    Any suggestion is appricated .
    Thanks in advance!
    Edited by: user11262117 on Jan 26, 2011 5:28 PM

    Hi Anuj,
    Thanks for your reply!
    I found that the data source is registered on server soa_server1 as follows:
    Binding Name: jdbc.soademoDatabase
    Class: weblogic.jdbc.common.internal.RmiDataSource_1033_WLStub
    Hash Code: 80328036
    toString Results: ClusterableRemoteRef(8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1 [8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1/291])/291
    Binding Name: jdbc.SOADataSource
    Class: weblogic.jdbc.common.internal.RmiDataSource_1033_WLStub
    Hash Code: 92966755
    toString Results: ClusterableRemoteRef(8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1 [8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1/285])/285
    I don't know how to determine which server the DBAdapter is targetted to.
    But I found the following information:
    Under Deoloyment->DBAdapter->Monitoring->Outbound Connection Pools
    Outbound Connection Pool Server State Current Connections Created Connections
    eis/DB/SOADemo AdminServer Running 1 1
    eis/DB/SOADemo soa_server1 Running 1 1
    eis/DB/soademoDatabase AdminServer Running 1 1
    eis/DB/soademoDatabase soa_server1 Running 1 1
    The DbAdapter is related to the following files:
    C:\ Oracle\ Middleware\ home_11gR1\ Oracle_SOA1\ soa\ connectors\ DbAdapter. rar
    C:\ Oracle\ Middleware\ home_11gR1\ Oracle_SOA1\ soa\ DBPlan\ Plan. xml
    I unzipped DbAdapter.rar, opened weblogic-ra.xml and found that there's only one data source is registered:
    <?xml version="1.0"?>
    <weblogic-connector xmlns="http://www.bea.com/ns/weblogic/90">
    <enable-global-access-to-classes>true</enable-global-access-to-classes>
    <outbound-resource-adapter>
    <default-connection-properties>
    <pool-params>
    <initial-capacity>1</initial-capacity>
    <max-capacity>1000</max-capacity>
    </pool-params>
    <properties>
    <property>
    <name>usesNativeSequencing</name>
    <value>true</value>
    </property>
    <property>
    <name>sequencePreallocationSize</name>
    <value>50</value>
    </property>
    <property>
    <name>defaultNChar</name>
    <value>false</value>
    </property>
    <property>
    <name>usesBatchWriting</name>
    <value>true</value>
    </property>
    <property>
    <name>usesSkipLocking</name>
    <value>true</value>
    </property>
    </properties>
              </default-connection-properties>
    <connection-definition-group>
    <connection-factory-interface>javax.resource.cci.ConnectionFactory</connection-factory-interface>
    <connection-instance>
    <jndi-name>eis/DB/SOADemo</jndi-name>
              <connection-properties>
                   <properties>
                   <property>
                   <name>xADataSourceName</name>
                   <value>jdbc/SOADataSource</value>
                   </property>
                   <property>
                   <name>dataSourceName</name>
                   <value></value>
                   </property>
                   <property>
                   <name>platformClassName</name>
                   <value>org.eclipse.persistence.platform.database.Oracle10Platform</value>
                   </property>
                   </properties>
              </connection-properties>
    </connection-instance>
    </connection-definition-group>
    </outbound-resource-adapter>
    </weblogic-connector>
    Then I decided to use eis/DB/SOADemo for testing.
    For JDeveloper project, after I deployed to weblogic server, it works fine.
    But for OSB project referencing wsdl, jca and mapping file from JDeveloper project, still got the same error as follows:
    BEA-380001: Invoke JCA outbound service failed with application error, exception:
    com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/DBAdapterTest/DBReader [ DBReader_ptt::DBReaderSelect(DBReaderSelect_inputParameters,PersonTCollection) ] - WSIF JCA Execute of operation 'DBReaderSelect' failed due to: Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/SOADataSource].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.SOADataSource'. Resolved 'jdbc'; remaining name 'SOADataSource'.
    ; nested exception is:
    BINDING.JCA-11622
    Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/SOADataSource].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.SOADataSource'. Resolved 'jdbc'; remaining name 'SOADataSource'.
    You may need to configure the connection settings in the deployment descriptor (i.e. DbAdapter.rar#META-INF/weblogic-ra.xml) and restart the server. This exception is considered not retriable, likely due to a modelling mistake.
    It almost drive me crazy!!:-(
    What's the purpose of 'weblogic-ra.xml' under the folder of 'C:\Oracle\Middleware\home_11gR1\Oracle_OSB1\lib\external\adapters\META-INF'?
    ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

  • I am getting ORA-01403: no data found error while calling a stored procedur

    Hi, I have a stored procedure. When I execute it from Toad it is successfull.
    But when I call that from my java function it gives me ORA-01403: no data found error -
    My code is like this -
    SELECT COUNT(*) INTO L_N_CNT FROM TLSI_SI_MAST WHERE UPPER(CUST_CD) =UPPER(R_V_CUST_CD) AND
    UPPER(ACCT_CD)=UPPER(R_V_ACCT_CD) AND UPPER(CNSGE_CD)=UPPER(R_V_CNSGE_CD) AND
    UPPER(FINALDEST_CD)=UPPER(R_V_FINALDEST_CD) AND     UPPER(TPT_TYPE)=UPPER(R_V_TPT_TYPE);
         IF L_N_CNT >0 THEN
              DBMS_OUTPUT.PUT_LINE('ERROR -DUPlicate SI-1');
              SP_SEL_ERR_MSG(5,R_V_ERROR_MSG);
              RETURN;
         ELSE
              DBMS_OUTPUT.PUT_LINE('BEFORE-INSERT');
              INSERT INTO TLSI_SI_MAST
                   (     CUST_CD, ACCT_CD, CNSGE_CD, FINALDEST_CD, TPT_TYPE,
                        ACCT_NM, CUST_NM,CNSGE_NM, CNSGE_ADDR1, CNSGE_ADDR2,CNSGE_ADDR3,
                        CNSGE_ADDR4, CNSGE_ATTN, EFFECTIVE_DT, MAINT_DT,
                        POD_CD, DELVY_PL_CD, TRANSSHIP,PARTSHIPMT, FREIGHT,
                        PREPAID_BY, COLLECT_BY, BL_REMARK1, BL_REMARK2,
                        MCC_IND, NOMINATION, NOTIFY_P1_NM,NOTIFY_P1_ATTN , NOTIFY_P1_ADDR1,
                        NOTIFY_P1_ADDR2, NOTIFY_P1_ADDR3, NOTIFY_P1_ADDR4,NOTIFY_P2_NM,NOTIFY_P2_ATTN ,
                        NOTIFY_P2_ADDR1,NOTIFY_P2_ADDR2, NOTIFY_P2_ADDR3, NOTIFY_P2_ADDR4,
                        NOTIFY_P3_NM,NOTIFY_P3_ATTN , NOTIFY_P3_ADDR1,NOTIFY_P3_ADDR2, NOTIFY_P3_ADDR3,
                        NOTIFY_P3_ADDR4,CREATION_DT, ACCT_ATTN, SCC_IND, CREAT_BY, MAINT_BY
                        VALUES(     R_V_CUST_CD,R_V_ACCT_CD,R_V_CNSGE_CD,R_V_FINALDEST_CD,R_V_TPT_TYPE,
                        R_V_ACCT_NM,R_V_CUST_NM ,R_V_CNSGE_NM, R_V_CNSGE_ADDR1,R_V_CNSGE_ADDR2, R_V_CNSGE_ADDR3,
                        R_V_CNSGE_ADDR4,R_V_CNSGE_ATTN,     R_V_EFFECTIVE_DT ,SYSDATE, R_V_POD_CD,R_V_DELVY_PL_CD,R_V_TRANSSHIP ,R_V_PARTSHIPMT , R_V_FREIGHT,
                        R_V_PREPAID_BY ,R_V_COLLECT_BY ,R_V_BL_REMARK1 ,R_V_BL_REMARK2,R_V_MCC_IND,
                        R_V_NOMINATION,R_V_NOTIFY_P1_NM, R_V_NOTIFY_P1_ATTN, R_V_NOTIFY_P1_ADD1, R_V_NOTIFY_P1_ADD2,
                        R_V_NOTIFY_P1_ADD3, R_V_NOTIFY_P1_ADD4, R_V_NOTIFY_P2_NM, R_V_NOTIFY_P2_ATTN, R_V_NOTIFY_P2_ADD1,
                        R_V_NOTIFY_P2_ADD2, R_V_NOTIFY_P2_ADD3, R_V_NOTIFY_P2_ADD4, R_V_NOTIFY_P3_NM, R_V_NOTIFY_P3_ATTN,
                        R_V_NOTIFY_P3_ADD1, R_V_NOTIFY_P3_ADD2, R_V_NOTIFY_P3_ADD3, R_V_NOTIFY_P3_ADD4,
                        SYSDATE,R_V_ACCT_ATTN,R_V_SCC_IND,R_V_USER_ID,R_V_USER_ID
                        DBMS_OUTPUT.PUT_LINE(' SI - REC -INSERTED');
         END IF;

    Hi,
    I think there is a part of the stored procedure you did not displayed in your post. I think your issue is probably due to a parsed value from java. For example when calling a procedure from java and the data type from java is different than expected by the procedure the ORA-01403 could be encountered. Can you please show the exact construction of the call of the procedure from within java and also how the procedure possible is provided with an input parameter.
    Regards, Gerwin

  • TIME_OUT error while executing an ABAP report

    Hi All,
    I have an ABAP report in which the code looks like below
    LOOP AT pernr_tab . " *Internal table consists of  around 25k records* 
       IF last_run_date IS INITIAL. " Initial Load
          PERFORM fill_data USING pernr_tab-pernr last_run_date CHANGING lv_flag.
          lv_flag = 'x'.
          initial_count = initial_count + 1.
        ELSE.
          PERFORM fill_data USING pernr_tab-pernr last_run_date CHANGING lv_flag.
          IF pernr_tab-status EQ '3' AND pernr_tab-begda GE today. " Future Hire records during delta load
            lv_flag = 'x'.
          ENDIF.
          IF lv_flag EQ 'x'.
            delta_change_count = delta_change_count + 1.
          ENDIF.
        ENDIF.
        IF lv_flag = 'x'. "initial load or changes in some infotype for delta load
          PERFORM append_data.
          CLEAR: p0000_tab[],p0000,p0001_tab[],p0001,p0002_tab[],p0002,p0006_tab[],p0006,p0006_phone,
    p0007_tab[],p0007,p0027_tab[],p0027,p0032_tab[],p0032,p0105_tab_usr[],p0041_tab[],p0041.
          CLEAR: p0105_usr,p0105_tab_ven[],p0105_ven,p0105_tab_mgr[],p0105_mgr,p0709_tab[],p0709,pos_id,
    functional_title,p0521_tab[],p0521,p2001_tab[],p2001,corporate_band_code,corporate_title_code,wa_pernr_tab.
        ENDIF.
      ENDLOOP.
    inside the loop i am fetching data from 10 different infotypes and adding into another internal table.When I execute this program it is giving TIME_OUT error.Is there any way i can resolve this issue(I dont want to run this program in the background)
    Thanks
    Bala Duvvuri
    Edited by: Bala Duvvuri on Jul 13, 2010 8:25 AM
    Edited by: Thomas Zloch on Jul 13, 2010 4:58 PM - line width reduced

    fill_data
    reading 10 infotypes
    call function 'HR_READ_INFOTYPE'
      exporting
    *     TCLAS                 = 'A'
        pernr                 = p_pernr_tab_pernr
        infty                 = '0000'
       begda                 = '18000101'
       endda                 = '99991231'
    *   bypass_buffer         = 'X'
      tables
        infty_tab             = p0000_tab
    exceptions
       infty_not_found       = 1
       others                = 2
      if sy-subrc <> 0.
      endif.
      if p0000_tab[] is not initial.
        sort p0000_tab by endda  descending.
        read table p0000_tab into p0000 index 1.
        if p0000-aedtm ge last_run_date.
          lv_flag = 'X' ." Infotype 0 has been changed
        endif.
      endif.
    append_data
    assigning 50 different fields from above infotypes to ls_emp_data
    ls_emp_data-functional_title = functional_title.
      ls_emp_data-last_update = today.
      if ( p2001-awart = '5006' and p2001-begda <= today and p2001-endda >= today ) or
    ( p0521-atzph = 'AB' AND p0521-begda <= today AND p0521-endda >= today ).
        ls_emp_data-absent = text-007.
      ELSE.
        ls_emp_data-absent = text-008.
      ENDIF.
      ls_emp_data-corporate_band_code = corporate_band_code.
      ls_emp_data-corporate_title_code = corporate_title_code.
      APPEND ls_emp_data TO lt_emp_data.
      CLEAR:ls_emp_data,lv_flag.
    Thanks
    Bala Duvvuri
    Edited by: Thomas Zloch on Jul 14, 2010 2:48 PM - code tags added

  • PST is not an Outlook data file error coming while opening the Outlook

    Hi,
    Suddenly my outlook is showing the .pst file is not an Outlook data file error while starting. No big changes was made to the system to cause this. I had upgraded the IE version to 9 the same day. Done lots of trouble shooting but still not able fix my pst
    file to work. My all mails are in that pst and there is no backup for it.
    I am using Office Outlook 2007. I have already tried the following  to make it work bu failed each time.
    1. Tried to repair the .pst file using scanpst tool in the system, but closed due to an unexpected error at the Phase 6.
    2. Successfully completed the repair using the scanpst of another Outlook 2010 system and added the pst back to my systems Outlook but it had only an Inbox folder in it which was empty.
    3. Tried to repair the using other third party software's but none came successful.
    Please let me know anything can be done for to make my old pst work again.
    Sanjay 
    Avatar of Sanjay Oyitty Sanjay Oyitty

    Hi,
    What's the size of that .pst file?
    Right click on the file and click Properties, make sure there's no
    Read-only Attribute. Then we can run Scanpst.exe again to check the result again:
    http://support.microsoft.com/kb/272227
    Regards,
    Melon Chen
    TechNet Community Support

  • TIME_OUT error with /sapapo/rtsinput_cube

    We are in the process of cutting over to a new
    planning area ZPA2 with data from a backup InfoCube
    (ZIC21), and the /sapapo/rtsinput_cube program
    is failing with a TIME_OUT error.
    The overall job finishes with a success message,
    although data is only partially loaded in to the
    planning area. When you look at the spool, the first
    steps are green, then the job times out for the
    subsequent steps.
    We ran this first with 3 key figures being loaded
    from cube to planning area for approx 1/5 of 464000
    CVCs for 5 years in monthly buckets only.
    This errored with SAPSQL_INVALID_FIELDNAME message as
    well as the TIME_OUT error message, and data was
    loaded partially in the ZPA2 planning area (only 2
    months).
    We then reduced the time horizon from 5 years of
    months to be only 1 year in months.
    This errored with the TIME_OUT error message. Again,
    data was partially loaded in to the planning area
    (about 7 of the 12 months was loaded).
    The prime issue is with the constant time out.
    Has anyone any idea how to solve this problem ?

    not sure what version you are using but if you are on the latesr versions, its a good idea to use parallel processing - you can find it on the additonal settings tab
    you can also set it to copy only the specified period (previously it copied the whole bucket in the planning area to cover the time period)
    if you are in versions that dont support this - then you can create parallel jobs that get triggered by an event or use a process chain and run this in parallel. As long as you dont overlap the periods it should be fine
    you should also ask your basis team to check the time out set up for the application and get it extended (you can see this in rz11 i think and choosing the correct parameter - would leave it to the basis folks to help in this) Alternately check your processing power and see if you can get it increased

  • TIME_OUT error while trying to execute a report

    Hi all,
    When I try to exceute a Z report using a particular variant, it displays the output.
    However, when another person tries to run the same Z report using the same variant, it gives a TIME_OUT error.
    ABAP runtime errors    TIME_OUT
    Time limit exceeded.
    Please let me know the reason for this.
    Thanks and regards,
    Anishur

    Hi Anishur,
    Is the TIME OUT occuring on the same client as you are running the report?  And if so, does the user have the same authorisation as you?  The TIME OUT will occur if the program is handling more data, if you are seeing different performance it will be down to the amount of data process.
    As others have suggested, look at improving the performance of the report.
    Regards,
    Nick

  • Misterious "ORA-01403: no data found" error

    Hi,
    I've been searching through the forum and the Internet with no results, so I'm posting here hoping you can help me.
    Here is the situation: I had a page 30 with a tree branching to a form on a view.This worked fine until one day that suddenly, we got the "ORA-01403: no data found" error when loading the page. I guess that's normal, since the form shows the data filtered in the tree (in the tree we click on an element and pass its ROWID to the form), and when we load the page no ROWID is selected. So, the first question would be, how can I manage this situation? My first option was to have a condition in the Automatic Fetch for the Form, so that it's executed only when the ROWID item is not null. It worked a couple of times. After a while, it crashed again.
    Tired of trying things, we decided to copy this page 30 to another page (page #80). Surprisingly, page #80 worked meanwhile #30 crashed. And they are completely the same, the data has not changed either.
    Could it be that page #30 somehow is corrupted? It's really strange that the same exact copy of the page works with one ID but not with another one.
    Has someone experienced this situation? it makes me feel like the app is unsecure and unstable. Any ideas, please?
    Apex version: 4.1.1 on APEX listener.
    Thanks,
    Elena.

    fac586 wrote:
    Elena.mtc wrote:
    Thanks for your response. The form is a single row (a form on a table or view). Here is what I've found out thanks to your tips:
    It affects to all the three users that I have access with.Try it with a new user that has never accessed the page before.
    I used the debug mode (new to me, so I'm kind of slow on this, still) and wrote an exception message on the Automatic Fetch Row and I managed to know the ROWID that the fetch process is using in the where clause. Of course this ROWID does not exist in the table, so the first question that comes to my mind is "how come and from where is the page getting this ROWID?". I don't know and can't comment because I can't see the app. ;-)
    If the ROWID is stored in an item, check the Maintain session state property isn't set to Per User.I changed this and seemed to work. But it crashed later. I made a computation process to set to null the rowid's when loading the page. After reset, I deleted the computations (due to application functionality) and it worked. However, I do not trust this page anymore. So we created a new fresh app and copied this page. Luckily we are on development environment :)
    >
    So, I made a computation to initialize to null the item that saves the ROWID, and now it worked.
    This page that I'm testing on is #30, but the valid page is #80 (it has changes made on the look and feel of the page). So, I deleted page #30, and copied page #80 (which works perfectly) into #30. And it crashes again (obvius, since it doesn't have the computation that gives null to the rowid item). What's "interesting" is that I still see the exception message I wrote in the older version of page #30. So, it seems that some values need to be purged and that the delete is not clean, but how? I'm not entirely clear on what you've been doing there, but it does indeed all sound rather odd.
    I suspect that you wouldn't be able to reproduce this problem on apex.oracle.com. However, it would probably be helpful if you could upload the app and provide developer credentials to the workspace so we could see how it works and if there's anything obvious (or not) that might cause this problem.Unluckily I cannot do this. Thank you anyway for the help, it's hard to explain when something so weird happens.
    Thanks!
    Elena.

  • Getting time_out Error in report bcos of SQL Statement!

    Hi Experts,
    I am getting Time_out error  in my big program because of Select statement.I have checked in SQL Trace AND  fuond the message
    "SQL0104N  An unexpected token "vakey" was found following "umh
      kschl".  Expected tokens may include:  ",".  SQLSTATE=42601"
    I have given the following code and SQL statement where the error is getting occured.Please help.
    TYPES: BEGIN OF ty_vakey1,
               vakey TYPE vakey,
             END OF ty_vakey1,
             ty_t_vakey1 TYPE HASHED TABLE OF ty_vakey1 WITH UNIQUE KEY vakey.
      DATA: ih_vakey1 TYPE ty_t_vakey1,
            wa_vakey1 TYPE ty_vakey1,
            wa_vakey  TYPE ty_vakey.
      LOOP AT i_vakey INTO wa_vakey.
        wa_vakey1-vakey = wa_vakey-vakey.
        COLLECT wa_vakey1 INTO ih_vakey1.
        CLEAR wa_vakey1.
      ENDLOOP.
      IF ih_vakey1 IS NOT INITIAL.
    Here it is required to select the Key field and so cannot be used in Where condition
        SELECT knumh                                             "Condition record number
               kschl                                             "Pricing conditions
               vakey                                             "Variable Key
               datab                                             "Valid-From Date
               datbi                                             "Valid To Date
               FROM konh
               INTO TABLE i_konh
               FOR ALL ENTRIES IN ih_vakey1
               WHERE vakey EQ ih_vakey1-vakey .           "#EC CI_NOFIELD
    Here #EC CI_NOFIELD is used because the key field of the table konh ( ie knumh)
    is to be selected and so is not used in the WHERE condition above
        IF sy-subrc NE c_0.
    No data found for specified criteria
          MESSAGE i036.
          LEAVE LIST-PROCESSING.
        ENDIF.
    Thanks&Regards,
    Sarkar

    Check if i_konh definition matches the fields in the SELECT statement.
    Also, since you're not using the primary key when selecting from KONH table, such program by design is going to run forever and may time out eventually. Usually KONH is a very large table. Not sure what's the purpose of this program, but I'm pretty sure there must be some other way to program this. There are actually some standard FMs that read the conditions.
    A secondary index might help, but only as a last resort.

  • ICal wont sync due to inconsistent data

    I get an error message saying... Calendars could not be synced due to inconsistent date. Confirm that your computers data is valid *+(not sure how to do this)+* then reset Calendars on .Mac in system preferences. *+( I have done this and had no luck)+*
    I have also gone into isync and reset my sysc history and then re-synced... having all my computer data be replaced by the .Mac data. i get the same error.
    I have 3 machines that should all be synced. I have just done a clean install on all machines in the following order:
    a mac mini, a Powermac G5, and my Macbook. The macbook is the problem machine. the other 2 systems are synced perfectly.
    Any advice?

    Dude,
    Thanks!!!
    I did two clean installations on my 2 MBPs to leopard starting Friday night and Finishing tonight. I opted to not migrate data because I wanted to leave all my messy Tiger stuff and start again.
    My first MPB must be the dominant one on .Mac because
    MBP-2 >> .Mac >> MPB-1 didn't work .. if it weren't for your reply I wouldn't have tried the other way around and would have just given up until next weekend.
    Thanks.

  • SQL Server 2005 huge data migration. Taking more than 18 hours ...

    Hi,
    We are migrating an SQL server 2005 DB to Oracle 10 g 64 bit 10.2.0.5. The source sql server DB has few tables (<10) and only few stored procedure(<5). However one table of source is very huge, it has 70 million records in it. I started the migration process; it was running till 18 hours passed away then I cancelled it. I am using SQL Developer 3.1.0.7 with online migration.
    Kindly let me know if I choose offline migration will it reduce the data move time. Is there some other tool by oracle that can migrate huge data (records in millions) quickly. Is there some configuration in SQL Developer that can catalyses the process?
    Kindly help.
    Best Regards!
    Irfan

    I tried using SQL Loader to move data. The file exported by SQL server for the transaction table is of size around 17 GB. The sql loader stopped giving error on file size limit. The error is given below:
    SQL*Loader: Release 10.2.0.5.0 - Production on Thu Mar 22 15:15:28 2012
    Copyright (c) 1982, 2007, Oracle. All rights reserved.
    SQL*Loader-510: Physical record in data file (E:/mssqldata/transaction.txt) is longer than the maximum(1048576)
    SQL*Loader-2026: the load was aborted because SQL Loader cannot continue.
    C:\Users\Administrator\Desktop\wuimig\datamove\wuimig\2012-03-22_14-20-29\WUI\dbo\control>
    Please help me out how do I move such huge data to oracle from sql server.
    Best Regards!

  • Invoice is incomplete due to export data

    dear all
    my invoice is incomplete due to export data
    so i could not generate the accounting data for the invoice can any one suggest me wat will be the solution and wat i ve to fill in export data...
    the following are the error comming in eaach step
    <b>Document 90000003 has been saved (foreign trade data incomplete)
    Message no. VF072
    Diagnosis
    The billing document you have edited contains foreign trade data. These are however not complete.
    This means that the document cannot be transferred to accounting (FI).
    System Response
    The system issues a message.
    Procedure
    Maintain the incomplete foreign trade data. Then you can transfer the billing document to accounting (FI).</b>
    please suggest me the way
    with regards
    subrat

    the incomplition comming for export dta is as follows
    <b>business transaction type for foreign trade
    export import procedure for foreign trade
    comodity code import code for foreign trade</b>
    but wen i m clicking the drop down menu the fields having no values
    so wer to maintain data for this fields
    please help me
    with regards
    subrat

Maybe you are looking for

  • Error while loading the data from excel to database.

    Hi, I am using PL/SQL developer to load the data from excel to database. I will set the data source in the control panel and will proceed through ODBC importer in pl/sql developer to import the data. What exactly the error is when i click the filenam

  • Is there any way to prevent web.xml from any change like java class?

    hi all, Is there any way to prevent web.xml from any change after making EAR(WAR)? One can easily make a change in web.xml and redeploy the application to get the result. Now we want to restrict the web.xml as java class for any change after making E

  • Problem with Date in EDI posting

    PO created and updated by the EDI 945 from mexico SAP system. Once the EDI is sent,it changes the arrival date of the PO to the ship date.  An example is STO 506613882.  When created this PO, it had a delivery date of 02-01-08.  It was updated by the

  • Query to Handle special characters in the conditions

    Hi all, We have a table for colour codes and there related information but the colour codes have special characters in them like AL&.MPD CH(SB00 ECA&BC1 TD..0023 0O'DON4 i need to check if these exist in the table or not but when i query like select

  • Why can't I backup my catalog in PSE6?

    Every time I try to backup my catalog in PSE6 I get the message"The exception unknown software exception (0xc000000d) occurred in the application at location 0x78138aa0." And the programme shuts down. There was some suggestion that deleting preferenc