Variant taking date sydate for scheduling

Hello Gurus,
i want to create a variant in selection screen of Bex report
that take sys-date for a date variable
coz the reports here are scheduled for broadcasting daily and month end date
Now for this there is a feature in EP
to set a variant for query data selection for scheduling
My problem is this variant shud pick date dynamically so that user gets data for the date.
Can this b done any way around ?

Hi Sonal,
Use SAP Exit variable or
Create new Customer Exit variable and populate default value as current date.
[How to... Create & Populate User Exit variable |http://sapbwneelam.blogspot.com/2007/10/how-to-create-populate-user-exit.html]
Hope it Helps
Srini

Similar Messages

  • Saving the data fetched for scheduled Crystal Reports

    We have requirement to schedule report generation . An rpt will be provided as input to scheduler and whenever the schedule is triggered , the rpt should run to fetch data from database (without viewing the report ) and save rpt with fetched data on the file system in rpt format. Is there an API in SDK that provides this functionality?

    "Scheduling" is available only in CR Server or BOE suite of products. You can "export" the report to disk instead. This will run the report to fetch data and save it to disk with data. Here is an example code:
    import com.crystaldecisions.reports.sdk.*;
    import com.crystaldecisions.sdk.occa.report.lib.*;
    import com.crystaldecisions.sdk.occa.report.exportoptions.*;
    import java.io.*;
    public class JRCExportReport {
         static final String REPORT_NAME = "JRCExportReport.rpt";
         static final String EXPORT_FILE = "C:\\myExportedReport.pdf";
         public static void main(String[] args) {
              try {
                   //Open report.               
                   ReportClientDocument reportClientDoc = new ReportClientDocument();               
                   reportClientDoc.open(REPORT_NAME, 0);
                   //NOTE: If parameters or database login credentials are required, they need to be set before.
                   //calling the export() method of the PrintOutputController.
                   //Export report and obtain an input stream that can be written to disk.
                   //See the Java Reporting Component Developer's Guide for more information on the supported export format enumerations
                   //possible with the JRC.
                   ByteArrayInputStream byteArrayInputStream = (ByteArrayInputStream)reportClientDoc.getPrintOutputController().export(ReportExportFormat.PDF);
                   //Release report.
                   reportClientDoc.close();
                   //Use the Java I/O libraries to write the exported content to the file system.
                   byte byteArray[] = new byte [byteArrayInputStream.available()];
                   //Create a new file that will contain the exported result.
                   File file = new File(EXPORT_FILE);
                   FileOutputStream fileOutputStream = new FileOutputStream(file);
                   ByteArrayOutputStream byteArrayOutputStream = new ByteArrayOutputStream(byteArrayInputStream.available());
                   int x = byteArrayInputStream.read(byteArray, 0, byteArrayInputStream.available());
                   byteArrayOutputStream.write(byteArray, 0, x);
                   byteArrayOutputStream.writeTo(fileOutputStream);
                   //Close streams.
                   byteArrayInputStream.close();
                   byteArrayOutputStream.close();
                   fileOutputStream.close();
                   System.out.println("Successfully exported report to " + EXPORT_FILE);
              catch(ReportSDKException ex) {
                   ex.printStackTrace();
              catch(Exception ex) {
                   ex.printStackTrace();
    Edited by: Aasavari Bhave on Oct 29, 2009 2:26 PM
    Edited by: Aasavari Bhave on Oct 29, 2009 2:29 PM

  • SQL Server 2014 supported for scheduled refresh?

    The following link, which lists the supported data sources for scheduled refresh in Power BI, lists all versions of SQL Server from 2005 to 2012:
    https://support.office.com/en-gb/article/Supported-Data-Sources-and-Data-Types-cb69a30a-2225-451f-a9d0-59d24419782e?ui=en-US&rs=en-GB&ad=GB
    However 2014 is not listed. Is this an oversight or is SQL Server 2014 really not supported?
    Thanks,
    Chris
    Check out my MS BI blog I also do
    SSAS, PowerPivot, MDX and DAX consultancy and run
    public SQL Server and BI training courses in the UK

    Chris, have you heard about this?
    I'm checking with our tech writing team. 
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • In V_V2 the role of Sort Item by date of creation and Sort by delivery date of earliest schedule line for the Sales Orders when Material is out of Stock.

    Hi All,
    Good Morning,
    I am working on once Incident for Re-Scheduling Program V_V2 where the user has requested us to change the criteria for determining the priority of Sales Document based on Sort Item by date of creation. Currently they are using the Sort by delivery date of earliest schedule line in the Variant.
    I would like to know the role these both options play in the Re-Scheduling program V_V2. I would like to know the impact of both of them and the kind of testing I need to perform to study the behaviour of Sales Orders especially when the material is out of stock?
    Please explain with help of examples.
    Thanks
    Farhan.

    This is not that simple that an example would explain how this transaction works but I'll give you some examples.
    Material A has SO1 and SO2 of 10KG each, stock is 20KG so both quantities are confirmed at today. A stock recount determines a reduction to 5KG. This program would change confirmed quantity of SO1 to 5KG and SO2 to 0KG if:
    Creation date of SO1 is older then SO2 in strategy 1
    1st delivery date of SO1 is smaller then SO2 in strategy 2
    Material B has SO1 and SO2 with 10KG each but no quantity confirmed, stock is 0KG. A prodution adds 15KG to stock. The program would change confirmed quantities of SO1 to 10KG and SO2 to 5KG if:
    Creation date of SO1 is older then SO2 in strategy 1
    1st delivery date of SO1 is smaller then SO2 in strategy 2
    I'm assuming a specific configuration, there are many other scenarios possible.
    regards,
    Edgar

  • GR for scheduling agreement even before actaul delivery date

    Hi,
    Is is possible to to GR for schedule line even before the actual delivery date. ?
    GobinathanG

    Hi,
    yes it is. Use Transaction MB01 and activate the selection box 'Suggest Zero Lines'.
    Carola

  • [SD] table for Schedule line date

    If we display a sales order through Va03 and Go to tab Item overview. There is a excel sheet kind of display showing
    all items. The 12th col is 'First Date' and the f1 help says its 'Schedule line date'.
    I need to display this date in my report.
    <i>F1 Help Says :</i>
    <b>Schedule line date</b>
    The delivery date for the schedule line. The delivery date can be the
    Date requested by the customer, or (if the requested date cannot be met)
    Earliest date proposed by the system after automatic availability checks and delivery scheduling
    <b>Use</b>
    The system uses the delivery date for each schedule line together with predefined time estimates for delivery activities (such as picking and packing) to determine the
    Material availability date
    Loading date
    Goods issue date
    Transportation planning date
    <b>Dependencies</b>
    If you want the system to automatically determine the above dates, the Delivery scheduling and Transportation scheduling fields must be active in the sales document definition ( Tools --> Customizing, Sales and Distribution --> Basic Functions --> Delivery scheduling and transportation scheduling --> Define scheduling by sales document type).
    <b>Note</b>
    If there is more than one schedule line for the item, the system displays the earliest delivery date in the item and overview screens.

    Hi flora
    You can get he details of the schedule line data in the VBEP table.
    EDATU is the schedule line date
    regards
    kishore

  • Query taking long time for EXTRACTING the data more than 24 hours

    Hi ,
    Query taking long time for EXTRACTING the data more than 24 hours please find the query and explain plan details below even indexes avilable on table's goe's to FULL TABLE SCAN. please suggest me.......
    SQL> explain plan for select a.account_id,round(a.account_balance,2) account_balance,
    2 nvl(ah.invoice_id,ah.adjustment_id) transaction_id,
    to_char(ah.effective_start_date,'DD-MON-YYYY') transaction_date,
    to_char(nvl(i.payment_due_date,
    to_date('30-12-9999','dd-mm-yyyy')),'DD-MON-YYYY')
    due_date, ah.current_balance-ah.previous_balance amount,
    decode(ah.invoice_id,null,'A','I') transaction_type
    3 4 5 6 7 8 from account a,account_history ah,invoice i_+
    where a.account_id=ah.account_id
    and a.account_type_id=1000002
    and round(a.account_balance,2) > 0
    and (ah.invoice_id is not null or ah.adjustment_id is not null)
    and ah.CURRENT_BALANCE > ah.previous_balance
    and ah.invoice_id=i.invoice_id(+)
    AND a.account_balance > 0
    order by a.account_id,ah.effective_start_date desc; 9 10 11 12 13 14 15 16
    Explained.
    SQL> select * from table(dbms_xplan.display);
    PLAN_TABLE_OUTPUT
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)|
    | 0 | SELECT STATEMENT | | 544K| 30M| | 693K (20)|
    | 1 | SORT ORDER BY | | 544K| 30M| 75M| 693K (20)|
    |* 2 | HASH JOIN | | 544K| 30M| | 689K (20)|
    |* 3 | TABLE ACCESS FULL | ACCOUNT | 20080 | 294K| | 6220 (18)|
    |* 4 | HASH JOIN OUTER | | 131M| 5532M| 5155M| 678K (20)|
    |* 5 | TABLE ACCESS FULL| ACCOUNT_HISTORY | 131M| 3646M| | 197K (25)|
    | 6 | TABLE ACCESS FULL| INVOICE | 262M| 3758M| | 306K (18)|
    Predicate Information (identified by operation id):
    2 - access("A"."ACCOUNT_ID"="AH"."ACCOUNT_ID")
    3 - filter("A"."ACCOUNT_TYPE_ID"=1000002 AND "A"."ACCOUNT_BALANCE">0 AND
    ROUND("A"."ACCOUNT_BALANCE",2)>0)
    4 - access("AH"."INVOICE_ID"="I"."INVOICE_ID"(+))
    5 - filter("AH"."CURRENT_BALANCE">"AH"."PREVIOUS_BALANCE" AND ("AH"."INVOICE_ID"
    IS NOT NULL OR "AH"."ADJUSTMENT_ID" IS NOT NULL))
    22 rows selected.
    Index Details:+_
    SQL> select INDEX_OWNER,INDEX_NAME,COLUMN_NAME,TABLE_NAME from dba_ind_columns where
    2 table_name in ('INVOICE','ACCOUNT','ACCOUNT_HISTORY') order by 4;
    INDEX_OWNER INDEX_NAME COLUMN_NAME TABLE_NAME
    OPS$SVM_SRV4 P_ACCOUNT ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME ACCOUNT_NAME ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT CUSTOMER_NODE_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT ACCOUNT_TYPE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_ACCOUNT_TYPE ACCOUNT_TYPE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_INVOICE INVOICE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_PREVIOUS_INVOICE PREVIOUS_INVOICE_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME_ID ACCOUNT_NAME ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME_ID ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 I_LAST_MODIFIED_ACCOUNT LAST_MODIFIED ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_INVOICE_ACCOUNT INVOICE_ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ACCOUNT ACCOUNT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ACCOUNT SEQNR ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_INVOICE INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADINV INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA CURRENT_BALANCE ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA ACCOUNT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_LMOD LAST_MODIFIED ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADINV ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_PAYMENT PAYMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADJUSTMENT ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_APPLIED_DT APPLIED_DATE ACCOUNT_HISTORY
    OPS$SVM_SRV4 P_INVOICE INVOICE_ID INVOICE
    OPS$SVM_SRV4 U_INVOICE CUSTOMER_INVOICE_STR INVOICE
    OPS$SVM_SRV4 I_LAST_MODIFIED_INVOICE LAST_MODIFIED INVOICE
    OPS$SVM_SRV4 U_INVOICE_ACCOUNT ACCOUNT_ID INVOICE
    OPS$SVM_SRV4 U_INVOICE_ACCOUNT BILL_RUN_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_BILL_RUN BILL_RUN_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_INVOICE_TYPE INVOICE_TYPE_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_CUSTOMER_NODE CUSTOMER_NODE_ID INVOICE
    32 rows selected.
    Regards,
    Bathula
    Oracle-DBA

    I have some suggestions. But first, you realize that you have some redundant indexes, right? You have an index on account(account_name) and also account(account_name, account_id), and also account_history(invoice_id) and account_history(invoice_id, adjustment_id). No matter, I will suggest some new composite indexes.
    Also, you do not need two lines for these conditions:
    and round(a.account_balance, 2) > 0
    AND a.account_balance > 0
    You can just use: and a.account_balance >= 0.005
    So the formatted query isselect a.account_id,
           round(a.account_balance, 2) account_balance,
           nvl(ah.invoice_id, ah.adjustment_id) transaction_id,
           to_char(ah.effective_start_date, 'DD-MON-YYYY') transaction_date,
           to_char(nvl(i.payment_due_date, to_date('30-12-9999', 'dd-mm-yyyy')),
                   'DD-MON-YYYY') due_date,
           ah.current_balance - ah.previous_balance amount,
           decode(ah.invoice_id, null, 'A', 'I') transaction_type
      from account a, account_history ah, invoice i
    where a.account_id = ah.account_id
       and a.account_type_id = 1000002
       and (ah.invoice_id is not null or ah.adjustment_id is not null)
       and ah.CURRENT_BALANCE > ah.previous_balance
       and ah.invoice_id = i.invoice_id(+)
       AND a.account_balance >= .005
    order by a.account_id, ah.effective_start_date desc;You will probably want to select:
    1. From ACCOUNT first (your smaller table), for which you supply a literal on account_type_id. That should limit the accounts retrieved from ACCOUNT_HISTORY
    2. From ACCOUNT_HISTORY. We want to limit the records as much as possible on this table because of the outer join.
    3. INVOICE we want to access last because it seems to be least restricted, it is the biggest, and it has the outer join condition so it will manufacture rows to match as many rows as come back from account_history.
    Try the query above after creating the following composite indexes. The order of the columns is important:create index account_composite_i on account(account_type_id, account_balance, account_id);
    create index acct_history_comp_i on account_history(account_id, invoice_id, adjustment_id, current_balance, previous_balance, effective_start_date);
    create index invoice_composite_i on invoice(invoice_id, payment_due_date);All the columns used in the where clause will be indexed, in a logical order suited to the needs of the query. Plus each selected column is indexed as well so that we should not need to touch the tables at all to satisfy the query.
    Try the query after creating these indexes.
    A final suggestion is to try larger sort and hash area sizes and a manual workarea policy.alter session set workarea_size_policy = manual;
    alter session set sort_area_size = 2147483647;
    alter session set hash_area_size = 2147483647;

  • GRs created for schedule line of future date

    While making Goods Receipts using MIGO transaction, we are seeing tha GRs can be done for scheduling agreements having schedule line of future dates. Is there some way of restricting the GR upto to current date only. Meaning suppose if I have the following schedule lines:
    category of delivery date         delivery-date           schedule quantity
    M                                               01.2010                   100
    M                                               02.2010                    200
    M                                               03.2010                    150
    Currently the system is allowing to do GRs of quantity = 450 (100 + 200 + 150) in the month of February, 2010.
    But I want that while doing a GR in February, 2010 the system should allow to do GR of quantity 300 only (100 + 200). Please let me know how we can do so in the system.
    Regards,
    Pratima.

    Dear Veer,
    Your answer has helped in the MIGO transaction. Please let me know if something similar is possible for inbound deliveries also. Meaning like in MIGO transaction, I should not be able to make inbound deliveries (in xns VL31N) for schedule lines of future dates / months.
    Regards,
    Pratima.

  • Wrong data in MCE1 & MCE3 reports for scheduling agreements

    Hi All,
    I have got problem while executing MCE1 & MCE3 reports.It shows wrong PO value and GR Qty. While executing , I am giving inputs as P.Org,Vendor and Period in T.Code MCE1. After executing ,I select Info Structure S011.But the output is not matching with Actual GR quantity and PO Value if you see Purchase Order History.The same case with report MCE3 report,where Info Structure S012 is being selected.
    I have checked in MC26 , where data field for period determination is set as Purchase Order date.
    So Can anybody assist me on this ...How to get the exact data (PO value,Invoice amount,Order qty)which should match with PO history??
    Further to this, Is there any Standard report to get the value of materials bought via consignment??
    Please let me know if there is any customization setting involved??
    Thanks in advance..
    Somu

    1> On what logic is the system picking that calculation schema?
    As per "Define Schema Determination" in spro Purchasing document should pick the Pricing Procedure .
    After creation of scheduling agreement have chacked your values in Scheduling agreement (Purchase ORG ,Schema grp in Vendor)
    2> What settings can I do to make it pick that calculation schema only which has been assigned with POrg Sch Gp and Ve Sch Gp?
    Check in Vendor mater is there any Plant specific Purcahsing data maintained.If it is plant specific Check purchasing data for that Plant.

  • Database Variant to Data.vi not working for the Date datatype with LV 8.2?

    I'm moving a large body of LV database code from LV 7.1 to 8.2 and find that the Database Variant to Data.vi is not working correctly when used with the Date datatype. It works fine with 8.0, and the common Variant to Data works also. Am I missing something? Thanks in advance for any assistance. Wes

    Thanks for the prompt reply Crystal,
    The data is stored in an Oracle database using the DATE type. I'm querying many rows along with other columns and converting each of the values as necessary for each column with the 'Database Variant to Data' vi. Only conversion to Timestamp is no longer working as of version 8.2. I recognize that plain Variant to Data works but I have many (100's) of VIs to change if that is the only solution (not the end of the world). Most often the dates are originally generated in the database using PL/SQL procedures calling SYSDATE which look like: 5/1/2006 11:56:26 AM (in TOAD anyway) which I then need to read into LV as type Timestamp.
    Regards, Wes.

  • Labview How to specify 1d array of clusters as data types for variant to data

    Hi, I'm new to labview. Can anyone tell me how to specify 1d array of clusters as data types for variant to data?

    First of all, you should be sure that there is such a data type within the variant; otherwise, you will run into errors.
    I recommend you to create the cluster and create a type definition from it. Then drop an array shell from the array palette and drop the cluster type into that array.
    Connect that constant to the data type input of the Variant To Data function.
    Norbert
    CEO: What exactly is stopping us from doing this?
    Expert: Geometry
    Marketing Manager: Just ignore it.

  • Changing time-out for scheduled data refresh

    Using a Power Query connection, is it possible to extend the time-out time for scheduled data refreshes? The amount of data to be retrieved is rather limited, but there's thousands of rows (NAV server).
    If not, any suggestions to how to reduce latency?
    Thanks.

    Thorm,
    Is this still an issue?
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • Dynamic date parameters for a scheduled report that runs every day

    I have a report that has input parameters for a start date and
    end date. The report will be scheduled to run once a day at a
    scheduled time. I need the report to automatically set the start
    date and end dates when the report runs at the scheduled time.
    When the report runs at the scheduled time, the end date needs
    to be set to that time and the start date needs to be set to 24
    hrs before the end date. I want the start and end dates to roll
    forward automatically when the scheduled report runs each day. I
    don't want any user intervention to change the start and end
    dates every day. How can I do this in a report ?
    Also, since the report is scheduled to run each day, I want the
    output to be saved to a unique file name each time the report is
    run as each day's report will report on the data only for that
    day. How can I get the report to save to a unique file name
    automatically each day ? Again, I don't want any user
    intervention where they would set the output file name manually.

    Try adding the following in the beforereport or beforepform
    trigger, before the RETURN clause (TRUNC gets rid of the time
    component of the date):
    :p_date_from := TRUNC(sysdate) - 1;
    :p_date_to := TRUNC(sysdate);
    :desname := 'REPORT'||TO_CHAR(sysdate, 'DDMMYYYY')||'.txt'
    Paul Williams

  • Taking much time for loading data

    Dear All
    While loading data in <b>PRD system</b> (for master data and transaction data), it's taking much time. For ex., 2LIS_05_QOITM, I am loading delta data only in <b>PSA</b> from R/3. Some times (yesterday) it's taking 2 mins and some times (today) it's taking 5 hrs not yet completed (it's yellow). Actually yesterday, we went to SM58 in R/3 side and executed that LUW for some other data sources. Now also we can do but we don't want to like that. Because we are expecting permanent soulution. Could u please advice me. I am getting below message in status tab
    Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the source system.
    I am loading data through Process chain and user is <b>BWREMOTE (authorized user).</b>
    Please help me.
    Thanks a lot in advance
    Raja

    Dear Karthik
    No I couldn't resolve till now
    But Everything is fine.
    Now status is yellow only (209 from 209). Now what i want to do.
    Im getting below message <b>in status tab</b>
    Missing data packages for PSA Table
    Diagnosis
    Data packets are missing from PSA Table . BI processing does not return any errors. The data transport from the source system to BI was probably incorrect.
    Procedure
    Check the tRFC overview in the source system.
    You access this log using the wizard or following the menu path "Environment -> Transact. RFC -> Source System".
    Error handling:
    If the tRFC is incorrect, resolve the errors listed there.
    Check that the source system is connected properly to BI. In particular, check the remote user authorizations in BI.
    <b>In detail tab</b>, I am getting below message
    Info IDoc 2 : sent, not arrived ; IDoc ready for dispatch (ALE service)
    Thanks in advance
    Raja

  • Impdp taking long time for only few MBs data...

    Hi All,
    I have one query related to impdp. I have one expdp file and size is 47M. When I restore this dmp using impdp it will take more time. Also initially table_data loaded finsih very fast but then later on alter function/procedure/view taking a lot time almost 4 to 5 hrs.
    I have no idea why its taking long time... Earlier I can see one DB link is failed and it given error for "TNS name could not resolved" so I create DB link also before run impdp but the same result. Can any one suggest what could be the cause to taking long time for only 47MB data?
    Note - Both expdp and impdp database version is 11.2.0.3.0. If I import the same dmp file into 11.2.0.1.0 then its done in few mins.
    Thanks...

    Also Read
    Checklist For Slow Performance Of DataPump Export (expdp) And Import (impdp) [ID 453895.1]
    DataPump Import (IMPDP) is Very Slow at Object/System/Role Grants, Default Roles [ID 1267951.1]

Maybe you are looking for