Fill Internal Table for Table Header EQUI - Long Time

Good morning,
I have installed DMIS (2006_1_620) and DMIS_CNT (2006_1_620), both at the level of SP = 13.
I am running the subactivity "Fill Internal Table for Table Header EQUI" and have 1 day + 10 hours and continue working.
Also note that the table CNVTDMS_05_equi  is increased million in millions and now has 6 million records.
It is normal to take a long time? (EQUI table has 240 million records and therefore will take many days).
Any suggestions?
Thanks,
Hugo

Hello Hugo,
I know this is old , but did you ever find a solution to this.  I am having the same problem.  I considered adding an index to table QMIH, with Keys MANDT and BEQUI.
I also only have one bckgrd proc running in sender systems with this statement when I drill in using SM50:
SELECT
/*+
  FIRST_ROWS (1)
FROM
  "QMIH"
WHERE
  "MANDT"=:A0 AND "BEQUI"=:A1 AND ROWNUM <=:A2 #
This has been running for over 8hrs now.
Any thoughts?
Thanks,
NICK

Similar Messages

  • Tables for BOM Header and Components

    Dear Friends,
    Please let me know which tables provide a link between the header Material and corresponding Component Materials of a BOM?
    I have seen STKO and STPO but couldn't find such a link in them. STPO gives the component material but I couldn't get any link to Header material.
    Regards,
    Alok.

    Hello Alok,
    Table for BOM header             - STKO
    Table for BOM item                - STPO
    Table for material to BOM link - MAST
    Check if MARC table has a material in it and the same material is available in MAST also, it means that BOM has been created for that material..
    If Material is available in MARC and not available in MAST, it means BOM has not been created for that material.
    Reward Points, if useful.
    Regards,
    Nitin.

  • RSRV: Oracle statistics info table for table /BIC/FCube1 may be obsolete

    Hi,
    I run RSRV on cube1 and got several yellow lines such as:
    Oracle statistics info table for table /BIC/FCube1 may be OBSOLETE
    Oracle statistics info table for table /BIC/SSalesOrg may be OBSOLETE
    Oracle statistics info table for table /BIC/DCube1P may be OBSOLETE
    DB statistics for Infocube Cube1 must be recreated.
    I read here on SDN that running Correct Error in RSRV is only a temporay fix and that the best solution is to fix it from the database level with BRCONNECT.
    But the DBA says she has already run BRCONNECT yet there was no change in these several of the lines which came out as Yellow.  ... still same OBSOLETE messages.
    1. Any additional sugestions to fix these problems at the database level?
    2. In case I decide to fix with Correct Error in RSRV, what issues can I encounter with the cube?
    Can this lead to a failure of the cube?
    Will users encounter any issues with report?
    Does fixing the OBSOLETE in the error message in RSRV have any hazzards?
    Thanks

    Hi,
    it is years of data but how do you decide that the data is huge to warrant creating a new cube?
    You noted that
    "verify if it makes sense to copy the data into a new cube "
    How do I verify that?
    Is creating a new cube the only solution to this OBSOLETE problem?
    Why is it referring to only particular tables as OBSOLETE and doesn't that indicate that this is not a problem with the overall cube?
    Thanks

  • JBO-28004: Could not lock row in control table for table null

    hello.
    we have deployed two application on a server.
    both have code like this:
         select = "SELECT * from tab";
    vo = e.getDBTransaction().createViewObjectFromQueryStmt(select);
                        vo.executeQuery();
                        RowSet rs = vo.getRowSet();
                        if(rs != null)
                             Row lastRow = vo.getRowSet().last();
    the last line throws sometimes an exception (.last()).
    JBO-28004: Could not lock row in control table for table null
    at oracle.jbo.PCollException.throwException(PCollException.java:34)
    at oracle.jbo.pcoll.OraclePersistManager.commit(OraclePersistManager.java:229)
    at oracle.jbo.pcoll.OraclePersistManager.dropTable(OraclePersistManager.java:499)
    at oracle.jbo.pcoll.OraclePersistManager.createTable(OraclePersistManager.java:692)
    at oracle.jbo.pcoll.OraclePersistManager.insert(OraclePersistManager.java:1492)
    at oracle.jbo.pcoll.PCollNode.passivateElem(PCollNode.java:542)
    at oracle.jbo.pcoll.PCollNode.passivate(PCollNode.java:657)
    at oracle.jbo.pcoll.PCollection.passivateLRULeafNode(PCollection.java:351)
    at oracle.jbo.pcoll.PCollection.checkActiveLeafLimit(PCollection.java:404)
    at oracle.jbo.pcoll.PCollection.nodeRecentlyUsed(PCollection.java:252)
    at oracle.jbo.pcoll.PCollNode.<init>(PCollNode.java:73)
    at oracle.jbo.pcoll.PCollNode.checkForSplit(PCollNode.java:1553)
    at oracle.jbo.pcoll.PCollNode.addObject(PCollNode.java:1622)
    at oracle.jbo.pcoll.PCollNode.addObject(PCollNode.java:1610)
    at oracle.jbo.pcoll.PCollection.addElement(PCollection.java:753)
    at oracle.jbo.server.QueryCollection.fetch(QueryCollection.java:1247)
    at oracle.jbo.server.QueryCollection.getRowCount(QueryCollection.java:1065)
    at oracle.jbo.server.ViewRowSetImpl.getRowCount(ViewRowSetImpl.java:1444)
    at oracle.jbo.server.ViewRowSetIteratorImpl.last(ViewRowSetIteratorImpl.java:1183)
    at oracle.jbo.server.ViewRowSetImpl.last(ViewRowSetImpl.java:2246)
    at oracle.jbo.server.ViewObjectImpl.last(ViewObjectImpl.java:4352)
    at com.omv.emis.extemis.bc.EmsFreigabeImpl.checkFullMonth(EmsFreigabeImpl.java:327)
    at com.omv.emis.extemis.bc.EmsFreigabeImpl.doDML(EmsFreigabeImpl.java:293)
    at oracle.jbo.server.EntityImpl.postChanges(EntityImpl.java:3410)
    at oracle.jbo.server.DBTransactionImpl.doPostTransactionListeners(DBTransactionImpl.java:2274)
    at oracle.jbo.server.DBTransactionImpl.postChanges(DBTransactionImpl.java:2216)
    at oracle.jbo.server.DBTransactionImpl.commitInternal(DBTransactionImpl.java:1511)
    at oracle.jbo.server.DBTransactionImpl.commit(DBTransactionImpl.java:1677)
    at EmsFreigabeView1_BrowseEdit._jspService(_EmsFreigabeView1__BrowseEdit.java:134)
    at com.orionserver.http.OrionHttpJspPage.service(OrionHttpJspPage.java:56)
    at oracle.jsp.runtimev2.JspPageTable.service(JspPageTable.java:317)
    at oracle.jsp.runtimev2.JspServlet.internalService(JspServlet.java:465)
    at oracle.jsp.runtimev2.JspServlet.service(JspServlet.java:379)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
    at com.evermind.server.http.ResourceFilterChain.doFilter(ResourceFilterChain.java:64)
    at oracle.security.jazn.oc4j.JAZNFilter.doFilter(JAZNFilter.java:283)
    at com.evermind.server.http.ServletRequestDispatcher.invoke(ServletRequestDispatcher.java:563)
    at com.evermind.server.http.ServletRequestDispatcher.forwardInternal(ServletRequestDispatcher.java:309)
    at com.evermind.server.http.HttpRequestHandler.processRequest(HttpRequestHandler.java:767)
    at com.evermind.server.http.HttpRequestHandler.run(HttpRequestHandler.java:259)
    at com.evermind.server.http.HttpRequestHandler.run(HttpRequestHandler.java:106)
    this exception does not occur every time just sometimes.
    (no other user/application uses the specific table, so there is no other lock)
    maybe someone knows an error source.
    thx in advance.

    Gina:
    Could you run you app with diagnostic turned on?
    If you're launching your middle tier from command prompt, you should invoke java as
    java -Djbo.debugoutput=console ...
    If you're launching your middle tier from w/i JDev..
    1. Select the project.
    2. Do right mouse click and select "Project Settings..."
    3. On the Settings dialog, select Configurations/Runner.
    4. In the righthand side pane, you should see a textbox for "Java
    Options". Please add the following JVM switch:
    -Djbo.debugoutput=console
    Then, rerun. The run command should include
    -Djbo.debugoutput=console as in
    "D:\JDev9i\jdk\bin\javaw.exe" -Djbo.debugoutput=console -classpath ...
    Then, you should see a lot of diagnostic messages. It will also give you complete stack trace including that of the detail exception.
    When you get the exception diagnostic output, please post that to this thread.
    Thanks.
    Sung

  • HT1351 for syncing it takes long time and not completed what to do???

    for syncing it takes long time and not completed what to do???

    Debbie:
    deborahfromwindsor wrote:
    he advises restarting by inserting the OSX disc and pressing down the C button to reboot from there then selecting disk utility, hard disk and repair.... Does he mean me to hold down the C key on the alpha keyboard or the ctrl key?
    Should I just ask for my money back??? If it is a simple repair do I just literally push the disc in, push the power button and hold down the C button?
    That's where I would begin, too, with
    Repair Disk
    Insert Installer disk and Restart, holding down the "C" key until grey Apple appears.
    Go to Installer menu (Panther and earlier) or Utilities menu (Tiger) and launch Disk Utility.
    Select your HDD (manufacturer ID) in the left panel.
    Select First Aid in the Main panel.
    (Check S.M.A.R.TStatus of HDD at the bottom of right panel, and report if it saysanything but Verified)
    Click Repair Disk on the bottom right.
    If DU reports disk does not need repairs quit DU and restart.
    If DU reports errors Repair again and again until DU reports disk is repaired.
    If DU reports errors it cannot repair you will need touse autility like TechTool Pro,Drive Geniusor DiskWarrior
    First we need to determine if the issue you are experiencing with the computer is software or hardware based. Once we have gotten things sorted out there should be time enough to make you decision about keeping or returning it.
    cornelius

  • Table for PO-header node text

    Hi all,
    In which table the po-header node text will gets saved.
    Thanks in advance.
    Thanks,
    Asmitha.

    Hi
    No where you will find the entire Long text of this type
    All texts will be stored with the Paramters like OBJECT,ID,NAME,LANG in STXH tables
    You have to use READ_TEXT fun module to fetch this text by passing the above 4 parameters.
    Double click on the text, from the Menu GOTO-> Header
    will see the above 4 paramters
    pass them to fetch this text
    READ_TEXT
    READ_TEXT provides a text for the application program in the specified work areas.
    The function module reads the desired text from the text file, the text memory, or the archive. You must fully specify the text using OBJECT, NAME, ID, and LANGUAGE. An internal work area can hold only one text; therefore, generic specifications are not allowed with these options.
    After successful reading, the system places header information and text lines into the work areas specified with HEADER and LINES.
    If a reference text is used, SAPscript automatically processes the reference chain and provides the text lines found in the text at the end of the chain. If an error occurs, the system leaves the function module and triggers the exception REFERENCE_CHECK.
    Function call:
    CALL FUNCTION 'READ_TEXT'
    EXPORTING CLIENT = SY-MANDT
    OBJECT = ?...
    NAME = ?...
    ID = ?...
    LANGUAGE = ?...
    ARCHIVE_HANDLE = 0
    IMPORTING HEADER =
    TABLES LINES = ?...
    EXCEPTIONS ID =
    LANGUAGE =
    NAME =
    NOT_FOUND =
    OBJECT =
    REFERENCE_CHECK =
    WRONG_ACCESS_TO_ARCHIVE =
    Export parameters:
    CLIENT
    Specify the client under which the text is stored. If you omit this parameter, the system uses the current client as default.
    Reference field: SY-MANDT
    Default value: SY-MANDT
    OBJECT
    Enter the name of the text object to which the text is allocated. Table TTXOB contains the valid objects.
    Reference field: THEAD-TDOBJECT
    NAME
    Enter the name of the text module. The name may be up to 70 characters long. Its internal structure depends on the text object used.
    Reference field: THEAD-TDNAME
    ID
    Enter the text ID of the text module. Table TTXID contains the valid text IDs, depending on the text object.
    Reference field: THEAD-TDID
    LANGUAGE
    Enter the language key of the text module. The system accepts only languages that are defined in table T002.
    Reference field: THEAD-TDSPRAS
    ARCHIVE_HANDLE
    If you want to read the text from the archive, you must enter a handle here. The system uses it to access the archive. You can create the handle using the function module ACHIVE_OPEN_FOR_READ.
    The value '0' indicates that you do not want to read the text from the archive.
    Reference field: SY-TABIX
    Default value: 0
    Import parameters:
    HEADER
    If the system finds the desired text, it returns the text header in this parameter.
    Structure: THEAD
    Table parameters:
    LINES
    The table contains all text lines that belong to the text read.
    Structure: TLINE
    Exceptions:
    ID
    The text ID specified in the parameter ID does not exist in table TTXID. It must be defined there together with the object of the text module.
    LANGUAGE
    The parameter LANGUAGE contains a language key that does not exist in table T002.
    NAME
    The parameter NAME contains the name of a text module that does not correspond to the SAPscript conventions.
    Possible errors:
    The field contains only blanks.
    The field contains the invalid characters ‘*’ or ‘,’.
    OBJECT
    The parameter OBJECT contains the name of a text object that does not exist in table TTXOB.
    NOT_FOUND
    The system did not find the specified text module.
    REFERENCE_CHECK
    The text module to be read has no text lines of its own but refers to the lines of another text module. This reference chain can include several levels. For the current text, the chain is interrupted, that is, one of the text modules referred to in the chain no longer exists.
    WRONG_ACCESS_ TO_ARCHIVE
    The exception WRONG_ACCESS_TO_ARCHIVE is triggered if an archive is accessed using an incorrect or non-existing archive handle or an incorrect mode (that is, read if the archive is open for writing or vice versa).
    <b>Reward points for useful Answers</b>
    Regards
    Anji

  • Setup tables for invoices header data (2LIS_13_VDHDR)

    Hi experts,
    I am using 2LIS_13_VAHDR and 2LIS_13_VAITM to transfer data to SAP BW.  Through the t-code SBIW, I I run the job to fill setup tables with invoices data. the job is completed succefully. when I check the extractors in RSA3, the extractor 2LIS_13_VAHDR doesn't bring any data while 2LIS_13_VAITM bring a set of items !!
    I m wondering why I dont get invoices header data!
    Thank's a lot for your help.
    Abdess,

    Hi,
    Did you delete the old initialization request from BW and also delete old setup table.
    Because the issue is strange, If you are able to see entries in item then header should also fetch data.
    Repeat the steps again if you have time.
    Use T-Code LBWG to delete setup table.
    Delete MCXE13 entries in LBWQ.
    Use T-Code OLI9BW to fillup setup table for 2lis_13.
    Thanks,
    Shakthi Raj Natarajan.

  • Using Word Easy Table Under Report Generation takes long time to add data points to table and generate report

    Hi All,
    We used report generation tool kit to generate the report on word and with other API 's under it,we get good reports .
    But when the data points are more (> 100 on all channels) it take a long time  to write all data and create a table in the word and generate report.
    Any sugegstions how to  make this happen in some seconds .
    Please assist.

    Well, I just tried my suggestion.  I simulated a 24-channel data producer (I actually generated 25 numbers -- the first number was the row number, followed by 24 random numbers) and generated 100 of these for a total of 2500 double-precision values.  I then saved this table to Excel and closed the file.  I then opened Word (all using RGT), wrote a single text line "Text with Excel", inserted the previously-created "Excel Object", and saved and closed Word.
    First, it worked (sort of).  The Table in Word started on a new page, and was in a very tiny font (possibly trying to fit 25 columns on a page?  I didn't inspect it very carefully).  This is probably "too much data" to really try to write the whole table, unless you format it for, say, 3 significant figures.
    Now, timing.  I ran this four times, two duplicate sets, one with Excel and Word in "normal" mode, one in "minimized".  To my surprise, this didn't make a lot of difference (minimized was less than 10% faster).  Here are the approximate times:
         Generate the data -- about 1 millisecond.
         Write the Excel Report -- about 1.5 seconds
         Write the Word Report -- about 10.5 seconds
    Seems to me this is way faster than trying to do this directly in Word.
    Bob Schor

  • Table for Table descriptions

    Hi All,
    Does anyone have any idea - is there any table that stores the descriptions of the tables?
    Example: MARA: General Material Data
                   VBAK: Sales Document: Header Data
    Can anyone tell me where can i get these descriptions?
    Actually my requirement is i need to get the descriptions for the transaction VK15 key combination?
    Like    SalesOrg/Material
              Customer/Material etc.. The corresponding tables for these are A902 ... etc.. and these are descriptions for these tables.
    Can you please suggest?
    Thanks.
    Regards.

    Hi SM,
      For table text :
    Table :DD02T : R/3 DD: SAP table texts
    Text field: DDTEXT  :Short Text Describing R/3 Repository Objects
      For Table:
    DD02L : SAP Tables
    With luck,
    Pritam.

  • Updating a table using cursor is taking long time in oracle

    Hi,
    I am working on the oracle database 11gR2. I am trying update a table column which is newly added through cursor. below is the cursor.
    BEGIN
         FOR lcur_tab IN (select l.LOGIN_ID as login_id, lt.ERROR_CODE as error_code from LOGIN l INNER JOIN LOGIN_TASK lt ON (l.LAST_LOGIN_TASK_ID = lt.LOGIN_TASK_ID) )
         LOOP
                   UPDATE ACCOUNT SET LOGIN_ERROR_CODE = lcur_tab.error_code where ACCOUNT_ID IN (SELECT ACCOUNT_ID FROM LOGIN_ACCOUNT where LOGIN_ID = lcur_tab.login_id ) ;
         end LOOP;
         commit;
    END;
    In the cursor we are trying to copy the Login table error_code value to LOGIN_ERROR_CODE of ACCOUNT Table, for the accounts associated with that Login. Here newly added column is LOGIN_ERROR_CODE in account table. this account table have 11million rows in it. when i used the above cursor it took 3hours and still running. So we stopped after that. Is there any way i can change this cursor syntax to update those column values. How can we do this type of updates?
    Also i am planning to run thus update in the background . I don't have much knowledge on the PL/SQL stuff. So please help on this.
    Thanks in advance.
    Edited by: Hari on Mar 21, 2012 10:06 PM

    Is there a reason that you don't want to do this with a single `UPDATE` statement? That should be quite a bit more efficient.
    Something like
    UPDATE account a
       SET login_error_code = (
          select lt.error_code
            from login_task lt
                 join login l on (l.last_login_task_id = lt.login_task_id)
                 join login_account la on (la.login_id = l.login_id)
           where la.account_id = a.account_id )
    WHERE EXISTS (
          select 1
            from login_task lt
                 join login l on (l.last_login_task_id = lt.login_task_id)
                 join login_account la on (la.login_id = l.login_id)
           where la.account_id = a.account_id );Justin

  • Using Table.Join formula takes extremly long time to get results.(Why Query Folding doesn't work?)

    Hi,
    I built a query with 4 tables inside (load from Oracle DB and two of them are quite big, more than millions of rows). After filtering, I tried to build relationships between tables using Table.Join formula. However, the process took extremly long time to
    bring out results (I ended the process after 15 mins' processing). There's a status bar kept updating while the query was processing, which is showed as  . I suppose
    this is because the query folding didn't working, so PQ had to load all the data to local memory first then do the opertion, instead of doing all the work on the source system side. Am I right? If yes, is there any ways to solve this issue?
    Thanks.
    Regards,
    Qilong 

    Hi Curt,
    Here's the query that I'm refering,
    let
        Source = Oracle.Database("reporting"),
        AOLOT_HISTS = Source{[Schema="GEN",Item="MVIEW$_AOLOT_HISTS"]}[Data],
        WORK_WEEK = Source{[Schema="GEN",Item="WORK_WEEK"]}[Data],
        DEVICES = Source{[Schema="GEN",Item="MVIEW$_DEVICES"]}[Data],
        AO_LOTS = Source{[Schema="GEN",Item="MVIEW$_AO_LOTS"]}[Data],
        Filter_WorkWeek = Table.SelectRows(WORK_WEEK, each ([WRWK_YEAR] = 2015) and (([WORK_WEEK] = 1) or ([WORK_WEEK] = 2) or ([WORK_WEEK] = 3))), 
        Filter_AlotHists = Table.SelectRows(AOLOT_HISTS, each ([STEP_NAME] = "BAKE" or [STEP_NAME] = "COLD TEST-IFLEX" or [STEP_NAME] = "COLD TEST-MFLEX") and ([OUT_QUANTITY] <> 0)),
        #"Added Custom" = Table.AddColumn(Filter_AlotHists, "Custom", each Table.SelectRows(Filter_WorkWeek, (table2Row) => [PROCESS_END_TIME] >= table2Row[WRWK_START_DATE] and [PROCESS_END_TIME] <= table2Row[WRWK_END_DATE])),
        #"Expand Custom" = Table.ExpandTableColumn(#"Added Custom", "Custom", {"WRWK_YEAR", "WORK_WEEK", "WRWK_START_DATE", "WRWK_END_DATE"}, {"WRWK_YEAR", "WORK_WEEK",
    "WRWK_START_DATE", "WRWK_END_DATE"}),
        Filter_AolotHists_byWeek = Table.SelectRows(#"Expand Custom", each ([WORK_WEEK] <> null)),
        SelectColumns_AolotHists = Table.SelectColumns(Filter_AolotHists_byWeek,{"ALOT_NUMBER", "STEP_NAME", "PROCESS_START_TIME", "PROCESS_END_TIME", "START_QUANTITY", "OUT_QUANTITY", "REJECT_QUANTITY",
    "WRWK_FISCAL_YEAR", "WRWK_WORK_WEEK_NO"}),
        Filter_Devices= Table.SelectRows(DEVICES, each ([DEPARTMENT] = "TEST1")),
        SelectColumns_Devices = Table.SelectColumns(Filter_Devices,{"DEVC_NUMBER", "PCKG_CODE"}),
        Filter_AoLots = Table.SelectRows(AO_LOTS, each Text.Contains([DEVC_NUMBER], "MC09XS3400AFK") or Text.Contains([DEVC_NUMBER], "MC09XS3400AFKR2") or Text.Contains([DEVC_NUMBER], "MC10XS3412CHFK") or Text.Contains([DEVC_NUMBER],
    "MC10XS3412CHFKR2")),
        SelectColumns_AoLots = Table.SelectColumns(Filter_AoLots,{"ALOT_NUMBER", "DEVC_NUMBER", "TRACECODE", "WAFERLOTNUMBER"}),
        TableJoin = Table.Join(SelectColumns_AolotHists, "ALOT_NUMBER", Table.PrefixColumns(SelectColumns_AoLots, "AoLots"), "AoLots.ALOT_NUMBER"),
        TableJoin1 = Table.Join(TableJoin, "AoLots.DEVC_NUMBER", Table.PrefixColumns(SelectColumns_Devices, "Devices"), "Devices.DEVC_NUMBER")
    in
        TableJoin1
    Could you please give me some hints why it needs so long to process?
    Thanks.

  • Find table for table maintenance

    Say you’re in transaction mm03 to display a material.  How do you find the table you can perform table maintenance on to edit selections for field MVGR5?

    Good question.  Here is the way that I found it just know.   I went to SE11 and looked at the MVKE table and saw that the check table for MVGR5 is TVM5,  then I went to the intial screen of SE11 and changed the radiobutton to view, and did F4 help, here I hit the expand view icon and entered TVM5 in as the primary table and checked "Maintainence view",  it returned  V_TVM5 as the maintenance view.  I entered this view name into SM30 and that's it.
    I don't think that it is always this cut and dry.
    Regards
    RIch Heilman

  • Table for table maintenance

    Hi SDN's,
    Which table gives the information if a Z table has table maintenance or not?
    Which means i need the list of all Z tables which do not have a table maintenance generated?
    Regards
    Pratyusha

    Hi,
       For checking if table maintenance is allowed:
    You can query the table DD02L for TABNAME = Z* where MAINFLAG = 'X' (Display/Maintenance Allowed) or SPACE (Display/Maintenance Allowed with Restrictions).
    If you want to check if there is a table maintenance is actually defined:
    Query the table TVDIR with TABNAME = Z* (Both Z Tables and Z Views)
    Regards
    Kiran Sure

  • Crystal reports for Eclipse (CR4E)  - taking long time to load

    HI
    We have used your Crystal reports for Eclipse (CR4E) but it is taking long time to load if the record count is more than 3 million
    Our architecture is we are using Oracle DB and web sphere application server both are running on SUN server,
    If we buy original version it will work fast?
    Or do you  have alternative solution for this, this one to implement one of our major client

    You have 3 million records, this is going to take time to process.  Also, you don't define what is meant by a long time.  Are we talking a couple of minutes, or something more like 20 minutes? 
    The real question is, do you really need the report to bring back 3 million records?  Could you do any filtering of the data to reduce the number of records being brought back?  Also, 3 million records could vary.  is each record a single field, or is it multiple.   The more fields, the larger the result set.
    Another thing you want to look at, is just how long is the query taking to run against the database.

  • Fill Internal Header Table for Table VBUK corrupted

    Hi all
    I had Cluster on each table ( VBUK, VBFCL... )
    SP13 all need notes are implemented,
    Parameter P_ROLL = 'X'
                     P_CUST = 'Y'
    but system dumpt all 6 hours Remote error
    one question does für VBUK fill also gave a Paramter p_full ?
    thanks for feedback or other inputs
    L

    TDMS experts:
    I remember that I indeed created an index for VBFA per that famous OSS note.
    However because I have changed to a different project so that I cannot verify.
    Additionally, this link also almost explicitly says that index on VBFA is feasible:
    Re: Performance issue during TD05X_FILL_VBUK_2 and TIME OUT
    So why in the previous posting Lou says:
    1256679 -> VBFA is a cluster table, i can't create a Index ...
    I appreciate if you experts can share your experience on this.
    Regards!

Maybe you are looking for