REG MVIEW

Hello All,
I have created one maintenance view for a particular table, in the view field block i have defined some fields.
When i execute this view, i can see all the fields except one field.
Can anybody help me tht y only one field is not displayed ???
Thanks,
Vibha

solved by my own

Similar Messages

  • Reg: Check View/MView query -

    Hi Experts,
    I'm trying to search for all references of a Table (in ALL_SOURCE). But not able to check into the query of VIEWs and MVIEWs.
    ALL_VIEWS and ALL_MVIEWS have LONG columns called TEXT and QUERY respectively.
    Any pointers how can I search into these columns?
    Help much appreciated!
    Ranit B.

    You can use all_dependencies
    SELECT *
      FROM all_dependencies
    WHERE type = 'VIEW'
       AND referenced_owner = 'TABLE_OWNER'
       AND referenced_name = 'YOUR_TABLE';Or you can have a look at Long to Varchar2 conversion....

  • FAST enabled MView takes ages to refresh

    Hi All,
    I have a materlialized view which is based on a single fact table ( OAS_BALANCE_UNIONED 11,445,156 rows) and a bunch of dimension tables each containing limited number of rows (a typical start schema).
    The conditioins to make the mview fast refreshable have been met so the mview created successfully with the FAST option.
    The mview then worked just fine after than and had been refreshing FAST (6 minutes) for the whole day till other mviews started to refresh in the same time. Some of those mviews read from some of the base tables of the mview in question. Since then, the mivew has started to take ages (3 hours) to refresh. There is surely something wrong because the COMPLETE refresh was faster than that.
    Later, I stopped all the refresh jobs of the other mviews. Yet, the FAST mview still has the same problem of taking ages to refresh.
    When I monitored the session that refreshes the mview, I noticed in the "Long Ops" that
    most of the time is spent on a step named as "Hash Join" (it appears in the Message column). Hash join occurrs because of the outer join.
    Why my FAST mview suddenly started to take so long time to refresh?
    How can I make my FAST mivew really refresh fast?
    Technical Details are below:
    OS: Windows 2003 32-bit
    DB version: Oracle 10g R1 ( 10.1.0.2.0 )
    CREATE MATERIALIZED VIEW FACT_BAL_MV
    BUILD DEFERRED
    REFRESH FAST ON DEMAND
    WITH PRIMARY KEY
    AS
    SELECT B.cmpcode  BAL_KEY,
           B.CMPCODE,
           B.YR,
           B.PERIOD,
           B.BALCODE,
           B.CURCODE,
           B.REPBASIS,
           TO_NUMBER (B.YR || LPAD (B.PERIOD, 2, 0)) DIM_PERIOD_MV,
           D_CMP.DIM_KEY DIM_COMPANY,
           D_CUR.DIM_KEY DIM_CURRENCY,
           D_EL_1.DIM_KEY DIM_EL_1,
           D_EL_2.DIM_KEY DIM_EL_2,
           D_EL_3.DIM_KEY DIM_EL_3,
           D_EL_4.DIM_KEY DIM_EL_4,
           GRP_1_P_EL.DIM_KEY DIM_GRP_1_P_EL,
           D_GRP_1_B_EL.DIM_KEY DIM_GRP_1_B_EL,
           D_GRP_1_K_EL.DIM_KEY DIM_GRP_1_K_EL,
           D_GRP_2_D_EL.DIM_KEY DIM_GRP_2_D_EL,
           D_GRP_3_E_EL.DIM_KEY D_GRP_3_E_EL,
           B.ROWID X_OB_ROWID,
           D_CMP.ROWID X_CMP_ROWID,
           D_CUR.ROWID X_CUR_ROWID,
           D_EL_1.ROWID X_EL_1_ROWID,
           D_EL_2.ROWID X_EL_2_ROWID,
           D_EL_3.ROWID X_EL_3_ROWID,
           D_EL_4.ROWID X_EL_4_ROWID,
           GRP_1_P_EL.ROWID X_GRP_1_P_EL_ROWID,
           GRP_1_P.ROWID X_GRP_1_P_ROWID,
           D_GRP_1_B_EL.ROWID X_GRP_1_B_EL_ROWID,
           D_GRP_1_B.ROWID X_GRP_1_B_ROWID,
           D_GRP_1_K_EL.ROWID X_GRP_1_K_EL_ROWID,
           D_GRP_1_K.ROWID X_GRP_1_K_ROWID,
           D_GRP_2_D_EL.ROWID X_GRP_2_D_EL_ROWID,
           D_GRP_2_D.ROWID X_GRP_2_D_ROWID,
           D_GRP_3_E_EL.ROWID X_GRP_3_E_EL_ROWID,
           D_GRP_3_E.ROWID X_GRP_3_E_ROWID
      FROM CLVE_STAGING.OAS_BALANCE_UNIONED B,
           FINANCE.DW_DIM_COMPANY D_CMP,
           FINANCE.DW_DIM_CURRENCY D_CUR,
           FINANCE.DW_DIM_EL_1 D_EL_1,
           FINANCE.DW_DIM_EL_2 D_EL_2,
           FINANCE.DW_DIM_EL_3 D_EL_3,
           FINANCE.DW_DIM_EL_4 D_EL_4,
           FINANCE.DW_DIM_GRP_1_P_EL GRP_1_P_EL,
           FINANCE.DW_DIM_GRP_1_P GRP_1_P,
           FINANCE.DW_DIM_GRP_1_B_EL D_GRP_1_B_EL,
           FINANCE.DW_DIM_GRP_1_B D_GRP_1_B,
           FINANCE.DW_DIM_GRP_1_K_EL D_GRP_1_K_EL,
           FINANCE.DW_DIM_GRP_1_K D_GRP_1_K,
           FINANCE.DW_DIM_GRP_2_D_EL D_GRP_2_D_EL,
           FINANCE.DW_DIM_GRP_2_D D_GRP_2_D,
           FINANCE.DW_DIM_GRP_3_E_EL D_GRP_3_E_EL,
           FINANCE.DW_DIM_GRP_3_E D_GRP_3_E
    WHERE     B.CMPCODE = D_CMP.CODE
           AND (B.CMPCODE = D_CUR.CMPCODE(+) AND B.CURCODE = D_CUR.CODE(+))
           AND (B.CMPCODE = D_EL_1.CMPCODE(+) AND B.EL1 = D_EL_1.EL_CODE(+))
           AND (B.CMPCODE = D_EL_2.CMPCODE(+) AND B.EL2 = D_EL_2.EL_CODE(+))
           AND (B.CMPCODE = D_EL_3.CMPCODE(+) AND B.EL3 = D_EL_3.EL_CODE(+))
           AND (B.CMPCODE = D_EL_4.CMPCODE(+) AND B.EL3 = D_EL_4.EL_CODE(+))
           AND (GRP_1_P_EL.CMPCODE = GRP_1_P.CMPCODE(+)
                AND GRP_1_P_EL.GRP_CODE = GRP_1_P.GRP_CODE(+))
           AND (B.CMPCODE = GRP_1_P_EL.CMPCODE(+)
                AND B.EL1 = GRP_1_P_EL.EL_CODE(+))
           AND (D_GRP_1_B_EL.CMPCODE = D_GRP_1_B.CMPCODE(+)
                AND D_GRP_1_B_EL.GRP_CODE = D_GRP_1_B.GRP_CODE(+))
           AND (B.CMPCODE = D_GRP_1_B_EL.CMPCODE(+)
                AND B.EL1 = D_GRP_1_B_EL.EL_CODE(+))
           AND (D_GRP_1_K_EL.CMPCODE = D_GRP_1_K.CMPCODE(+)
                AND D_GRP_1_K_EL.GRP_CODE = D_GRP_1_K.GRP_CODE(+))
           AND (B.CMPCODE = D_GRP_1_K_EL.CMPCODE(+)
                AND B.EL1 = D_GRP_1_K_EL.EL_CODE(+))
           AND (D_GRP_2_D_EL.CMPCODE = D_GRP_2_D.CMPCODE(+)
                AND D_GRP_2_D_EL.GRP_CODE = D_GRP_2_D.GRP_CODE(+))
           AND (B.CMPCODE = D_GRP_2_D_EL.CMPCODE(+)
                AND B.EL2 = D_GRP_2_D_EL.EL_CODE(+))
           AND (D_GRP_3_E_EL.CMPCODE = D_GRP_3_E.CMPCODE(+)
                AND D_GRP_3_E_EL.GRP_CODE = D_GRP_3_E.GRP_CODE(+))
           AND (B.CMPCODE = D_GRP_3_E_EL.CMPCODE(+)
                AND B.EL3 = D_GRP_3_E_EL.EL_CODE(+));
    CREATE INDEX FINANCE.DW_FACT_BAL_MV_INX
    ON FACT_BAL_MV (X_OB_ROWID);
    SQL> select count(X_OB_ROWID) from FACT_BAL_MV ;
    COUNT(X_OB_ROWID)
             11444816
    Elapsed: 00:00:27.85
    SQL> select bytes/1024/1024/1024 GB from
    user_segments where segment_name='FACT_BAL_MV';
            GB
    4.4296875search id: redcorolla

    Hard to say why the refresh is now slow. Possibilties include
    * system resources implemented by the fast refresh
    * something changed on the query making the background query run slowly
    Look at the query first. 18 tables, with almost every join an outer join. Some of the tables appear more than once in the query.
    Get an executioion plan to see how the query is being executed.

  • Mail can't send email from 123-reg accounts

    Hi all,
    I've got an issue with Mac Mail can’t send email from 123-reg accounts.
    I know there is a lot of articles on simular matters and some I have tried out with no luck, others are just speaking a whole other language.
    Any help would be greatly recived, here's my situation below:
    Running a iMac 2.8GHz Intel Core i5 on OS X 10.8.3
    I’m using 3 accounts on Macs Mail client, one is my iCloud account that is working fine and the other two are my 123-reg accounts that I use for business purposes. I’ve been working like this for a couple of years now and about 6 months ago I upgrade my operating system from Snow Leopard to Mountain Lion.
    2 or 3 months ago I randomly got a problem where I was unable to send any emails from my 123-reg account, I could only receive emails. I checked my settings and they were all correct and unchanged so I attempted to delete them and recreate the account but with no such luck.
    I then decided to check if it was a problem at 123-regs end and attempted to log into web mail. This worked and was able to both send and receive emails so eliminated that possibility.
    I also have these accounts set up on my iPhone, so checked if there was any issues and everything was working fine. I then checked my iphone details and made sure the same was on my desktop and still no such luck.
    I’ve tried out some of the articles I’ve found on here (the ones written in plain English) still no results. I’m not very IT illiterate and just wondered if anyone else has got any ideas, whether there is something specific with 123-reg accounts and Mac Mail?
    Many thanks.

    Thanks
    I just had another hunt around looking at posts and discovered that it was because I recently changed to BT Infinity and they had a block on.
    I didn't even think to question the ISP.
    https://discussions.apple.com/message/21173608#21173608

  • Use CONTEXT index on mview or use mview's rewrite directly

    Please let me explain what I mean in Subject. I have 9 tables. Each of these tables has about 40,000 rows and one table has 2 million rows. Using first approach, I can build a join-only materialized view on top of nine table's mview log. then query on these nine tables directly. Advantage for doing that is use rewrite.
    <p>
    for second approach, I build a CONETXT index on several columns on the mview, and then query through CONTEXT index on mview directly. This is pretty much like Barbara did on CREATE INDEX book_idx ON book_search
    [http]Indexing multiple columns of multiple tables using CTXCAT
    but she used CTXCAT instead of CONTEXT index.
    <p>
    My question is will second approach better than first one and why. Unlike basic join several tables which gives you predictable performance, I often feel that CONTEXT index performance is unpredictable when tables have more than several thousands rows (maybe I did something wrong, but still looking for document regarding performance) .
    <p>
    I will appreciate someone could show hints on the issue.
    <p>
    Message was edited by:
    qwe15933
    Message was edited by:
    qwe15933
    Message was edited by:
    qwe15933
    Message was edited by:
    qwe15933

    The best method to find out what is best for any individual situation is to test and compare. In general, Oracle Text is best at searching unstructured data, such as large documents and has a lot of features that enable you to do different kinds of searches. If you only have structured data, with each column only containing one short thing and you only need simple searches, then you probably don't need Text. There are also a few things that can be done to indexes and/or queries to optimize performance. It would help to have an idea what your typical data is like and what sorts of searches you anticipate needing.

  • Reg: Message and BPM

    Hi
    I'm having a scenario where i use a JDBC adapter to extract data from a DB. As a result of the query say, 10 rows are returned as message to XI server. I have a transformation (BPM) set and the receiver(target system) is a file. When i open the file to see its contents i can see only the first record been transfered. The mapping part used for the transformation node is having IF condition. But all the fetched records satisfies the conditions in the mapping.
    Pls let me know the corrective step.
    reg: Prabhu

    Hi,
    1)Check the input XML i.e all 10 records are coming into XI . This you can check in SXMB_MONI
    2) Then you can test the mapping in the integration Reposiotry .. So now you can get if any mapping problem. For this , check is the occurence of target strcuture is 1..n or 0.n
    /people/michal.krawczyk2/blog/2005/09/16/xi-how-to-test-your-mapping-in-real-life-scenarios
    3) If mapping is correct , then check the RWB->Message Monitoring->Message Display Tool and check the payload.
    4) If this is correct, then check the File COntent Conversion of the Receiver File Adapter.
    Hope this helps,
    Regards,
    Moorthy

  • Reg: How to Stop Polling of JDBC Adapter without Scheduling the adapter

    Dear Gurus,
    Here I am having one requirement. My clients wants to send data from JDBC adapter (ORcale System table) to R3 system via RFC.
    His Requirement::
    1. He is not telling the time of data flow from Oracle server so that based on that we can schedule the adapter in the Communication Channel monitoring (Availablitiy Time Planning) or Else we can Schedule by deciding the setting of the polling interval time.
    2. He is telling that When ever he waants to send the data he will place one dummy file in the File Adapter FTP location so that it will became an Indication for u to poll the jdbc adapter and to transfer the data to R3.
    3. Untill he keeps the file or gives indication he doesn;t want to communicate with Oracle server due to some security and it s a most important data base he doesn't want to disturb that Oracle Server as so many business are running  on that.....
    We Proposed::
    1. Atleast you need to tell the scheduling time or poll interval time so that we can schedule our adapter.
    but he s not accepting for this
    2. Atleast U need to give access for Data base to enter one more extra field like STATUS CODE so that we will add one number and we keep on Update in the Update table and based on that Update table statement it will poll.
    but he s not even accepting for this
    3. Finally we prposed that to create another table in the Oracle SYstem as Dulplicate Table which is similar to Standard Orginal table when ever he wants to pick the data please keep that data in this Duplicate TABle so that JDBC adapter will pick the data from thsi TABLE instead of picking the data from that standard table so that it will not effect any standard table data in the table.
    but he s not even accepting for this
    We have done some R & D:::
    1. WE approached even through BPM and via switch conditions is one scenario
       FILE RECEIVE >SWITCH CONDITION> RECEIVE AND SEND or else EXIT
    2. Using correlation in anotehr scenario means correlating File adapter and JDBC and based on one dynamica value it will goes to SEND STEP ( RECE IVE --> RECEIVE --> SEND STEPS )
    Even though we know this...concept that...we jsut tried::
    In BPM we can control the flow in XI 3.0 but we cannot Stop the Polling of JDBC adapter at backend because one the data comes from FILE adapter it will keep on HOLD untill it receives the JDBC from Oracle then based on the condition or Correlation it will goes futher SEND step means after that file adapter is picking file or not ...what ever it may be JDBC will polls at backend and brings that data to BPM"
    Hence sugest me How to Stop Polling of JDBC Adapter without Scheduling the adapter or else using STATUS CODE Update statements in JDBC Tables 
    Regards:
    Amar Srinivas Eli

    Hi! All,
    Finally I decided to do the scenario in two steps:
    1: FILE REQ --> JDBC REQ -->JDBC RES --> FILE RECV
    2: FILE RECV --> RFC
    But I am getting issue while doing first scenario
    Desgn :
    I have created 2 Synchronous interfaces :
    1) FILE 2 JDBC REQ
    In this a) out put message is FILE  Req
              b) Input msage:; FILE RES
    2} JDBC2FILE RECV
            a) Output mesage;; JDBC REQ
            b) Input Msge :: JDBC Response
    Mappings:
    1) File REQ-->JDBC REQ
    2) JDBC RES-->FILE RES
    Interface mappings:
    1: FILE 2 JDBC REQ--> JDBC 2 FLE RECV
    CONFIGURATION ::
    1: One Seder File CC
    2: Two reciever CC's one is for JDBC RECEIVER and other s FILE RECEIVER
    3; One Sender Agreement
    4: 2 Recver agreements
    5: One Interface Determination and
    6: One RECCV Determination
    My Question;;
    1. First let confirm whether my development steps are right or not ?
    2: Another thing s I am not sure reg Configuration Steps means
    whetehr one interface determination and one Receiver Determinations are required or not as these are synchronous Interfaces
    3: main Issue is::::
    If my scenario s FILE2RFC2FILE then I will get RFC response automatically but here issue is this is JDBC
    My reqquirement is By sending one Field from fILE to JDBC REQ it needs to send entire TAbLE records as a Response to file as XML
    without having Sender JDBC how can I send the JDBC Res to FILE and If that is the case then again JDBC adapter is polling which is contradict to the client requuirement which i explained above.
    pleas suggest me the Detailed steps mainly Colloboration agreements and logical routings and
    also explain in detail if i can  go for BPM
    Also give cleear blogs but before giving make sure that it contains detailed screen shots because aIready gone thorugh
    Scenario File-JDBC-RFC
    File<-->JDBC Sync coomunication.
    https://www.sdn.sap.com/irj/scn/wiki?path=/display/xi/file-rfc-file(Without+BPM)
    /people/luis.melgar/blog/2008/05/13/synchronous-soap-to-jdbc--end-to-end-walkthrough
    Regards::
    Amar Srinivas Eli

  • Excise Reg at Depot

    Hi all,
    We have Excise REg no for Plant. But plant does not have Excise reg no.
    Material is shipped to Depot with excise invoice. What is the procedure at depot for GR.
    Do i need to capture the excise invoice?. From Depot, there is no excise invoice generated for customers.
    thanks.

    Hi,
    Refer following link;
    [J95 - Sales at Depot|http://help.sap.com/bp_bblibrary/500/HTML/J95_SaleDepot_EN_IN.htm]

  • How to set reg.cgi for VideoPhoneLabs

    i start with stratus .i try set it to work with my dedicated apache server and sql but have no clue how to do it or where to put this file on my server.
    i realy have probleme with reg.cgi
    for now i did www/cgi-bin/reg.cgi and in the xml i set my websiteurl = "http://88..../cgi-bin/"
    also how to complete reg.cgi for have it to talk with my db ?
    here the reg.cgi file.
    help please ...
    #! /usr/bin/python --
    reg.cgi by Michael Thornburgh.
    This file is in the public domain.
    IMPORTANT: This script is for illustrative purposes only. It does
    not have user authentication or other access control measures that
    a real production service would have.
    This script should be placed in the cgi-bin location according to
    your web server installation. The database is an SQLite3 database.
    Edit the location of the database in variable "dbFile".
    Create it with the following schema:
    .schema
    CREATE TABLE registrations (
        m_username VARCHAR COLLATE NOCASE,
        m_identity VARCHAR,
        m_updatetime DATETIME,
        PRIMARY KEY (m_username)
    CREATE INDEX registrations_updatetime ON registrations (m_updatetime ASC);
    # CHANGE THIS
    dbFile = '.../registrations.db'
    import cgi
    import sqlite3
    import xml.sax.saxutils
    query = cgi.parse()
    db = sqlite3.connect(dbFile)
    user = query.get('username', [None])[0]
    identity = query.get('identity', [None])[0]
    friends = query.get('friends', [])
    print 'Content-type: text/plain\n\n<?xml version="1.0" encoding="utf-8"?>\n<result>'
    if user:
        try:
            c = db.cursor()
            c.execute("insert or replace into registrations values (?, ?, datetime('now'))", (user, identity))
            print '\t<update>true</update>'
        except:
            print '\t<update>false</update>'
    for f in friends:
        print "\t<friend>\n\t\t<user>%s</user>" % (xml.sax.saxutils.escape(f), )
        c = db.cursor()
        c.execute("select m_username, m_identity from registrations where m_username = ? and m_updatetime > datetime('now', '-1 hour')", (f, ))
        for result in c.fetchall():
            eachIdent = result[1]
            if not eachIdent:
                eachIdent = ""
            print "\t\t<identity>%s</identity>" % (xml.sax.saxutils.escape(eachIdent), )
            if f != result[0]:
                print "\t\t<registered>%s</registered>" % (xml.sax.saxutils.escape(result[0]), )
        print "\t</friend>"
    db.commit()
    print "</result>"

    Persistent binding is effectively provided by STMS (MPxIO) - is there anything in particular you're wanting to do that STMS doesn't provide?

  • Fast Refresh mview performance

    Hi,
    I'm actually facing with Fast Refresh Mview performace problems and i would like to know what is the possible improvments for the fast refresh procedure:
    - base table is 1 500 000 000 rows partitions by day , subparttion by has (4)
    - mlog partition by hash (4) with indexes on pk and snaptime
    - mview partition by day subpartition by hash (4)
    10 000 000 insertions / day in base table/mlog
    What improvment or indexes can i add to improve fast refresh ?
    Thanks for help

    Hi,
    Which DB version are you using?
    Did you have a look at the MV refresh via Partition Change Tracking (PCT)?
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14223/advmv.htm#sthref575
    If it's possible to use PCT, it would probably improve a lot the performance of the refresh of your MV.
    Regards
    Maurice

  • Mview Fast Refresh Problem

    Hello,
    We run a fast refresh for a particular mview. I have the mview log created on the primary db (base table) and the mview created on another reporting database.
    The fast refresh seems to be fine throughout the day but at the end of a day it fails to do the fast refresh and i'm forced to do a complete refresh everyday which fixes it but this cycle continues until the end of day when it breaks again.
    If you notice in the next date column the refresh date turns to 01-JAN-00, not sure why this happens and this is causing the problem.
    JOB REFRESH_JOB LAST_DATE NEXT_DATE NEXT_SEC BROKEN
    70 CON_SYS"."MV_BRICKNW" 04-AUG-10 01-JAN-00 00:00:00 Y
    75 CON_SYS"."MV_BRICKNWCAN" 04-AUG-10 01-JAN-00 00:00:00 Y
    Could someone please tell me what the problem is.

    The Refresh Date gets reset when the job is Broken (which you see as BROKEN='Y'). The job is Broken when 16 consecutive attempts have failed (which you would see in FAILURE_COUNT).
    You have to find out why it fails -- fails 16 times. There seems to be an issue with the date/time logic OR one of the source objects probably gets dropped or becomes unavailable close to the end of the day.
    Hemant K Chitale

  • Mview fast refresh taking longer than complete refresh

    Hi there,
    We have a mview which takes about an hour to refresh in complete mode. Created all the mview logs and refreshing via fast method but this takes longer.
    When all the mview logs empty works fine as fast refresh and far far quicker as would expect. Tried upadting one of records on base table called todays_date which simply stores todays date to get around fast refresh restriction with sysdate.
    Anybody any idea why so slow?
    Also wondering if better approach. Thought about just having an aggregate table and processing the inserts in MLOG$_TRANS_TABLE via a merge statement. Also heard about change data capture.
    Any thoughts/suggestions welcome.
    Many Thanks

    Well, search with your post subject on this forum. There are a lot of threads discussing the same.

  • MVIEW refresh takes long time

    Materialized view takes long time to refresh but when i tried select & insert into table it's very fast.
    i executed SQL and it takes ust 1min ( total rows is 447 )
    but while i refresh the MVIEW it takes 1.5 hrs ( total rows is 447 )
    MVIEW configration :-
    CREATE MATERIALIZED VIEW EVAL.EVALSEARCH_PRV_LWC
    TABLESPACE EVAL_T_S_01
    NOCACHE
    NOLOGGING
    NOCOMPRESS
    NOPARALLEL
    BUILD DEFERRED
    REFRESH FORCE ON DEMAND
    WITH PRIMARY KEY
    Not sure why so much diffrence

    infant_raj wrote:
    Materialized view takes long time to refresh but when i tried select & insert into table it's very fast.
    i executed SQL and it takes ust 1min ( total rows is 447 )
    but while i refresh the MVIEW it takes 1.5 hrs ( total rows is 447 )A SELECT does a consistent read.
    A MV refresh does that and also writes database data.
    These are not the same thing and cannot be directly compared.
    So instead of pointing at the SELECT execution time and asking why the MV refresh is not as fast, look instead WHAT the refresh is doing and HOW it is doing that.
    Is the execution plan sane? What events are the top ones for the MV refresh? What are the wait states that contributes most to the processing time of the refresh?
    You cannot use the SELECT statement's execution time as a direct comparison metric. The work done by the refresh is more than the work done by the SELECT. You need to determine exactly what work is done by the refresh and whether that work is done in a reasonable time, and how other sessions are impacting the refresh (it could very well be blocked by another session).

  • Fast Refreshable MVIEW

    Trying to create a fast refresh materialized view based on a simple join of two tables over a database link oracle gives a 'COMPLEX QUERY' error (ORA-12015).
    Logging on the origin tables should be ok as creating the MVIEW based on the same select but in the database where the origin tables are poses no problem. The logging tables are created based on primary key and include rowid.
    The select on which the MVIEW should be created is someting like:
    select
    tab1.rowidn id1,
    tab2.rowidn id2,
    tab1.key_idn key1,
    tab2.key_idn key2
    from table1@other_db tabl,table2@other_db tab2
    where tab1.key_idn = tab2.key_idn)
    Does this mean it is impossible to create a fast refresh materialized view over a database link when a join condition is involved?

    Replace the ANSI Joins with Conventional Joins.
    Change this :
    SELECT * FROM V_FIRST VF
    LEFT JOIN R4
    ON VF.ID=R4.ID
    LEFT JOIN R5
    ON  VF.ID=R5.IDto this :
    SELECT VF.* FROM V_FIRST VF, R4, R5
    WHERE  VF.ID=R4.ID
    AND  VF.ID=R5.IDHemant K Chitale

  • Mview's taking a looong time to refresh. Any ideas?

    I have the following mview below. We are running the refresh using the FORCE option nightly. There are times when the refresh takes hours and we have to kill the job. Then, we just do a drop/recreate of the mview and it completes in < 20 minutes. Anyone have any ideas as to why?
    The merchant table has about 800k rows.
    The merchant_transaction table has about 18 million rows.
    I have included the explain plan of this query below.
    CREATE MATERIALIZED VIEW CARS.MER_MERTXN_MV
    TABLESPACE MVIEW_TB
    NOCACHE
    LOGGING
    NOCOMPRESS
    NOPARALLEL
    BUILD IMMEDIATE
    REFRESH FORCE
    NEXT NULL
    WITH PRIMARY KEY
    AS
    SELECT mt.merchant_num,
    mt.transaction_ref,
    mt.transaction_type,
    mt.transaction_status_code,
    mt.transaction_date,
    mt.sale_date,
    NVL(mt.transaction_credit_amt,0),
    NVL(mt.transaction_debit_amt,0),
    NVL(mt.open_credit_amt,0),
    NVL(mt.open_debit_amt,0),
    NVL(mt.reserve_bal,0),
    NVL(mt.collection_bal,0),
    NVL(mt.acct_bal,0),
    m.coll_assigned,
    m.acct_location_code,
    m.acct_type_code,
    m.marker_bank_code,
    m.cost_center_code,
    m.portfolio_code,
    m.rowid,
    mt.rowid
    FROM merchant_transactions mt, merchant m
    WHERE mt.merchant_num = m.merchant_num;
    EXPLAIN PLAN
    Operation     Object Name     Rows     Bytes     Cost     Object Node     In/Out     PStart     PStop
    SELECT STATEMENT Optimizer Mode=ALL_ROWS          18 M          212259                     
    HASH JOIN          18 M     2G     212259                     
    TABLE ACCESS FULL     CARS.MERCHANT     791 K     40 M     9206                     
    TABLE ACCESS FULL     CARS.MERCHANT_TRANSACTIONS     18 M     1G     96231

    1. Yes, but they aren't being used as per explain plan. I added a hint which helped. The problem is when we do a complete refresh. The fast refresh works fine. However, we have times when there is so much data added, it decides to do a complete. We are running FORCE. Does it DELETE the data or TRUNCATE the data when it does a complete? It seems its doing a DELETE.
    2. no rows
    3. 10.1. No parallelism being used.

Maybe you are looking for