Table access using DB link

Hi all,
I have database A and database B. I have a table in A that I want to grant insert access to B. I have a DB link(Blink) created on B. I executed the following statement with errors though. Can someone please advise.
GRANT INSERT, SELECT ON A.table1 TO scott@Blink;

Hi,
SunS wrote:
I have user userB in database B who connects to A. When I execute the following statement, I get an error "user or role userB does not exist"
GRANT INSERT, SELECT ON A.table1 TO userB;Exactly.
Now the question is: What user on database A is associated with that link? When someone on database B uses a link to database A, they connect as some particular user on database A. Who is that user on database A?
Say userB on database B is supposed to do things like this
SELECT ... FROM table1@alink;What did you say after "CONNECT TO" when you created the link? If you don't remember, then, on database B, say
SELECT  username
FROM    all_db_links
WHERE   dblink  = 'ALINK';That's the user on database A that needs privileges.

Similar Messages

  • ARP Table Access using SNMP

    Dear All,
    In Cisco Switches and routers we can enable access to the ARP Table using SNMP by using the following commands:
    no snmp-server view MibExcluded at excluded (Mac Address Table)
    no snmp-server view MibExcluded ip.21 no excluded (Routing Table)
    snmp-server view MibExcluded ip.22 excluded (ARP Table)
    how i can enable the ASA/FWSM to allow access to the ARP Table using SNMP?
    best Regards

    The at and ipNetToMediaTable are not supported on the ASA or FWSM. There have been numerous requests for support, but I do not know of an ETA for these objects.
    The bug ID for the enhancement request is CSCsq35214.

  • Table refresh Using Network Link

    Hi All,
    I am using 11g 11.1.0.6 database and I am usning network link to refresh few tables. The requirement is something like , on production , we have META schema
    and on Test we have meta_qa schema. meta on prod has five tables , for example tab1... tab5. now, i need to refresh these tables in meta_qa on test.
    we are also using one job , which is doing the same thing , but here the only difference is , we don't have meta_qa schema on prod database.
    regards

    Thanks Justin for the inputs, but I still not out of the woods.
    userid=x/x
    TABLES=(USER_SECURITY_INBOX,USER_ACCESS_SECURITY,WH_USERS_OPID,WH_USERS_OPID_INFO,WH_USERS_REDBRICK_INFO)
    EXCLUDE=CONSTRAINT
    table_exists_action=truncate
    remap_schema=META:META_QA
    NETWORK_LINK=CPMPROD
    LOGFILE=data_pump_dir:meta_reload_imp.log
    content=DATA_ONLY
    JOB_NAME=RELOADAnd the log file returns following ----
    Import: Release 11.1.0.6.0 - 64bit Production on Friday, 30 March, 2012 15:20:05
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - 64bit Production
    With the Partitioning and OLAP options
    Starting "ORAADMIN"."RELOAD":  PARFILE=/db/cpmtest/oracle/scripts/imp/ctl/cpmtest_meta_qa_reload_dp_imp.ctl
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    ORA-39166: Object WH_USERS_REDBRICK_INFO was not found.
    ORA-39166: Object WH_USERS_OPID_INFO was not found.
    ORA-39166: Object WH_USERS_OPID was not found.
    ORA-39166: Object USER_ACCESS_SECURITY was not found.
    ORA-39166: Object USER_SECURITY_INBOX was not found.
    ORA-31655: no data or metadata objects selected for job
    Job "ORAADMIN"."RELOAD" completed with 6 error(s) at 15:20:12

  • Doubt Regarding Table Accessed via DB Link

    I have the following SQL:
    select * from tab1@db_link tab1, tab2
    where tab1.col1=tab2.col2
    and tab1.col2=99;
    Can I ensure that Oracle brings data corresponding to:
    select * from tab1@db_link tab1
    where tab1=99;
    and then joins the data with tab2.
    What could be fastest way to access data in this manner? Basically i need to tune a similar query.

    No not much difference. On both the cases oracle goes for the same plan. Here is a small test case
    select t.*
      from t@to_inlabnew t, t1
    where t.object_id = t1.object_id
       and t.object_type = 'TABLE'
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.03       0.16          0          0          1           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch      171      1.09       1.68          0          6          0        2544
    total      173      1.12       1.84          0          6          1        2544
    Rows     Execution Plan
          0  SELECT STATEMENT   GOAL: ALL_ROWS
       2544   HASH JOIN
       1000    INDEX   GOAL: ANALYZED (FAST FULL SCAN) OF 'T1_IDX' (INDEX)
    199792    REMOTE OF 'T' (REMOTE)
                   [TO_INLABNEW.REGRESS.RDBMS.DEV.US.ORACLE.COM]
                  SELECT "OWNER","OBJECT_NAME","SUBOBJECT_NAME","OBJECT_ID",
                  "DATA_OBJECT_ID","OBJECT_TYPE","CREATED","LAST_DDL_TIME",
                  "TIMESTAMP","STATUS","TEMPORARY","GENERATED","SECONDARY" FROM
                    "T" "T" WHERE "OBJECT_TYPE"='TABLE'
    select t.*
      from (select * from t@to_inlabnew where object_type = 'TABLE') t, t1
    where t.object_id = t1.object_id
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch      171      1.09       1.46          0          6          0        2544
    total      173      1.09       1.46          0          6          0        2544
    Rows     Execution Plan
          0  SELECT STATEMENT   GOAL: ALL_ROWS
       2544   HASH JOIN
       1000    INDEX   GOAL: ANALYZED (FAST FULL SCAN) OF 'T1_IDX' (INDEX)
    199792    REMOTE OF 'T' (REMOTE)
                   [TO_INLABNEW.REGRESS.RDBMS.DEV.US.ORACLE.COM]
                  SELECT "OWNER","OBJECT_NAME","SUBOBJECT_NAME","OBJECT_ID",
                  "DATA_OBJECT_ID","OBJECT_TYPE","CREATED","LAST_DDL_TIME",
                  "TIMESTAMP","STATUS","TEMPORARY","GENERATED","SECONDARY" FROM
                    "T" "T" WHERE "OBJECT_TYPE"='TABLE'

  • Indx table access using export statement

    Hi ,
    Can any one help me to resolve this.
    we are getting an error in the program enhancement include.
    at line no 82 and its job is failing as this is running in back ground.
    |>>>>>|
    export l_t_zzpstyp to database indx(D7) ID 'IWSZZPSTYP'.                           
    full code
    Line |SourceCde                                                                              
    |
    |   52|          
    <fs1> TYPE ANY.                                                         
    |
    |   53|clear:l_t_zzpstyp,l_t_zzwerks,l_t_zzbsakz,
    ""ST20499/D7AK9A0VR5                     
    |
    |   54| 
    l_wa_zzpstyp,l_wa_zzwerks,l_wa_zzbsakz. ""ST20499/D7AK9A0VR5                     
    |
    |   55|IF NOT i_ekko-ekorg IS INITIAL.                                                        
    |
    |   56|                                                                                       
    |
    |   57|  CALL FUNCTION 'Z_EXIT_CHECK_PURCHORG'                                                
    |
    |   58|
    EXPORTING                                                                          
    |
    |   59| 
    ekorg     
    = i_ekko-ekorg                                                    
    |
    |   60| 
    exitname  
    = 'ZXM06U41_D7'                                                   
    |
    |   61|
    IMPORTING                                                                          
    |
    |   62| 
    exit_active
    = zxm06u41d7_active                                               
    |
    |   63|
    EXCEPTIONS                                                                         
    |
    |   64| 
    exit_not_found = 1                                                               
    |
    |   65| 
    OTHERS    
    = 2.                                                              
    |
    |   66|  IF sy-subrc = 0 AND zxm06u41d7_active = 'X'.                                         
    |
    |   67|                                                                                       
    |
    |   68|*}   INSERT                                                                            
    |
    |   69|                                                                                       
    |
    |   70|
    DATA: xuekpo LIKE ekpo.                                                            
    |
    |   71|
    DATA: xuekko LIKE ekko.                                                            
    |
    |   72|                                                                                       
    |
    |   73|
    CLEAR: xuekpo, xuekko.                                                             
    |
    |   74|*{Begin of change                      
    ST20499/D7AK9A0VR5                         
    |
    |   75|
    clear:l_t_zzpstyp,l_t_zzwerks,l_t_zzbsakz,                                         
    |
    |   76|     
    l_wa_zzpstyp,l_wa_zzwerks,l_wa_zzbsakz.                                      
    |
    |   77|
    xuekpo-pstyp = i_ekpo-pstyp.                                                       
    |
    |   78|
    EXPORT zzpstyp = xuekpo-pstyp TO MEMORY ID 'ZZPSTYP'.                              
    |
    |   79|*Import ID 'IWSZZPSTYP' in requirement routine  RV61B922                               
    |
    |   80|
    l_wa_zzpstyp-ZZPSTYP = i_ekpo-pstyp.                                               
    |
    |   81|
    append l_wa_zzpstyp to l_t_zzpstyp.                                                
    |
    |>>>>>|
    export l_t_zzpstyp to database indx(D7) ID 'IWSZZPSTYP'.                           
    |
    |   83|                                                                                       
    |
    |   84|
    xuekpo-werks = i_ekpo-werks.                                                       
    |
    |   85|
    EXPORT zzwerks = xuekpo-werks TO MEMORY ID 'ZZWERKS'.                              
    |
    |   86|*Import ID 'IWSZZWERKS' in requirement routine  RV61B922                               
    |
    |   87|
    l_wa_zzwerks-zzwerks = i_ekpo-werks.                                               
    |
    |   88|
    append l_wa_zzwerks to l_t_zzwerks.                                                
    |
    |   89|
    export l_t_zzwerks to database indx(D7) ID 'IWSZZWERKS'.                           
    |
    |   90|                                                                                       
    |
    |   91|
    xuekko-bsakz = i_ekko-bsakz.                                                       
    |
    |   92|*{   REPLACE   
    D7EK905213                                   
    2                
    |
    |   93|*\EXPORT ZZBSAKZ = XUEKKO-BSAKZ TO MEMORY ID 'ZZBSAKZ'.                                
    |
    |   94|
    EXPORT zzbsakz = xuekko-bsakz TO MEMORY ID 'ZZBSAKZ'.                              
    |
    |   95|*Import ID 'IWSZZBSAKZ' in requirement routine  RV61B922                               
    |
    |   96|
    l_wa_zzbsakz-zzbsakz = i_ekko-bsakz.                                               
    |
    |   97|
    append l_wa_zzbsakz to l_t_zzbsakz.                                                
    |
    |   98|
    export l_t_zzbsakz to database indx(D7) ID 'IWSZZBSAKZ'.                           
    |
    |   99|*}End of change                       
    ST20499/D7AK9A0VR5                          
    |
    |  100|  ENDIF.                                               
    " d7 active                
    |
    |  101|ENDIF.  "i_ekko-ekorg populated                  

    Hi,
    An error, huh..? Of the "Siri, I have an Abap run time error, what do I do" kind?
    Enter %pc in the command field while in dump display, save the dump as text file and attach it to message, please.
    cheers
    Janis

  • Accessing Tables using DB link .

    Hello Friends,
    I have a development and production env.
    I created a public DB link in Development say mylink .
    Trying to access the developement table from production using the following command say to access table x
    From PROD accessing DEV table x using dblink
    select * from x@mylink ;
    Got the error as :
    ORA-02019: connection description for remote database not found
    02019. 00000 - "connection description for remote database not found"
    Any idea on this why this error is coming ..
    thanks/kumar

    I have a development and production env.
    I created a public DB link in Development say mylink .
    Trying to access the developement table from production using the following command say to access table x
    From PROD accessing DEV table x using dblink
    select * from x@mylink ;You are trying to execute the query in PROD database. This query uses a DB link which points to the DEV database.
    Very simple to understand : Since you are using the DB LINK in PROD database, obviously it should be in PROD area... right?
    Just imagine you are using an object (the 'DBLINK') in PROD and it doesn't exist here. Got it?
    SELECT * FROM x@mylink;
    Here,
    x -- table in DEV database
    mylink -- DB LINK in PROD database (pointing to DEV)
    Please try this.
    HTH
    Edited by: ranit B on Dec 25, 2012 12:24 AM

  • 20 Index Restriction on Remote Tables (i.e. using Database Links)

    The Oracle Database Administrator's Guides for 10g and 11g document a performance restriction that "No more than 20 indexes are considered for a remote table." If I go back to the 8i documentation it says "In cost-based optimization, no more than 20 indexes per remote table are considered when generating query plans. The order of the indexes varies; if the 20-index limitation is exceeded, random variation in query plans may result."
    Does anyone have more details on this performance restriction? In particular I am trying to answer these questions:
    1) Are the 20 indexes which are considered by the CBO still random in 10g?
    2) Can I influence which indexes are considered with index hints or will my hints only be considered if they are for one of the "random" 20 indexes which are being considered by the CBO?
    3) Are there any other approaches or work-arounds to this restriction assuming you need to select from a large remote table with more than 20 indexes (and need to perform the selection using 1 of those indexes to get adequate performance) or do we need to abandon database links for this table?
    Thanks in advance for your input.

    So, here's my simple test.
    SQL>
    SQL> create table gurnish.indexes20plus ( n1 number, n2 number, n3 number, n4 number, n5 number, n6 number, n7 number,
    2 n8 number, n9 number, n10 number, n11 number, n12 number, n13 number, n14 number, n15 number, n16 number,
    3 n17 number, n18 number, n19 number, n20 number, n21 number, n22 number, n23 number, n24 number,
    4 n25 number, n26 number, n28 number);
    create index xin1 on indexes20plus (n1);
    Table created.
    SQL> SQL> create index xin2 on indexes20plus (n2);
    create index xin3 on indexes20plus (n3);
    Index created.
    SQL> SQL>
    Index created.
    SQL> SQL> create index xin4 on indexes20plus (n4);
    Index created.
    SQL> SQL>
    Index created.
    SQL> SQL> create index xin5 on indexes20plus (n5);
    create index xin6 on indexes20plus (n6);
    Index created.
    SQL> SQL>
    Index created.
    SQL> SQL> create index xin7 on indexes20plus (n7);
    Index created.
    SQL> SQL> create index xin8 on indexes20plus (n8);
    Index created.
    SQL> SQL> create index xin9 on indexes20plus (n9);
    Index created.
    SQL>
    SQL> create index xin10 on indexes20plus (n10);
    Index created.
    SQL> SQL> create index xin11 on indexes20plus (n11);
    create index xin12 on indexes20plus (n12);
    create index xin13 on indexes20plus (n13);
    Index created.
    SQL> SQL>
    Index created.
    SQL> SQL>
    Index created.
    SQL> SQL> create index xin14 on indexes20plus (n14);
    Index created.
    SQL> SQL> create index xin15 on indexes20plus (n15);
    Index created.
    SQL>
    SQL> create index xin16 on indexes20plus (n16);
    Index created.
    SQL>
    SQL> create index xin17 on indexes20plus (n17);
    Index created.
    SQL> SQL> create index xin18 on indexes20plus (n18);
    Index created.
    SQL> SQL> create index xin19 on indexes20plus (n19);
    Index created.
    SQL> SQL> create index xin20 on indexes20plus (n20);
    Index created.
    SQL> SQL> create index xin21 on indexes20plus (n21);
    Index created.
    declare
    i number;
    begin
    for i in 1..100
    loop
    dbms_random.seed(i+100);
    insert into indexes20plus values (dbms_random.value(1,5),dbms_random.value(1,21),dbms_random.RANDOM, dbms_random.RANDOM,dbms_random.value(1,20),
    dbms_random.value(1,4),dbms_random.value(1,6), dbms_random.value(1,7),dbms_random.value(1,9),dbms_random.value(1,10),
    dbms_random.value(1,11),dbms_random.value(1,12),dbms_random.value(1,13),dbms_random.value(1,14),dbms_random.value(1,1),
    dbms_random.value(1,1),dbms_random.value(1,19),dbms_random.value(1,122),dbms_random.value(1,20),dbms_random.value(1,20)
    ,dbms_random.value(4,20),dbms_random.value(1,20),dbms_random.value(1,20),dbms_random.value(1,20),dbms_random.value(1,20)
    ,dbms_random.value(4,20),dbms_random.value(4,20));
    end loop;
    commit;
    end;
    SQL> set autotrace traceonly
    SQL> l
    1* select * from gurnish.indexes20plus@lvoprds where n1 = 4
    SQL> /
    no rows selected
    Execution Plan
    Plan hash value: 441368878
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU
    )| Time | Inst |
    | 0 | SELECT STATEMENT REMOTE | | 1 | 351 | 1 (0
    )| 00:00:01 | |
    | 1 | TABLE ACCESS BY INDEX ROWID| INDEXES20PLUS | 1 | 351 | 1 (0
    )| 00:00:01 | LVPRD |
    |* 2 | INDEX RANGE SCAN | XIN1 | 1 | | 1 (0
    )| 00:00:01 | LVPRD |
    Predicate Information (identified by operation id):
    2 - access("A1"."N1"=4)
    Note
    - fully remote statement
    - dynamic sampling used for this statement
    Statistics
    0 recursive calls
    0 db block gets
    0 consistent gets
    0 physical reads
    0 redo size
    1897 bytes sent via SQL*Net to client
    481 bytes received via SQL*Net from client
    1 SQL*Net roundtrips to/from client
    0 sorts (memory)
    0 sorts (disk)
    0 rows processed
    SQL> select * from gurnish.indexes20plus@lvoprds where n21 = 4;
    no rows selected
    Execution Plan
    Plan hash value: 2929530649
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU
    )| Time | Inst |
    | 0 | SELECT STATEMENT REMOTE | | 1 | 351 | 1 (0
    )| 00:00:01 | |
    | 1 | TABLE ACCESS BY INDEX ROWID| INDEXES20PLUS | 1 | 351 | 1 (0
    )| 00:00:01 | LVPRD |
    |* 2 | INDEX RANGE SCAN | XIN21 | 1 | | 1 (0
    )| 00:00:01 | LVPRD |
    Predicate Information (identified by operation id):
    2 - access("A1"."N21"=4)
    Note
    - fully remote statement
    - dynamic sampling used for this statement
    Statistics
    1 recursive calls
    0 db block gets
    0 consistent gets
    0 physical reads
    0 redo size
    1897 bytes sent via SQL*Net to client
    481 bytes received via SQL*Net from client
    1 SQL*Net roundtrips to/from client
    0 sorts (memory)
    0 sorts (disk)
    0 rows processed
    SQL>

  • Accessing large partitioned tables over a database link - any gotchas?

    Hi,
    We are in the middle of a corporate acquisition and I have a question about using database links to efficiently access large tables. There are two geographically distinct database instances, both on Oracle 10.2.0.5 sitting on Linux boxes.
    The primary instance (PSHR) contains a PeopleSoft HR and Payroll system and sits in our data centre.
    The secondary instance (HGPAY) runs a home grown payroll application and sits in a different data centre to PSHR.
    The requirement is to allow PeopleSoft (PSHR) to display targeted (one employee at a time) payroll data from the secondary instance.
    For example in HGPAY
    CREATE TABLE MY_PAY_DATA AS
    SELECT TO_CHAR(A.RN, '00000000') "EMP" -- This is an 8 digit leading 0 unique identifier
    , '20110' || to_char(B.RN) "PAY_PRD" -- This is a format of fiscal year plus fortnight in year (01-27)
    , C.SOME_KEY -- This is the pay element being considered - effectively random
    , 'XXXXXXXXXXXXXXXXX' "FILLER1"
    , 'XXXXXXXXXXXXXXXXX' "FILLER2"
    , 'XXXXXXXXXXXXXXXXX' "FILLER3"
    FROM ( SELECT ROWNUM "RN" FROM DUAL CONNECT BY LEVEL <= 300) A
    , (SELECT ROWNUM "RN" FROM DUAL CONNECT BY LEVEL <= 3) B
    , (SELECT TRUNC(ABS(DBMS_RANDOM.RANDOM())) "SOME_KEY" FROM DUAL CONNECT BY LEVEL <= 300) C
    ORDER BY PAY_PRD, EMP
    HGPAY.MY_PAY_DATA is Range Partitioned on EMP (approx 300 employees per partition) and List Sub-Partitioned on PAY_PRD (3 pay periods per sub-partition). I have limited the create statement above to represent one sub-paritition of data.
    On average each employee generates 300 rows in this table each pay period. The table has approx 180 million rows and growing every fortnight.
    In PSHR
    CREATE VIEW PS_HG_PAY_DATA (EMP, PAY_PRD, SOME_KEY, FILLER1, FILLER2, FILLER3)
    AS SELECT EMP, PAY_PRD, SOME_KEY, FILLER1, FILLER2, FILLER3 FROM MY_PAY_DATA@HGPAY
    PeopleSoft would then generate SQL along the lines of
    SELECT * FROM PS_HG_PAY_DATA WHERE EMP = ‘00002561’ AND PAY_PRD = ‘201025’
    The link between the data centres where PSHR and HGPAY sit is not the best in the world, but I am expecting tens of access requests per day rather than thousands, so I believe the link should have sufficient bandwidth to meet the requirements.
    I have tried a quick test on two production sized test instances and it works in that it presents the data, when I look at the explain plan I can see that the remote database is only presenting the relevant sub-partition over to PSHR rather than the whole table. Before I pat myself on the back with a "job well done" - is there a gotcha that I am missing in using dblink to access partitioned big tables?

    Yes, that's about right. A lot of this depends on exactly what happens in various "oops" scenarios-- are you, for example, just burning some extra CPU until someone comes to the DBA and says "my query is slow" or does saturating the network have some knock-on effect on critical apps or random long-running queries prevent some partition maintenance operations.
    In my mind, the simplest possible solution (assuming you are using a fixed username in the database link) would be to create a profile on HGPAY for the user that is defined for the database link that set a LOGICAL_READS_PER_CALL value that was large enough to handle any "reasonable" request and low enough to quickly kill any session that tried to do something "stupid". Obviously, you'd need to define "stupid" in your environment particularly where the scope of a "simple reconciliation report" is undefined. If there are no political issues and you can adjust the profile values over time as you encounter new reports that slowly increase what is deemed "reasonable" this is likely the simplest approach. If you've got to put in a change request to change the setting that has to be reviewed by the change control board at its next quarterly meeting with the outsourced DBA vendor, on the other hand, you could turn a 30 minute report into 30 hours of work spread over 30 days. In the ideal world, though, that's where I'd start.
    Getting more complex, you can use Resource Manager to kill queries that run too long on the wall clock. Since the network is almost certainly going to be the bottleneck, it's probably unlikely that the CPU throttling is going to do much good-- you can probably saturate the network with a very small amount of CPU. Network throttling in my mind is an extra step up in complexity again depending on the specifics of your particular situation and what you're competing with.
    Justin

  • Accessing db using DB link

    Hi,
    I am trying to access one remote Oracle database from another using database link. I have created dB link and started TNS listener in both systems as told in forums. Database link is created but on testing I get the following error “ The Database Link is not active.ORA-12541: TNS:no listener”
    Can anyone help me to resolve this? I am new to Oracle. Any pointers to this is appreciated.

    I have a development and production env.
    I created a public DB link in Development say mylink .
    Trying to access the developement table from production using the following command say to access table x
    From PROD accessing DEV table x using dblink
    select * from x@mylink ;You are trying to execute the query in PROD database. This query uses a DB link which points to the DEV database.
    Very simple to understand : Since you are using the DB LINK in PROD database, obviously it should be in PROD area... right?
    Just imagine you are using an object (the 'DBLINK') in PROD and it doesn't exist here. Got it?
    SELECT * FROM x@mylink;
    Here,
    x -- table in DEV database
    mylink -- DB LINK in PROD database (pointing to DEV)
    Please try this.
    HTH
    Edited by: ranit B on Dec 25, 2012 12:24 AM

  • How to test issue with accessing tables over a DB link?

    Hey all,
    Using 3.1.2 on XE, I have a little app. The database schema for this app only contains views to the actual tables, which happen to reside over a database link in a 10.1.0.5.0 DB.
    I ran across an issue where a filter I made on a form refused to work (see [this ApEx thread| http://forums.oracle.com/forums/message.jspa?messageID=3178959] ). I verified that the issue only happens when the view points to a table across a DB link by recreating the table in the local DB and pointing the view to it. When I do this, the filter works fine. When I change the view back to use the remote table, it fails. And it only fails in the filter -- every other report and every other tool accessing the remote table via the view works fine.
    Anyone know how I can troubleshoot this? For kicks, I also tried using a 10.2.0.3.0 DB for the remote link, but with the same results.
    TIA,
    Rich
    Edited by: socpres on Mar 2, 2009 3:44 PM
    Accidental save...

    ittichai wrote:
    Rich,
    I searched metalink for your issue. This may be a bug in 3.1 which will be fixed in 4.0. Please see Doc ID 740581.1 Database Link Tables Do NoT Show Up In Table Drop Down List In Apex. There is a workaround mentioned in the document.
    I'm not sure why I never thought of searching MetaLink, but thanks for the pointer! It doesn't match my circumstances, however. The Bug smells like a view not being queried in the APEX development tool itself -- i.e. the IDE's coding needs changing, not necessarily those apps created with the IDE.
    I'm working on getting you access to my hosted app...
    Thanks,
    Rich

  • 403 Error when access Table Storage using SAS token

    I have Azure Mobile Service which has a custom API to generate a sas token for accessing Table Storage from Windows Store app.
    I get following error in Windows Store app while accessing table storage using sas token:
    Server failed to authenticate the request. Make sure the value of Authorization header is formed correctly including the signature.
    Example of sas token generated:
    se=2014-09-12T03%3A10%3A00Z&sp=rw&spk=MicrosoftAccount%3A005d92ef08ec5d83081afed1e08641d2&epk=MicrosoftAccount%3A005d92ef08ec5d83081afed1e08641d2&sv=2014-02-14&tn=Folders&sig=91c7S1QM0byNdM80JncwRribXqsWS1iKmOH8cRvHWhQ%3D
    Azure Mobile Services API Code that generates sas token:
    exports.get = function(request, response) {
    var azure = require('azure-storage');
    var accountName = 'myAccountName';
    var accountKey = 'myAccountKey';
    var host = accountName + '.table.core.windows.net';
    var tableService = azure.createTableService(accountName, accountKey, host);
    var sharedAccessPolicy = {
    AccessPolicy: {
    Permissions: 'rw', //Read and Write permissions
    Expiry: dayFromNow(1),
    StartPk: request.user.userId,
    EndPk: request.user.userId
    var sasToken = tableService.generateSharedAccessSignature('myTableName', sharedAccessPolicy);
    response.send(statusCodes.OK, { sasToken : sasToken });
    function dayFromNow(days){
    var result = new Date();
    result.setDate(result.getDate() + days);
    return result;
    Windows Store app code that uses sas token:
    public async Task TestSasApi()
    try
    var tableEndPoint = "https://myAccount.table.core.windows.net";
    var sasToken = await this.MobileService.InvokeApiAsync<Azure.StorageSas>("getsastoken", System.Net.Http.HttpMethod.Get, null);
    StorageCredentials storageCredentials = new StorageCredentials(sasToken);
    CloudTableClient tableClient = new CloudTableClient(new Uri(tableEndPoint), storageCredentials);
    var tableRef = tableClient.GetTableReference("myTableName");
    TableQuery query
    = new TableQuery().Where(TableQuery.GenerateFilterCondition("PartitionKey",
    QueryComparisons.Equal,
    this.MobileService.CurrentUser.UserId));
    TableQuerySegment seg = await tableRef.ExecuteQuerySegmentedAsync(query, null);
    foreach (DynamicTableEntity ent in seg)
    string str = ent.ToString();
    catch (Exception ex)
    string msg = ex.Message;
    Exception:
    Any help is appreciated.
    Thanks in advance!
    Thanks, Vinod Shinde

    Hi Mekh,
    Thanks for the links. I checked them and mostly they are due to date time on client and server.
    But this is not the case in this scenario.
    here is the Request and Response from Fiddler.
    Request:
    GET
    https://myaccount.table.core.windows.net/Folders?se=2014-09-13T02%3A33%3A26Z&sp=rw&spk=MicrosoftAccount%3A005d92ef08ec5d83081afed1e08641d2&epk=MicrosoftAccount%3A005d92ef08ec5d83081afed1e08641d2&sv=2014-02-14&tn=Folders&sig=YIwVPHb2wRShiyE2cWXV5hHg0p4FwQOGmWBHlN3%2FRO8%3D&api-version=2014-02-14&$filter=PartitionKey%20eq%20%27MicrosoftAccount%3A005d92ef08ec5d83081afed1e08641d2%27
    HTTP/1.1
    Accept: application/atom+xml, application/xml
    Accept-Charset: UTF-8
    MaxDataServiceVersion: 2.0;NetFx
    x-ms-client-request-id: b5d9ab61-5cff-498f-94e9-437694e9256c
    User-Agent: WA-Storage/4.2.1 (Windows Runtime)
    Host: todoprime.table.core.windows.net
    Response:
    HTTP/1.1 403 Server failed to authenticate the request. Make sure the value of Authorization header is formed correctly including the signature.
    Content-Length: 437
    Content-Type: application/xml
    Server: Microsoft-HTTPAPI/2.0
    x-ms-request-id: 22c0543b-0002-0049-7337-da39f4000000
    Date: Thu, 11 Sep 2014 02:33:28 GMT
    <?xml version="1.0" encoding="utf-8" standalone="yes"?>
    <error xmlns="http://schemas.microsoft.com/ado/2007/08/dataservices/metadata">
      <code>AuthenticationFailed</code>
      <message xml:lang="en-US">Server failed to authenticate the request. Make sure the value of Authorization header is formed correctly including the signature.
    RequestId:22c0543b-0002-0049-7337-da39f4000000
    Time:2014-09-11T02:33:29.6520060Z</message>
    </error>
    Do you see anything different in this request/response?
    Thanks, Vinod Shinde

  • Accessing partitioned table thru remote databse link

    Dear All
    I want to use create table command in my database .I want access partitioned table of remote datbase thru database link.Pl let me know the scripts.
    Thanks
    Parth
    NB : I was using the foll but actually the data given is wrong.
    select * from Data@prod "partition(part_feb03)";

    Don't able use partition table with the database link, sorry
    Regards
    Hector

  • Logical error in creating tables using db link in solaris

    Hi,
    While creating table using the syntax create table newtab...as select * from tab@dblink .. I am facing a problem. The newtab table created is having a structure different (from datalength point of view of varchar2 and char datatypes) from the source. The data length is getting tripled i.e. if a column a is varchar2(20) in tab then it is becoming --- a varchar2(60) in newtab. This is happening in solaris environment when the two databases are in 2 different servers. Please let me know if there are any patches to resolve the problem.
    Thanks
    Arnab

    ORA-02019: connection description for remote database not foundHave you used this database link successfully for some other queries?
    The error posted seems to indicate that the DB Link is not functional at all. Has it worked for any other type of DML operation or is this the first time you ever tried to use the link?

  • How to fix Action Link Issue in Table and Pivot Table when used Section

    My OBIEE Version: OBIEE 11.1.1.6.5
    Issue Description: Action links (Navigate to BI Content) doesn't work in PIVOT TABLE or TABLE object when used Section area (one or more columns in Section area).
    If you want to recreate this issue please follow next steps:
    1.     First of all you have to create one analyse with min 3 columns (for example Time.“Calendar date“, Product.Product, „Base Fact“. Revenue -> from Sample Sales Lite )
    2.     From column properties -> Set value Action Link Interaction on some column for example Product.Product (choose Navigate to BI Content -> any created BI report )
    3.     Click Results Tab choose Pivot Table and assigned one dimension column to Section area
    4.     Click preview icon
    5.     When you click on action link nothing happens
    6.     If you drug and drop column from Section Area to Excluded, Table or Pivot table Area then Action Link works perfectly
    How to fix this issue?
    I'll be grateful for any help!

    This is Bug:15922681 - ACTION LINK NAVIGATE TO BI CONTENT DOES NOT WORK WITH ATTRIBUTE IN SECTION
    Upgrade to 11.1.1.6.6 will fix this issue.

  • How can I access a web link using forms

    Hi,
    I need to access a Application using a link provided. I need to pass the parameter like userid, password and connection string to that link.
    How can I achieve this in forms????
    I use forms 11.1.1.4 and Window 7 and IE 8
    Thanks,
    maddy
    Edited by: maddy on 27-Nov-2012 21:19

    WEB.SHOW_DOCUMENT ('http://www.oracle.com','_blank');The url can be generated dynamically from your code by replacing it with variables.
    WEB.SHOW_DOCUMENT ('http://myServer/someApp?' || myParameter || '=' || myParamaterValue,'_blank');

Maybe you are looking for

  • For Open Items

    I needed to make a account receivable report that should include Sales engineer (Partner name)  and Sales order No and Invoice No and Invoice Data. Only for open items Which tables have to join Please Guide me.

  • Changing seq. aspect ratio: crashes FCP

    Hi all, Despite my best intentions, my project contains sequences with different settings, which is causing problems on export. A quick overview of sequence contents and field dominance settings: - SEQUENCE FIELD DOMINANCE: NONE - PSD frame-graphic o

  • Back up from iCloud

    I had my iphone 5s stolen on the 17th Jan (not sure what iOS it was up to?) stupidly I hadn't backed up on icloud since the 6th Jan, still, not too long passed so was hopefully little was lost. Before getting my new phone I logged into icloud to run

  • Can't create a new file

    When trying to create a new file, I get the following prompt and AI hangs and stops opening existing files. "Can't create a new illustration. Problems occurred while saving the print file. -1" Anyone know what's going on? - CS6 AI v16.2.0 - MacBook P

  • Keeping song info. while copying from Ipod to computer

    My laptop totally failed and is now in pieces. I want to transfer all my songs to our desktop from my Ipod. I know how to do that but I was wondering if there is a program that maintains all the information from the songs, I really don't want to go t