Transportable Tablespace and Packages/Functions etc

We are currently experimenting with transportable tablespaces. Is it possible to take functions/packages etc or are you limited to tables/data/triggers.
Thanks
Chris

Transport tablespace is available in version 9.2. In that version you need to put the tablespace in read only mode, export the tablespace metadata, copy the tabalespace files, and then import the metadata. Now, I don't believe this will transport the other objects. Someone else chime in if I am incorrect.
What you can do is transport the tabelspace and then run an export rows=n
and then import this empty dump (that will include the procedures/functions.etc) with ignore=y

Similar Messages

  • ORA-00904 and packaged function in report

    Oracle 9.2.0.7
    APEX 2.0
    Is APEX not able to deal with SQL that calls a packaged function?
    I ask because I'm able to execute this query:
    SELECT s.doc_id,
    s.last_name || ', ' || s.first_name full_name,
    s.p_id,
    s.create_date
    , rwa_wfrole.get_role_list(s.proj_id, s.p_id) wfrole_label
    FROM signatures_vw s
    WHERE s.proj_id = 182
    AND s.status = 'APPROVED'
    ORDER BY s.create_date, s.last_name, s.first_name
    in SQL*Plus, but not in the SQL Workshop or in a report region. In the latter case I get "ORA-00904 invalid identifier" for "rwa_wfrole.get_role_list"... which is a packaged function. Actually, to be more precise, it's a package in a remote database. In the APEX instance, I have a synonym pointing to it.
    Thanks for insight.
    -John

    I can't reproduce this in 10g. I did find bug 4177810 (Fixed in 9.2.0.8 Server Patch Set) that looks a bit like this problem.
      Description
        An unexpected ORA-904 can occur for some internally generated SQL
        when the select list contains non-column constants and the select
        appears in the FROM clause. This can occur for SQL produced for
        parallel query or for remote / distributed queries.
        Workaround:
          Rewrite the query to avoid constant expressions in the FROM clause select lists.***************************
    There are also recommendations dating way back that you should include the schema name in the Create Synonym statement.
    Scott

  • Transportable tablespace and partitioned objects

    Hi,
    I have to transport tablespaces which have partitioned objects from one server to another. If anyone has a proven process to perform this operation, please share with us. Following is a sample of violation report sample I got for 1 tablespace:
    VIOLATIONS
    Partitioned table ARIES.A_BI_BOARD_USAGE is partially contained in the transport
    able set: check table partitions by querying sys.dba_tab_partitions
    Partitioned table ARIES.A_TESTING_SESSION is partially contained in the transpor
    table set: check table partitions by querying sys.dba_tab_partitions
    Default Partition (Table) Tablespace A_DATA_SML_000000 for A_BI_BOARD_USAGE not
    contained in transportable set
    Default Partition (Table) Tablespace A_DATA_SML_000000 for A_TESTING_SESSION not
    contained in transportable set
    Thanks
    Satish

    Hi,
    I am trying to transport the whole tablespace. What happened is that during duplication process, somehow few tablespaces in target were in offline mode and so duplication did not ceate the files properly on auxiliary server. Now, I need to get the 3 tablespaces from production to integration enviornment. These tablespaces have lot of partitioned objects. So I am trying to find a way to transport these tablespaces.
    Thanks
    Satish

  • Transportable Tablespace and Table Statistics

    We are loading a large datawarehouse of about 2Tb, with tables up to 100Gb full of raster images. The database is 9.2.0.7 on Solaris, and we are using Transportable Tablespaces to move datafiles and tablespaces from data integration server to production server.
    During the data integration, we calculate statistics on large tables, and it takes forever. We are wondering if these statistics are moved along with the metadata from data integration server to production server. In other words, should we calculate statistics before the tablespace export or after the tablespaces have been set online on the production server.
    Thanks,

    Hi,
    Upto my knowledge you dont need to calculate statistics.
    Even with export & import command there is a option for statistics.
    So when ever you export oracle by default export the statistics of that table.
    Hope this helps.
    Regards
    MMU

  • Stuck with transported tablespace and ASM

    Hi,
    I'm on 10.2.0.5.4 and playing with TTS feature .
    So I've made TTS on source DB 10.2.0.5 via rman tts feature (auxilary instance) , and copied to cloned database with ASM (same os solaris10).
    Now my situation is as follows :
    on target I can see TTS1 tablespace
    FILE_NAME                                          TABLESPACE_NAME                STATUS
    +DG1/sol10p/datafile/users.262.763395469           USERS                          AVAILABLE
    +DG1/sol10p/datafile/sysaux.260.763395447          SYSAUX                         AVAILABLE
    +DG1/sol10p/datafile/undotbs1.259.763395423        UNDOTBS1                       AVAILABLE
    +DG1/sol10p/datafile/system.258.763395387          SYSTEM                         AVAILABLE
    +DG1/sol10p/datafile/example.261.763395463         EXAMPLE                        AVAILABLE
    +DG1/sol10p/datafile/assm.257.763395311            ASSM                           AVAILABLE
    /u01/tts/tts1.274.763488147                        TTS1                           AVAILABLE
    /u01/tts/tts1.272.763487799                        TTS1                           AVAILABLEbut when trying to copy those files into ASM I'm getting
    RMAN> report schema;
    Report of database schema
    List of Permanent Datafiles
    ===========================
    File Size(MB) Tablespace           RB segs Datafile Name
    1    510      SYSTEM               ***     +DG1/sol10p/datafile/system.258.763395387
    2    310      UNDOTBS1             ***     +DG1/sol10p/datafile/undotbs1.259.763395423
    3    250      SYSAUX               ***     +DG1/sol10p/datafile/sysaux.260.763395447
    4    62       USERS                ***     +DG1/sol10p/datafile/users.262.763395469
    5    100      EXAMPLE              ***     +DG1/sol10p/datafile/example.261.763395463
    6    2048     ASSM                 ***     +DG1/sol10p/datafile/assm.257.763395311
    List of Temporary Files
    =======================
    File Size(MB) Tablespace           Maxsize(MB) Tempfile Name
    1    512      TEMP                 512         +DG1/sol10p/tempfile/temp.265.763396585
    RMAN> backup as copy datafile '/u01/tts/tts1.274.763488147' format '+DG1';
    Starting backup at 03-OCT-11
    using channel ORA_DISK_1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of backup command at 10/03/2011 20:00:34
    RMAN-20201: datafile not found in the recovery catalog
    RMAN-06010: error while looking up datafile: /u01/tts/tts1.274.763488147Please advice .
    Regards
    GregG

    I'm not using catalog (connecting as nocatalog) Hmm....
    The files are listed in V$DATAFILE and DBA_DATA_FILES but not in an RMAN REPORT SCHEMA ?
    That's curious. V$DATAFILE and RMAN REPORT SCHEMA would both be reading from the controlfile.
    What was the sequence of actions to transport the tablespace ? Have you tried reconnecting with RMAN ?
    Hemant K Chitale

  • Transportable tablespaces for migration and upgrade

    Hello All,
    I'm trying to find out some theory about transportable tablespace and upgrade in a single step. I've to move a DB from a Windows server to a Linux server. Win is running 10.1 and Linux 10.2 . It's possible to use RMAN to migrate the whole database and then on new node run startup upgrade and catupgrd.sql to upgrade catalog ?
    Should work?
    Thanks
    Raj

    The following doc for TDB says :-
    "Note: TDB requires the same Oracle Database software version on the source and target systems; hence, a database upgrade cannot be accomplished simultaneous with the platform migration. However, best practices for a database upgrade, as documented in the Oracle Upgrade Companion, also apply to any planned database maintenance activity, including platform migration."
    http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_10gR2_PlatformMigrationTDB.pdf
    There is also an Oracle Whitepaper called "Database Upgrade Using Transportable Tablespaces: Oracle Database 11g Release 1" but I am still trying to find a copy.

  • Export tables, sequence, and package question

    Hi all,
    I've export table like: exp username/password file=export.dmp log=export.log tables=A statistics=none.
    The export statement above just export only the table "A" structure not the table "A" data. So, how can I move all the data from table "A" to table "B".
    How do I export the sequence, and package and its data also?
    Thank you very much
    Kevin

    Can you create a DBlink in order to do the insert into B select * from A.
    Also, you can use the SQL*Plus Copy command.
    Here are some links for COPY command:
    <br>Copying Data
    <br>copy command
    <br>copy command vs sql*loader
    <br><br>
    Here is a link for moving code:
    <br>exporting packages,function etc. from one user to another.

  • Transport Tablespace using RMAN from 10.2.0.2 to 10.2.0.3 ??

    Hello all,
    Is it possible to do RMAN Transport Tablespace from 10.2.0.2 to 10.2.0.3 ??
    TIA,
    J J

    Are the two database on same platform?
    Yes, it's possible to move from 10.2.0.2 to 10.2.0.3
    You should check Oracle document,
    You cannot export transportable tablespaces and then import them into a database at a lower release level. The target database must be at the same or higher release level as the source database.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_export.htm#i1006840

  • How to assign a Transport request to a FUNCTION in PE04.

    Hi to All,
    I had a query related to FUNCTION PE04. This is a Z----- funtion (Full and FInal settlement of a leaving employee)  created by a some earlier in my place and was transported to Quality from Development. But the problem is i couldnt able to see it in Quality/Testing box. When i checked the Transport request in Dev and Quality its showing as transported. And this FUNCTION is working fine in Development box.
    Now im trying to assing a new Transport request to the function but i couldnt able to do so.
    Any suggestion welcome.
    Thanks,
    Mak,

    Resolved, Problem is from the basis side.

  • Error ORA-39125 and ORA-04063 during export for transportable tablespace

    I'm using the Oracle Enterprise Manager (browser is IE) to create a tablespace transport file. Maintenance...Transport Tablespaces uses the wizard to walk me through each step. The job gets created and submitted.
    The 'Prepare' and 'Convert Datafile(s)' job steps complete successfully. The Export step fails with the following error. Can anyone shed some light on this for me?
    Thank you in advance!
    =======================================================
    Output Log
    Export: Release 10.2.0.2.0 - Production on Sunday, 03 September, 2006 19:31:34
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Username:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Starting "SYS"."GENERATETTS000024": SYS/******** AS SYSDBA dumpfile=EXPDAT_GENERATETTS000024.DMP directory=EM_TTS_DIR_OBJECT transport_tablespaces=SIEBEL job_name=GENERATETTS000024 logfile=EXPDAT.LOG
    Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
    Processing object type TRANSPORTABLE_EXPORT/TABLE
    Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
    ORA-39125: Worker unexpected fatal error in KUPW$WORKER.UNLOAD_METADATA while calling DBMS_METADATA.FETCH_XML_CLOB [TABLE_STATISTICS]
    ORA-04063: view "SYS.KU$_IOTABLE_VIEW" has errors
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.KUPW$WORKER", line 6241
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    2CF48130 14916 package body SYS.KUPW$WORKER
    2CF48130 6300 package body SYS.KUPW$WORKER
    2CF48130 2340 package body SYS.KUPW$WORKER
    2CF48130 6861 package body SYS.KUPW$WORKER
    2CF48130 1262 package body SYS.KUPW$WORKER
    2CF0850C 2 anonymous block
    Job "SYS"."GENERATETTS000024" stopped due to fatal error at 19:31:44

    More information:
    Using SQL Developer, I checked the view SYS.KU$_IOTABLE_VIEW referred to in the error message, and it does indeed report a problem with that view. The following code is the definition of that view. I have no idea what it's supposed to be doing, because it was part of the default installation. I certainly didn't write it. I did, however, execute the 'Test Syntax' button (on the Edit View screen), and the result was this error message:
    =======================================================
    The SQL syntax is valid, however the query is invalid or uses functionality that is not supported.
    Unknown error(s) parsing SQL: oracle.javatools.parser.plsql.syntax.ParserException: Unexpected token
    =======================================================
    The SQL for the view looks like this:
    REM SYS KU$_IOTABLE_VIEW
    CREATE OR REPLACE FORCE VIEW "SYS"."KU$_IOTABLE_VIEW" OF "SYS"."KU$_IOTABLE_T"
    WITH OBJECT IDENTIFIER (obj_num) AS
    select '2','3',
    t.obj#,
    value(o),
    -- if this is a secondary table, get base obj and ancestor obj
    decode(bitand(o.flags, 16), 16,
    (select value(oo) from ku$_schemaobj_view oo, secobj$ s
    where o.obj_num=s.secobj#
    and oo.obj_num=s.obj#),
    null),
    decode(bitand(o.flags, 16), 16,
    (select value(oo) from ku$_schemaobj_view oo, ind$ i, secobj$ s
    where o.obj_num=s.secobj#
    and i.obj#=s.obj#
    and oo.obj_num=i.bo#),
    null),
    (select value(s) from ku$_storage_view s
    where i.file# = s.file_num
    and i.block# = s.block_num
    and i.ts# = s.ts_num),
    ts.name, ts.blocksize,
    i.dataobj#, t.bobj#, t.tab#, t.cols,
    t.clucols, i.pctfree$, i.initrans, i.maxtrans,
    mod(i.pctthres$,256), i.spare2, t.flags,
    t.audit$, t.rowcnt, t.blkcnt, t.empcnt, t.avgspc, t.chncnt, t.avgrln,
    t.avgspc_flb, t.flbcnt, t.analyzetime, t.samplesize, t.degree,
    t.instances, t.intcols, t.kernelcols, t.property, 'N', t.trigflag,
    t.spare1, t.spare2, t.spare3, t.spare4, t.spare5, t.spare6,
    decode(bitand(t.trigflag, 65536), 65536,
    (select e.encalg from sys.enc$ e where e.obj#=t.obj#),
    null),
    decode(bitand(t.trigflag, 65536), 65536,
    (select e.intalg from sys.enc$ e where e.obj#=t.obj#),
    null),
    (select c.name from col$ c
    where c.obj# = t.obj#
    and c.col# = i.trunccnt and i.trunccnt != 0
    and bitand(c.property,1)=0),
    cast( multiset(select * from ku$_column_view c
    where c.obj_num = t.obj#
    order by c.col_num, c.intcol_num
    ) as ku$_column_list_t
    (select value(nt) from ku$_nt_parent_view nt
    where nt.obj_num = t.obj#),
    cast( multiset(select * from ku$_constraint0_view con
    where con.obj_num = t.obj#
    and con.contype not in (7,11)
    ) as ku$_constraint0_list_t
    cast( multiset(select * from ku$_constraint1_view con
    where con.obj_num = t.obj#
    ) as ku$_constraint1_list_t
    cast( multiset(select * from ku$_constraint2_view con
    where con.obj_num = t.obj#
    ) as ku$_constraint2_list_t
    cast( multiset(select * from ku$_pkref_constraint_view con
    where con.obj_num = t.obj#
    ) as ku$_pkref_constraint_list_t
    (select value(ov) from ku$_ov_table_view ov
    where ov.bobj_num = t.obj#
    and bitand(t.property, 128) = 128), -- IOT has overflow
    (select value(etv) from ku$_exttab_view etv
    where etv.obj_num = o.obj_num)
    from ku$_schemaobj_view o, tab$ t, ind$ i, ts$ ts
    where t.obj# = o.obj_num
    and t.pctused$ = i.obj# -- For IOTs, pctused has index obj#
    and bitand(t.property, 32+64+512) = 64 -- IOT but not overflow
    -- or partitioned (32)
    and i.ts# = ts.ts#
    AND (SYS_CONTEXT('USERENV','CURRENT_USERID') IN (o.owner_num, 0) OR
    EXISTS ( SELECT * FROM session_roles
    WHERE role='SELECT_CATALOG_ROLE' ));
    GRANT SELECT ON "SYS"."KU$_IOTABLE_VIEW" TO PUBLIC;

  • Standby not in sync because test on primary of transportable tablespace etc

    This environment is new build environment , have not in use yet.
    db version is 11.2.0.3 in linux, both primary/standby are configured in RAC two nodes and storage are in ASM storage.
    primary db had tested by migration data using transportable tablespace methods. So the datafiles imported were put in a local filesystem which I have to switch to ASM afterwards.
    In such way, the standby db got the impression that datafiles should be in local filesystem and it got invalidated.
    Here is the error from standby log:
    ORA-01157: cannot identify/lock data file 4 - see DBWR trace file
    ORA-01110: data file 4: '/stage/index.341.785186191'
    and datafiles names are all in local filesystem in standby: /stage/...../
    However in primary datafile names are all in ASM instance:
    +DAT/prd/datafile/arindex.320.788067609
    How to resolve such situation?
    Thanks

    My steps for transportable tablespaces after scp the datafiles to linux are import the transportable datafiles into the database. and then using rman to copy it to "+DATA' in asm. and then switch dababase to copy.
    Those procedure all worked on primary.
    The standby was built before the transportable tablespace when there are only system, users, tools etc basic tablespaces. it worked and in sync with primary. However everytime the primary tested data migration with the transportable tablespaces, it will not work.
    so what is the right way to perform this migration? how can I made standby in sync with primary and to read the datafiles in ASM storage?
    Is there a way before the RMAN command to transfer the filesystem datafile to ASM, can I copy those local filesystem datafiles to ASM storage ? then do a import transportable datafiles with those on ASM?If you added any datafiles manually then those will be created in standby also as per the settings,
    But here you performing TTS, of course DML/DDLs can be transfered to standby by archive logs, but what about actual datafiles? Here in this case files exist in primary & those files not exist in standby
    AFAIK, you have to perform couple of steps. Once after your complete migration done in primary do as follows. (Do test it)
    1) Complete your migration on primary
    -- Check all data file status and location.
    2) Now Restore standby controlfile & newly migrated tablespaces in standby Database
    -- here you can directly restore to ASM using RMAN, because you are taking backup in primary using RMAN. So RMAN can restore them directly in ASM file system.
    3) Make sure you have all the datafiles where you want and take care no one data file is missing. Crosscheck all tablespace information in primary & standby databases.
    4) Now start *MRP*
    Here is my above plan, But I suggest to test it...
    One question, After *successful migration* are you able to SYNC standby with primary earlier ?

  • To generate Script of  procedures/functions and packages with grants

    Hi
    is there wome way for to generate Source Code Script of the some functions, procedures and Packages in file
    I must to generate Script withall grants
    Tia

    Datapump export with Content=Metadata_only.
    Use SQLFILE option to generate the Schema DDL which
    includes
    tables,indexes,grants,stats,procedures,triggers,views
    etc...Sorry, I'm using 10g , but I don't know to use DataPump, is there other method ?
    Thanks

  • Exp/imp procedures, functions and packages question

    Hi
    I've a 9i R2 version Oracle database. I would like to export procedures, functions and packages from a schema. How do I do that?
    Is there any script or command lines can provide?
    Thanks

    Hello user12259190.
    You can do an export of the user itself, excluding table data as inH:\>exp
    Export: Release 10.2.0.1.0 - Production on Tue Dec 22 11:22:52 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Username: db_user@db_sid
    Password:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, Data Mining and Real Application Testing options
    Enter array fetch buffer size: 4096 >
    Export file: EXPDAT.DMP >
    (2)U(sers), or (3)T(ables): (2)U > 2
    Export grants (yes/no): yes > no
    Export table data (yes/no): yes > no
    Compress extents (yes/no): yes > no
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    server uses UTF8 character set (possible charset conversion)
    Note: table data (rows) will not be exported
    Note: grants on tables/views/sequences/roles will not be exported
    . exporting pre-schema procedural objects and actions
    . exporting foreign function library names for user DB_USER
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions for user DB_USER
    About to export DB_USER's objects ...
    . exporting database links
    . exporting sequence numbers
    . exporting cluster definitions
    . about to export DB_USER's tables via Conventional Path ...
    . . exporting table  TABLE_NAMEs
    EXP-00091: Exporting questionable statistics.
    . exporting synonyms
    . exporting views
    . exporting stored procedures
    . exporting operators
    . exporting referential integrity constraints
    . exporting triggers
    . exporting indextypes
    . exporting bitmap, functional and extensible indexes
    . exporting posttables actions
    . exporting materialized views
    . exporting snapshot logs
    . exporting job queues
    . exporting refresh groups and children
    . exporting dimensions
    . exporting post-schema procedural objects and actions
    . exporting statistics
    Export terminated successfully with warnings.Unfortunately, you can't export just the objects you want to unless they are tables.
    Using import (imp) you can list the content of your packages, procedures, functions, views, etc. and perhaps that will give you what you need.
    Another choice would be to useSELECT * FROM user_source ORDER BY 2, 1, 3;to list the code.
    Hope this helps,
    Luke

  • 2.1.0.62: Problem with Package.Functions and Unit Tests

    I like the new Sqldeveloper - I startet trying Unit Tests as described here: Link: [http://www.oracle.com/technology/obe/11gr2_db_prod/appdev/sqldev/sqldev_unit_test/sqldev_unit_test.htm#t4]
    It worked with a test-procedure. Now i am trying to test my package functions, but all i get is this:
    Die folgende Prozedur wurde ausgeführt.
    Ausführungsaufruf
    BEGIN
    :1 := "PKG_MYPACK"."CREATEFUNCTION"(IN_PROGRAMMEID=>:2,
    IN_AMOUNT=>:3,
    IN_SWS=>:4);
    END;
    Bind-Variablen verwendet
    1 INTEGER OUT (null)
    2 INTEGER IN 1
    3 INTEGER IN 10
    4 INTEGER IN 11
    Ausführungsergebnisse
    ERROR
    Ungültige Konvertierung angefordert
    oracle.jdbc.driver.SQLStateMapping.newSQLException(SQLStateMapping.java:70)
    oracle.jdbc.driver.DatabaseError.newSQLException(DatabaseError.java:110)
    oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:171)
    oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:227)
    oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:439)
    oracle.jdbc.driver.OraclePreparedStatement.setObjectCritical(OraclePreparedStatement.java:7723)
    oracle.jdbc.driver.OraclePreparedStatement.setObjectInternal(OraclePreparedStatement.java:7496)
    oracle.jdbc.driver.OraclePreparedStatement.setObjectInternal(OraclePreparedStatement.java:7978)
    oracle.jdbc.driver.OracleCallableStatement.setObject(OracleCallableStatement.java:4063)
    oracle.jdbc.driver.OraclePreparedStatementWrapper.setObject(OraclePreparedStatementWrapper.java:221)
    oracle.dbtools.raptor.datatypes.strategies.callablestatement.CallableBindingDatum.customBindIN(CallableBindingDatum.java:135)
    oracle. ...
    what can i do?

    Created
    Bug 8976245 - EA1: UNIT_TEST: INVALID CONVERSION ERROR USING INTEGER PARAMETER
    and have asked bug responder to keep you updated on status here in the forum.
    Bad news is that any INTEGER parameter for which you specify a non-null value will fail.
    Possibly helpful news is that if you create a 'clone' of you function using NUMBER as the data type, you can continue to experiment with how unit testing may be of use to you.
    Sorry no instant answer. :(
    Brian
    SQL Developer Team

  • How to use type, packages, functions, and procedures in another schema ?

    I have two target schema in one OWB project, such as A and B. In a mapping of A, I would like to use some types, packages, functions, and procedures from B. I have tried the method of synonym as suggested, but I could not find the metadata of these when importing ... The only type of synonym I can import is the synonym for table. Is there a bug for synonym?
    If I cannot use synonym for this issue, is there another way to solve the problem?

    Now, in some instances you will absolutely need to create the second module as Carsten describes, however it should also be noted that you can reference objects in things like Expressions even if you have not loaded up the metadata. It is only when you need strong binding that it becomes neccessary to import objects. For everything else, as long as the reference will resolve at compile-time then you are good to go.
    For example, I have a function in one target schema (S1) and a private synonym to it in another(s2). A mapping in the S2 schema has an expression object that uses the synonym to the function in the expression property for a couple of the output attirbutes. The synonym has not been loaded into metadata - indeed OWB has no knowledge of its existance. But it resolves at compile time so the mapping validates and generates successfully.
    Mike

Maybe you are looking for

  • 8.1 Pro install now have no audio and video play

    I have an older G60-443cl and changed out the HDD for a SDD.  Installed a new version of Windows 8.1 Pro.  Everything works but my audio and video.  The audio drivers are installed and working properly (according to my system) but even on the speaker

  • I cannot cut and paste .rtf from a MS Word document to a web page.

    I am running a MAC desktop with version 10.5. I use MS Office for MAC. When I try and cut and paste from a .rtf file to an amateur writing site, I always get a timeout. Help please!

  • Imac 8,1 does not boot stuck on blue screen

    imac8,1 mid (2008) 20" stuck on blue screen after boot ,It starts in safemode, below is a list of thing i have check so far 1 Apple hardware test (says every thing is ok) 2 Disk utility repair (was successfull) 3 reset PRAM ,SMC 3 Boot for OSX dvd an

  • Adjusting my backup plan with Lion and TM Plus Clone

    Hi all, I am rethinking my backup plan for when I rollout Lion in my business. Currently in SL, we have clones scheduled to run every nite doing incremental backups and then we rotate the disks out periodically for offsite storage. Now I am thinking

  • Auto Invoice Error

    Hey,i am getting this error while running the Auto Invoice Import "Please correct the revenue account assignment" Note found in metalink was not helpful. Please let me know your inputs