Export/import masters *with* adjustments?

Hi,
I've a Mac Pro with my main Aperture installation, and a Mac Book with an Aperture installation for when I'm travelling. I would like to export masters and the adjustments from the Aperture library in the Mac Book and import them in the Aperture library in the Mac Pro.
Any advice? Aperture 2.1.

OK, I figured it out myself. Here's what I did. If you do this and manage to mess things up, it's your responsibility.
I have an external hard drive with a vault that contains the backups of the images on my Mac Book. I connected that drive to my Mac Pro. I showed the vault's contents with Finder, and similarly the Mac Pro's library contents. I copied the X.approject that contains the files I want to import/export with adjustments and other metadata from the vault to Mac Pro's library. Then I started Aperture by pressing Option/Command, and rebuilt the library.
It worked fine, I have the ratings and white balances etc. copied with the images.

Similar Messages

  • Exporting master images WITH adjustments

    Hello,
    I have a simple question about exporting images.
    Let's say I have imported a bunch of raw images into my Aperture Library (but they are referenced files, located in the folder specified by me). Then I have worked on them. And now I want to backup my work - all my adjustements. But I don't want to export them as new versions (jpgs), but just export the raw files together with adjustments I have just made.
    How do I do that?
    Any help would be really appreciated.
    Best greetings,
    Tomasz
    iMac G5 17'' Flat Panel   Mac OS X (10.4.8)  

    This is where the Adobe DNG could shine.
    http://www.adobe.com/products/dng/index.html
    Worth the read ... this can save the changes that one
    makes.
    DNG does NOT help in this case.
    There is a huge misunderstanding that this is part of what DNG can or should do.
    But if you think about it, what good are the adjustments being stored in DNG if you have to use a particular program to open it anyway?
    Think about Aperture's Edge Sharpen for a second. Lets say you store that value in a DNG. Fine, what other program is going to able to reproduce that result EXACTLY to how you were previewing it at 100% on your monitor?
    Lighroom is trying to do something along these lines by passing editing commands off to Bridge through DNG. But here you run into another problem - it constrains what editing any one program can do. If Lightroom is limited to only ever having editing commands that are the same as what Bridge offers, and no other program on earth supports them, then what have you really accomplished? Will unknown editing commands simply be dropped without warning?
    That's why I think simply exporting projects, which hold master images alongside sidecar files (very like XMP) that describe edits are about as good as you are going to get. If you want to truly preserve editing work and you care about quality, nothing beats a TIFF file where 100% of the pixels are exactly as you reviewed them during editing. I personally trust Aperture enough to back up master images along with edits, and am fine with that.

  • Can't export to DNG with adjustments

    HI all. I'm having a problem exporting a DNG file as an original or as DNG with adjustments. All the exports is the original DNG file. I do get a clean export with adjustments when I export to jpg. any thoughts?
    here's the details: software is lightroom 4.1 and camera raw 7.1. File was NEF converted to DNG on import. numerous adjustment to tone, w/brushes, etc. xmp written to file. tried dng export preset as well as through the export module. exported to various versions of camera raw. all with no luck. opened in CS 5 and Preview, with the same results: no adjustments to DNG.
    I have won a free high-end print and the printer says DNG will be better. Any thoughts??? have searched the forum to no avail.....
    Thanks in advance!
    David

    My understanding is tht you have to direct Lr to save the develop settings to the dng files.  Open - Edit - Catalog Settings - Metadata - Check the box "automatically write changes to xmp".

  • Export/import text with all styles (indesign CS3 (XP SP2))

    I'm looking for the best way to export text so authors can edit their text using MS Word, Excel or xml. (they don't want to use Indesign). after they are done editting I want to import the text back into Indesign (with par, char, obj styles).
    I know how to import/export to .doc with styles, but this doesn't work for table, object styles, and page breaks.
    Xml tags or plugin/software might be a solution.
    I've searched the forums, google and read the help file, but still don't know what the best/quickest solution might be. ... ?
    thanks in advance

    1. Those programs don't support every feaature of InDesign, so you are out of luck.
    2. If they don't want to use InDesign, ask them to use InCopy, that is the only one programm which supports the whole InDesgin-functionlity
    3. Page breaks should never be done manually, never in Word, neither with InDesign. They should be created via paragraph styles in both program, but with the import options is a checker to get Word's page breaks.

  • Export / import tablespace with all objects (datas, users, roles)

    Hi, i have a problem or question to the topic export / import tablespace.
    On the one hand, i have a database 10g (A) and on the other hand, a database 11g (B).
    On A there is a tablespace called PRO.
    Furthermore 3 Users:
    PRO_Main - contains the datas - Tablespace PRO
    PRO_Users1 with a role PRO_UROLE - Tablespace PRO
    PRO_Users2 with a role PRO_UROLE - Tablespace PRO
    Now, i want to transfer the whole tablespace PRO (included users PRO_MAIN, PRO_USER1, PRO_User2 and the role PRO_UROLE) from A to B.
    On B, I 've created the user PRO_Main and the tablespace PRO.
    On A , i execute following statement:
    expdp PRO_Main/XXX TABLESPACES=PRO DIRECTORY=backup_datapump DUMPFILE=TSpro.dmp LOGFILE=TSpro.log
    On B:
    impdp PRO_Main/XXX TABLESPACES=PRO DIRECTORY=backup_datapump DUMPFILE=TSpro.dmp LOGFILE=TSpro.log
    Result:
    The User PRO_Main was imported with all the datas.
    But i 'm missing PRO_USER1, PRO_User2 and the role PRO_UROLE...
    I assume, i 've used wrong parameters in my expd and / or impdp.
    It would be nice, if anybody can give me a hint.
    Thanks in advance.
    Best Regards,
    Frank

    When you do a TABLESPACE mode export by specifying just the tablespaces, then all that gets exported are the tables and their dependent objects. The users, roles, and the tablespace definitions themselves don't get exported.
    When you do a SCHEMA mode export by specifying the schemas, you will get the schema definitions (if the schema running the export is privied) and all of the objects that the schema owns. The schema does not own roles or tablespace definitions.
    In your case, you want to move
    1. schemas - which you already created 1 on your target database
    2. roles
    3. everything in the tablespaces owned by multiple schemas.
    There is no 1 export/import command that will do this. This is how i would do this:
    1 - move the schema definitions
    a. you can either create these manually or
    b1. expdp schemas=<your list of schemas> include=user
    b2 impdp the results from b1.
    2. move the roles
    expdp full=y include=role ...
    remember, this will include all roles. If you want to limit what gets exported, then use:
    include=role:"in ('ROLE1', 'ROLE2', ETC.)
    impdo the roles just exported
    3. move the user information
    a. If you want to move all of the schema's objects like functions, packages, etc, then you need to use a schema mode
    export
    expdp user/password schemas=a,b,c ...
    b. If you want to move only the objects in those tablespaces, then use the tablespace export
    expdp user/password tablespaces=tbs1, tbs2, ...
    c. import the dumpfile generated in step 3
    impdp user/password ...
    Hope this helps.
    Dean

  • Export/import table with XMLTYPE data_type and fine_grained policy

    Hi friends!
    I'm trying to export a table with XMLTYPE and faine-grained policy.
    Source: HP-UX - Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    Target: Linux 2.6.18-238.el5 - Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    I do it thru exp/imp old utilities as mention in metalink ID 1318012.1.
    One of the things that surprised me after exp/imp is that the number of objects increased:
    Objects at Source:
    exp \"/ as sysdba\" owner=xml log=xml file=xml.dmp
    Type
    Number of Objects
    TYPE
    431
    TRIGGER
    6
    TABLE
    17
    PROCEDURE
    2
    LOB
    120
    INDEX
    17
    FUNCTION
    1
    On Target:
    imp \"/ as sysdba\" file=xml.dmp fromuser=xml touser=xml log=xml.log
    Type
    Number of Objects
    TYPE
    431
    TABLE
    32
    PROCEDURE
    2
    LOB
    429
    INDEX
    478
    FUNCTION
    1
    Why is this happening? Is it normal?
    Other problem that I found is triggers are not imported, why!?!?
    Thank you very much for your help!
    José

    Hi!
    The new squema was created empty. If I run the following query:
    select '10g', TABLE_NAME,COUNT(*)
    from  DBA_LOBS@DSN_HP
    where OWNER='XML'
    group by TABLE_NAME
    union
    select '11g', table_name,count(*)
    from  DBA_LOBS
    where OWNER='XML'
    group by TABLE_NAME
    order by 2,1
    As a result:
    Version
    Table_name
    Total
    10g
    ACTION_TABLE
    1
    11g
    ACTION_TABLE
    1
    10g
    Document1767_TAB
    14
    11g
    Document1767_TAB
    13
    10g
    Document1852_TAB
    14
    11g
    Document1852_TAB
    13
    10g
    Document1941_TAB
    16
    11g
    Document1941_TAB
    15
    10g
    Document2016_TAB
    14
    11g
    Document2016_TAB
    13
    10g
    Document2087_TAB
    13
    11g
    Document2087_TAB
    12
    10g
    IBT_XML_RECIBIDOS
    1
    11g
    IBT_XML_RECIBIDOS
    1
    10g
    LINEITEM_TABLE
    2
    11g
    LINEITEM_TABLE
    2
    10g
    PURCHASEORDER
    7
    11g
    PURCHASEORDER
    7
    10g
    PurchaseOrder1145_TAB
    9
    11g
    PurchaseOrder1145_TAB
    7
    10g
    RICARDO
    13
    10g
    RICARDO2
    1
    11g
    RICARDO2
    1
    10g
    RITNTFER
    1
    11g
    RITNTFER
    1
    10g
    RITNTFRE_08
    13
    11g
    SYS_NT3+LEU6vbfGLgQ18DLgrURw==
    69
    11g
    SYS_NT3+LEU6vffGLgQ18DLgrURw==
    76
    11g
    SYS_NT3+LEU6vjfGLgQ18DLgrURw==
    63
    11g
    SYS_NT3+LEU6vpfGLgQ18DLgrURw==
    1
    11g
    SYS_NT3+LEU6vqfGLgQ18DLgrURw==
    2
    11g
    SYS_NT3+LEU6vTfGLgQ18DLgrURw==
    65
    11g
    SYS_NT3+LEU6vXfGLgQ18DLgrURw==
    66
    10g
    TESTCLOB
    1
    11g
    TESTCLOB
    1
    There are many new tables created...I suppose because the differences between versions...
    Any ideas?
    Thanks a lot!
    José

  • ?BUG? App Export / Import error with javascript

    Greetings -
    Anyone have an idea on how to get around this or if I'm doing something wrong?
    I have a page button to collapse a tree, with a URL Target of:
    javascript:apex.widget.tree.collapse_all($v(P2130_TREE_ID));
    Page and button work fine, have never had a problem.  Until I went to do an application import and found that the SQL file for the import was corrupted due to the above line.  SQL file has:
         p_button_redirect_url=>'javascript:apex.widget.tree.collapse_all($v(P'||to_char(2130+www_flow_api.g_id_offset)||''_TREE_ID));',
    It appears to have an extra (2x) single quote right before _TREE_ID.  I have this logic in several places in my application and every place caused the same error.
    Application copy also fails with same error.
    Thanks,
    Scott

    Tom -
    I assume this is the prior post you where talking about : https://forums.oracle.com/thread/2401714
    However, this doesn't cover my problem.  I have tried all the follow, and they all break export / import and copy functionality due to mismatched quotes:
    collapse_all($v(P2130_TREE_ID));
    collapse_all($v('P2130_TREE_ID'));
    collapse_all($v("P2130_TREE_ID"));
    collapse_all('&P2130_TREE_ID.');
    collapse_all("&P2130_TREE_ID.");
    These all allow the collapse button to work.
    NOW, if I change my ID tag to be something like 'P_SCOTT_TREE' then everything works just fine (import/export/copy).  It's having a number in the tag that is cause the problems because the export is casting it to_char and somehow throwing in an extra quote.
    Thanks,
    Scott

  • How to export/import data with different nls_lang???

    hello
    now i try to run my forms by APPLICATION SERVER to view the forms as web forms i success but data appear unreadable that it is arabic and i use the nls WE8ISO8859P1 so i change the nls to arabic like AR8MSWIN1256 but i faild so i create new database with charset UTF8 and nls_lang is AR8MSWIN1256 and try to insert new data when i do query about data ......the new data appear readable but the old unreadable so what can i do???
    i think that if i do export data and import it with new (charset and nls_lang) it will be success....but how can i do it??
    thank u
    regards

    Hi did you already exported data from SQL SERVER? if not using SQL*LOADER you cannot export data. SQL*LOADER is only mean for importing data from flat files(usually text files) into ORACLE tables.
    for importing data into oracle tables using sql*loader use below steps
    1) create a sql*loader control file.
    it looks like as follows
    LOAD DATA
    INFILE 'sample.dat'
    BADFILE 'sample.bad'
    DISCARDFILE 'sample.dsc'
    APPEND
    INTO TABLE emp
    TRAILING NULLCOLS
    or for sample script of control file search google.
    2) at command prompt issue following
    $ sqlldr test/test
    enter control file=<give control file name which you create earlier>
    debug any errors (if occured)

  • Copy Folder with Joins. Export/Import Folder with joins. In EUL.

    Ok, I've got a custom folder which has been made up by dragging items from 2 or 3 other folders into it.
    It then has some joins of it's own, quite a few.
    When trying to create a workbook from it, it takes 9 mins to run a query.
    I need to work out what is slowing it down. If I create the same workbook against the folder which has the majority of the items in the custom folder, it runs instantly.
    So I suspect it is one of the joins causing it.
    My plan was to duplicate the folder, then remove joins until I find out which one is causing it.
    However, if I cut n paste the folder, I get a copy without the joins.
    If I export the folder and import it I get a copy without the joins.
    Question then - how can I get a copy of a folder WITH the joins ?
    I'm slightly concerned that when I export my EUL from the dev database and import it into the live database that I'm not going to get any joins since the export and import into the dev database is not retaining the joins.
    Anyone ?

    Hi,
    The preferences for Disco Plus are set in the pref.txt file on the apps server and for Disco Desktop in the Windows Registry. I think the defaults are set on so unless you have changed them this is unlikely to help.
    I think I read somewhere that the 11g optimiser will remove unused outer joins or where there is a foreign key constraint. I may have made that last bit up as I cannot find a reference to it, but it may be worth exploring.
    To speed things up you could look at why this join is slowing things down. It could be that you need an index on the join column.
    The join actually is used, in that it has to check in the other table that a record exists. This is why Discoverer cannot remove the join from the complex folder query. If it did and there where no matching records in the other table then you would get a different result.
    Rod West

  • Full Export/Import Errors with Queue tables and ApEx

    I'm trying to take a full export of an existing database in order to build an identical copy in another database instance, but the import is failing each time and causing problems with queue tables and Apex tables.
    I have used both the export utility and Data Pump (both with partial and full exports) and the same problems are occurring.
    After import, queue tables in my schema are unstable. They cannot be dropped using the queue admin packages as they throw ORA-24002: QUEUE_TABLE <table> does not exist exceptions. Trying to drop the tables causes the ORA-24005: must use DBMS_AQADM.DROP_QUEUE_TABLE to drop queue tables error
    As a result, the schema cannot be dropped at all unless manual data dictionary clean up steps (as per metalink) are done.
    The Apex import fails when creating foreign keys to WWV_FLOW_FILE_OBJECTS$PART. It creates the table ok, but for some reason the characters after the $ are missing so the referencing tables try to refer to WWV_FLOW_FILE_OBJECTS$ only.
    I am exporting from Enterprise Edition 10.2.0.1 and importing into Standard edition 10.2.0.1, but we are not using any of the features not available in standard, and I doubt this would cause the issues I'm getting.
    Can anyone offer any advice on how I can resolve these problems so a full import will work reliably?

    Thanks for the lead!
    After digging around MetaLink some more, it sounds like I'm running into Bug 5875568 (MetaLink Note:5875568.8) which is in fact related to the multibyte character set. The bug is fixed in the server patch set 10.2.0.4 or release 11.1.0.6.

  • Pb export/import keywords with psd file

    While working with export and import of metadata, I found the following strange behavior with LR1.3.1:
    0) Initial keywords
    1) In LR, I modified keywords of a psd file
    2) Export of metadata
    3) Remove all keywords of the picture in LR
    4) Then, reload of keywords.
    => I should get keywords defined at the step 1
    and I always get initial keywords (step 0). It appears than something is not working as I would expect when importing, just exported data.
    Any help to understand this behavior or workaround would be welcome.
    Thanks all,
    PH

    As Curt stated, the keywords are written to the file itself and will not be cleared using the clear and import function in Keywords.
    Using the clear and import option is useful if you want to get rid of the default keywords or have a better new list for import at hand. If you already have assigned some of these old keywords to files and they are not present in your current list then they will show as italic in the keyword panel (usually at the bottom under "Other keywords") once a file with this old keyword is selected in Bridge.
    When a keyword is in italic then you have the option to make it again persistent in your new list using right mouse click menu or leave it as it is. Using the find option in the menu should provide you with the result of all present keywords in the files you are searching in, regardless if they are in the list or not.

  • Exporting/importing photos with Faces tags

    I have an iMac where I keep all photos, movies, etc. and an Air. I loaded all my photos to my Air for a long flight when I tagged them all with Faces and Locations. Now I cannot move them back onto my iMac (where I want to store them for space reasons) with any of that info. Please tell me I have not wasted 8 hours.

    Welcome to the Apple Discussions.
    The only way to do this is with iPhoto Library Manager
    Regards
    TD

  • Export masters with folders

    Hi,
    I'm having a lot of problems with exporting masters with folders (albums).
    I have project like:
    Project
    ->Album1
    ->Album2
    ->Album3
    If I want to export that, Aperture would export only photos with out folder structure.
    I chose "Project Name / Folder Name" in "folder naming presets" but he created folder only for project and not for albums.
    All I want to do is select an project, export masters and have the same folder structure in finder.
    What I'm doing wrong?
    Is not Project->Albums right workflow ?
    I have some two months projects with hounders of Albums and more than 10000 images. It would be very time consuming creating all this folders and arranging images in finder again.
    I would be really glad if someone could help me with this.
    thx and regards,
    Gregor

    I also wanted to do the same for a long time, because that used to be the way I worked before using aperture, but ultimately, it doesn't really matter that much. If you export your masters with xmp sidecars, you can easily search them with any other dam software, or even find the xmp's (and hence the image) with spotlight.
    I think the reason why you can't use album names while exporting is that albums are an arbitrary collection of images. You can have an album in one project containing images of another one, for example. There's also no way to avoid duplicates on export when exporting albums.
    That being said, it would be possible to create an applescript to do what you want, without too much fuss.
    j

  • Export complete database with InterMedia tables

    Hi all,
    I need to move my developement database for one schema to the same schema on another server. So I started up the "exp.exe" app and run trough the processes.
    I then moved the *.dmp file to the server and started "imp.exe". All tables, grants and alike are imported successfully, except the tables which are created with Intermedia.
    Is there a problem with exporting/importing tables with Intermedia columns? If so, is there a workaround? I need to have the other database rolling next week.
    Thank you.

    I only received your reply by email notification today....
    Yes, we are storing in the ORDSYS schema. This is how we have been told to store media's. From the manual it notes "Note that ALL Intermedia objects and procedures are defined in the ORDSYS schema".
    Better to store them in the schema that the user is using? Will then everyhting still work (Image manipulation, etc.)?

  • Export ,import efficiency

    Hi,
    I have a basic question here. Which is more efficient exp/imp or the expdp/impdp and why? I would like to read about it. good documents are welcomed.
    Thanks
    Kris

    Hi ,
    Definetly expdp/impdp(Datapump export import) is much better than original exp/imp which is more
    used in oracle 9i Databases.
    Top 10 difference between exp/imp(export/import) and expdp/impdp(Datapump export and import) are:
    1)Data Pump Export and Import operate on a group of files called a dump file set
    rather than on a single sequential dump file.
    2)Data Pump Export and Import access files on the server rather than on the client.
    This results in improved performance. It also means that directory objects are
    required when you specify file locations.
    3)The Data Pump Export and Import modes operate symmetrically, whereas original
    export and import did not always exhibit this behavior.
    For example, suppose you perform an export with FULL=Y, followed by an import using SCHEMAS=HR. This will produce the same results as if you performed an
    export with SCHEMAS=HR, followed by an import with FULL=Y.
    4)Data Pump Export and Import use parallel execution rather than a single stream of
    execution, for improved performance. This means that the order of data within
    dump file sets and the information in the log files is more variable.
    5)Data Pump Export and Import represent metadata in the dump file set as XML
    documents rather than as DDL commands. This provides improved flexibility for
    transforming the metadata at import time.
    6)Data Pump Export and Import are self-tuning utilities. Tuning parameters that
    were used in original Export and Import, such as BUFFER and RECORDLENGTH,
    are neither required nor supported by Data Pump Export and Import.
    7)At import time there is no option to perform interim commits during the
    restoration of a partition. This was provided by the COMMIT parameter in original
    Import.
    8)There is no option to merge extents when you re-create tables. In original Import,
    this was provided by the COMPRESS parameter. Instead, extents are reallocated
    according to storage parameters for the target table.
    9)Sequential media, such as tapes and pipes, are not supported.
    10)The Data Pump method for moving data between different database versions is
    different than the method used by original Export/Import. With original Export,
    you had to run an older version of Export (exp) to produce a dump file that was
    compatible with an older database version. With Data Pump, you can use the
    current Export (expdp) version and simply use the VERSION parameter to specify the target database version
    For more details and options:
    exp help=y
    imp help=y
    expdp help=y
    impdp help=y
    Fine manuals for referring:
    http://www.oracle-base.com/articles/10g/OracleDataPump10g.php
    Hope it helps.
    Best regards,
    Rafi.
    http://rafioracledba.blogspot.com

Maybe you are looking for