Problems with data merge (CS4) - puts all fields in one

Data merge is acting very weird, I have a standard textfile of six columns exported to .csv. When I import it in InDesign it puts all fields in one place (see image) and it becomes totally useless. Has anyone had this problem and if so, how did you solve it?
Thank you!

It is a regular textfile (http://www.pixentral.com/show.php?picture=1k0bmos4r2l7lCtegA1CGHIcw98jrM
) divided into columns...
28 apr 2009 kl. 23.00 skrev Peter Spier:
I took a look at the image, now that it's out of the queue, and the 
fields seem to be seaparted by semi-colons. How did you make the file?
>
ID recognizes comma-sparated (CSV) or tab-delimited text files.
>
Peter
>

Similar Messages

  • Strange Behavior with Data Merge (CS4) - Advice?

    This is the most bizarre behavior and I may have to chalk it up to a full moon or something since I can't reproduce it and I cannot find any hints as why this happened. This is my last attempt at figuring a solution...maybe someone has experienced this??? Anyway, here's what happened...
    I do a data merge for addressing envelopes. Been using the InDesign data merge feature for a couple of years now, no problems.
    My address list is approximately 1200. It's in a .csv file with the following fields: fname, lname, company, address1, address2, city, state, zip
    For my last mailing, approximately 300 addresses were merged wrong. Here's the crazy thing it did: instead of using the address1 field for the current record, it used the address1 field from the NEXT record, and ONLY the address1 field.
    Here's more bizarre-ness. The NEXT record was merged correctly.
    For example:
    Record 1 (address1 should be 555 Cherry Lane, Apt 5)
    John Smith
    Smith Company
    123 Main Street, Apt 5
    Small Town, CA 12345
    Record 2 (John Smith's address used THIS address1, but NOT address2 or any other field! Jane Doe's address is correct)
    Jane Doe
    Doe Company
    123 Main Street, Suite 300
    Big Town, TX 56789
    I've tested. I've looked for commonality in the records, where it's placed, if there was something unusual about it's format. I come up empty. The incorrect data seem randomly interspersed.
    One thing that I have to do is break up the actual merge. Either my computer memory or InDesign can't handle 1200 merge records. So, I merge approximately 200 records at a time (I make 6 print files). I don't know if this has anything to do with it (like I said, the incorrect data is interspersed all through out), but, thought I'd mention it.
    Any advice is appreciated. I'm hesitant to use it again until I can understand why this happened.
    thanks much,
    Kia

    CS was soemtimes flakey in this way with Data Merge.
    Are you updated to the latest patch? Most issues were resolved at some point.
    That said, I recall another user with this problem, and it never resolved with his particular file. I examined the data file and couldn't find any issues, but it behaved identically on my system tot he way it did on his so I can only conclude there was still some bug, or something about the data file that we were unable to detect. I seem to recall that the same data file merged correctly in CS3, but I don't recall if I saved it long enough to test in CS5. I've not had any issues with my own files in any version.
    Open the file in a plain text editor and look for any odd characters (I'd start the search where the merge first fails). Also do a Save As from the text editor to make sure there is no possibility of formatting from your spread sheet having found its way in.

  • Problems with data merge

    I am trying to do a data merge in InDesign CS5 to make an employee directory. I have placed the employee information into a google docs spreadsheet and downloaded it as a csv file. I was partly successful once in getting the file into the document, but it wasn’t complete. More specifically, not all the titles of the columns loaded into the data merge box. One was missing and it was the photos column. In the spreadsheet I titled the photos column like this: @photos
    I have tried many times since, but now nothing happens at all when I select the data source. I’ve even rebooted my computer, but it didn’t help. Thanks for any help. Merissa

    If Google Docs is like Excel, you may need to title that column as    '@photos  in your spreadsheet. (the apostrophe before the @ prevents the spreadsheet from seeing the @ as an operator.)

  • IDOC: Problem with data filter for IDOC extension field

    Hallo!
    I've created an idoc extension for the basic type DEBMAS06 that works fine. Now I want to use a data filter for one field ( company code ) of my segment. Every segment with a company code different from 100 should be filtered and not send to the other client. But what happend is that for all customers that have at least one company code different from 100, all segments including the one with cc 100 were deleted and a error "Segment ... does not exist for message type DEBMAS" appeared on the screen.
    Does anyone have any ideas about this problem?

    Not sure about changes to be made at the filtering options.
    An alternative would be sending the data to XI as it is and perform the mapping to remove the unnecessary segments.
    Disadvantage: Unnecessary processing of segment would be done at XI.
    Advantage: The integration logic would be completely handled by XI.
    Regards,
    Prateek

  • Problem with Change Document FM  - Not all fields are tracked in CDPOS

    Hello everybody,
    Using transaction SCDO and following necessary steps described in forum, I have created a change document ZFBR to track changes of a table ZFBR. The problem is, the generated FM ZFBR_WRITE_DOCUMENT does not track changes of some fields of the table (inserts are always ok, problem is with updates). After some debugging I realized that the FM DDIF_NAMETAB_GET was returning the table DFIES_TAB where the LOGFLAG was empty for some fields (which are not tracked) so that changes of these fields were not visible in table CDPOS. However, fields with LOGFLAG = 'X' are tracked succesfully (changes are visible in table CDPOS).
    I did not specify any fields to be get logged or not during the steps in SCDO (can we do such a specification anywhere?), does anyone know why the field LOGFLAG has the value '' (space)?
    Best regards;
    Ozcan.

    Hi again,
    The problem has been solved.
    The data elements used in a table are the reason for the problem I have described above. When displaying the details of a data element, under tab "Further Characteristics" , the "Change Document" checkbox has to be selected to enable change document tracking. I have copied the problematic data elements with Z prefix and made the necessary enabling. I have replaced the data elements with their Z equivalents.
    Best regards.
    Ozcan.

  • Help please, i have a problem with DATA MERGE !

    why my coloumns are in one line ? after clicking "select data source" i have this problem, in csv file is everything allright !

    hmm, i noticed one thing, when i save file in .CSV, and after i save the same file in .XLSX, in ID it opens .CSV like this
    but when i save file in excel as CSV again, and re-open it in ID then it shows me that error

  • Having a problem with dates when I send my numbers doc to excel. dates are all out and that they have to cut and paste individual entries onto their spreadsheet. Any idea how I can prevent this

    having a problem with dates when I send my numbers doc to excel. dates are all out and that they have to cut and paste individual entries onto their spreadsheet. Any idea how I can prevent this.
    I'm using Lion on an MBP and Numbers is the latest version

    May you give more details about what is wrong with your dates ?
    M…oSoft products aren't allowed on my machines but I use LibreOffice which is a clone of Office.
    When I export from Numbers to Excel and open the result with LibreOffice, the dates are correctly treated.
    To be precise, dates after 01/01/1904 are correctly treated. dates before 01/01/1904 are exported as strings but, as it's flagged during the export process, it's not surprising.
    Yvan KOENIG (VALLAURIS, France) mardi 3 janvier 2012
    iMac 21”5, i7, 2.8 GHz, 12 Gbytes, 1 Tbytes, mac OS X 10.6.8 and 10.7.2
    My iDisk is : http://public.me.com/koenigyvan
    Please : Search for questions similar to your own before submitting them to the community
    For iWork's applications dedicated to iOS, go to :
    https://discussions.apple.com/community/app_store/iwork_for_ios

  • Has anyone else had a problem with data? after updating my phone to the new update, i have all this data, and keep getting notifications that my data is getting high. i never had this problem before.

    has anyone else had a problem with data? after updating my phone to the new update, i have all this data, and keep getting notifications that my data is getting high. i never had this problem before.

    fair enough.  No need for any unnecessary posts either.  You issue had been addressed ad nauseum already in this forum had you bothered to search this forum (as forum etiquette would dictate) before posting.
    In any case, I hope that your problem was solved.

  • Problem with date format from Oracle DB

    Hi,
    I am facing a problem with date fields from Oracle DB sources. The date format of the field in DB table is 'Date base type is DATE and DDIC type is DATS'.
    I mapped the date fields to Date characters in BI. Now the data that comes to PSA is in weird format. It shows like -0.PR.09-A
    I have changing the field settings in DataSource  to internal and external and also i have tried mapping these date fields to text fields with out luck. All delivers the same format.
    I have also tried using conversion routines like, CONVERSION_EXIT_IDATE_INPUT to change format. It also delivers me the same old result.
    If anybody of you have any suggestions or if anybody have you experienced such probelms, Please share your experience with me.
    Thanks in advance.
    Regards
    Varada

    Thanks for all your reply. I can only the solutions creating view in database. I want some solution to be done in BI. I appreciate if some of you have idea in it.
    The issue again in detail
    I am facing an issue with date fields from oracle data. The data that is sent from Oracle is in the format is -0.AR.04-M. I am able to convert this date in BI with conversion routine in BI into format 04-MAR-0.
    The problem is,  I am getting data of length 10 (Output format) in the format -0.AR.04-M where the month is not in numericals. Since it is in text it is taking one character spacing more.
    I have tried in different ways to convert and increased the length in BI, the result is same. I am wondering if we can change the date format in database.
    I am in puzzle with the this date format. I have checked other Oracle DB connections data for date fields in BI, they get data in the format 20.081.031 which will allow to convert this in BI. Only from the system i am trying creating a problem.
    Regards
    Varada

  • Importing CSV file with Data Merge Fails

    Specs
    See pasted text from CSV at http://pastebin.com/mymhugpN
    I am using InDesign CS6 (8.0.1)
    I created the CSV by downloading it from a Google Spreadsheet as a CSV. I confirm with the Terminal that the character encoding is utf-8 usnig the file command.
    Problem detailed
    I am trying to import a CSV file (utf-8) with Data Merge via the Select Data Source... command with Show Import Options checked. When viewing the Data Source Import Options dialog, I set the following options—Delimiter:Comma, Encoding:Unicode, Platform:Macintosh. I leave Preserve Spaces in Data Source unchecked. It fails to import any variables and produces no error message. I have tried other CSV files as well (created TextEdit, Espresso, etc.) and it seems that InDesign will not import any files if Unicode is specified as the encoding, no matter which other options are specified.
    Can anyone else confirm this?
    Importing as ACSII works, but obviously does not display my content correctly.

    Mike is having some trouble posting in this thread (and I am too), but he sent me a PM with what he wanted to say:
    OK. I think I might have a positive answer for you.
    I was getting lost in the upper ASCII characters you showed. In your test file I never could see any--a case of not seeing the trees for the forest.
    Your quote marks are getting dropped in your test file. Now, this may or may not affect other factors but it does in some further testing. I believe ID has an issue with dropping quote marks even in a plain ASCII file if the marks are at the beginning of a sentence and the file is tab delimited. Call it a bug.
    Because of all the commas and quote marks in your simple file, I think you should be exporting from Google Docs' spreadsheet as a tab-delimited file. This exported file has to be opened in a text editor capable of saving it out as a UTF-16 BE (Big Endian) type of file.
    Also, I think you are going to have to use proper quote marks throughout, or change them in the exported tab-delimited file. Best to have a correct source, though.
    Here is your sample ZIPped up. I think it works properly. But then again, I think I might be bleary-eyed by now.
    http://www.wenzloffandsons.com/temp/merge_psalms_utf-16.zip
    Take care, Mike

  • Automatic printing to location, with data merge?

    Hello all,
    I am quite new to this, and i have a question...
    I am using Data Merge with large number of values to fill in fields in indesign...but for each value line, printing has to be done to different folder. Is there any way to set printing location with data merge? 

    Try this:
    Save your file as a UTF-16 BE (Big Endian) file.
    Import showing options in ID. I'm on a PC, choose the below regardless.
    Should look like this in ID.
    Apologies to Farsi-speaking people everywhere. I pulled some text out of a Farsi text file to make up this tab-delimited merge file. I don't speak it, so it is likely servely non-sensical.
    The text editor being used here (first screen shot) is the OpenSource NotePad++. I am also not using the ME version of ID.
    Take care, Mike

  • Problem with date format dd/mm/yyyy. But I need to convert yyyy-mm-dd.

    Dear friends,
    I have the problem with date format. I receiving the date with the format dd/mm/yyyy. But I can upload to MySQL only in the format of yyyy-mm-dd.
    how should I handle this situation, for this I've created these code lines.But I have some problem with these line. please help me to solve this problem.
    String pattern = "yyyy-mm-dd";
    SimpleDateFormat format = new SimpleDateFormat(pattern);
    try {
    Date date = format.parse("2006-02-12");
    System.out.println(date);
    } catch (ParseException e) {
    e.printStackTrace();
    System.out.println(format.format(new Date()));
    this out put gives me Tue Apr 03 00:00:00 IST 2007
    But I need the date format in yyyy-mm-dd.
    regards,
    maza
    thanks in advance.

    Thanks Dear BalusC,
    I tried with this,
    rs.getString("DATA_SCAD1")// where the source from .xls files
    String pattern = "yyyy-MM-dd";
    SimpleDateFormat format = new SimpleDateFormat(pattern);
    try {
    Date date = format.parse("DATA_SCAD1");
    System.out.println(date);
    } catch (ParseException e) {
    e.printStackTrace();
    System.out.println(format.format(new Date()));
    this out put gives me Tue Apr 03 00:00:00 IST 2007
    But I want to display the date format in yyyy-mm-dd.
    regards,
    maza

  • I have one problem with Data Guard. My archive log files are not applied.

    I have one problem with Data Guard. My archive log files are not applied. However I have received all archive log files to my physical Standby db
    I have created a Physical Standby database on Oracle 10gR2 (Windows XP professional). Primary database is on another computer.
    In Enterprise Manager on Primary database it looks ok. I get the following message “Data Guard status Normal”
    But as I wrote above ”the archive log files are not applied”
    After I created the Physical Standby database, I have also done:
    1. I connected to the Physical Standby database instance.
    CONNECT SYS/SYS@luda AS SYSDBA
    2. I started the Oracle instance at the Physical Standby database without mounting the database.
    STARTUP NOMOUNT PFILE=C:\oracle\product\10.2.0\db_1\database\initluda.ora
    3. I mounted the Physical Standby database:
    ALTER DATABASE MOUNT STANDBY DATABASE
    4. I started redo apply on Physical Standby database
    alter database recover managed standby database disconnect from session
    5. I switched the log files on Physical Standby database
    alter system switch logfile
    6. I verified the redo data was received and archived on Physical Standby database
    select sequence#, first_time, next_time from v$archived_log order by sequence#
    SEQUENCE# FIRST_TIME NEXT_TIME
    3 2006-06-27 2006-06-27
    4 2006-06-27 2006-06-27
    5 2006-06-27 2006-06-27
    6 2006-06-27 2006-06-27
    7 2006-06-27 2006-06-27
    8 2006-06-27 2006-06-27
    7. I verified the archived redo log files were applied on Physical Standby database
    select sequence#,applied from v$archived_log;
    SEQUENCE# APP
    4 NO
    3 NO
    5 NO
    6 NO
    7 NO
    8 NO
    8. on Physical Standby database
    select * from v$archive_gap;
    No rows
    9. on Physical Standby database
    SELECT MESSAGE FROM V$DATAGUARD_STATUS;
    MESSAGE
    ARC0: Archival started
    ARC1: Archival started
    ARC2: Archival started
    ARC3: Archival started
    ARC4: Archival started
    ARC5: Archival started
    ARC6: Archival started
    ARC7: Archival started
    ARC8: Archival started
    ARC9: Archival started
    ARCa: Archival started
    ARCb: Archival started
    ARCc: Archival started
    ARCd: Archival started
    ARCe: Archival started
    ARCf: Archival started
    ARCg: Archival started
    ARCh: Archival started
    ARCi: Archival started
    ARCj: Archival started
    ARCk: Archival started
    ARCl: Archival started
    ARCm: Archival started
    ARCn: Archival started
    ARCo: Archival started
    ARCp: Archival started
    ARCq: Archival started
    ARCr: Archival started
    ARCs: Archival started
    ARCt: Archival started
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    ARC1: Becoming the heartbeat ARCH
    Attempt to start background Managed Standby Recovery process
    MRP0: Background Managed Standby Recovery process started
    Managed Standby Recovery not using Real Time Apply
    MRP0: Background Media Recovery terminated with error 1110
    MRP0: Background Media Recovery process shutdown
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[1]: Assigned to RFS process 2148
    RFS[1]: Identified database type as 'physical standby'
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[2]: Assigned to RFS process 2384
    RFS[2]: Identified database type as 'physical standby'
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[3]: Assigned to RFS process 3188
    RFS[3]: Identified database type as 'physical standby'
    Primary database is in MAXIMUM PERFORMANCE mode
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: No standby redo logfiles created
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[4]: Assigned to RFS process 3168
    RFS[4]: Identified database type as 'physical standby'
    RFS[4]: No standby redo logfiles created
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: No standby redo logfiles created
    10. on Physical Standby database
    SELECT PROCESS, STATUS, THREAD#, SEQUENCE#, BLOCK#, BLOCKS FROM V$MANAGED_STANDBY;
    PROCESS STATUS THREAD# SEQUENCE# BLOCK# BLOCKS
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    RFS IDLE 0 0 0 0
    RFS IDLE 0 0 0 0
    RFS IDLE 1 9 13664 2
    RFS IDLE 0 0 0 0
    10) on Primary database:
    select message from v$dataguard_status;
    MESSAGE
    ARC0: Archival started
    ARC1: Archival started
    ARC2: Archival started
    ARC3: Archival started
    ARC4: Archival started
    ARC5: Archival started
    ARC6: Archival started
    ARC7: Archival started
    ARC8: Archival started
    ARC9: Archival started
    ARCa: Archival started
    ARCb: Archival started
    ARCc: Archival started
    ARCd: Archival started
    ARCe: Archival started
    ARCf: Archival started
    ARCg: Archival started
    ARCh: Archival started
    ARCi: Archival started
    ARCj: Archival started
    ARCk: Archival started
    ARCl: Archival started
    ARCm: Archival started
    ARCn: Archival started
    ARCo: Archival started
    ARCp: Archival started
    ARCq: Archival started
    ARCr: Archival started
    ARCs: Archival started
    ARCt: Archival started
    ARCm: Becoming the 'no FAL' ARCH
    ARCm: Becoming the 'no SRL' ARCH
    ARCd: Becoming the heartbeat ARCH
    Error 1034 received logging on to the standby
    Error 1034 received logging on to the standby
    LGWR: Error 1034 creating archivelog file 'luda'
    LNS: Failed to archive log 3 thread 1 sequence 7 (1034)
    FAL[server, ARCh]: Error 1034 creating remote archivelog file 'luda'
    11)on primary db
    select name,sequence#,applied from v$archived_log;
    NAME SEQUENCE# APP
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00003_0594204176.001 3 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00004_0594204176.001 4 NO
    Luda 4 NO
    Luda 3 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00005_0594204176.001 5 NO
    Luda 5 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00006_0594204176.001 6 NO
    Luda 6 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00007_0594204176.001 7 NO
    Luda 7 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00008_0594204176.001 8 NO
    Luda 8 NO
    12) on standby db
    select name,sequence#,applied from v$archived_log;
    NAME SEQUENCE# APP
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00004_0594204176.001 4 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00003_0594204176.001 3 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00005_0594204176.001 5 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00006_0594204176.001 6 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00007_0594204176.001 7 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00008_0594204176.001 8 NO
    13) my init.ora files
    On standby db
    irina.__db_cache_size=79691776
    irina.__java_pool_size=4194304
    irina.__large_pool_size=4194304
    irina.__shared_pool_size=75497472
    irina.__streams_pool_size=0
    *.audit_file_dest='C:\oracle\product\10.2.0\admin\luda\adump'
    *.background_dump_dest='C:\oracle\product\10.2.0\admin\luda\bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='C:\oracle\product\10.2.0\oradata\luda\luda.ctl'
    *.core_dump_dest='C:\oracle\product\10.2.0\admin\luda\cdump'
    *.db_block_size=8192
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_file_name_convert='luda','irina'
    *.db_name='irina'
    *.db_unique_name='luda'
    *.db_recovery_file_dest='C:\oracle\product\10.2.0\flash_recovery_area'
    *.db_recovery_file_dest_size=2147483648
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=irinaXDB)'
    *.fal_client='luda'
    *.fal_server='irina'
    *.job_queue_processes=10
    *.log_archive_config='DG_CONFIG=(irina,luda)'
    *.log_archive_dest_1='LOCATION=C:/oracle/product/10.2.0/oradata/luda/ VALID_FOR=(ALL_LOGFILES, ALL_ROLES) DB_UNIQUE_NAME=luda'
    *.log_archive_dest_2='SERVICE=irina LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=irina'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_max_processes=30
    *.log_file_name_convert='C:/oracle/product/10.2.0/oradata/irina/','C:/oracle/product/10.2.0/oradata/luda/'
    *.open_cursors=300
    *.pga_aggregate_target=16777216
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=167772160
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='C:\oracle\product\10.2.0\admin\luda\udump'
    On primary db
    irina.__db_cache_size=79691776
    irina.__java_pool_size=4194304
    irina.__large_pool_size=4194304
    irina.__shared_pool_size=75497472
    irina.__streams_pool_size=0
    *.audit_file_dest='C:\oracle\product\10.2.0/admin/irina/adump'
    *.background_dump_dest='C:\oracle\product\10.2.0/admin/irina/bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='C:\oracle\product\10.2.0\oradata\irina\control01.ctl','C:\oracle\product\10.2.0\oradata\irina\control02.ctl','C:\oracle\product\10.2.0\oradata\irina\control03.ctl'
    *.core_dump_dest='C:\oracle\product\10.2.0/admin/irina/cdump'
    *.db_block_size=8192
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_file_name_convert='luda','irina'
    *.db_name='irina'
    *.db_recovery_file_dest='C:\oracle\product\10.2.0/flash_recovery_area'
    *.db_recovery_file_dest_size=2147483648
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=irinaXDB)'
    *.fal_client='irina'
    *.fal_server='luda'
    *.job_queue_processes=10
    *.log_archive_config='DG_CONFIG=(irina,luda)'
    *.log_archive_dest_1='LOCATION=C:/oracle/product/10.2.0/oradata/irina/ VALID_FOR=(ALL_LOGFILES, ALL_ROLES) DB_UNIQUE_NAME=irina'
    *.log_archive_dest_2='SERVICE=luda LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=luda'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_max_processes=30
    *.log_file_name_convert='C:/oracle/product/10.2.0/oradata/luda/','C:/oracle/product/10.2.0/oradata/irina/'
    *.open_cursors=300
    *.pga_aggregate_target=16777216
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=167772160
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='C:\oracle\product\10.2.0/admin/irina/udump'
    Please help me!!!!

    Hi,
    After several tries my redo logs are applied now. I think in my case it had to do with the tnsnames.ora. At this moment I have both database in both tnsnames.ora files using the SID and not the SERVICE_NAME.
    Now I want to use DGMGRL. Adding a configuration and a stand-by database is working fine, but when I try to enable the configuration DGMGRL gives no feedback and it looks like it is hanging. The log, although says that it succeeded.
    In another session 'show configuration' results in the following, confirming that the enable succeeded.
    DGMGRL> show configuration
    Configuration
    Name: avhtest
    Enabled: YES
    Protection Mode: MaxPerformance
    Fast-Start Failover: DISABLED
    Databases:
    avhtest - Primary database
    avhtestls53 - Physical standby database
    Current status for "avhtest":
    Warning: ORA-16610: command 'ENABLE CONFIGURATION' in progress
    It there anybody that experienced the same problem and/or knows the solution to this?
    With kind regards,
    Martin Schaap

  • I am still having problems with my music videos freezing (all purchased from iTunes) since the update.  Have restarted, rebooted etc. and it is worse than before.  Help!  Spent a lot of money on these and they won't play properly.

    I am still having problems with my music videos freezing (all purchased on iTunes) since the update.  The first day of the update, no problems...played 6 hours of music videos and no freezing.  Today...every single video froze!!! So frustrating!  I rebooted, restarted, powered down and no help.   I would appreciate any suggestions.  Thanks.

    I don't know how to do this without using a mouse point.
    This may help, http://www.computerhope.com/issues/ch000542.htm
    As for question #2, you call tell how iTunes is sorting songs by the up or down arrow next to a header name.
    Even the original order that the songs were entered or date added or last played can be sorted.
    Much easier with a mouse or touch pad.

  • Assistance with Data Merge for product sell sheets

    Last year I started creating product sell sheets for our sales team. The sell sheet is built containg these elements: A product photo, a data table displaying product numbers, descriptions etc.. Underneath the photo and table is a text box. I have created a lot of these sell sheets for a variety of product categories. The prices are updated fairly often, maybe once every month or two. I created all of these using excel and Illustrator. Other than photoshop, Adobe products were forign to me and never looked into what indesign was capable of until now.
    I am wondering if the data merge feature would be a good fit for what I am doing. I would be more than happy to share my project files. I am asking for a little bit of hand holding here to expedite this process. Anyone out there willing to take a look and help me out??
    Project Files:
    PDF Sell Sheet created in Illustrator (would like to have the same look and feel but have the data automated in Indesign)
    Excel Data (Looking to reformat the way the data is displayed so that it would work well with Data Merge)

    I don' think data merge will lend itself well to that sort of layout. It's just not sophisticated enough. For datamerge, all the data for each module has to be in one row on your excel sheet. Example: The setup on top will yield the example below. The image has a placeholder in the far left.
    For your example, I think your best bet is to set up table styles. (table styles won't hold column widths, but there is a script that you can use for that: it's referenced in this thread. If you make good use of paragraph styles and table styles, it will still go much more quickly in InDesign.
    http://forums.adobe.com/message/4929669

Maybe you are looking for

  • Create a report based on Workflow entities status.

    Hello guys! I want to make a report that shows me all entities grouped by workflow status (Not Started, First Pass, Under Review, etc.). It is possible? All answers will be accepted!! Regards! Fabian

  • File Content conversion at the sender adapter without Key field

    Hi All, I have a requirement ....We are using Message Transform Bean at the sender adapter (we are not going for FCC as we are tranforming a decryption) ..its is a flat file... we dnot have a key field for one of the node .....we have total 5 nodes u

  • Bank account formats - checks

    Dear all, Does someone have any idea how I can see how the format of the bank accounts are checked in SAP? (for different bank countries)? I need to clean a list with bank accounts so their format is accepted by SAP when I upload them. Is there any p

  • Defining default field for cursor in an infotype

    What defines which field on an infotype (e.g. it0032) that the cursor defaults to.  I'm having problems due to the cursor defaulting to the "from" date field in our development clients, and it defaulting to the "previous personnel number" field in ou

  • Problem with javaplugin1.3.1_01

    I am using javaplugin1.3.1_01. My applet window opens from the login page. I keep this applet window open and I try to open a new applet window (same applet) from a new login page.But the new applet window doesnot open up, instead the previous applet