Problem in handling multiple source of Complex CSV files in ODI

Hi,
I have a requirement where I have multiple CSV files.
In Details:
when I have one CSV file I use the following syntax in JDBC URL in creation of Data Server under Complex file of Topology navigator.
jdbc:snps:complexfile?f=C:\pqr\pqr.csv&d=C:\pqr\pqr.xsd&re=Root
When there are multiple CSV files what would be the syntax as? will it be like
jdbc:snps:complexfile?f=C:\pqr\*.csv&d=C:\pqr\pqr.xsd&re=Root (I am getting error as unexpected token * when I am using * and clicking on Test Connection Button).
Please help..
Thanks.

Hi,
first of all according to documentation use backslash
http://docs.oracle.com/cd/E21764_01/integrate.1111/e12644/appendix_complex_driver.htm#CCHDJGHF
in second place under my opinion you are not able to use mask to identify more file. A workaround is to use a variable to store the file name. BUT i've never used Complex Files, so i'm speculating about it.

Similar Messages

  • Problem in handling complex CSV file

    Hi,
    I am facing a strange problem in handling a complex CSV file.
    The content of the file is as follows. It is getting executed through an XSD and the target is a table.
    First time when the interface is executed ( for e.g session id - 19001) it is getting completed successfully.
    Now when making changes in the CSV file suppose adding ( for e.g.-
    DEPT,D05,Books,ABC Store
    CUST,C01,Sayantan,[email protected]
    CUST,C02,Shubham,[email protected])
    or deleting ( for e.g.-
    CUST,C02,Sarbajit,[email protected])
    and running the interface I am not getting the added D05 in the target table nor my C02 data is getting removed i.e.- the updated data in CSV file is not getting fetched and I am getting the same records as i got when i ran interface of session id 19001.
    I am not getting why it is happening??
    The CSV file used in session id -19001 is:
    DEPT,D01,Retail,World Mart
    CUST,C01,Anindya,[email protected]
    CUST,C02,Rashni,[email protected]
    DEPT,D02,Food,Food Bazar
    CUST,C01,Abhijit,[email protected]
    CUST,C02,Anirban,[email protected]
    CUST,C03,Sharmistha,[email protected]
    DEPT,D03,Water,SW
    CUST,C01,Nirmalya,[email protected]
    DEPT,D04,Clothes,City style
    CUST,C02,Sarbajit,[email protected]
    CUST,C03,Abhishek,[email protected]

    Here's what you can do to handle CSV files using HSQL.
    Say the CSV file contains order data. Each order record contains data about a particular order like order number, product code, quantity, ordered by, date etc. Now lets take a hypothetical requirement which says, the user needs to know how many orders were ordered per product.
    Option 1]  At the basic level, read the order file, parse each line and write the logic to get count of orders per product.
    Option 2] Load this CSV file into a database like MySQL and write database queries to get the orders per product.
    So where does this leave us? Have we run out of options? I am sure we would have tried & used the above two options . But I wanted a different approach. Following questions were lingering in my mind:
    1] Why can’t I write SQL queries against the CSV file itself. After all it’s like any RDBMS table.
    2] Why should I load the CSV file into some database before I query it?
    3] Why can’t I create a database table & attach the CSV file to it?
    Continue Reading Here: Handling CSV Files

  • Power shell script to add multiple aliases with a .csv file.

    I have an email address lets say [email protected] and I need 1000 aliases added to the email. For example [email protected], [email protected] etc etc. and I have a .csv file with 990 columns with just project1, project2.
    Is there a way to run a command in powershell exchange to import all the aliases I need in a few commands? Can someone point me in the right direction if I am in the wrong forum also? Thanks
    No sure if this is correct or not:
    > $mbx = Get-Mailbox Project
    > Import-CSV "C:\SomePath\wherever.csv" | foreach { $mbx.EmailAddresses += $._SmtpAddress }
    > Set-Mailbox project -EmailAddresses $mbx.EmailAddresses

    Help Import-Csv -FULL
    $aliases=Import-Csv aliases.csv
    ¯\_(ツ)_/¯

  • Create Response groups multiple reponse groups using CSV file

    Hi Champs,
    We have good amount of response groups has to be created, I try to write the script but failed. Can any one help me to create Response groups through script.
    In below script agents should be taken from CSV file.
    Import-Module Lync
    $serviceId="service:"+(Get-CSService | ?{$_.Applications -like "*RGS*"}).ServiceId;
    $ag = New-CsRgsAgentGroup -Name "agent group" -Parent $serviceId;
    $ag.Description = "Contain the agents";
    $ag.ParticipationPolicy = "Formal"
    $ag.AgentAlertTime = "20"
    $ag.AgentsByUri.Add("sip:[email protected]")
    $ag.AgentsByUri.Add("sip:[email protected]")
    $ag.AgentsByUri.Add("sip:[email protected]")
    Set-CsRgsAgentGroup $ag
    Regards
    Vijendhar

    You also need to create queue and workflow. Please check how to create Response Group using Lync Server Management Shell at
    http://blogs.technet.com/b/csps/archive/2010/09/15/rgscreateresponsegroup.aspx.
    Lisa Zheng
    TechNet Community Support

  • Remove column name in generated CSV file through ODI

    Hi,
    I want to remove the column names in CSV file which I am creating through database.
    I have tried putting "heading to zero in the the datastore", which I created for csv file, but it is not working.
    Please assist.
    Thanks,
    Sudeep Naik

    Okay let me explain.After creating your csv file you open it.If 1st row contains column name then during reverse you use heading as 1. Regarding c1,c2 these are the column name defined by odi.Its not going to come as your data.The way i told you just make it by giving header 1 then reverse it.After that just drag it to your interface.Then on interface right click on source table and click on view data.Here you can find if your column name is coming as data or not.Hope you are clear now.
    Thanks.

  • Merge/Concat Multiple Columns in a Flat-File wthin ODI

    Hi,
    I was wondering if there is a quick way to merge/concat multiple columns in a Flat-File from within ODI, a functionality kinda similar to Essbase load rule files for join/split columns etc.
    Please let me know.
    Thanks

    Looks like the CONCAT function works only with Memory Engine. Could someone please confirm this? If this is true, I guess applying the patch is the only resolution.
    Thanks

  • Problem when handling multiple messages using sequential convoy

    Hi Guys,
    We receive two files from FTP server. One is Order file and other one is notes file which is optional one. If the note flag in order file is 1 then we should process notes file as well. Here we implemented sequential convoy and it was working fine when we
    receive order file first and then notes file. it is throwing error when we receive notes file first as we set initialize correlation for Order file.  Please share your ideas what to do when we receive notes file first than order file?. Thanks!!

    Hi,
    i added debug points and ran it. i dropped one order file and one note file but it is initializing active instance for both order and note files. Here is the order of logs.
    Initialized Order
    Initialized Note
    followed Note

  • Comma problem in receiver file FCC csv file

    Hi,
    We had a problem with CSV conversions in file adapter.Our scenario is: Proxy>Xi>FCC(csv)
      <?xml version="1.0" encoding="utf-8" ?>
    - <n0:RDM_OrganizationalStructure_MT_OB xmlns:n0="ness.com:RDM_ORGANIZATIONALSTRUCTURE" xmlns:prx="urn:sap.com:proxy:DV3:/1SAI/TAS9B35FF35521A19F6CBEE:701:2009/02/10">
    - <row>
      <KTEXT>Information,Technology</KTEXT>
      <KTEXT>"Information, for internal use only" </KTEXT>
       </row>
      </n0:RDM_OrganizationalStructure_MT_OB>
    In O/P csv file the filed KTEXT is spiting as Information onecolumn and Technology in the next column.
    for avoiding this i conceited with " (double-quote) on both the sides to the KTEXT filed. so the Information,Technolg is appending in the same column. Its fine,
    but for the second record from the source side itself the its getting as "Information, for internal use only"
    so this it is spliting and appending in the next column. in 1st column Information and in the 2nd for internal use only. i want to append this in 1st column as "Information, for internal use only"
    how to resolve this issue for appending the record in the same column?
    and also i want to know why " is not inserting in csv file.
    My Fcc conversion parameters are
    record  Structure: row
    row.addHeaderLine: 1
    row.headerLine: Code,Name,Description
    row.fieldSeparator: ,
    row.endSeparator: 'nl'
    pls advice how to overcome this problem
    Thanks
    Vankadoath

    Hi Vanka,
    In O/P csv file the filed KTEXT is spiting as Information onecolumn and Technology in the next column. for avoiding this i conceited with " (double-quote) on both the sides to the KTEXT filed. so the Information,Technolg is appending in the same column. Its fine,
    To make it in the same column, you put it into double quotes.
    "Information,Technology" -->> Information,Technology
    but for the second record from the source side itself the its getting as "Information, for internal use only" . so this it is spliting and appending in the next column. in 1st column Information and in the 2nd for internal use only. i want to append this in 1st column as "Information, for internal use only"
    As this already has double quotes, you need to use 2 more double quotes to escape that.
    """Information, for internal use only""" -->> "Information, for internal use only"
    Regards,
    Sunil Chandra

  • How to handle multiple streams

    Hi
    I wanted to handle multiple source Streams with Player. do i need to create a each player for each Stream ?? how can i do that ??
    Thanks in Advance,
    Karthikeyan R

    It's not at all obvious what a SequenceOutputStream would do. The way to write lots of stuff over a socket is to wrap multiple output streams round sock.getOutputStream and flush each one when you've finished with it. (There may be issues with compressed / bit output streams).

  • Handling Multiple Result Sets

    Hi
    I have a problem while handling multiple result sets. To fix that problem i need to upgrade my JDBC version2.0 to 4.0.
    My Problem is that i dont know which jars need to be downloaded to upgrade my JDBC version.

    Mallika_23 wrote:
    Is that any more ideas ???1. Learn how google works.
    2. Find the drivers
    3. Read the documentation
    4. Ask questions here once you have actually read the documentation.

  • HANDLING MULTIPLE ENTRIES IN THE INTERNAL TABLE

    Hi guys,
              I got some problem with handling multiple entries from database table. I am retriving four fields and an amount field from the database table and creatinf a file to upload in the application server.But the file has to be taken like below.
          If the four fields which i am retrieving are repeated then we have to take the net amount and make all the entries as one record. else if any of the four fields vary then i have to take it as a seperate record in the internal table. So how can we do this !! I have tried AT NEW and some other logics too but could get the exact one.
      Pls help me out ASAP....

    I think I may have misunderstood your problem.  If you have an exactly match(all four fields), only then you need to collect,  right?
    Try this example.
    report zrich_0002.
    data: begin of itab occurs 0,
          field1(10) type c,
          field2(10) type c,
          field3(10) type c,
          field4 type p decimals 2,
          end of itab.
    data: begin of icheck occurs 0,
          field1(10) type c,
          field2(10) type c,
          field3(10) type c,
          field4 type p decimals 2,
          end of icheck.
    data: begin of itab2 occurs 0,
          field1(10) type c,
          field2(10) type c,
          field3(10) type c,
          field4 type p decimals 2,
          end of itab2.
    start-of-selection.
      itab-field1 = 'A'.
      itab-field2 = 'B'.
      itab-field3 = 'C'.
      itab-field4 = '123.45'.
      append itab.
      itab-field1 = 'A'.
      itab-field2 = 'B'.
      itab-field3 = 'C'.
      itab-field4 = '123.45'.
      append itab.
      itab-field1 = 'A'.
      itab-field2 = 'B'.
      itab-field3 = 'C'.
      itab-field4 = '123.45'.
      append itab.
      itab-field1 = 'A'.
      itab-field2 = 'B'.
      itab-field3 = 'C'.
      itab-field4 = '678.90'.
      append itab.
      loop at itab.
        read table icheck with key field1 = itab-field1
                                   field2 = itab-field2
                                   field3 = itab-field3
                                   field4 = itab-field4.
        if sy-subrc = 0.
          move-corresponding itab to itab2.
          collect itab2.
        else.
          move-corresponding itab to icheck.
          append icheck.
          move-corresponding itab to itab2.
          append itab2.
        endif.
      endloop.
      check sy-subrc  = 0.
    Regards,
    Rich Heilman

  • AJAX handling multiple responses

    Hi,
    I have a problem in handling multiple responses from server.
    I am sending requests simultaniously (response for the first requset comes sending other request) but i am getting only the response for the last request all previous request responses are lost.
    how to handle thi situation.is there a way to handle?
    pls help me asap.

    since it is for sure using a single connection to the server, it is quite normal
    can't you manke the request sequentialy?

  • Problem in posting multiple documents in FB70 through BDC...

    Hi,
       I am facing a problem while posting multiple documents from a flat file in FB70 thourgh BDC batch input by using session method. However, the session is getting terminated after posting the first document. Please clarify my problem.
    Below is a sample of the data in my flat file -
    Customer,Invoice Date,Posting Date,Amount,G/L Account,Amount in doc. Curr.
    34,03.06.2009,03.06.2009,100,211000,100
    34,03.06.2009,03.06.2009,200,211000,200
    34,03.06.2009,03.06.2009,300,211000,300
    And below is the sample code of my program.
    REPORT ZANI2_BDC_FB70
           NO STANDARD PAGE HEADING LINE-SIZE 255.
    TYPES: BEGIN OF TY_STR_FB70,
             ACCNT(20)      TYPE C,     "Customer
             BLDAT(20)      TYPE C,     "Invoice Date
             BUDAT(20)      TYPE C,     "Posting Date
             WRBTR(20)      TYPE C,     "Basic Data - Amount
             HKONT(20)      TYPE C,     "G/L Account
             ITEM_WRBTR(20) TYPE C,
             "Line Item Amount In Document Currency
           END OF TY_STR_FB70,
          BEGIN OF TY_STR_FILEDATA,
             FIELDS TYPE STRING,
          END OF TY_STR_FILEDATA.
    *   data definition
    *       Batchinputdata of single transaction
    DATA:   BDCDATA LIKE BDCDATA    OCCURS 0 WITH HEADER LINE,
    *       messages of call transaction
            MESSTAB LIKE BDCMSGCOLL OCCURS 0 WITH HEADER LINE.
    DATA: IT_STR_FILEDATA TYPE STANDARD TABLE OF TY_STR_FILEDATA,
          WA_STR_FILEDATA TYPE TY_STR_FILEDATA,
          IT_STR_FB70 TYPE STANDARD TABLE OF TY_STR_FB70,
          WA_STR_FB70 TYPE TY_STR_FB70,
          G_STR_FILENAME TYPE STRING,
          G_FLG_ERROR TYPE I,
          G_INDEX TYPE I,
          G_STR_INDEX TYPE STRING,
          G_STR_FIELDNAME TYPE STRING.
    SELECTION-SCREEN BEGIN OF BLOCK INPUT.
    PARAMETERS: P_FILE TYPE IBIPPARMS-PATH,
                P_SES  TYPE APQI-GROUPID.
    SELECTION-SCREEN END OF BLOCK INPUT.
    AT SELECTION-SCREEN ON VALUE-REQUEST FOR P_FILE.
      CALL FUNCTION 'F4_FILENAME'
        EXPORTING
          PROGRAM_NAME  = SYST-CPROG
          DYNPRO_NUMBER = SYST-DYNNR
          FIELD_NAME    = 'P_FILE'
        IMPORTING
          FILE_NAME     = P_FILE.
    Continued in the next thread..

    ...Continued from the previous thread.
    PERFORM OPEN_GROUP.
        G_INDEX = 1.  
        LOOP AT IT_STR_FB70 INTO WA_STR_FB70.
           PERFORM BDC_DYNPRO      USING 'SAPMF05A' '1200'.
          PERFORM BDC_FIELD       USING 'RF05A-BUSCS'
                                         'R'.
         "Customer
          PERFORM BDC_FIELD       USING 'INVFO-ACCNT'
                                         WA_STR_FB70-ACCNT.  
          "Invoice Date
          PERFORM BDC_FIELD       USING 'INVFO-BLDAT'
                                        WA_STR_FB70-BLDAT.                               
          "Posting Date
          PERFORM BDC_FIELD       USING 'INVFO-BUDAT'
                                        WA_STR_FB70-BUDAT.                                
         "Basic data - Amount
          PERFORM BDC_FIELD       USING 'INVFO-WRBTR'
                                        WA_STR_FB70-WRBTR.     
          PERFORM BDC_FIELD       USING 'INVFO-WAERS'
                                        'EUR'.
          G_STR_INDEX = G_INDEX.
          IF G_INDEX < 10.
            CONCATENATE '0' G_STR_INDEX INTO G_STR_INDEX.
          ENDIF.
          CONDENSE G_STR_INDEX NO-GAPS.
          CLEAR G_STR_FIELDNAME.
          CONCATENATE  'ACGL_ITEM-WRBTR(' G_STR_INDEX ')' INTO
          G_STR_FIELDNAME.
          PERFORM BDC_FIELD       USING 'BDC_CURSOR'
                                        G_STR_FIELDNAME.
                                        "'ACGL_ITEM-WRBTR(01)'.
          "G/L Acct
          CONCATENATE  'ACGL_ITEM-HKONT(' G_STR_INDEX ')' INTO
          G_STR_FIELDNAME.
          PERFORM BDC_FIELD       USING  G_STR_FIELDNAME       
                                         WA_STR_FB70-HKONT.    
            "Amount .
          CONCATENATE  'ACGL_ITEM-WRBTR(' G_STR_INDEX ')' INTO
          G_STR_FIELDNAME.
          PERFORM BDC_FIELD       USING G_STR_FIELDNAME   
                                        WA_STR_FB70-ITEM_WRBTR.
          "PERFORM BDC_DYNPRO      USING 'SAPMF05A' '1200'.
          PERFORM BDC_FIELD       USING 'BDC_OKCODE'
                                        '=BU'.
          CLEAR WA_STR_FB70.
        ENDLOOP.
        IF G_FLG_ERROR IS INITIAL.
          PERFORM BDC_TRANSACTION USING 'FB70'.
        ENDIF.
        IF G_FLG_ERROR IS INITIAL.
          PERFORM CLOSE_GROUP.
          MESSAGE 'Session created successfully.' TYPE 'I'.
        ENDIF.
      ENDIF.
    Continued in the next thread..

  • How can I read, millions of records and write as *.csv file

    I have to return some set of columns values(based on current date) from the database (could be million of record also) The dbms_output can accomodate only 20000 records. (I am retrieving thru a procedure using cursor).
    I should write these values to a file with extn .csv (comma separated file) I thought of using a utl_file. But I heard there is some restriction on the number of records even in utl_file.
    If so, what is the restriction. Is there any other way I can achive it? (BLOB or CLOB ??).
    Please help me in solving this problem.
    I have to write to .csv file, the values from the cursor I have concatinated with "," and now its returning the value to the screen (using dbms_output, temporarily) I have to redirect the output to .csv
    and the .csv should be in some physical directory and I have to upload(ftp) the file from the directory to the website.
    Please help me out.

    Jimmy,
    Make sure that utl_file is properly installed, make sure that the utl_file_dir parameter is set in the init.ora file and that the database has been re-started so that it will take effect, make sure that you have sufficient privileges granted directly, not through roles, including privileges to the file and directory that you are trying to write to, add the exception block below to your procedure to narrow down the source of the exception, then test again. If you still get an error, please post a cut and paste of the exact code that you run and any messages that you received.
    exception
        when utl_file.invalid_path then
            raise_application_error(-20001,
           'INVALID_PATH: File location or filename was invalid.');
        when utl_file.invalid_mode then
            raise_application_error(-20002,
          'INVALID_MODE: The open_mode parameter in FOPEN was
           invalid.');
        when utl_file.invalid_filehandle then
            raise_application_error(-20002,
            'INVALID_FILEHANDLE: The file handle was invalid.');
        when utl_file.invalid_operation then
            raise_application_error(-20003,
           'INVALID_OPERATION: The file could not be opened or
            operated on as requested.');
        when utl_file.read_error then
            raise_application_error(-20004,
           'READ_ERROR: An operating system error occurred during
            the read operation.');
        when utl_file.write_error then
            raise_application_error(-20005,
                'WRITE_ERROR: An operating system error occurred
                 during the write operation.');
        when utl_file.internal_error then
            raise_application_error(-20006,
                'INTERNAL_ERROR: An unspecified error in PL/SQL.');

  • Error while refreshing a report using local csv file

    Hi,
    I'm using BI 4.1 SP02.
    While using Rich client, I've created a report with some merged queries, while one of the queries is a local CSV file - saved on AD in some server, and not on the repository inside BO.
    While trying to refresh the report with the Rich client, it all went great.
    Now, while using BI Launchpad java based app, I can't refresh the report - I get the following error:
    "An Internal error occurred while calling 'processDPCommandsEx' API. (Error: ERR_WIS_30270) (WIS 30270)"
    Should I be able to refresh a report without the Rich if it contains a local file (which is possible to EDIT only with rich) ?
    If so, then did someone ran into this error?
    Thank you,
    Or.

    First of all, thanks for both of the replies.
    Second,
    my problem is unlikely have to do something with permissions from one reasons -
    when the report is using XLS\XLSX on same folder(with same name prefix) - the report is running without any problem.
    Only problem is while refreshing without Rich while the source is network CSV file.
    Any suggestions?
    Thanks.

Maybe you are looking for

  • Delete file in the filesystem

    Hello, I am developing an application with the apex. I created a section where you can upload files to the linux file system to a specific folder, not upload to the database, I do not want me to take unnecessary space. I can also download the file li

  • Viewer plays by default instead of Canvas

    Final Cut 5.1.4 My Editor may have clicked on something by mistake and I don't know how to reset it. Here's the problem: When she clicks on the Canvas after having played a clip in the Viewer, if she plays from the Spacebar, it will play the Viewer w

  • Arranging playlists in itunes

    never had a problem before. I just imported a bunch of new tunes. i even rebooted. now when I try to move songs up or down in a playlist, nothing happens.

  • Regarding static iview

    hi all,   I need some clarifications regarding iviews. As per my understanding, whenever we navigate to different iviews/applications in portal, the whole is refreshed. So, if I place an application in portal through iview in the default framework pa

  • Overlay real time and date for video playback

    I have made a mock CCTV film with Final Cut as an art piece and I want to play it with the current real time and date overlaid. Is there an application that does that?