One file to mutliple metadata records ?

I have a requirement to connect one file for content management to multiple metadata records - for example ...
We have a (single) loan document which applied to many properties. Each property has its own associated metadata, such as the development, type (mall, office, etc) which apply to the property, but not directly to he loan.
Given a loan stored under properties A, B and C, the document should be accessible through searches for A, B OR C; it would be something like the metadata record in the db was a reference to a file, shared by other metadata records.
Has anyone had to handle this before ? Can you give me a clue as to how to get started >
TIA,
Mark

Sounds to me like a use case for Related Links - you can create parent-child or peer-to-peer links between content items. Once a link is created (e.g. loan as a parent, properties as children), you reach properties from the loan or vice versa.
Another thing that might be useful for you is metadata-only items (that is, content-less items storing metadata only). There was a blog article by Kyle Hatlestad or Bex Huff on this topic. You should be able to find it even searching this forum.

Similar Messages

  • One file to be sent to one of two message types.

    Hi,
    I have gone through many other messages but didn't get an answer. Please read the issue.
    I have one file with 4 (suppose) records out of which 2 records go to one message type and the rest to other.
    I have two message mappings for each message type (same basic type), one MI and two OM's.
    On the confic side I have created all required paths.
    When I put a file through its been picked up and in the message monitoring it shows as if they are scheduled to be sent but they are not creating any IDOC's. From past few days they are in the scheduled mode only.
    Can anyone please help me. If I am not clear, I will explain in detail.
    Thank you, Regards,
    Mohana.

    Mohana,
    I think there are few entries in your queues. Since I cannot understand whether it is an inbound or outbound please go to SMQ1 and SMQ2. If you see any entries check all of them and delete them. So once deleted test it once again.
    Regards,
    ---Satish

  • How to add the records of 2 internal table records into one file

    hello experts,
    My scenario is...
    I am retrieving the data for the for the credit, debit and trailer records of the customer into 3 different internal tables and finally i have to append all those records into one file first debit records then credit records finally the trailer record.... how to do that can anyone give some idea plzzzzzzzzz..
    Plz its bit urgent..
    Thanks a lot for your anticipation
    SRI

    Hello,
    Do like this.
    " Assume u have three itab.
    "Itab1 - debit
    "Itab2 - credit
    "Itab3 - Credit.
    REPORT ZV_TEST_SERVER .
    *PARAMETERS: P_FILE TYPE STRING."RLGRAP-FILENAME.
    CALL FUNCTION 'GUI_DOWNLOAD'
      EXPORTING
    *   BIN_FILESIZE                  =
        FILENAME                      = P_FILE
    *   FILETYPE                      = 'ASC'
       APPEND                        = 'X'  " Check here
    *   WRITE_FIELD_SEPARATOR         = ' '
    *   HEADER                        = '00'
    *   TRUNC_TRAILING_BLANKS         = ' '
    *   WRITE_LF                      = 'X'
    *   COL_SELECT                    = ' '
    *   COL_SELECT_MASK               = ' '
    *   DAT_MODE                      = ' '
    * IMPORTING
    *   FILELENGTH                    =
      TABLES
        DATA_TAB                      = ITAB1
    * EXCEPTIONS
    *   FILE_WRITE_ERROR              = 1
    *   NO_BATCH                      = 2
    *   GUI_REFUSE_FILETRANSFER       = 3
    *   INVALID_TYPE                  = 4
    *   NO_AUTHORITY                  = 5
    *   UNKNOWN_ERROR                 = 6
    *   HEADER_NOT_ALLOWED            = 7
    *   SEPARATOR_NOT_ALLOWED         = 8
    *   FILESIZE_NOT_ALLOWED          = 9
    *   HEADER_TOO_LONG               = 10
    *   DP_ERROR_CREATE               = 11
    *   DP_ERROR_SEND                 = 12
    *   DP_ERROR_WRITE                = 13
    *   UNKNOWN_DP_ERROR              = 14
    *   ACCESS_DENIED                 = 15
    *   DP_OUT_OF_MEMORY              = 16
    *   DISK_FULL                     = 17
    *   DP_TIMEOUT                    = 18
    *   FILE_NOT_FOUND                = 19
    *   DATAPROVIDER_EXCEPTION        = 20
    *   CONTROL_FLUSH_ERROR           = 21
    *   OTHERS                        = 22
    IF SY-SUBRC <> 0.
    * MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    *         WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    CALL FUNCTION 'GUI_DOWNLOAD'
      EXPORTING
    *   BIN_FILESIZE                  =
        FILENAME                      = P_FILE
    *   FILETYPE                      = 'ASC'
       APPEND                        = 'X'  " Check here
    *   WRITE_FIELD_SEPARATOR         = ' '
    *   HEADER                        = '00'
    *   TRUNC_TRAILING_BLANKS         = ' '
    *   WRITE_LF                      = 'X'
    *   COL_SELECT                    = ' '
    *   COL_SELECT_MASK               = ' '
    *   DAT_MODE                      = ' '
    * IMPORTING
    *   FILELENGTH                    =
      TABLES
        DATA_TAB                      = ITAB2  " Check here
    * EXCEPTIONS
    *   FILE_WRITE_ERROR              = 1
    *   NO_BATCH                      = 2
    *   GUI_REFUSE_FILETRANSFER       = 3
    *   INVALID_TYPE                  = 4
    *   NO_AUTHORITY                  = 5
    *   UNKNOWN_ERROR                 = 6
    *   HEADER_NOT_ALLOWED            = 7
    *   SEPARATOR_NOT_ALLOWED         = 8
    *   FILESIZE_NOT_ALLOWED          = 9
    *   HEADER_TOO_LONG               = 10
    *   DP_ERROR_CREATE               = 11
    *   DP_ERROR_SEND                 = 12
    *   DP_ERROR_WRITE                = 13
    *   UNKNOWN_DP_ERROR              = 14
    *   ACCESS_DENIED                 = 15
    *   DP_OUT_OF_MEMORY              = 16
    *   DISK_FULL                     = 17
    *   DP_TIMEOUT                    = 18
    *   FILE_NOT_FOUND                = 19
    *   DATAPROVIDER_EXCEPTION        = 20
    *   CONTROL_FLUSH_ERROR           = 21
    *   OTHERS                        = 22
    IF SY-SUBRC <> 0.
    * MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    *         WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    CALL FUNCTION 'GUI_DOWNLOAD'
      EXPORTING
    *   BIN_FILESIZE                  =
        FILENAME                      = P_FILE
    *   FILETYPE                      = 'ASC'
       APPEND                        = 'X'  " Check here
    *   WRITE_FIELD_SEPARATOR         = ' '
    *   HEADER                        = '00'
    *   TRUNC_TRAILING_BLANKS         = ' '
    *   WRITE_LF                      = 'X'
    *   COL_SELECT                    = ' '
    *   COL_SELECT_MASK               = ' '
    *   DAT_MODE                      = ' '
    * IMPORTING
    *   FILELENGTH                    =
      TABLES
        DATA_TAB                      = ITAB3 " Check here
    * EXCEPTIONS
    *   FILE_WRITE_ERROR              = 1
    *   NO_BATCH                      = 2
    *   GUI_REFUSE_FILETRANSFER       = 3
    *   INVALID_TYPE                  = 4
    *   NO_AUTHORITY                  = 5
    *   UNKNOWN_ERROR                 = 6
    *   HEADER_NOT_ALLOWED            = 7
    *   SEPARATOR_NOT_ALLOWED         = 8
    *   FILESIZE_NOT_ALLOWED          = 9
    *   HEADER_TOO_LONG               = 10
    *   DP_ERROR_CREATE               = 11
    *   DP_ERROR_SEND                 = 12
    *   DP_ERROR_WRITE                = 13
    *   UNKNOWN_DP_ERROR              = 14
    *   ACCESS_DENIED                 = 15
    *   DP_OUT_OF_MEMORY              = 16
    *   DISK_FULL                     = 17
    *   DP_TIMEOUT                    = 18
    *   FILE_NOT_FOUND                = 19
    *   DATAPROVIDER_EXCEPTION        = 20
    *   CONTROL_FLUSH_ERROR           = 21
    *   OTHERS                        = 22
    IF SY-SUBRC <> 0.
    * MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    *         WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    If useful reward.
    Vasanth

  • Recording entire cd as one file

    How do you record an entire cd as one file or join the tracks into one file. Is that even possible with iTunes?

    Click here and follow the instructions.
    (30536)

  • I'm recording video on JVC HD camera.Every 13:54 it starts a new file.I have one file that won't play.I get this message:Quick time player can't open (names file) because the movie's file format isn't recognized.  Only happens this one file.Ideas?

    I can't open one Quick Time movie file.  It says: Quick time player can't open (names file) because the movie's file format isn't recognized.  Only happens on this one file.  All the others of a two hour shoot are fine. 
    Any ideas what's happened and is there some way to open that file and retrieve what's on it? 

    Thanks this was useful. I have been ripping my audiobooks and I was finding that they would get uploaded to iCloud in the process, which in itself was fine, but then iTunes would not let me change the media type to Audiobook, which is annoying as it effects how iTunes treats the media. Like showing you the skip back 10 seconds button.
    The general outline here helped me work around the issue and change the media type before the media was re-uploaded to iCloud.
    Sean

  • How to load a flat file with lot of records

    Hi,
    I am trying to load a flat file with hundreds of records into an apps table. when i create the process and deploy it onto the console it asks for an input in an html form. why does it ask for an input when i have specified the input file directory in my process? is there any way around tis where in it just reads all the records from the flat file directly??is custom queues anyway related to what I am about to do?any documents on this process will be greatly appreciated.If any one can help me on this it will be great. thank you guys....

    After deploying it, do you see if it is active and the status is on from the BPEL console BPEL Process tab? It should not come up to ask for input unless you are clicking it from the Dashboard tab. Do not click it from the Dashboard. Instead you should put some files into the input driectory. Wait few seconds you should see the instances of the BPEL process is created and start to process the files asynchrously.

  • Processing one file at a time in Biztalk

    We are changing the way we're processing files in that where we used to receive hundreds of small files throughout the day, we are now going to have these files batched into larger files that get dropped to our incoming Biztalk directory. I'm concerned about
    having the server pick up multiple large files at one time. Is there a way to tell a Biztalk application to only process one file at a time, and not to pick up the next file until the previous one has finished? Note, I do not have a way to build this into
    the Orchestration, as this is a Biztalk application that is actually installed by our Trizetto(QNXT) product. It uses Biztalk for XML serialization. 
    Thanks.

    I wouldn't worry about the size of the messages unless they are approaching 1GB. BizTalk can process roughly 400 msgs/sec. Also if you're worried about processing multiple large files try load balancing. With BizTalk it is quite "easy" to load
    balance. In a nut shell you have to have the dll of the application installed on all the servers you want to be apart of load balancing and BizTalk takes care of the rest. Your large files would more than likely split up and processed on different servers.
    http://msdn.microsoft.com/en-us/library/aa578057(v=bts.80).aspx
    The link above is for high availability but could also be used to get a better understanding of load balancing. Also if you use de-batching to split up the 3000-4000 records that would help with processing them. If an orchestration receives a file with 4000
    records in them that instance alone processes the whole file. Whereas if you split up the messages each message gets it's own instance of the orchestration which leads to better performance. The only drawback is that it takes more resources so your environment
    needs to be fairly large. Also if one server is busy another server with the same dll can take the de-batched messages and process them.
    Here is a link to my blog post about de-batching. It is very basic but gives a good base for de-batching.
    http://camartin.azurewebsites.net/post/BizTalk-Debatching

  • File Adapter - Reading 2 files from sender and one file to receiver

    All - I am looking for a best solution for reading two files from sender and generating one file to the receiver.
    Here is the scenario
    Sender file s - A is a master file. B is a detail file; B will have a one to many records for each record of File A.. that is the relation ship between 2 files..  I am looking for a better solution to
    Example
    Input:
    File A will have a record like "summary1 ABC 1234"
    File B will have a record like "Detailfor Summary1 CDF 8679"
    Output:
    File C shall have a record "consolidatedofSummary&Detail ABC 1234 8679"
    1) read records of master row and detail row from these 2 files
    2) do the mapping to the receiver output format which it is creating for records with some harcoded values and based on the values from master and detail rows of sender file
    if I can accomblish reading sender files and do the mapping to the receiver , assume that mapping to the receiver format would not be cumbersome.
    please help me...
    Thanks

    Hi Gopal,
    Ur approach to the interface is perfect.... Go ahead.
    For selecting multiple files in source use Advanced Selection for files.. and specify the directory paths and also the file name shema...
    This will fetch the 2 files into the integration engine... then u map the 2 source files to one target file(2:1 mapping).
    Specify accordingly the receiver side FCC and test it end to end...
    If u face any prblm post here... some1 will help u out..
    Babu

  • Can the File Adapter handle Multiple Record Layouts in COBOL?

    I am trying to parse flat files that have a record type field in the first four characters on each record. As a test I tried to use the following record definitions:
    01 CONTRACT-HEADER.
    05 RECORD-TYPE PIC X(4).
    05 TRANSACTION-TYPE PIC X.
    05 CONTRACT-NUMBER PIC X(32).
    05 CONTRACT-RELEASE-NBR PIC X(30).
    05 CONTRACT-FILLER PIC X(2025).
    01 CONTRACT-LINEITEM.
    05 RECORD-TYPE PIC X(4).
    05 TRANSACTION-TYPE PIC X.
    05 LINE-FILLER PIC X(1664).
    I also tried smaller records with simple test data. When I press the "test" button in the native file adapter the XML is never generated if I have more than one record type as above. Suppose that the contract-header record type always has the string "X001" in it, while the contract-lineitem always has "X015" in the record-type. I chose Us-ascii and chose $eol as a delimiter. The file on my windows computer has carriage return line feed after each record. I know that the fixed length converter works, but because the records really have 50 fields on each record type and are long, I don't think I can actually use that one in one sitting. Has anyone ever gotten two different record lenght cobol definitions to work? Do I have to hand edit the XML to specify the delimiter? I don't see any delimter specified in the generated XML before I attempt the test.

    Rahul,
    I recommned you should use "|" as "," can be embedded in a text string. I guess you can pass a directive to use to deliver "|" delimited file :). If you want to handle both then you might have to write a shell script or bat to pass delimiter as an argument which will edit and replace your control file.
    Regards

  • FTWR File Size Worksheet no records shown?!

    This is regarding DART (Data Retention Tool).
    In FTW1A I extracted data. In AL11 I can see the data extraction file created in FTW1A. In FTWR “File Size Worksheet” no records are shown for the company code for which I extracted data in FTW1A. Does anybody have a clue what went wrong? How to get records shown  in FTWR?

    The size indication you are looking at is probably for megapixels. You can see the actual filesize in Lightroom in the metadata panel if you set it to "all" but I don't think you can show it on the badge in the Library grid display. A 6MP 16-bit uncompressed tiff should be about 36 megs, so close enough..

  • Lightroom - Photoshop Integration breaks for more than one file

    This message is for those of us who process lots of images. I just spend a n afternoon working with Lightroom 1.0 to process the 900+ images taken at a sporting event over the weekend. My task is to find the keepers, and then run a custom action in Photoshop on every one of those files to prepare the files for delivery to the client.
    Its pretty clear anyone doing serious noise reduction or sharpening work needs to have PS right now, but even if those limitations are addressed, let's assume there's always going to be something that needs to be done in Photoshop.
    The problem is the current integration of Photoshop and Lightroom seems to be based on a single file, i.e. there's no batch capabilities, but I need to be able to set up and run recorded Photoshop actions on 900 files! (OK, only half were keepers, but you get the idea.)
    The workaround is to export all of the files out of Lightroom into a directory, and then run a batch process to create a second directory of the processed files. I now have 900 files that aren't in Lightroom's database! Unfortunately Lightroom doesn't recognize files on the hard drive, and so if you're using Lightroom as a file manager, in six months I'll never know those post-processed files are even there.
    Sure, I can delete the first directory, but I'll definitely want to keep the final files around for later. The only option is to then run another import on the final directory, which creates more down time.
    The problem is the chances of getting out of sync are huge, plus there's no association between the original and processed versions as happens when you edit from within Lightroom. (In that case the original and processed are stacked.)
    Now, the 1.0 version is amazing, and I consider the program to be one of the best programs ever designed. This is just a request to consider tackling this problem in version 2.0.

    Yes, that's a big improvement in 1.0. The only problem is it only works for one file :-(
    Overall I've found stacking to be less useful in other contexts, unfortunately. For me at least its been easier to just go through all of the photos initially to rank them, and then only show the candidates to send to clients.

  • File import skips last record without blank line

    The following code works great for importing a csv file and inserting the records into tables..
    but the last record is not read unless there is a blank line after it..
    what needs to be changed so that the last line works with or without a carriage return after it..
    create or replace PACKAGE BODY "HTMLDB_TOOLS"
    AS
    TYPE varchar2_t IS TABLE OF VARCHAR2(32767) INDEX BY binary_integer;
    -- Private functions --{{{
    PROCEDURE delete_collection ( --{{{
    -- Delete the collection if it exists
    p_collection_name IN VARCHAR2
    IS
    BEGIN
    IF (htmldb_collection.collection_exists(p_collection_name))
    THEN
    htmldb_collection.delete_collection(p_collection_name);
    END IF;
    END delete_collection; --}}}
    PROCEDURE csv_to_array ( --{{{
    -- Utility to take a CSV string, parse it into a PL/SQL table
    -- Note that it takes care of some elements optionally enclosed
    -- by double-quotes.
    p_csv_string IN VARCHAR2,
    p_array OUT wwv_flow_global.vc_arr2,
    p_separator IN VARCHAR2 := ','
    IS
    l_start_separator PLS_INTEGER := 0;
    l_stop_separator PLS_INTEGER := 0;
    l_length PLS_INTEGER := 0;
    l_idx BINARY_INTEGER := 0;
    l_quote_enclosed BOOLEAN := FALSE;
    l_offset PLS_INTEGER := 1;
    BEGIN
    l_length := NVL(LENGTH(p_csv_string),0);
    IF (l_length <= 0)
    THEN
    RETURN;
    END IF;
    LOOP
    l_idx := l_idx + 1;
    l_quote_enclosed := FALSE;
    IF SUBSTR(p_csv_string, l_start_separator + 1, 1) = '"'
    THEN
    l_quote_enclosed := TRUE;
    l_offset := 2;
    l_stop_separator := INSTR(p_csv_string, '"', l_start_separator + l_offset, 1);
    ELSE
    l_offset := 1;
    l_stop_separator := INSTR(p_csv_string, p_separator, l_start_separator + l_offset, 1);
    END IF;
    IF l_stop_separator = 0
    THEN
    l_stop_separator := l_length + 1;
    END IF;
    p_array(l_idx) := (SUBSTR(p_csv_string, l_start_separator + l_offset,(l_stop_separator - l_start_separator - l_offset)));
    EXIT WHEN l_stop_separator >= l_length;
    IF l_quote_enclosed
    THEN
    l_stop_separator := l_stop_separator + 1;
    END IF;
    l_start_separator := l_stop_separator;
    END LOOP;
    END csv_to_array; --}}}
    PROCEDURE get_records(p_blob IN blob,p_records OUT varchar2_t) --{{{
    IS
    l_record_separator VARCHAR2(2) := chr(13)||chr(10);
    l_last INTEGER;
    l_current INTEGER;
    BEGIN
    -- Sigh, stupid DOS/Unix newline stuff. If HTMLDB has generated the file,
    -- it will be a Unix text file. If user has manually created the file, it
    -- will have DOS newlines.
    -- If the file has a DOS newline (cr+lf), use that
    -- If the file does not have a DOS newline, use a Unix newline (lf)
    IF (NVL(dbms_lob.instr(p_blob,utl_raw.cast_to_raw(l_record_separator),1,1),0)=0)
    THEN
    l_record_separator := chr(10);
    END IF;
    l_last := 1;
    LOOP
    l_current := dbms_lob.instr( p_blob, utl_raw.cast_to_raw(l_record_separator), l_last, 1 );
    EXIT WHEN (nvl(l_current,0) = 0);
    p_records(p_records.count+1) := utl_raw.cast_to_varchar2(dbms_lob.substr(p_blob,l_current-l_last,l_last));
    l_last := l_current+length(l_record_separator);
    END LOOP;
    END get_records; --}}}
    -- Utility functions --{{{
    PROCEDURE parse_textarea ( --{{{
    p_textarea IN VARCHAR2,
    p_collection_name IN VARCHAR2
    IS
    l_index INTEGER;
    l_string VARCHAR2(32767) := TRANSLATE(p_textarea,chr(10)||chr(13)||' ,','@@@@');
    l_element VARCHAR2(100);
    BEGIN
    l_string := l_string||'@';
    htmldb_collection.create_or_truncate_collection(p_collection_name);
    LOOP
    l_index := instr(l_string,'@');
    EXIT WHEN NVL(l_index,0)=0;
    l_element := substr(l_string,1,l_index-1);
    IF (trim(l_element) IS NOT NULL)
    THEN
    htmldb_collection.add_member(p_collection_name,l_element);
    END IF;
    l_string := substr(l_string,l_index+1);
    END LOOP;
    END parse_textarea; --}}}
    PROCEDURE parse_file( --{{{
    p_file_name IN VARCHAR2,
    p_collection_name IN VARCHAR2,
    p_headings_item IN VARCHAR2,
    p_columns_item IN VARCHAR2,
    p_ddl_item IN VARCHAR2,
    p_table_name IN VARCHAR2 DEFAULT NULL
    IS
    l_blob blob;
    l_records varchar2_t;
    l_record wwv_flow_global.vc_arr2;
    l_datatypes wwv_flow_global.vc_arr2;
    l_headings VARCHAR2(4000);
    l_columns VARCHAR2(4000);
    l_seq_id NUMBER;
    l_num_columns INTEGER;
    l_ddl VARCHAR2(4000);
    BEGIN
    IF (p_table_name is not null)
    THEN
    l_ddl := 'insert into '||p_table_name||' '||
    'select '||v(p_columns_item)||' '||
    'from htmldb_collections '||
    'where seq_id > 0 and collection_name='''||p_collection_name||'''';
    execute immediate l_ddl;
    RETURN;
    END IF;
    BEGIN
    select blob_content into l_blob from wwv_flow_files
    where name=p_file_name;
    EXCEPTION
    WHEN NO_DATA_FOUND THEN
    raise_application_error(-20000,'File not found, id='||p_file_name);
    END;
    get_records(l_blob,l_records);
    IF (l_records.count < 2)
    THEN
    raise_application_error(-20000,'File must have at least 2 ROWS, ONE HEADER and ONE OR MORE DATA, id='||p_file_name);
    END IF;
    -- Initialize collection
    htmldb_collection.create_or_truncate_collection(p_collection_name);
    -- Get column headings
    csv_to_array(l_records(1),l_record);
    l_num_columns := l_record.count;
    if (l_num_columns > 50) then
    raise_application_error(-20000,'Max. of 50 columns allowed, id='||p_file_name);
    end if;
    -- Get column headings and names
    FOR i IN 1..l_record.count
    LOOP
    l_headings := l_headings||':'||l_record(i);
    l_columns := l_columns||',c'||lpad(i,3,'0');
    END LOOP;
    l_headings := ltrim(l_headings,':');
    l_columns := ltrim(l_columns,',');
    htmldb_util.set_session_state(p_headings_item,l_headings);
    htmldb_util.set_session_state(p_columns_item,l_columns);
    -- Save data into specified collection
    FOR i IN 2..l_records.count
    LOOP
    csv_to_array(l_records(i),l_record);
    l_seq_id := htmldb_collection.add_member(p_collection_name,'dummy');
    FOR i IN 1..l_record.count
    LOOP
    htmldb_collection.update_member_attribute(
    p_collection_name=> p_collection_name,
    p_seq => l_seq_id,
    p_attr_number => i,
    p_attr_value => l_record(i)
    END LOOP;
    END LOOP;
    DELETE FROM wwv_flow_files WHERE name=p_file_name;
    END;
    BEGIN
    NULL;
    END;

    found someone that helped me.. here was the fix:
    l_current := dbms_lob.instr( p_blob, utl_raw.cast_to_raw(l_record_separator), l_last, 1 );
    -- START of new code, lines above as per existing
    IF nvl(l_current,0) = 0 AND LENGTH(TRIM(utl_raw.cast_to_varchar2(dbms_lob.substr(p_blob,32767,l_last)))) > 0 THEN
    p_records(p_records.count+1) := utl_raw.cast_to_varchar2(dbms_lob.substr(p_blob,32767,l_last));
    END IF;
    -- END of new code,, lines below as per existing
    EXIT WHEN (nvl(l_current,0) = 0);

  • How effective splitting one file in to multiple in PI

    Hi Forum,
    Is it possible to split one file in to 20 or more using PI.
    If yes,how effective it is.

    Yes. And there are various ways for doing this.
    1. In the adapter engine while picking the flat file using option "Recordset per Message"
    2. Using multi-mapping message split.
    3. Using BPM loops.
    The effectiveness would depend upon the size of message and no of records it contain. e.g. if the size is small, say only 20 records, then splitting the message in 20 parts would not be a good option. But if the size is too large with say thousands of messages, then splitting would turn out to be effective.
    Regards,
    Prateek

  • File info or metadata

    I cannot seem to find a way to enter the job name into the "file info" or metadata in Photoshop CS. If the job name was entered in a previous version of photoshop such as PS7 it will appear in the file info under the advanced tab underthe XMP Basic job ticket properties but cannot be edited. As a work around I am using PS 7 to enter my job name and then using CS for everything else. There must be a better way. Am I missing something?

    Ahem, well this is interesting.  I haven't found a way to write a custom panel to display the JobName in CS and make it editable - perhaps a limitation to the panel language.<br /><br />One option is to create a template to apply to multiple files - either through the flyout menu in File Info, Replace in the Advanced panel or via the File Browser > Edit > Replace Metadata.<br /><br />Below is a template that I created.  Best way to create your own is to save a template with the JobName property present, then go in an edit the file to remove the other template data and retain the jobName section as I did below.<br /><br />Edit the jobName value between the tags and rename the template - then use the Replace command to apply it to files in CS.<br /><br />Not a pretty work around.  Otherwise I would suggest using another field going forward.<br /><br />Gunar<br /><br /><?xpacket begin='Ôªø' id='W5M0MpCehiHzreSzNTczkc9d'?><x:xmpmeta xmlns:x='adobe:ns:meta/' x:xmptk='XMP toolkit 3.0-29, framework 1.6'><br /><rdf:RDF xmlns:rdf='http://www.w3.org/1999/02/22-rdf-syntax-ns#' xmlns:iX='http://ns.adobe.com/iX/1.0/'><br /><br /> <rdf:Description rdf:about=''<br />  xmlns:stJob='http://ns.adobe.com/xap/1.0/sType/Job#'<br />  xmlns:xapBJ='http://ns.adobe.com/xap/1.0/bj/'><br />  <xapBJ:JobRef><br />   <rdf:Bag><br />    <rdf:li rdf:parseType='Resource'><br />     <stJob:name>new info goes into here</stJob:name><br />    </rdf:li><br />   </rdf:Bag><br />  </xapBJ:JobRef><br /> </rdf:Description><br /><br /></rdf:RDF><br /></x:xmpmeta><br /><?xpacket end='r'?>

  • Writing to XMP file or one file per folder?

    I am about to start using LR2. Used to using PS and Bridge. I use DNG and tiffs, and jpgs. I am used to having one file per folder for bridge's info. I do NOT want an XMP file per picture file. How does LR2 do this? I have read a few posts about XMP files and such. Is there a setting to make one file per folder, or does LR2 make XMP files for each picture?

    >XMP sidecars are only created for proprietary raw files (eg CR2, NEF, etc) and only when you save metadata out to file.
    Good, so since I do not use proprierary RAW's, I am safe. Thanks. I do plan to save metadata to the files, so in the future, other programs can use it.

Maybe you are looking for

  • Podcast syncing & Podcast cover view in iTunes 11

    Hello, One of the features for iTunes 11 was podcast syncing. It was supposed to sync what podcasts you had between your computers and mobile devices, and also sync position. Was quite looking forward to this feature. Where is it and how do you make

  • How to start with Oracle SOA

    Hi All, I want to learn Oracle SOA,Can anyone please help me how to start lerning Oracle SOA. I have worked one year as java programmer befor. Thanks and regards, Deekay.

  • Excise tab in not appearing in GR (new plant configured)

    Hi Experts, Excise tab is not appearing while GR posting. I have configured a new plant in Sandbox server for testing purpose. Have done all required settings as under: - New Plant configured - Complete Excise related configuration done in Logistics

  • Internet disconnects

    I have three WiFi Macs and one connected directly to a Airport express in WDS mode. I've got another APX also in WDS mode in my loft, and then my main APX connected to my DSL modem. Lately Airport has been randomly dropping the Internet (will display

  • Color PDFs print all red

    I am using Acrobat 8 Standard edition on Vista, and every time I attempt to print a color file on an HP Color Laserjet printer, it prints all red. The printer's installed correctly, attached via USB, and prints fine from other applications. It's only