Processing large file in multiple threads

I have a large plain text file (~1GB) that contains lots (~500,000) of chunks of text, separated by the record separator: "\n//\n". In my application, I have a class that reads the file and turns the records into objects, which is fairly time-consuming (the bottleneck is definitely in the processing, not the IO). The app is running on a muti-core machine, so I want to split the processing up over several threads.
My plan is to write a "Dispatcher" class, which reads the file from disk and maintains a queue of records, and a "Processor" class, which requests records from the Dispatcher object, turns the records into objects, and add()s them to an array. Is this the correct way to do it? Can anyone point me in the direction of a tutorial or example of this type of multi-threaded file processing?

public class TaskExecutionr {
    private static final int NTHREADS = 10; // Probably want to do System.getNumberOfProcessors+1 here
    private static final Executor exec
            = Executors.newFixedThreadPool(NTHREADS);
    private static Collection result = new ConcurrentLinkedQueue();
    public static void main(String[] args) throws IOException {
        FileReader reader =
        while (true) {
            final String text = reader.read(500,000); // read 500.000 lines to a string.
            Runnable task = new Runnable() { // create a new runnable that deals with the text
                public void run() {
                    handle(text);
            exec.execute(task); // Place the task in the Executor
    private static void handle(Socket connection) {
        // make your objects here
        result.add(<your new object>);
}http://www.javaconcurrencyinpractice.com/listings/TaskExecutionWebServer.java
Edited by: Fireblaze-II on Jun 19, 2008 5:48 AM

Similar Messages

  • Problem while processing large files

    Hi
    I am facing a problem while processing large files.
    I have a file which is around 72mb. It has around more than 1lac records. XI is able to pick the file if it has 30,000 records. If file has more than 30,000 records XI is picking the file ( once it picks it is deleting the file ) but i dont see any information under SXMB_MONI. Either error or successful or processing ... . Its simply picking and igonring the file. If i am processing these records separatly it working.
    How to process this file. Why it is simply ignoring the file. How to solve this problem..
    Thanks & Regards
    Sowmya.

    Hi,
    XI pickup the Fiel based on max. limit of processing as well as the Memory & Resource Consumptions of XI server.
    PRocessing the fiel of 72 MB is bit higer one. It increase the Memory Utilization of XI server and that may fali to process at the max point.
    You should divide the File in small Chunks and allow to run multiple instances. It will  be faster and will not create any problem.
    Refer
    SAP Network Blog: Night Mare-Processing huge files in SAP XI
    /people/sravya.talanki2/blog/2005/11/29/night-mare-processing-huge-files-in-sap-xi
    /people/michal.krawczyk2/blog/2005/11/10/xi-the-same-filename-from-a-sender-to-a-receiver-file-adapter--sp14
    Processing huge file loads through XI
    File Limit -- please refer to SAP note: 821267 chapter 14
    File Limit
    Thanks
    swarup
    Edited by: Swarup Sawant on Jun 26, 2008 7:02 AM

  • Processing Large Files using Chunk Mode with ICO

    Hi All,
    I am trying to process Large files using ICO. I am on PI 7.3 and I am using new feature of PI 7.3, to split the input file into chunks.
    And I know that we can not use mapping while using Chunk Mode.
    While trying I noticed below points:
    1) I had Created Data Type, Message Type and Interfces in ESR and used the same in my scenario (No mapping was defined)Sender and receiver DT were same.
    Result: Scenario did not work. It created only one Chunk file (.tmp file) and terminated.
    2) I used Dummy Interface in my scenario and it worked Fine.
    So, Please confirm if we should always USE DUMMY Interfaces in Scenario while using Chunk mode in PI 7.3 Or Is there something that I am missing.
    Thanks in Advance,
    - Pooja.

    Hello,
    While trying I noticed below points:
    1) I had Created Data Type, Message Type and Interfces in ESR and used the same in my scenario (No mapping was defined)Sender and receiver DT were same.
    Result: Scenario did not work. It created only one Chunk file (.tmp file) and terminated.
    2) I used Dummy Interface in my scenario and it worked Fine.
    So, Please confirm if we should always USE DUMMY Interfaces in Scenario while using Chunk mode in PI 7.3 Or Is there something that I am missing.
    According to this blog:
    File/FTP Adapter - Large File Transfer (Chunk Mode)
    The following limitations apply to the chunk mode in File Adapter
    As per the above screenshots, the split never cosiders the payload. It's just a binary split. So the following limitations would apply
    Only for File Sender to File Receiver
    No Mapping
    No Content Based Routing
    No Content Conversion
    No Custom Modules
    Probably you are doing content conversion that is why it is not working.
    Hope this helps,
    Mark
    Edited by: Mark Dihiansan on Mar 5, 2012 12:58 PM

  • Process large file using BPEL

    My project have a requirement of processing large file (10 MB) all at once. In the project, the file adapter reads the file, then calls 5 other BPEL process to do 10 different validations before delivering to oracle database. I can't use debatch feature of adapter because of Header and detail record validation requirement. I did some performace tuing (eg: auditlevel to minimum, logging level to error, JVM size to 2GB etc..) as per performance tuing specified in Oracle BPEL user guide. We are using 4 CPU, 4GB RAM IBM AIX 5L server. I observed that the Receive activity in the begining of each process is taking lot of time, while other transient process are as per expected.
    Following are statistics for receive activity per BPEL process:
    500KB: 40 Sec
    3MB: 1 Hour
    Because we have 5 BPEL process, so lot of time is wasted in receive activity.
    I did't try 10 MB so far, because of poor performance figure for 3 MB file.
    Does any one have any idea how to improve performance of begining receive activity of BPEL process?
    Thanks
    -Simanchal

    I believe the limit in SOA Suite is 7MB if you want to use the full payload and perform some kind of orchastration. Otherwise you need to do some kind of debatching, which you stated will not work.
    SOA Suite is not really designed for your kind of use case as it needs to parocess this file in memory, when any transformation occurs it can increase this message between 3 - 10 times. If you are writing to a database why can you read the rows one by one?
    If you are wanting to perform this kind of action have a look at ODI (Oracle Data Integrator). I Also believe that OSB (Aqua Logic) can handle files upto 200MB this this can be an option as well, but it may require debatching.
    cheers
    James

  • Processing large files on Mac OS X Lion

    Hi All,
    I need to process large files (few GB) from a measurement. The data files contain lists of measured events. I process them event by event and the result is relatively small and does not occupy much memory. The problem I am facing is that Lion "thinks" that I want to use the large data files later again and puts them into cache (inactive memory). The inactive memory is growing during the reading of the datafiles up to a point where the whole memory is full (8GB on MacBook Pro mid 2010) and it starts swapping a lot. That of course slows down the computer considerably including the process that reads the data.
    If I run "purge" command in Terminal, the inactive memory is cleared and it starts to be more responsive again. The question is: is there any way how to prevent Lion to start pushing running programs from memory into the swap on cost of useless harddrive cache?
    Thanks for suggestions.

    It's been a while but I recall using the "dd" command ("man dd" for info) to copy specific portions of data from one disk, device or file to another (in 512 byte increments).  You might be able to use it in a script to fetch parts of your larger file as you need them, and dd can be used to throw data from and/or to standard input/output so it's easy to get data and store in temporary container like a file or even a variable.
    Otherwise if you can afford it, and you might with 8 GB or RAM, you could try and disable swapping (paging to disk) alltogether and see if that helps...
    To disable paging, run the following command (in one line) in Terminal and reboot:
    sudo launchctl unload -w /System/Library/LaunchDaemons/com.apple.dynamic_pager.plist
    To re-enable paging, run the following command (in one line) in Terminal:
    sudo launchctl load -w /System/Library/LaunchDaemons/com.apple.dynamic_pager.plist
    Hope this helps!

  • Mac OSX desktop dropping connection with multiple copy processes & large files

    The servers are 6.5 SP3 running NFAP, the MAC OSX is 10.4.2 updated. The
    volume the macs are using is part of a cluster. The users mount the volumes
    on their macs and everying is for the most part fine. If they grab a bunch
    of files and copy them from desktop to server it's fine as long as it's only
    a single copy process. The users are part of the hi-res department and the
    files can be 1GB or larger. If they drag one or more large files, and then
    while that's copying they drag some more files, so both copy processes are
    running at once....quite often the volume will dismount from the desktop and
    you will get unable to copy because some resource is unavailable. Sometimes
    the finder crashes, sometimes not. Often the files that were partially
    copied get locked and the users needs to reboot their Mac in order to delete
    them. I'm getting pretty desperate hear, anyone have an idea what's going
    on. I don't know if this is a Tiger thing or a large file thing or a
    multiple copy stream thing, a netware thing or a mac thing.....we have
    hundreds of other users running OSX 10.3 and earlier who are not reporting
    this problem, but they also don't copy files that size. Someone please tell
    me they have seen this before....thanks very much. Oh, before going to 6.5
    and NFAP the servers were 5.1 with Prosoft server and they never had the
    problem.
    Jake

    Thanks for your help, I have incidents open now with Apple and Novell, I
    hope one of them can provide something for us. We tried applying 6.5 SP4 to
    a test server....the problem still happened but was "better", the copy
    operations still quit but with SP4 applied the volume did not dismount....or
    if it did it remounted automatically because it was still connected after
    OKing through the copy errors.
    "Jeffrey D Sessler" <[email protected]> wrote in message
    news:[email protected]...
    >I tried two 2GB files. No problems at all but I'm in a 100% end-to-end
    >Gigabit environment. My server storage is also a very-fast SAN.
    >
    > Best,
    > Jeff
    >
    >
    > "Jacob Shorr" <[email protected]> wrote in message
    > news:[email protected]...
    >> Jeffrey,
    >>
    >> Have you tried the exact same test, dragging say two 500MB files in
    >> seperate
    >> copy operations? I hear what you're saying about the 10/100 link, but we
    >> don't run gigabit to the desktops, and we're not going to anytime soon.
    >> Even if that could resolve the issue we need something kind of other fix
    >> for
    >> our infrastructure. I will look into any errors on the switch.
    >>
    >> "Jeffrey D Sessler" <[email protected]> wrote in message
    >> news:[email protected]...
    >>> Well, considering that I'm not seeing the issue on my 10.4.2 machines
    >>> against my 6.5Sp3 servers, I'm not sure what you should do at this
    >>> point.
    >>> Since you say that the 10.3 machines don't have an issue, it makes it
    >> sound
    >>> to me like this is an Apple issue.
    >>>
    >>> The logs point at a communication issue... Is there anyway to get that
    >>> Mac
    >>> on to a Gigabit connection to see if you can duplicate it?
    >>>
    >>> The other option is to wait for 10.4.3 to be released and see if the
    >> problem
    >>> goes away.
    >>>
    >>> Again, on only a 10/100 link, one copy of a large file _will_ saturate
    >>> the
    >>> link.Perhaps 10.4.2 has an issue with this?
    >>>
    >>> Also, when you're doing the copy, what to the error counters in the
    >> switches
    >>> say?
    >>>
    >>> Jeff
    >>>
    >>> "Jacob Shorr" <[email protected]> wrote in message
    >>> news:[email protected]...
    >>> > There are definately no mis-matches. This has been checked and
    >> re-checked
    >>> > a
    >>> > dozen times. It's only on 10.4......we can replicate it on every 10.4
    >>> > machine, and we cannot replicate it on any machine that is 10.3. What
    >>> > should I do to go about getting this fixed, should I be contacting
    >>> > Apple
    >>> > or
    >>> > Novell? The speed is always good until it actually decides to drop
    >>> > and
    >>> > cut
    >>> > off.
    >>> >
    >>> >
    >>> > "Jeffrey D Sessler" <[email protected]> wrote in message
    >>> > news:7jj%[email protected]...
    >>> >> Looks like communication between the Mac and the Netware server is
    >>> > dropping.
    >>> >> AFP in 10.3 and 10.4 support auto-reconnection but I'm sure that it
    >> will
    >>> >> fail the copy process.
    >>> >>
    >>> >> I'd first check to make sure that there are not any mis-matches on
    >>> >> the
    >>> >> switch e.g. the Mac is set to Auto (as it should be) but someone has
    >> set
    >>> > the
    >>> >> switch to a forced mode. Both should be auto. A duplex miss-match
    >>> >> could
    >>> >> cause the Mac not to see the heart beat back from the Novell server.
    >>> >>
    >>> >> Like I said, if the workstation is only on 10/100, a single copy
    >> process
    >>> > on
    >>> >> a G5 Mac will saturate that link. Adding more concurrent copies will
    >> only
    >>> >> result in everything slowing down and taking longer, or you'll get
    >>> >> the
    >>> >> dropped connections.
    >>> >>
    >>> >> Best,
    >>> >> Jeff
    >>> >>
    >>> >>
    >>> >> "Jacob Shorr" <[email protected]> wrote in message
    >>> >> news:Ybc%[email protected]...
    >>> >> > Take a look at the last entries in the system log right after it
    >>> > happened,
    >>> >> > let me know if it means anything to you. Thanks.
    >>> >> >
    >>> >> > Sep 29 13:26:10 yapostolides kernel[0]: AFP_VFS afpfs_mount:
    >>> >> > /Volumes/FP04SYS11, pid 210
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >> doing
    >>> >> > reconnect on /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >>> > connect
    >>> >> > to
    >>> >> > the server /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >>> > Opening
    >>> >> > session /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >>> > Logging
    >>> >> > in
    >>> >> > with uam 2 /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >>> >> > Restoring
    >>> >> > session /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS
    >>> >> > afpfs_MountAFPVolume:
    >>> >> > GetVolParms failed 0x16
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >>> >> > afpfs_MountAFPVolume failed 22 /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides KernelEventAgent[43]: tid 00000000
    >>> >> > received
    >>> >> > VQ_DEAD event (32)
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_Reconnect:
    >>> > posting
    >>> >> > to
    >>> >> > KEA to unmount /Volumes/FP04SYS11
    >>> >> > Sep 29 13:31:13 yapostolides KernelEventAgent[43]: tid 00000000
    >>> >> > type
    >>> >> > 'afpfs', mounted on '/Volumes/FP04SYS11', from
    >>> >> > 'afp_0TQCV10QsPgy0TShVK000000-4340.2c000006', dead
    >>> >> > Sep 29 13:31:13 yapostolides KernelEventAgent[43]: tid 00000000
    >>> >> > found
    >> 1
    >>> >> > filesystem(s) with problem(s)
    >>> >> > Sep 29 13:31:13 yapostolides kernel[0]: AFP_VFS afpfs_unmount:
    >>> >> > /Volumes/FP04SYS11, flags 524288, pid 43
    >>> >> >
    >>> >> >
    >>> >> >
    >>> >> >
    >>> >> > "Jeffrey D Sessler" <[email protected]> wrote in message
    >>> >> > news:GH%[email protected]...
    >>> >> >> We move large files all the time under SP3 with no issues however,
    >>> > there
    >>> >> > are
    >>> >> >> several finder/copy/afp issues in Tiger that are do to be fixed in
    >>> >> >> 10.4.3.
    >>> >> >>
    >>> >> >> Also, if you have any type of network issue such as duplex
    >> mis-matches
    >>> > or
    >>> >> >> are running say, only a 10/100 network, a single Mac can not only
    >>> >> >> transfer
    >>> >> >> more than 10MB/sec (filling the network pipe) or generate so many
    >>> >> > collisions
    >>> >> >> (duplex mis-match) that you could drop communication to the
    >>> >> >> server.
    >>> >> >>
    >>> >> >> What type of server (speed, disks, raid level, NIC speed) and what
    >>> >> >> type
    >>> >> >> of
    >>> >> >> network (switched gigabit, switched 10/100, shared 10/100, etc.)
    >>> >> >>
    >>> >> >> How long does it take to copy that single 1GB file to the server?
    >>> >> >>
    >>> >> >> Does a single copy process always work?
    >>> >> >>
    >>> >> >> Jeff
    >>> >> >>
    >>> >> >> "Jacob Shorr" <[email protected]> wrote in message
    >>> >> >> news:[email protected]...
    >>> >> >> > The servers are 6.5 SP3 running NFAP, the MAC OSX is 10.4.2
    >> updated.
    >>> >> > The
    >>> >> >> > volume the macs are using is part of a cluster. The users mount
    >> the
    >>> >> >> > volumes
    >>> >> >> > on their macs and everying is for the most part fine. If they
    >> grab
    >>> >> >> > a
    >>> >> >> > bunch
    >>> >> >> > of files and copy them from desktop to server it's fine as long
    >>> >> >> > as
    >>> > it's
    >>> >> >> > only
    >>> >> >> > a single copy process. The users are part of the hi-res
    >> department
    >>> > and
    >>> >> >> > the
    >>> >> >> > files can be 1GB or larger. If they drag one or more large
    >>> >> >> > files,
    >>> > and
    >>> >> >> > then
    >>> >> >> > while that's copying they drag some more files, so both copy
    >>> > processes
    >>> >> > are
    >>> >> >> > running at once....quite often the volume will dismount from the
    >>> >> >> > desktop
    >>> >> >> > and
    >>> >> >> > you will get unable to copy because some resource is
    >>> >> >> > unavailable.
    >>> >> >> > Sometimes
    >>> >> >> > the finder crashes, sometimes not. Often the files that were
    >>> > partially
    >>> >> >> > copied get locked and the users needs to reboot their Mac in
    >>> >> >> > order
    >>> >> >> > to
    >>> >> >> > delete
    >>> >> >> > them. I'm getting pretty desperate hear, anyone have an idea
    >> what's
    >>> >> > going
    >>> >> >> > on. I don't know if this is a Tiger thing or a large file thing
    >> or
    >>> >> >> > a
    >>> >> >> > multiple copy stream thing, a netware thing or a mac
    >>> >> >> > thing.....we
    >>> > have
    >>> >> >> > hundreds of other users running OSX 10.3 and earlier who are not
    >>> >> > reporting
    >>> >> >> > this problem, but they also don't copy files that size. Someone
    >>> > please
    >>> >> >> > tell
    >>> >> >> > me they have seen this before....thanks very much. Oh, before
    >> going
    >>> > to
    >>> >> >> > 6.5
    >>> >> >> > and NFAP the servers were 5.1 with Prosoft server and they never
    >> had
    >>> >> >> > the
    >>> >> >> > problem.
    >>> >> >> >
    >>> >> >> > Jake
    >>> >> >> >
    >>> >> >> >
    >>> >> >>
    >>> >> >>
    >>> >> >
    >>> >> >
    >>> >>
    >>> >>
    >>> >
    >>> >
    >>>
    >>>
    >>
    >>
    >
    >

  • How to read from and write into the same file from multiple threads?

    I need to read from and write into a same file multiple threads.
    How can we do that without any data contamination.
    Can u please provide coding for this type of task.
    Thanks in advance.

    Assuming you are using RandomAccessFile, you can use the locking functionality in the Java NIO library to lock sections of a file that you are reading/writing from each thread (or process).
    If you can't use NIO, and all your threads are in the same application, you can create your own in-process locking mechanism that each thread uses prior to accessing the file. That would take some development, and the OS already has the capability, so using NIO is the best way to go if you can use JDK 1.4 or higher.
    - K
    I need to read from and write into a same file
    multiple threads.
    How can we do that without any data contamination.
    Can u please provide coding for this type of task.
    Thanks in advance.

  • Read ini file with multiple threads

    I have a state machine architecture, but I have multiple threads. For instance, one is dealing with listening for mulitple tcp connections and storing their refnums to an array in a functional global. Another thread is accessing these refnums and collecting data it's receiving from each tcp connection. This data is queued up and then dequeued in a third thread which does some modification and sends it out via serial. My question is, when you have a situation like this and have to read an ini file, where should you read it? It seems like the most logical place would be outside your loops so you can get all the tcp and serial info (port, baud rate, etc) then wire it it to your create listener or initialize serial connection despite them being in different threads. But then again, normal state machine architecture would want an "initialize" case. If you did this though which loop would you put the init case in? And you would then have to worry about synchronizing loops becuase you wouldn't want one to try and create a listener while another thread was still reading ini data which would include the port to listen on. Maybe I'm overthinking this haha. Suggestions? Edit: one more question. Does it seem overkill that I have a tcp loop listening for data and queuing it up and a separate loop sending out the processed data via serial? Or should I just have one state that gets tcp data stores it in a shift register, then another state that sends it out via serial, and returns the state machine to the tcp read state?
    Message Edited by for(imstuck) on 03-03-2010 01:13 PM
    Message Edited by for(imstuck) on 03-03-2010 01:17 PM
    CLA, LabVIEW Versions 2010-2013

    Most of the applications I work on at the moment are used for testing barcode and label printers. The test applications I design are focused on testing the printer's firmware, not the hardware. Within our applications we have three primary objects (Unfortunately they are not native LabVIEW objects yet. They were developed before native LVOOP.) The primary objects we use in our applications are a log object, a connection object (communication interface to the printer) and a printer object. In any single instance of a test we only have a single printer, a single connection to the printer and one or more discrete logs. Each instance of these objects represent a single, real physical entity. A singleton object is a virtual representation of the physical world. Let's take the log object since that is the most simple of the objects described above. Naturally for a given log file you have the log file name and path. We also provide other attributes such as the maximum size of a single file (we allow log files to span multiple files), whether it is a comma delimited file or if it contains raw data, if timestamps should be included with a log entry and so forth. Most of these attributes are static for a log file with the exception of the name and such things as whether the logging is actually enabled or disabled. If we split a wire and had multiple instances of the log file (the way native LVOOP actually works) the attribute for whether logging is currently enabled or disabled will only pertain to the specific instance, or specific wire for the that object. Since this truly represents a single item, one log file, we need that attribute to be shared for all references to the instance of the log object. Since we allow this we can set an attribute on the log object in any task and it will be reflected in any other task that is using it. Think of the way a action engine or functional global works. However, in this case we provide discrete methods for the various actions.
    I hope that made some sense. If not let me know since I just whipped up this response.
    Mark Yedinak
    "Does anyone know where the love of God goes when the waves turn the minutes to hours?"
    Wreck of the Edmund Fitzgerald - Gordon Lightfoot

  • Upload and Process large files

    We have a SharePoint 2013 OnPrem installation and have a business application that provides an option to copy local files into UNC path and some processing logic applied before copying it into SharePoint library. The current implementation is
    1. Users opens the application and  clicks “Web Upload” link from left navigation. This will open a \Layouts custom page to select upload file and its properties
    2. User specifies the file details and chooses a Web Zip file from his local machine 
    3. Web Upload Page Submit Action will
         a. call WCF  Service to copy Zip file from local machine to a preconfigure UNC path
         b. Creates a list item to store its properties along with the UNC path details
    4. Timer Job executes in a periodic interval to
         a. Query the List to see the items that are NOT processed and finds the path of ZIP file folder
         b. Unzip the selected file 
         c. Loops of unzipped file content - Push it into SharePoint library 
         d. Updates list item in “Manual Upload List”
    Can someone suggest a different design approach that can manage the large file outside of SharePoint context? Something like
       1. Some option to initiate file copy from user local machine to UNC path when he submits the layouts page
       2. Instead of timer jobs, have external services that grab data from a UNC path and processes periodic intervals to push it into SharePoint.

    Hi,
    According to your post, my understanding is that you want to upload and process files for SharePoint 2013 server.
    The following suggestion for your reference:
    1.We can create a service to process the upload file and copy the files to the UNC folder.
    2.Create a upload file visual web part and call the process file service.
    Thanks,
    Dennis Guo
    TechNet Community Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Dennis Guo
    TechNet Community Support

  • File Adapter Not Processing Large File

    Hi Everyone,
    We are experiencing a problem in an interface that includes file adapter. The scenario is as follows:
    1. A .csv file containing multiple records needs to be processed; for each line of the .csv file, an XML file should be genrated. <b>We are using the 'Recordsets Per Message' field in the Content Conversion Parameters to achieve this.</b> After the source .csv file is processed, it is deleted.
    We were testing with small 15-30 records in each input file. For the purpose of testing scalability, we increased the number of records in a file to nearly 300. But the file adapter did not pick up the file.Comm.Channel in RWB is showing green, MDT contains no entries and SXMB_MONI also shows no messages. What can be the problem? Is there any limit on the size of file that can be converted in this way?
    Awaiting your replies,
    Regards,
    Amitabha

    Amitabha,
    300 records should not be a problem at all. If you are not getting any error msg in CC monitoring, then it is time you take a look into VA logs.
    Ref: /people/michal.krawczyk2/blog/2005/09/07/xi-why-dont-start-searching-for-all-errors-from-one-place
    Regards,
    Jai Shankar

  • COMPUTE CRASHES WHEN PROCESSING LARGE FILES

    As far as basic operations, my G4 is running smoothly.
    However, whenever I need it to process significant files, such as exporting a 30minute video from FCP as a Quicktime movie, or using Compressor to encode an .M2V file, the computer crashes. Basically, if any task is going to take longer than fifteen minutes to complete, I know my computer won't make it.
    Thus far I've done a fresh install of the system software, reinstalled all applications, trashed prefs, run the pro application updtates, run disk utility, played with my work flow (internal vs. external drives), etc.
    I wonder if perhaps my processor is failing, if I need more memory (though my 768MB exceeds application minimum requirements), or if perhaps these G4s just aren't adequately equipped to run the newer pro application versions.
    Thanks in advance for any advice.

    I can't pull the dimm out due to the fact that I need a certain amount of memory installed to be able to run the software in the first place.
    I've run the hardware test disc that came with my computer and it has not detected any problems.
    I don't think heat is the issue as, according to the Temperature Monitor utility I downloaded, my computer remains consistant at around 58 degrees, even when performing difficult processes.
    According to my activity Monitor, when I'm processing one of these larger files, I'm using as much as 130% of the cpu, but it also can remains as low as 10% for extended periods. Both seem odd.
    Any thoughts?

  • Burning Large File to Multiple DVDs

    I am trying to burn a large group of files to multiple DVDs. I have Roxio Toast at work and all I do is drag all the items into the CD and it tells me how many DVDs it will require and asks me for the next one when the first is done. Pretty easy. I dont have toast at home and am wondering how to do that with the default built in burn software.
    I did a search, but my search skills are lacking and after pages of non-related seaching I figured I would just post the question.

    I'm also interested in this question and am confused.
    Are you talking about * Explanation 1) making duplicates of a CD without having to manually drag every single folder onto a dvd - or * Explanation 2) copying one set of data, too large for one DVD, across multiple DVDs (like span/concatenation).
    Itchandscratch - in the subject line you say 'large file' then 'group of files' in your explanation. If the files are in a group (comprised of smaller individual files) could you not say, halve them and burn onto two disks - or is the folder system of the group complex? But then you say
    "I have Roxio Toast at work and all I do is drag all the items into the CD and it tells me how many DVDs it will require and asks me for the next one when the first is done" Suggests not Explanation 1
    KJK555 You say "Drag and drop files and folders until image reaches DVD size.
    Close image." Which made me think we are talking about explnation 1
    Sorry if i'm being stupid. I'm interested but confused! : )

  • Processing Large Files in Adobe Premiere Elements 12

    Greetings,
    I am trying to process video files that are approximately 500 to 750mb in Adobe Premiere Elements 12.  I am running on Windows 7 Home Premium, Intel i5 processor 2300, 8gb DDR3 and Invidia GeForce GT520 Video Card.  It is choking so bad I cannot edit my videos at all. Reliable technical support here says that it is the limitations of the software and upgrading my hardware would be next to useless.  Can you please clarify, unfortunately Premiere Pro is out of my budget.

    Lucidity2014
    I have gone over your thread again. I do not believe that project settings are at the core of file size being a limitation in your workflow. And, at face value, the format of AVCHD.mov (1920 x 1080 @ 24 progressive frames per second) should be supported by Premiere Elements 12 on Windows 7 64 bit (and you have QuickTime latest version installed on your computer along with Premiere Elements 12).
    This is what I would like you to do to demonstrate whether or not, in your situation, the project settings are at the core of your file size limitation issue
    1. Open a new Premiere Elements 12 project to the Expert workspace. Go to File Menu/New/Project and Change Settings.
    2. In the Change Settings dialog, make sure that the project preset is set to
    NTSC
    DSLR
    1080p
    DSLR 1080p24
    Before you close out of there, make sure that you have a check mark next to "Force Selected Project Setting on This Project" in the new project dialog which is the last dialog you should see as you exit that area.
    3. Then back in the Expert workspace of the project, import your AVCHD.mov (500 to 750 MB/4 minutes 49 seconds) using the program's Add Media/Files and Folders.
    a. Do you see an orange line over the Timeline content when it is first dragged to the Timeline?
    b. Do the problems exit as before?
    Click on Start button. In the Search field above the Start button, type in System Information. In System Information, please tell us what you see for:
    Total Physical Memory
    Available Physical Memory
    Total Virtual Memory
    Available Virtual Memory
    Page File Space
    From you initial report, your installed RAM is supposed to be 8 GB.
    Previously you wrote
    The filepath is; C:\Users\Brian Ellis\Desktop\GBC Video for the Web\Video Footage
    I assumed the free hard drive location would be the same properties as the C: Drive (1.07 TB), but please advise if I should move off the desktop?
    Just in case, after you have ruled out the project settings factor, please change the file path so that the file is saved to Libraries/Documents or Libraries/Videos. Then start a new project, go through setting the project preset manually, and then in the project Add Media/Files and Folders from the new hard drive save location (Documents or Videos). Moving forward do you have an external hard drive for video storage?
    We will be watching for your progress.
    Thank you.
    ATR
    Add On Comment...To give others a view of the project settings that you have been using for the problem situation, please go to the Edit Menu/Project Settings/General and tell us what is there for Editing Mode, Timebase, Frame Size, and Pixel Aspect Ratio - even if the fields appear to be grayed out. That should answer everyone's questions.

  • After Effects CS5-best settings for rendering many large files after multiple passes with the clone

    I have a moderate level Quad core PC,specs below, with 24Gb of RAM.
    I am processing sequences of timelapse photos brought in from Adobe RAW as either a raw sequence at the original 5K res (Canon 5D2) or as 4K 16bit TIFFs. There are typically between 400-500 shots in each sequence.
    I use clone brush in AE CS5 to remove 'stray objects' such as birds, people or vehicles from frames.
    I can succesfully process these if I only have to clone out the odd object from a sequence, but occasionally I require to remove something on multiple frames. This might be a car that drove into shot and stayed for say 200 frames before driving away again.
    The clone tool lets me paint out the arrival and then the parked vehicle by successively using the previous frame as the source. However at some point, even with say 16-20 Gb set as the AE memory, the program will get choked and start to crawl. This knocks on to the final render, when I will get outragous times for rendering even single frames, say 3 hours per frame etc.
    I assume that AE is having to go back through many multiple source frames as the cloning that I have done is continually using the previous 'already cleaned up frame' as the source.So frame 150 is using frame 149, which used frame 148 etc.
    Observing the Resource Monitor as AE renders, I see not much Processor usage, tyically 20-30% but maxed out Memory. Lots of reading of files off disk, usually apparently 'random' Windows system files, but also lots or reads from the timelapse source files as well.
    I did get single render times down to 1 hour but then did more tweaking and haven't got back to anything better than 2.5 hours per frame since!
    Is there a better way of achieving what I want?
    My PC specs and set up: HP xw8600 with x5450 Quad core processor.
    24 Gb RAM set usually as 16 Gb for AE (3 cores) and 1 core with max 3Gb for other progs. Sometimes I have set this to up to 20Gb fror AE.
    Running Vista Business (yes I know Win7 would be better-but hey it's yet more money!).
    Nvidia GTX460 1Gb memory with Cuda 'hack'-which works well with PP CS5.
    Windows Page File set to about 20% more than Windows usually sets it to, on either separate SATA drive or dual with the Windows C drive.
    Source files on external USB 3 drive (pretty fast).
    AE render files on RAID 0 internal 7200 drive.
    AE output files to another external USB3 drive.
    Any suggestions for improving my set up would be appreciated, as I'm getting old watching AE render my projects!

    Is there a better way of achieving what I want?
    Don't use AE? Just kidding, but your own analysis very much covers the facts already. Yes, it's AE going back and forth in time and yes, it's the holding of those frames in memory which is chewing up your memory. Depending on the situation you may look into patching up your disturbances with masked still images rather than cloning across the sequence, but beyond that I don't see much potential to make things more efficient. It's just AE trying to be über-smart with caching and then making a mess. The only other thing is the RAW import, which may consume unnecessary memory. Batching the files in Photoshop and only using PSDs or TIFFs in AE may squeeze out an extra GB of RAM. On that note also consider doing the cleanup in PS. After all, PS Extended does support video/ image sequences to some degree...
    Mylenium

  • Processing delimited file with multiple format records

    Hi All,
    I have got file containing following records to be processed by ODI:
    1000022,0236,E,APHDQR,,264, 32.83,13-08-2010,13-08-2010,W,33,
    1000022,0236,D,APHDQR,,Z01, 47.05,13-08-2010,13-08-2010,W,33,
    1000022,0236,M,APHDQR,,159, 3.00,13-08-2010,13-08-2010,W,33,
    1000022,0236,A,APHDQR,B120,7140,026,7140, 338.80,106,33,5018676E,
    1000022,0236,B, 35.19,1,503,W,13-08-2010,13-08-2010,6 ,COR ,
    As you can see there are 5 types of records (Record type 'E','D','M - same format) and ( Record Type 'A' and 'B').
    I need to be able to process records with Record Type 'E','D','M' and IGNORE records with record type 'A','B'
    I tried placing filter at staging level (CL_RECORD IN ('D','E','M')) in the Interface definition. But it doesn't work and still tries to process records with record type 'A' and 'B'....:(
    Any ideas? and any other way I can process records ??
    Thanks.
    KS

    OK,
    then you don't have too much options...
    the problem is that before apply the filter ODI tries to upload the data into the C$ table and, as there are multiples formats, some columns is getting "wrong datatype" (or some error of this kind).
    My question about the initial character was because if they were always the same, you could use a feature from JDBC Driver to file but only works to initial characters.
    In your case, it will be necessary to create a temp table with 3 columns like:
    Filler1 - Before the record type column
    Rec_Type - the record type
    Fille2 - the other columns
    Filler1 and Filler2 columns should has the necessary lenght to get all data from record.
    Now you can filter the record type but will need a "substring" function to separate the embedded into filler1 and filler2.
    Make any sense?
    Cezar Santos
    http://odiexperts.com

Maybe you are looking for

  • MULTIPLE crashes on G5 IMacs

    I work at a veterinary hospital with 5 IMac G5's, a mac mini, and a G5Tower. We have had multiple crashes with all 5 of our IMacs. There seems to be no rhyme or reason to the crashes. Generally speaking, they occur after we shut them down. Upon resta

  • Need to upgrade my MacBook from Snow Leopard.  Should I go straight to Yosemite?

    I am a nervous person. The idea of upgrading always fills me with dread. I have way too much to do, to have computer problems. Yet my MacBook is getting sluggish (I can type faster than the words show up, now). It's probably because I still run 10.6.

  • Can't install OS 10.4 on new hard drive

    I installed a new WD 320Gb hard drive in our PowerBook G4 today but can't load the os. I connected the G4 as a target to another mac and partitioned the new drive and erased it with zero. Everything looks good there but . . . the g4 wont recognize th

  • E3000

    Hello, I'm trying to find more information E3000 specs.  From what I've seen on newegg it looks like it uses 802.11n Draft but everything I can find on the Cisco's website doesn't mention draft at all. Also I'm wondering if it can be configured to be

  • How to create a dual boot on OSX Mavericks?

    I've Mavericks and I want to install two more OS X (10.7 and 10.8) on the same machine. Is there any way to install earlier OS X along with the latest OSX ? I've created partitions and tried to install there but a message says something like this "Th