How Oracle Hold Big Files ?

hi,
i have some big-size pictures and i want to hold them in the oracle database and when a button pressed for ex.
display these..
how can i do this?
thanks for any reply
regards..

ok i hold the picture with the following code but how will i display the picture for ex. by the form developer?
create or replace procedure insert_img as
f_lob bfile;
b_lob blob;
begin
insert into waterfalls (falls_name,falls_photo)values ( 'MyGif', empty_blob() )
return falls_photo into b_lob;
f_lob := bfilename( 'MY_FILES', 'dog.gif' );
dbms_lob.fileopen(f_lob, dbms_lob.file_readonly);
dbms_lob.loadfromfile( b_lob, f_lob, dbms_lob.getlength(f_lob) );
dbms_lob.fileclose(f_lob);
commit;
end;
thanks for any reply..

Similar Messages

  • How to read big files

    Hi all,
    I have a big text file (about 140MB) containing data I need to read and save (after analysis) to a new file.
    The text file contains 4 columns of data (so each row has 4 values to read).
    When I try to read all the file at once I get a "Memory full" error messags.
    I tried to read only a certain number of lines each time and then write it to the new file. This is done using the loop in the attached picture (this is just a portion of the code). The loop is repeated as many times as needed.
    The problem is that for such big files this method is very slow and If I try to increase the number of lines to read each time, I still see the PC free memory decending slowly at the performance window....
    Does anybody have a better idea how to implement this kind of task?
    Thanks,
    Mentos.
    Attachments:
    Read a file portion.png ‏13 KB

    Hi Mark & Yamaeda,
    I made some tests and came up with 2 diffrenet aproaches - see vis & example data file attached.
    The Read lines aproach.vi reads a chunk with a specified number of lines, parses it and then saves the chunk to a new file.
    This worked more or less OK, depending on the dely. However in reality I'll need to write the 2 first columns to the file and only after that the 3rd and 4th columns. So I think I'll need to read the file 2 times - 1st time take first 2 columns and save to file, and then repeat the loop and take the 2 other columns and save them...
    Regarding the free memory: I see it drops a bit during the run and it goes up again once I run the vi another time.
    The Read bytes approach reads a specified number of bytes in each chunk until it finishes reading all the file. Only then it saves the chunks to the new file. No parsing is done here (just for the example), just reading & writing to see if the free memory stays the same.
    I used 2 methods for saving - With string subset function and replace substring function.
    When using replace substring (disabled part) the free memory was 100% stable, but it worked very slow.
    When using the string subset function the data was saved VERY fast but some free memory was consumed.
    The reading part also consumed some free memory. The rate of which depended on the dely I put.
    Which method looks better?
    What do you recommand changing?
    Attachments:
    Read lines approach.vi ‏17 KB
    Read bytes aproach.vi ‏17 KB
    Test file.txt ‏1 KB

  • How to exchange big files between mac users through the internet?

    How can we send and receive big files between Mac users (apart from Dropbox) ?

    If you aren't on the same local network (which I assume you aren't), then the easiest solution is to use Google.
    If you simply google "sending large files" you will get hundreds of suggestions/applications.
    Ie Dropbox, Google, YouSendIt, FTP, etc

  • How to split big file?

    Hi, folks,
    I have a big binary file (around 400M). How can I split it to several small files? If I just use the file read subvi in labview, it will show "not enough memory" error. How can I solve this problem?
    Any help will be appreciated!

    > I have a big binary file (around 400M). How can I split it to several
    > small files? If I just use the file read subvi in labview, it will
    > show "not enough memory" error. How can I solve this problem?
    >
    This is because you are attempting to read the entire file at once.
    File I/O allows you to read anywhere from one byte to the entire file.
    I don't know what is in your file, but the read characters from file has
    an input for how many characters. Or you can use the more advanced File
    Open followed by File Read with specific types wired up.
    Either way, determine how you will break it up. 40 10M files, 10 40M
    files, etc. Then make a loop where you create the new file, read the
    data from the old large file, then write it to the new file, then clos
    e
    the new file and go to the next iteration of the loop.
    Greg McKaskle

  • How to handle Big FIles in SAP PI Sender file adapter

    Hi all ,
    I have developed a interface , where it is File to Proxy, it is fine when i do with small and normal files
    The structure contain one  Header unbounded  detail and one  Trailer, how to handle when the file size is more than 40 MB
    Thanking you
    Sridhar

    Hi Sridhar Gautham,
    We can set a limit on the request body message length that can be accepted by the HTTP Provider Service on the Java dispatcher. The system controls this limit by inspecting the Content-Length header of the request or monitoring the chunked request body (in case chunked encoding is applied to the message). If the value of the Content-Length header exceeds the maximum request body length, then the HTTP Provider Service will reject the request with a 413 u201CRequest Entity Too Largeu201D error response. You can limit the length of the request body using the tting MaxRequestContentLength property of the HTTP Provider Service running on the Java dispatcher. By default, the maximum permitted value is 131072 KB (or 128MB).You can configure the MaxRequestContentLength property using the Visual Administrator tool. Proceed as follows:
           1.      Go to the Properties tab of the HTTP Provider Service running on the dispatcher.
           2.      Choose MaxRequestContentLength property and enter a value in the Value field. The length is specified in KB.
           3.      Choose Update to add it to the list of properties.
           4.      To apply these changes, choose  (Save Properties).
    The value of the parameter MaxRequestContentLength has to be set to a high value.
    The Visual administartor tool may be accessed using this link
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/40a08db9-59b6-2c10-5e8f-a4b2a9aaa3d2?quicklink=index&overridelayout=true
    In short  ICM parameters to reset values for this case are
    icm/HTTP/max_request_size_KB
    icm/server_port_ TIMEOUT
    rdisp/max_wprun_time
    zttp/max_memreq_MB
    Please look into this thread to know more about ICM parameters
    http://help.sap.com/saphelp_nw04/helpdata/en/61/f5183a3bef2669e10000000a114084/frameset.htm
    Second solution is that you must split the source file, so that each file is less than 5MB in size, then PI would not cause problem for file size between 1MB-5MB. you can insert header and trailer for individual smaller file obtained after split. All this can be done using scripts or conventional programing provided individual records within file are independent of each other. Finally you have to rename each new file created and put them in PI folder in sequential manner. All this can be achieved by simple shell script/batch file, a C code or java code. If you are going for a C or Java code you need a script to call them from PI communication channel parameter  "run operating system command before message processing".
    regards
    Anupam

  • How to transfer big files from imac to hard drive

    haveing diffulty of copying file from Imac to Hard drives. is there any free app which i can download and help me with my problem.

    To show HD on desktop
    go into FINDER at top then PREFERENCES  then GENERAL tab
    then check "hard disks" and "external disks"
    after that drag and drop files as you need to.  You dont need an application for same.

  • Navigating the Macintosh HD - How to find big files?

    Hi all,
    My Mac mini has swallowed a 7gb file by accident... it was a DVD image I planned to watch and delete, but I forgot the location of the file on the HD.
    The naming convention will make it difficult for me to search by name.
    Is there a way to search the drive by file size??
    Steve

    Indeed, easy to find that way. If, on the other hand, you're looking more in general than for a specific file, there's a useful utility to have around called WhatSize that can help track what files and folders are taking what amount of space. That can be helpful if trying to work out why available drive space has decreased! WhatSize can be downloaded from www.versiontracker.com

  • How to load big file.

    I have 20 G file. I am considering to use sql loader or external table loading.
    Which performance of method is better?

    Howerver, external table can use parallel run. Do you make sure sql loader is quickly? I use Sql loader,but spend 1 hour. Of course , I have not try external tabel loading.

  • Upload of very big files (300MB+)

    Hello all,
    I am trying to create application in HTMLDB to store files via webbrowser in DB(BLOB). I created all needed components and now stuck with the problem of uploading big files. Basicaly when file is over 100MB it becomes unreliable and for big files does not work at all. Can somebody help me to figure out how to upload big files into htmldb application. Any hints and sugestions are welcome. Examples will be even more appreciated.
    Sincerely,
    Ian

    Ian,
    When you say "big files does not work at all", what do you see in the browser? Is no page returned at all?
    When a file is uploaded, it takes some amount of time to simply transfer the file from the client to modplsql. If you're on a local Gbit network, this is probably fast. If you're doing this over a WAN or over the Internet, this is probably fairly slow. As modplsql gets this uploaded file, it writes it to a temporary BLOB. Once fully received, modplsql will then insert this into the HTML DB upload table.
    I suspect that the TimeOut directive in Apache/Oracle HTTP Server is kicking in here. The default setting for this is 300 (5 minutes).
    I believe the timeout is reset by modplsql during file transfer to avoid a timeout operation while data is still being sent. Hence, I believe the insertion of your large file into the file upload table is taking longer than the TimeOut directive.
    The easy answer is to consider increasing your TimeOut directive for Apache/Oracle HTTP Server.
    The not so easy answer is to investigate why it takes so long for this insert, and tune the database accordingly.
    Hope this helps.
    Joel

  • How to handle big java source files in JDEV 10.1.3

    Hi,
    I have some .java files that are like over 10,000 lines long. I know this is crazy but I didn't write it but I have to maintain it. I really dont feel like refactoring all these big source files right now.
    JDeveloper has lots of while-editing features that are probably grinding to a halt on this really big file. My CPU usage shows ~97% whenever I do anything in this file.
    Is there some preferences I could set or something I can disable just while editing this big file.
    Thanks,
    Simon.

    The delays between edits and the IDE stalls seemed
    about the same with this change in place.Ah... that's sad. It seems like there's another (un-configurable) timer mechanism for background parsing used by the structure pane.
    I didn't
    really understand what the numbers were controlling
    so I didn't try any other values. They control the frequency of a timer that is reset each time you press a key in the code editor. After the number of ms specified in these parameters has elapsed after the last keypress, the code in the editor is parsed. There are two settings, one for short files and one for long files (I forget what the threshold is between short and long).
    Maybe it isn't the parser since it still goes to 100%
    CPU (just less often) but I can still move the mouse
    around the file - no stalling. Here's something I'd find useful to investigate this problem: Run jdeveloper with jdev.exe so that you have a console. When the CPU hits 100%, switch to the console and press Ctrl+Break. Copy the stack dump (maybe take a couple of them) and email them to me (brian dot duff at oracle dot com).
    Dropping the structure window is a fairly simple work
    around for me. I can live without the structure
    window for this file although it is actually most
    useful as the file's size, and structure, increases.Yes, I can imagine the structure pane would be quite useful with a mammoth java file like the one you're working with :) Incremental find might be your friend in its absence (Ctrl+E, I think).
    If your considering an ER for the structure window
    (:-) then how about this too:
    "Cursur location within the file should be shown in
    the structure window"
    This already happens for Ant (build.xml) files but
    not for Java files - or at least not for me. The
    problem is that it is easy to get lost in a large
    file and the navigation assistance should be the same
    across all structured file types. Funnily enough, I was talking to one of the developers on the team that builds the infrastructure for the xml based parts of the product (including ant) about these kind of inconsistencies between the XML and Java structure panes just a couple of days ago... :) I'll file a bug for this (rather than an ER), since it's a UI inconsistency.
    Thanks,
    Brian

  • Suggestion needed for processing Big Files in Oracle B2B

    Hi,
    We are doing a feasibility study for Using Oracle AS Integration B2B over TIBCO. We are presently using TIBCO for our B2B transactions. Now since my client company planning to Implement Fusion Middleware (Oracle ESB and Oracle BPEL), we are also looking at Oracle AS Integration B2B for B2B transactions (On other words we are planning to replace TIBCO by Oracle Integration B2B if possible).
    I am really concern about one thing that is receiving and processing any "BIG FILE" (15 MB of size) from trading partner.
    Present Scenario: One of our trading partner is sending Invoice documents in a single file and that file size can grow upto 15 MB of size. In our existing scenario when we receive such big files from trading partner (through TIBCO Business Connect - BC), Tibco BC works fine for 1 or 2 files but it crashes once it received multiple files of such size. What exactly happening is Whatever Memory that TIBCO BC is consuming to receive one such big file, are not getting released after processing and as a result TIBCO BC throws "OUT OF MEMORY" error after processing some files.
    My questions:
         1. How robust the Oracle AS Integration B2B is, in terms of processing such big files?
         2. Is there any upper limit in terms of size that Oracle AS Integration B2B can handle for receiving and processing data?
         3. What is the average time required to receive and process such big file? (Lets say we are talking about 15MB of size).
         4. Is there any documentation availble that talks about any such big files through Oracle B2B?
    Please let me know if you need more information.
    Thanks in advance.
    Regards,
    --Kaushik                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

    Hi Ramesh,
    Thanks for your comment. We will try to do POC ASAP. I will definitely keep in touch with you during this.
    Thanks bunch.
    Regards,
    --Kaushik                                                                                                                                                                                                                                                                                                                               

  • Does anybody know how Oracle load large N-Triple file into Oracle 11g R1?

    Does anybody know how Oracle load large N-Triple(NT) file into Oracle 11g R1 by using sql*loader according to their benchmark results?
    Their benchmark results indicate they have over come the large data set problem.
    http://www.oracle.com/technology/tech/semantic_technologies/htdocs/performance.html
    It means they have loaded LUBM 8000(1.068 Billion+ Triples) into Oracle successfully, but there is no detailed steps provided. For instance, 32-bit or 64-bit platform they used, only one NT file being used corresponding to one dataset or several NT files?
    Is there any exception occured during the loading process if the NT file beyond 60GB? When using jena to generate NT file against LUBM(8000), the size of the NT file would definitely beyond 60GB.
    We are dividing such large NT file into several small ones? Is it a good approach? I'm hesitating to do so!

    A Linux 32-bit platform was used for bulk-load of LUBM-8000 1.106 billion (before duplicate elimination) RDF triples into Oracle.
    Multiple gzipped N-Triple files were used to hold the LUBM-8000 data. zcat was used on all these files together to send the complete data into a named pipe. SQL*Loader used this named pipe as the input data file to load the data into a staging table in Oracle. Once the staging table was loaded, the sem_apis.bulk_load_from_staging_table API was used to load the data into Oracle Semantic Store.
    (Additional details in http://www.oracle.com/technology/tech/semantic_technologies/htdocs/performance.html )
    Thanks.

  • Not enough space on my new SSD drive to import my data from time machine backup, how can I import my latest backup minus some big files?

    I just got a new 256GB SSD drive for my mac, I want to import my data from time machine backup, but its larger than 256GB since it used to be on my old optical drive. How can I import my latest backup keeping out some big files on the external drive?

    Hello Salemr,
    When you restore from a Time Machine back up, you can tell it to not transfer folders like Desktop, Documents. Downloads, Movies, Music, Pictures and Public. Take a look at the article below for the steps to restore from your back up.  
    Move your data to a new Mac
    http://support.apple.com/en-us/ht5872
    Regards,
    -Norm G. 

  • How do I find out the version of  Oracle Time Zone files?

    Hello there,
    How do I find out the version of Oracle Time Zone files? I'm in the process of applying the recent oracle patch and need to find out the version of oracle time zone in my db. Could you please help me?

    run
    SELECT version FROM v$timezone_file;

  • How to create xml file from Oracle and sending the same xml file to an url

    How to create xml file from Oracle and sending the same xml file to an url

    SQL/XML (XMLElement, XMLForest, XMLAgg, etc) and UTL_HTTP.
    Whether that works for you with the version of Oracle you have, your requirements, and needs is another story. A little detail goes a long way.

Maybe you are looking for

  • Trying to download PS Elements 12 and it says my copy is counterfeit.

    Hi there,      I'm trying to download Photoshop Elements 12 on my laptop and when it reaches 99% it says my copy is counterfeit. I was given this as a gift for my birthday in all of the proper packaging (never been opened). How can I download this on

  • After update software freezes and other troubles

    I installed the latest update from apple little over a week ago. I also started having these issues about the same time: I`ll just list my problems: -Firefox, Adobe CS5, Preview, Adobe Reader, iTunes and just about any other program freezes or crashe

  • Memory Card N500

    i wonder if somebody has the same problem like mine. i was putting some mp3 songs and some pictures in the memory card, somehow the songs won't play and the pictures didn't show in photos slide.  but nothing is wrong with the memory card.. it can sav

  • Camera shutter wont open

    I used the camara in my new phone a couple of times, no problem, I got it on Tuesday 7/9, no today I was going to take a pic and the front camera shutter does not open, or it shows a black screem. I have tried face time and the front camera works. I

  • CheckBoxControl

    hi how can i use a check box control to save to the database? i habe 5 check boxes,the user can select either 1 or more and then i have to save to the corresponding table to the database, pls help monali