CD Joliet File System (long filenames)

On the Solaris 8 platform on Intel, I have the following problem. When I mount CDs that were written using the Joliet (Windows long filenames) file system, I can only see the 8.3 short filenames. On other unix systems (linux,sco), I can see the long filenames. Is there a way to enable the long filenames ?

Solaris has no support for the "joliet" extension to iso9660. But
for many years, it has support for "rock-ridge" extensions. Rock-Ridge
gives you long filenames (>8.3 and >64 (!)), symlinks, user & group IDs,
etc, ...
Please use "rock-ridge" extensions when you make a CD for unix systems.
If you burn your own CDR, you can use the "mkisofs" command to
generate iso9660 filesystems with both "joliet" and "rock-ridge"
(mkisofs options -J -r/-R)

Similar Messages

  • Install CD-R of SOLARIS 8 / long filenames / ISO

    Hi all,
    First, a precision. I'm french and i don't know if i can explain my
    problem, 'cause my english ;((
    I've been burning solaris 8 apps to CD, with Easy CD Creator.
    When I want to access the install CD under Solaris, the filename appear as
    DOS names i.e. with 8 caracters and cut.
    So, my SUN Sparc 10 write :
    " bad magic number in disk label"
    "can't open disk label package"
    1/ The problem is it because the filename is too short and cut ?
    => if yes, Is there any way to preserve the long filenames when viewing CD
    content?
    => I told Easy CD Creator to not generate Joliet file system but ISO9660,
    but it do not allow long filenames under ISO9660 !! ;(((
    The resulting CD's are what i explain to you.
    The problem is it due to bad filenames ?
    2/ So, if that is the reason of my difficulty, How can i do to solve it ?
    Thanks for your help.
    Laurent
    [email protected]

    The solaris 8 CDs you've downloaded contains "ISO9660" images already,
    so if you're using options like "don't create joliet filenames" in
    your CDR software you're obviously generating another ISO9660
    filesystem layer around the solaris 8 ISO9660 image. That's wrong.
    Your supposed to write these bits directly to a CDR, without creating
    a new file tree structure. Check the documentation for your CDR
    software how to write a "raw cd image" to a CDR.
    Btw. the
    Solaris 8 download page contains some explicit instructions
    for burning the solaris 8 cds with "Easy CD Creator". In step 7. the
    page says:
    <pre>
    For example, if you use Easy CD Creator, add the extension .iso
    to the file name. Next, from the File Menu in the program, select
    Create CD from CD image. Then, select that image file.
    </pre>

  • Insufficient perms/long filename errors

    I formatted a WD external HDD to have an HFS+(for backup using SuperDuper) and 2 FAT partitions. I had been using the hard drive without any errors, and did a reformat of OS X recently. Then, I wiped the three partitions on external HDD. When trying to copy files that I downloaded a day or two ago (after OS X reformat), from OS X's downloads folder to the HFS partition, it shows an error that some files have long filenames and cannot be copied. Some other files show error of insufficient permission.
    Can it be fixed?

    I was trying to copy files downloaded to my Macbook HD to an non-journaled HFS+ partition and it worked. However, when I tried to copy the same files from the external HDD to a different folder on Macbook HD, it shows an error about insufficient permissions.
    For the partition in external drive, I've checked "ignore permissions" and my username shows up under the list of owners. I also did a repair/repair permissions from Disk utility in the OS X install DVD. The DVD and Macbook contains the same OS X version (10.5.5), so I don't think an out-of-date Disk Utility messed up things.
    I'd run the external HDD and Macbook with 10.5.5, without any problems before reinstalling OS X and problems cropped up.

  • JSP XML output to file system

    I hava a JSP application that uses the XDK to query the database and display XML rowset data to a JSP. This rowset data can be a few hundred rows or up to 50,000 rows,
    as we publish technical manuals, parts manuals and wiring manuals for aircraft.
    Getting the data to display within the JSP is not a problem. The problem is
    that the XML rowset data shares two purposes:
    1. Query by Example, returning
    XML formatted data for display within a browser.
    2. Use of the XML formatted
    data as an external entity file for our SGML authoring software to print the
    paper publication.
    The JSP I developed (I'm not sure I should take credit for it) uses the XMLData bean and it will open the browser File/SaveAs dialog in both Netscape (Unix) and IE5.5 (Windows). HOWEVER, when I create a second JSP to perform the same task on a different table the JSP errorpage shows with no error and no errors within the JDeveloper console. This has been kicking my butt for some time, and would very much appreciate a workable solution. I will share whatever I can share with the provider. Here is the JSP code.
    =============================================
    <?xml version="1.0"?>
    <%@ page language = "java" errorPage="errorpage.jsp" import = "java.io.*, java.util.*, oracle.jbo.*, javax.naming.*, oracle.jdeveloper.html.*, oracle.jbo.html.databeans.*" contentType="text/html;charset=ISO-8859-1" %>
    <jsp:useBean class="oracle.jbo.html.databeans.XmlData" id="h25EqXml" scope="request" >
    <%
    h25EqXml.setResultElement("equiptbl");
    h25EqXml.setReleaseApplicationResources(true);
    h25EqXml.setDisplayAttributes("Ein,Pnr,Nomen,Cage,Diagnbr,Sht,Effect");
    h25EqXml.initialize(pageContext,"h25Wires_JspApp_h25Wires_be_H25Wires_beModule.HU25A_Equipment");
    response.setContentType("text/sgml");
    response.setHeader("Content-location","attachment;filename="+"equiptbl.sgm");
    ServletOutputStream sos = response.getOutputStream();
    BufferedInputStream bis = new BufferedInputStream (new FileInputStream("equiptbl.sgm"));
    int data;
    while((data = bis.read()) != -1)
    sos.write(data);
    bis.close();
    sos.flush();
    sos.close();
    %>
    <jsp:forward page= "HU5A_Equipment_Browse.jsp" />
    </jsp:useBean>
    ============================================
    The code for the other pages is the same with the exception of the filename xxx.sgm and the bean id.
    I very much appreciate your help. Thanks!
    null

    Hello Paulo,
    thank you for your answer, but i'm using type 04 XML for SEPA.
    when i select Type XML, the fields "output to file system" and "filename" are hidden.
    i dont understand why ...

  • HowTo? Prevent DngConverter from truncating long filenames? Bug?

    I've noticed that when using DngConverter 8.3 in Windows-7-64 with .cr2 RAW files with long filenames, the end of the .dng filename can be truncated
    Here are some examples (cr2 before and dng after)
    ok:
    140306a_198950_DeckStarTrails_6d_35mm_Iso_3200_Spd_30_Ap_f2.CR2                 
    140306a_198950_DeckStarTrails_6d_35mm_Iso_3200_Spd_30_Ap_f2.dng
    Truncated:
    140306a_198951_DeckStarTrails_6d_35mm_Htp_Iso_200_Spd_121_Ap_f8.CR2
    140306a_198951_DeckStarTrails_6d_35mm_Htp_Iso_200_Spd_121_A.dng
    140306a_198965_DeckStarTrails_6d_35mm_Htp_Iso_200_Spd_61_Ap_f3-2.CR2
    140306a_198965_DeckStarTrails_6d_35mm_Htp_Iso_200_Spd_61_Ap.dng
    140306a_199079_DeckStarTrails_5d2_17mm_Iso_400_Spd_30_Ap_f5-6.CR2
    140306a_199079_DeckStarTrails_5d2_17mm_Iso_400_Spd_30_Ap_f5.dng
    140306a_199133_DeckStarTrails_5d2_17mm_DarkFrame_Htp_Iso_200_Spd_120_A p_f5.CR2
    140306a_199133_DeckStarTrails_5d2_17mm_DarkFrame_Htp_Iso_20.dng
    There seems to be a limit of 59 characters. I used Bridge to embed Metadata in the filename.
    Am I doing something wrong? Is this "as expected"? Or could this be a possible bug?
    Could this be because the total filename (with sub-directory name) was excessive? Here's the sub-directory:
    S:\_WorkInProgress\_RAW\2014
    A workaround is to use shorter filenames, but a limit of 59 chars seems rather restrictive.

    You may want to post in the DNG  forum:
    http://forums.adobe.com/community/dng?view=discussions

  • Framemaker uses $filename for short file name, can we edit this to change appearance? We do not want the short file name of long filename to include the .fm extension can this be removed or modified to make this happen?

    Framemaker uses <$filename> for short file name, can we edit this to change appearance? We do not want the short file name of long filename to include the .fm extension can this be removed or modified to make this happen? In compiling our books it would be helpful to not have this extension appear as it then requires us to create extra files without them.

    See: System Variables

  • Reloading iphoto library and streamlining iphoto file system.

    I had a hard-disk failure recently, and lost a lot of photos: basically all the full size files. Only the thumbnails remained on the system disk, so when I had the new drive installed, the engineer made these the iphoto library.
    Fortunately, I had most of my photos backed up on a separate harddrive (in the 'originals' folder). What is the best way of putting these into my iphoto library in place of the thumbnails already there? (ie, don't want to see both in iphoto).
    I'd put the iPhoto library on the separate hard-disk, but unbeknownst to me, about a year ago, iPhoto began saving imports on the system drive again, so these were the pictures I lost.
    On another point, since I installed iphoto 6, the location of photos has become confusing, with the whole photo library pre-iphoto 6 duplicated in a folders called 'Originals' and 'Data'. Neither of these maintain the original easily navigable and logical date hierarchy file structure, but have saved all photos in a folder marked 2007, by roll number. The date folders which remain only contain a 'data' folder containing .attr files. Not very practical.
    In short the whole thing is a mess: for some photos the system has saved 5 copies in various places.
    Is there anything that can be done to consolidate the iPhoto library and put it back on a date-based filing system?

    raggabishp
    To start at the end:
    Is there anything that can be done ... put it back on a date-based filing system?
    No, once you use iPhoto 6, that's the file system. But it's not that difficult to follow: A Note about the iPhoto Library Folder:
    In this folder there are various files, which are the Library itself and some ancillary files. Then you have three folders
    Originals are the photos as they were downloaded from your camera or scanner.
    (ii) Modified contains edited pics, shots that you have cropped, rotated or changed in any way.
    This allows the Photos -> Revert to Original command - very useful if you don't like the changes you've made.
    (iii) Data holds the thumbnails the the app needs to show you the photos in the iPhoto Window.
    Finding the Picture file is easy: There are three ways (at least) to get files from the iPhoto Window.
    1. Drag and Drop: Drag a photo from the iPhoto Window to the desktop, there iPhoto will make a full-sized copy of the pic.
    2. File -> Export: Select the files in the iPhoto Window and go File -> Export. The dialogue will give you various options, including altering the format, naming the files and changing the size. Again, producing a copy.
    3. Show File: Right- (or Control-) Click on a pic and in the resulting dialogue choose 'Show File'. A Finder window will pop open with the file already selected.
    Rolls in the iPhoto Window correspond exactly with the Roll Folders in the Originals Folder in the iPhoto Library Folder. You can move photos between Rolls, you can rename rolls, edit them, create them, as long as you do it via the iPhoto Window. Check out the Info Pane (wee 'i', lower left) the name and date fields are editable. Edit a Roll Name using the Info Pane, the Roll Folder in iPhoto Library Folder/Originals will also have the new name.
    So the structure is different, but - especially if you use the Film Rolls view - very straightforward.
    There is no easy way to rid yourself of these 'thumbs become originals'. If the thumbs have the same filenames as the Originals, then you could overwrite the thumbs with the Originals, but you would need to do that on a file by file basis, I'm afraid.
    Other than that, compare the full size pics with the thumbs and trash the duplicates is all I can suggest.
    Regards
    TD

  • Dfc: Display file system space usage using graph and colors

    Hi all,
    I wrote a little tool, somewhat similar to df(1) which I named dfc.
    To present it, nothing better than a screenshot (because of colors):
    And there is a few options available (as of version 3.0.0):
    Usage: dfc [OPTIONS(S)] [-c WHEN] [-e FORMAT] [-p FSNAME] [-q SORTBY] [-t FSTYPE]
    [-u UNIT]
    Available options:
    -a print all mounted filesystem
    -b do not show the graph bar
    -c choose color mode. Read the manpage
    for details
    -d show used size
    -e export to specified format. Read the manpage
    for details
    -f disable auto-adjust mode (force display)
    -h print this message
    -i info about inodes
    -l only show information about locally mounted
    file systems
    -m use metric (SI unit)
    -n do not print header
    -o show mount flags
    -p filter by file system name. Read the manpage
    for details
    -q sort the output. Read the manpage
    for details
    -s sum the total usage
    -t filter by file system type. Read the manpage
    for details
    -T show filesystem type
    -u choose the unit in which
    to show the values. Read the manpage
    for details
    -v print program version
    -w use a wider bar
    -W wide filename (un truncate)
    If you find it interesting, you may install it from the AUR: http://aur.archlinux.org/packages.php?ID=57770
    (it is also available on the archlinuxfr repository for those who have it enabled).
    For further explanations, there is a manpage or the wiki on the official website.
    Here is the official website: http://projects.gw-computing.net/projects/dfc
    If you encounter a bug (or several!), it would be nice to inform me. If you wish a new feature to be implemented, you can always ask me by sending me an email (you can find my email address in the manpage or on the official website).
    Cheers,
    Rolinh
    Last edited by Rolinh (2012-05-31 00:36:48)

    bencahill wrote:There were the decently major changes (e.g. -t changing from 'don't show type' to 'filter by type'), but I suppose this is to be expected from such young software.
    I know I changed the options a lot with 2.1.0 release. I thought it would be better to have -t for filtering and -T for printing the file system type so someone using the original df would not be surprised.
    I'm sorry for the inconvenience. There should not be any changes like this one in the future though but I thought it was needed (especially because of the unit options).
    bencahill wrote:
    Anyway, I now cannot find any way of having colored output showing only some mounts (that aren't all the same type), without modifying the code.
    Two suggestions:
    1. Introduce a --color option like ls and grep (--color=WHEN, where WHEN is always,never,auto)
    Ok, I'll implement this one for 2.2.0 release It'll be more like "-c always", "-c never" and "-c auto" (default) because I do not use long options but I think this would be OK, right?
    bencahill wrote:2. Change -t to be able to filter multiple types (-t ext4,ext3,etc), and support negative matching (! -t tmpfs,devtmpfs,etc)
    This was already planned for 2.2.0 release
    bencahill wrote:Both of these would be awesome, if you have time. I've simply reverted for now.
    This is what I would have suggested.
    bencahill wrote:By the way, awesome software.
    Thanks I'm glad you like it!
    bencahill wrote:P.S. I'd already written this up before I noticed the part in your post about sending feature requests to your email. I decided to post it anyway, as I figured others could benefit from your answer as well. Please forgive me if this is not acceptable.
    This is perfectly fine Moreover, I seem to have some troubles with my e-mail addressee... So it's actually better that you posted your requests here!

  • SAP GoLive : File System Response Times and Online Redologs design

    Hello,
    A SAP Going Live Verification session has just been performed on our SAP Production environnement.
    SAP ECC6
    Oracle 10.2.0.2
    Solaris 10
    As usual, we received database configuration instructions, but I'm a little bit skeptical about two of them :
    1/
    We have been told that our file system read response times "do not meet the standard requirements"
    The following datafile has ben considered having a too high average read time per block.
    File name -Blocks read  -  Avg. read time (ms)  -Total read time per datafile (ms)
    /oracle/PMA/sapdata5/sr3700_10/sr3700.data10          67534                         23                               1553282
    I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    2/
    We have been asked  to increase the size of the online redo logs which are already quite large (54Mb).
    Actually we have BW loading that generates "Chekpoint not comlete" message every night.
    I've read in sap note 79341 that :
    "The disadvantage of big redo log files is the lower checkpoint frequency and the longer time Oracle needs for an instance recovery."
    Frankly, I have problems undertanding this sentence.
    Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right ?
    But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    Thank you.
    Any useful help would be appreciated.

    Hello
    >> I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    The recommended ("standard") values are published at the end of sapnote #322896.
    23 ms seems really a little bit high to me - for example we have round about 4 to 6 ms on our productive system (with SAN storage).
    >> Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right?
    Correct.
    >> But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    A checkpoint is occured on every logswitch (of the online redologfiles). On a checkpoint event the following 3 things are happening in an oracle database:
    Every dirty block in the buffer cache is written down to the datafiles
    The latest SCN is written (updated) into the datafile header
    The latest SCN is also written to the controlfiles
    If your redologfiles are larger ... checkpoints are not happening so often and in this case the dirty buffers are not written down to the datafiles (in the case of no free space in the buffer cache is needed). So if your instance crashes you need to apply more redologs to the datafiles to be in a consistent state (roll forward). If you have smaller redologfiles more log switches are occured and so the SCNs in the data file headers (and the corresponding data) are closer to the newest SCN -> ergo the recovery is faster.
    But this concept does not really fit the reality because of oracle implements some algorithm to reduce the workload for the DBWR in the case of a checkpoint.
    There are also several parameters (depends on the oracle version) which control that a required recovery time is kept. (for example FAST_START_MTTR_TARGET)
    Regards
    Stefan

  • Verication file system failed to make partition

    I can't make a partition verification file system failed

    Hello,
    In Disk Utility, try Repair Disk first, you need another boot drive or the Install Disc if this is trying on your Boot Drive.
    "Try Disk Utility
    1. Insert the Mac OS X Install disc, then restart the computer while holding the C key.
    2. When your computer finishes starting up from the disc, choose Disk Utility from the Installer menu at top of the screen. (In Mac OS X 10.4 or later, you must select your language first.)
    *Important: Do not click Continue in the first screen of the Installer. If you do, you must restart from the disc again to access Disk Utility.*
    3. Click the First Aid tab.
    4. Select your Mac OS X volume.
    5. Click Repair Disk, (not Repair Permissions). Disk Utility checks and repairs the disk."
    http://docs.info.apple.com/article.html?artnum=106214
    Then try a Safe Boot, (holding Shift key down at bootup), run Disk Utility in Applications>Utilities, then highlight your drive, click on Repair Permissions, reboot when it completes.
    (Safe boot may stay on the gray radian for a long time, let it go, it's trying to repair the Hard Drive.)
    If perchance you can't find your install Disc, at least try it from the Safe Boot part onward.

  • Crystal Reports XI String [255] limit with the File System Data driver...

    I was trying to create a Crystal Reports XI report to return security permissions of files and folders.  I have been able to successfully connect and return data using the File System Data driver as the Data Source; however the String limit on the ACL NT Security Field is 255 characters.  The full string of data to be returned can be much longer than the 255 limit and I cannot find how to manipulate that parameter. 
    I am currently on Crystals XI and Crystal XI R2 and have applied the latest service packs but still see the issue.  My Crystal Reports Database DLL for File System data ( crdb_FileSystem.dll ) is at Product Version 11.5.10.1263.
    Is it possible to change string limits when using the File System Data driver as the Data Source?  If so, how can that be accomplished.  If not, is there another method to retrieve information with the Windows File System Data being the Data Source?  Meaning, could I reach my end game objective of reporting on the Windows ACL's with Crystal through another method?

    Hello,
    This is a known issue. Early versions you could not create folder structures longer than 255 characters. With the updates to the various OS's this is now possible but CR did not allocate the same space required.
    It's been tracked as an enhancement - ADAPT01174519 but set for a future release.
    There are likely other ways of getting the info and then putting it into an Excel file format and using that as the data source.
    I did a Google search and found this option: http://www.tomshardware.com/forum/16772-45-display-explorer-folders-tree-structure-export-excel
    There are tools out there to do this kind of thing....
    Thank you
    Don
    Note the reference to msls.exe appears to be a trojan: http://www.greatis.com/appdata/d/m/msls.exe.htm so don't install it.
    Edited by: Don Williams on Mar 19, 2010 8:45 AM

  • Store \ Retrieve files from file system

    Hi to all!
    I would like to implement a solution for storing files uploaded via apex user interface to servers file system. As well I would like this files to be retrievable by apex users. I designed the following solution:
    For upload:
    1. Through file browse item user chooses file to be uploaded
    2. File goes to custom table (as BLOB)
    -- so far i would use apex Upload\Download files tutorial
    3. File(BLOB) would then have to be written to file system to some directory and file id would have to be written to some db table which holds pointers to files on file system
    4. delete file(blob) from custom table (from step 2)
    For download:
    1. user chooses link from some report region(based on table giving file pointers to files residing on file system)
    2. file identified with chosen file pointer is then inserted into blob column of some custom table in db
    3. from custom table with download procedure fie is finally presented to user
    4. delete file(blob) from custom table (from step 2)
    Using apex tutorial for Upload\Download files it is straitforward to get the files from db table or into db table using blobs. But i have not seen any example of using BFILE or migrating files from db to file system and vice versa.
    So some Q arise:
    a) How can I implement step 3 under For upload section above
    b) How can I implement step 2 under For download section above
    c) Is there any way to directly upload file to file system via apex user interface or to directly download file from file system via some report region link column?
    Please help!!!
    Regards Marinero
    Message was edited by:
    marinero

    marinero,
    Here is a procedure that will copy an uploaded file to the file system:
      Procedure BLOB_TO_FILE(p_file_name In Varchar2) Is
        l_out_file    UTL_FILE.file_type;
        l_buffer      Raw(32767);
        l_amount      Binary_Integer := 32767;
        l_pos         Integer := 1;
        l_blob_len    Integer;
        p_data        Blob;
        file_name  Varchar2(256);
      Begin
        For rec In (Select ID
                              From HTMLDB_APPLICATION_FILES
                             Where Name = p_file_name)
        Loop
            Select BLOB_CONTENT, filename Into p_data, file_name From HTMLDB_APPLICATION_FILES Where ID = rec.ID;
            l_blob_len := DBMS_LOB.getlength(p_data);
            l_out_file := UTL_FILE.fopen('UPDOWNFILES_DIR', file_name, 'wb', 32767);
            While l_pos < l_blob_len
            Loop
              DBMS_LOB.Read(p_data, l_amount, l_pos, l_buffer);
              If l_buffer Is Not Null Then
                UTL_FILE.put_raw(l_out_file, l_buffer, True);
              End If;
              l_pos := l_pos + l_amount;
            End Loop;
            UTL_FILE.fclose(l_out_file);
        End Loop;         
      Exception
        When Others Then
          If UTL_FILE.is_open(l_out_file) Then
            UTL_FILE.fclose(l_out_file);
          End If;
      end; And here is a procedure that will download a file directly from the file system:
      Procedure download_my_file(p_file In Number) As
        v_length    Number;
        v_file_name Varchar2(2000);
        Lob_loc     Bfile;
      Begin
        Select file_name
          Into v_file_name
          From UpDownFiles F
         Where File_id = p_file;
        Lob_loc  := bfilename('UPDOWNFILES_DIR', v_file_name);
        v_length := dbms_lob.getlength(Lob_loc);
        owa_util.mime_header('application/octet', False);
        htp.p('Content-length: ' || v_length);
        htp.p('Content-Disposition: attachment; filename="' || SUBSTR(v_file_name, INSTR(v_file_name, '/') + 1) || '"');
        owa_util.http_header_close;
        wpg_docload.download_file(Lob_loc);
      End download_my_file;I could put a sample application on apex.oracle.com, but it wouldn't be able to access the file system on that server.

  • How to insert a JPG file from file system to Oracle 10g?

    I have developed a schema to store photos as BLOB which store the text description as CLOB original filename, file size.
    I also use ctxsys.context to index TEXT_DESCRIPTION in order to perform Oracle Text Search and it works.
    I would like to insert some JPG file from say C:\MYPHOTO\Photo1.jpg as a new record. How can I do this in SQL PLus and/or Loader?
    How can I retrieve the PHOTO_IMAGE back to the file system using SQL Plus and/or command line in DOS?
    See the following script:
    create user myphoto identified by myphoto;
    grant connect, resource, ctxapp to myphoto;
    connect myphoto/myphoto@orcl;
    PROMPT Creating Table PHOTOS
    CREATE TABLE PHOTOS
    (PHOTO_ID VARCHAR2(15) NOT NULL,
    PHOTO_IMAGE BLOB,
    TEXT_DESCRIPTION CLOB,
    FILENAME VARCHAR2(50),
    FILE_SIZE NUMBER NOT NULL,
    CONSTRAINT PK_PHOTOS PRIMARY KEY (PHOTO_ID)
    create index idx_photos_text_desc on
    PHOTOS(TEXT_DESCRIPTION) indextype is ctxsys.context;
    INSERT INTO PHOTOS VALUES
    ('P00000000000001', empty_blob(), empty_clob(),
    'SCGP1.JPG',100);
    INSERT INTO PHOTOS VALUES
    ('P00000000000002', empty_blob(), 'Cold Play with me at the concert in Melbourne 2005',
    'COLDPLAY1.JPG',200);
    INSERT INTO PHOTOS VALUES
    ('P00000000000003', empty_blob(), 'My parents in Melbourne 2001',
    'COLDPLAY1.JPG',200);
    EXEC CTX_DDL.SYNC_INDEX('idx_photos_text_desc');
    SELECT PHOTO_ID ,TEXT_DESCRIPTION
    FROM PHOTOS;
    SELECT score(1),PHOTO_ID ,TEXT_DESCRIPTION
    FROM PHOTOS
    WHERE CONTAINS(TEXT_DESCRIPTION,'parents',1)> 0
    ORDER BY score(1) DESC;
    SELECT score(1),PHOTO_ID ,TEXT_DESCRIPTION
    FROM PHOTOS
    WHERE CONTAINS(TEXT_DESCRIPTION,'cold play',1)> 0
    ORDER BY score(1) DESC;
    SELECT score(1),score(2), PHOTO_ID ,TEXT_DESCRIPTION
    FROM photos
    WHERE CONTAINS(TEXT_DESCRIPTION,'Melbourne',1)> 0
    AND CONTAINS(TEXT_DESCRIPTION,'2005',2)> 0
    ORDER BY score(1) DESC;

    Hi
    You can use the following to insert an image:
    create table imagetab(id number primary key,imagfile blob, fcol varchar2(10));
    create or replace directory imagefiles as 'c:\'
    declare
        v_bfile BFILE;
        v_blob  BLOB;
      begin
        insert into imagetab (id,imagfile,fcol)
        values (3,empty_blob(),'BINARY')
        return imagfile into v_blob;
        v_bfile := BFILENAME ('IMAGEFILES', 'MyImage.JPG');
        Dbms_Lob.fileopen (v_bfile, Dbms_Lob.File_Readonly);
        Dbms_Lob.Loadfromfile (v_blob, v_bfile, Dbms_Lob.Getlength(v_bfile));
        Dbms_Lob.Fileclose(v_bfile);
        commit;
      end;
    /

  • Unable to get the file system information for: \\****servername\E$\; error = 64 Unable to distribute content to DP

    One of our DPs has stopped loading content. 
    I've research for quite a bit and cannot find a clear cut reason to this.  This server only has a DP role, I verified sharing permissions, all looked good. This DP has been running just fine for the last year or so and all sudden it will no longer load
    packages.  The disk drive is still present I can still reach the hidden share \\servername.com\E$
    Verified that the SMSSIG$ folder is there and the last entry is from 4/23/2015 
    SCCM 2012 R2 
    OS 2008 R2 Standard
    Any help is greatly appreciated!
    Here's a snipit from the distmgr.log
    Start updating the package on server ["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\...
    Attempting to add or update a package on a distribution point.
    Will wait for 1 threads to end.
    Thread Handle = 0000000000001E48
    STATMSG: ID=2342 SEV=I LEV=M SOURCE="SMS Server" COMP="SMS_DISTRIBUTION_MANAGER" SYS=***.com SITE=1AB PID=2472 TID=8252 GMTDATE=Thu Apr 30 19:12:01.972 2015 ISTR0="SYSMGMT Source" ISTR1="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    ISTR2="" ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=2 AID0=400 AVAL0="CAS00087" AID1=404 AVAL1="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    SMS_DISTRIBUTION_MANAGER 4/30/2015 2:12:01 PM
    8252 (0x203C)
    The current user context will be used for connecting to ["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\.
    Successfully made a network connection to \\*****.com\ADMIN$.
    Ignoring drive \\*****.com\C$\.  File \\*****.com\C$\NO_SMS_ON_DRIVE.SMS exists.
    Unable to get the file system information for: \\*****.com\E$\; error = 64.
    Failed to find a valid drive on the distribution point ["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\
    Cannot find or create the signature share.
    STATMSG: ID=2324 SEV=E LEV=M SOURCE="SMS Server" COMP="SMS_DISTRIBUTION_MANAGER" SYS=sccmprdpr1sec2.mmm.com SITE=1AB PID=2472 TID=8252 GMTDATE=Thu Apr 30 19:12:55.206 2015 ISTR0="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    ISTR1="CAS00087" ISTR2="" ISTR3="30" ISTR4="94" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=2 AID0=400 AVAL0="CAS00087" AID1=404 AVAL1="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    Error occurred. Performing error cleanup prior to returning.
    Cancelling network connection to \\*****.com\ADMIN$.

    Error 64 is being returned which is simply "the network name is no longer available".
    There can be a number of reasons for this from SMB compatibility issues (2003 servers wont support SMB2), to the expected and actual computer name of the boxes don't match (tries to authenticate with server.tld.com when the actual name is srv-01.tld.com and
    you just put a C-name in). I'd start from the top:  Try opening said share from the Primary Site server as that's the box doing the work.  Verify the IP and computer name is legit and that no one has played ACL games between the two systems (remember
    RPC only initiates/listens on port 135 but established connections are up in the dynamic port range).
    At the end of the day it's an issues "underneath" SCCM, and not an SCCM problem specifically. 

  • .XMP files no longer appear as sidecars--how to clear metadata?

    Photoshop CS2 ver 9.0.2
    Bridge ver 1.0.4.6
    ACR ver 3.7
    eMac G4 800 MHz
    Mac OS X ver 10.4.9
    1 Meg RAM
    12 GB available disk space
    Bridge and Photoshop are set to save metadata in .XMP sidecar files, but sidecar files no longer appear. That is, up till a few days ago, .XMP files appeared when I opened raw files, but no longer. The drop-down menu is still set to save metadata in sidecar files.
    Cannot clear metadata in Bridge. I can select "Camera Raw Defaults" in ACR, but that doesn't clear cropping.
    1) How do I return already-stored metadata to .XMP files?
    2) How do I get PSCS2 to store future metadata as .XMP files?
    3) How do I clear metadata currently stored in ACR/PS database?
    4) What did I do to cause system to store metadata in ACR/PS database despite being set to save in sidecar files?
    Thanks.
    --HC

    Noel,
    Excellent point.  I have no idea.  Maybe this was only working for
    .PSDs/etc. (which contain the metadata embedded in the file itself[,
    right?]).  I may have been deluding myself this whole time.  It's
    possible that it never really worked the way I thought it did to begin with.
    But, more specifically, here's the current question:  If I have this
    file (which is a backup from long ago):
    J:\backup of M90\Documents and Settings\philipt\Application
    Data\Adobe\CameraRaw\Database
    ...assuming that drives are currently mapped (by drive letter) the same
    way they were when that backup was made), can I just replace the current
    file:
    C\Documents and Settings\philipt\Application Data\Adobe\CameraRaw\Database
    ...with the one from J: and expect it to work correctly?
    What's the worst that can happen if I do this?
    BTW, the file on J: is 2,569 Kb; the (new) one on C: is 2,754 Kb.  This
    doesn't seem to bode well, since the old one (J:) should have much, much
    more data in it (assuming that the "Database" file per se is where the
    metadata is stored): the older file has complex,
    multi-tiered data for 10s of 1000s of images.  "??"
    Thanks again for your willingness to help with this!!
    Aloha,
    -pt

Maybe you are looking for

  • I need assistance. PLEASE. Help me for the love of God HELP.

    I'm trying to update my 2gb Ipod nano with the updater. The program freezes when it trys to initialize it. I put it in Hard Disk mode and its the same thing. I need help.

  • IMessage delays. Up to 15 minutes.

    iPhone 5 user. Latest updates. Basically. incoming iMessages can be delayed by up to 15 minutes. Sometimes I'll get a bulk of 5-10 messages suddenly appear on my phone. Sometimes without a 'Text Tone' alert.. Regardless of being at home or out and ab

  • IPad Mini Retina

    Will the iPad Mini Retina 16GB and 32GB model prices reflect the same as Apple's?

  • How can I convert imovie into photo's

    How can I convert imovie into photo's

  • HTTP Send

    I'm trying to send a SOAP message from XI via HTTP adapter to an external web services client URL and I get an 'HTTP_TIMEOUT' message. The trace says 'HTTP-Client: exception during receive: HTTP_COMMUNICATION_FAILURE'. I can take the message and send