IPod formatted to UNIX File System (UFS)

Hello all, I had an old 15G iPod (the generation with the 4 button above the wheel) and I decided to reformat it and use it as a firewire drive, it's not used for music anymore. Anyway, I was about half way through the format and I noticed I had accidently sellected UNIX File System (UFS) instead of Mac OS Extended. Well I let lt finish and thought I'd just reformat it again to Mac OS Extended, only now the iPods not even being recognized by OSX? I looked in Disk Utility and nothing there either. iPod restore does nothing either, just says connect an iPod. There's an Apple on the screen and that's it? What gives? Will I be able to get this thing running again?

Thanks but no thanks! I got it going using info at this link:
http://ipodlinux.org/Installation
Got it in Disk Mode!!!

Similar Messages

  • Can i format ipod classic with NTFS file system on win7?

    can i format ipod classic with NTFS file system on win7?

    Not if you want to have an iPod as opposed to an ordinary external drive.
    (60693)

  • UNIX File System?

    I was able in the past to format a disk with UNIX File System and read and write to it using Disk Utility in OS X Tiger. Now, my new Leopard machine does not show that option. What happened to that? How do I get it back?
    Thanks,
    John

    I'm pretty sure UFS support is totally gonzo in leopard. you certainly [can not install Leopard|http://docs.info.apple.com/article.html?artnum=306516] on UFS formatted drives.

  • Uploaded Files stored in Oracle 10G database or in Unix File system

    Hey All,
    I am trying to understand best practices on storing uploaded files. Should you store within the database itself (this is the current method we are using by leveraging BLOB storage) or use a BFILE locator to use the files system storage (we have our DB's on UNIX) . . .or is there another method I should be entertaining? I have read arguments on both sides of this question. I wanted to see what answers forum readers could provide!! I understand there are quite a few factors but the situation I am in is as follows:
    1) Storing text and pdf documents.
    2) File sizes range from a few Kb to up to 15MB in size
    3) uploaded files can be deleted and updated / replaced quite frequently
    Right now we have an Oracle stored procedure that is uploading the files binary data into a BLOB column on our table. We have no real "performance" problems with this method but are entertaining the idea of using the UNIX file system for storage instead of the database.
    Thanks for the insight!!
    Anthony Roeder

    Anthony,
    First word you must learn here in this forum is RESPECT.
    If you require any further explanation, just say so.
    BLOB compared with BFILE
    Security:
    BFILEs are inherently insecure, as insecure as your operating system (OS).
    Features:
    BFILEs are not writable from typical database APIs whereas BLOBs are.
    One of the most important features is that BLOBs can participate in transactions and are recoverable. Not so for BFILEs.
    Performance:
    Roughly the same.
    Upping the size of your buffer cache can make a BIG improvement in BLOB performance.
    BLOBs can be configured to exist in Oracle's cache which should make repeated/multiple reads faster.
    Piece wise/non-sequential access of a BLOB is known to be faster than a that of a BFILE.
    Manageability:
    Only the BFILE locator is stored in an Oracle BACKUP. One needs to do a separate backup to save the OS file that the BFILE locator points to. The BLOB data is backed up along with the rest of the database data.
    Storage:
    The amount of table space required to store file data in a BLOB will be larger than that of the file itself due to LOB index which is the reason for better BLOB performance for piece wise random access of the BLOB value.

  • External Tables to Unix File System 10G R2

    Can anyone help with setting up an external table that reads a flat file from a Unix File system.
    I have sampled a file ok and created an external table and deployed it to the database ok but it can find the link through to the unix file system to read the file.
    I created the location as an FTP type and have referenced the path to the relevant directory /oracle02/app/OWB_files when creating the location. I have placed the relevant named file in the directory but when i try to look at the table in TOAD i get the following errors
    ORA-29913 error in executing ODCIEXTTABLEOPEN callout
    ORA-29400 data cartridge error
    KUP-04040 files 121123_PENS.txt in MLCC_FILES_LOCATION_0 not found
    Does anyone have a step by step guide for creating these and am i doing some thing wrong. Is choosing the FTP type in the location correct and is the path specified correctly. I can see much information on thisin the manual!
    Your assistance would be appreciated

    HI,
    You make sure that, the path should be shared one. We can do this using samba server.
    Regards,
    Gowtham Sen.

  • How to upload pdf file format in unix file directory?

    how to upload pdf file format in unix file directory?

    Hi Karthikeyan,
    If you have PDF internal table with you, You can follow the below method to upload into Unix directory.
        DATA:file_path TYPE char100 VALUE '/usr/sap/tmp/rep_out.PDF'.
        OPEN DATASET file_path FOR OUTPUT IN BINARY MODE.
        IF sy-subrc EQ 0.
          LOOP AT it_pdf.
            TRANSFER it_pdf TO file_path.
          ENDLOOP.
        ENDIF.
        CLOSE DATASET file_path.
    Thanks
    Venkat.O

  • Open API App (Windows based) fails to open FMB on Unix file system

    My Open API, Windows based app, can successfully open and 'get' properties of FMBs stored in the Windows file system. However, it fails to load the FMB when the FMB resides on a networked Unix server. The same FMBs on Unix can be opened by the Windows based FormBuilder (over the network). I can copy the FMB down to Windows and without re-compiling the FMB, my Open API app can 'load' the FMB and 'get' all the properties. What suggestions can you give for debugging / resolving this? I need to be able to 'Load' the FMBs (through the Open API), that reside in the Unix file system, from Windows.
    JJ

    Generally, this is why we will tell you that accessing net shares is not supported and in places where it might even be supported, we would still suggest that it is not recommended. Accessing via net shares (especially through Windows) is often problematic. There are various performance and connectivity issues that, unfortunately fool you into believing that the product you are using is flawed when the problem is really a connection issue with the share.
    In your case, because you are not exactly using an Oracle product (initially), Oracle can't offer much anyway, but I would recommend against using shares whenever possible. If you need to access a file, copy it locally first, perform whatever task on it, then return the updated file to its origin. This method protects you from things like net failure and instability as well as the performance issues associated with accessing files remotely.

  • Download PDF spool to unix file system

    Does anyone know how to programatically download a PDF document spool to the unix file system?
    I am trying to find a method to send PDF documents that have gone to spool to a unix file system. If anyone had any ideas on how to do this, please let me know.
    Thanks.

    Hi,
    For this define a logical file while using transaction FILE.
    In your code you will get the complete path of the file while using FM FILE_GET_NAME.
    Then download your PDF while using instruction
    OPEN DATASET...
    TRANSFERT ...
    Hope this help you .
    Best regards

  • Java script in HTMLDB to check if file exists in Unix file system

    How do I use javascript to check if file is exists in Unix file system. I would like to dispaly the columns only if file is exists.

    Hello,
    This is one of those features that the manuals do not cover.
    How to use and build AJAX features could be a whole book all by itself, and it's not really HTML DB specific feature even though we have built some hooks in application and javascript to make it easier.
    Take a look at this thread
    Netflix: Nice UI ideas
    and I've built some examples here
    http://htmldb.oracle.com/pls/otn/f?p=11933:11
    Or just search the forums for AJAX or XMLHTTP
    Carl

  • Unix file system & Oracle datafiles--urgent plz

    How i can chech my oracle db files on which unix file system? In HP/UX exvirnment??

    select * from dba_data_files
    AUTOEXTENSIBLE column gives you whether autoextend is on or not.
    Join with dba_free_space to get free space for each file.
    You can check the following link
    http://www.oracle.com/technology/oramag/code/tips2003/083103.html

  • Essbase unix file system best practice

    Is there such thing in essbase as storing files in different file system to avoid i/o contention? Like for example in Oracle, it is best practice to store index files and data files indifferent location to avoid i/o contention. If everything in essbase server is stored under one file directory structure as it is now, then the unix team is afraid that there may run into performance issue. Can you please share your thought?
    Thanks

    In an environment with many users (200+) or those with planning apps where users can run large long-running rules I would recommend you separate the application on separate volume groups if possible, each volume group having multiple spindles available.
    The alternative to planning for load up front would be to analyze the load during peak times -- although I've had mixed results in getting the server/disk SME's to assist in these kind of efforts.
    Some more advanced things to worry about is on journaling filesystems where they share a common cache for all disks within a VG.
    Regards,
    -John

  • [Solved] Canot format MicroSD. File system is read-only

    I formatted it for my special needs some time ago using fdisk.
    Don't remember the partition table, but there was 1 fat32 and 1 ext4 partition.
    Now I want to format it fat32 clean, but I get the errors.
    When I open GParted, it says
    GParted wrote:/dev/sdb contains GPT signatures, indicating that it has a GPT table.  However, it does not have a valid fake msdos partition table, as it should.  Perhaps it was corrupted -- possibly by a program that doesn't understand GPT partition tables.  Or perhaps you deleted the GPT table, and are now using an msdos partition table.  Is this a GPT partition table?
    If I click yes - it shows partitions, but creating new partition table fails, as it is read-only.
    If I click no - it does not show up in the device list.
    Here's a picture of the structure of the device.
    What also strange, I remember making 1 GiB ext4, and rest fat32. But it shows the opposite.
    So I tried formatting it with fdisk
    $ fdisk /dev/sdb
    o
    n
    w
    fdisk: cannot write disk label: Bad file descriptor
    I made sure they are unmounted
    umount: /dev/sdb1: not mounted
    umount: /dev/sdb2: not mounted
    Then I checked the filesystems. The second one appears to be read-only
    # fsck /dev/sdb1
    fsck from util-linux 2.23.2
    fsck.fat 3.0.22 (2013-07-19)
    /dev/sdb1: 18 files, 96978/698280 clusters
    # fsck /dev/sdb2
    fsck from util-linux 2.23.2
    e2fsck 1.42.8 (20-Jun-2013)
    fsck.ext4: Read-only file system while trying to open /dev/sdb2
    Disk write-protected; use the -n option to do a read-only
    check of the device.
    # fsck -n /dev/sdb2
    fsck from util-linux 2.23.2
    e2fsck 1.42.8 (20-Jun-2013)
    /dev/sdb2: clean, 860/65664 files, 56633/262144 blocks
    Tried the zero trick, but failed as well.
    # dd if=/dev/zero of=/dev/sdb
    dd: failed to open ‘/dev/sdb’: Read-only file system
    # dd if=/dev/zero of=/dev/sdb2
    dd: failed to open ‘/dev/sdb2’: Read-only file system
    Tried to disable the write-protection
    # hdparm -r0 /dev/sdb
    /dev/sdb:
    setting readonly to 0 (off)
    readonly = 0 (off)
    After that, fdisk still fails to format.
    In Windows 8, it mounts only the fat32 part.
    diskpart can't erase partition table as well.
    Last edited by Doctor Drive (2013-10-09 08:28:13)

    Finally I found the solution
    MicroSD adapter has a "lock" feature.
    When I inserted the adapter, the lock always triggered to locked state.
    So I fixed it in unlocked state using the ducktape.
    http://kb.sandisk.com/app/answers/detai … -or-locked

  • SOlaris 10 file system UFS

    Hello All.
    I'mnot able to mount ufs on optical disk that have 8192 block size if the kernel pagesize is less then 8192. Our driver that manage UDO disk is able to work properly on Sparc with UFS but not on AMD.
    Is there any possibility to use UFS on AMD with 8192 block size ?

    We are trying to implement ufs on 8192 block size
    optical media and we already succed using Solaris
    8/9/10 on sparc architecture where PAGESIZE is 8192.
    We have our private SCSI driver to pilot optical disk
    but no problem, UFS works fine, we simply specify
    fragsize=8192 during newfs.
    If we do the same on AMD where the PAGESIZE is 4192
    we get no error during write but file cannot be read
    back. It seems that UFS failes because the PAGESIZE <
    Physical block size.I see what you're saying. Yes, I think you'll have problems with trying to make UFS work on a system where pagesize < phys block size.
    Sorry for the initial misunderstanding. I have no suggestions for how you might overcome that, or how much code makes that assumption.
    Darren

  • Seems to be a isssue with DPM and a Volume that is formated ReFS (Resilient File System) on Server 2012

    Hi
    Here is the issue(s) I found in using DPM 2012 to Backup Server 2012 with Volumes are formated ReFS. (Short Answser - Looks like a Protection jobs fail if you make a volume on the priotected server that is ReFS, does not matter if its somthing
    your protecting)
    1. When creating Protecting groups for Volumes that are formated ReFS, they are not listed under all Volumes.  All the NTFS volumes show up.
    2.  Hyper-V -- If Virtual serves is on a ReFS Volume, as soon as your done creating a protection group.    You get a Replica inconsitant.   
    3.  If you modify a protection group where everything is working fine. Add a ReFS Volume put a new vm there and then added it to the protecion group it fails.   Every item comes back replica inconstant after you a consitency check.
    What made this so much fun figuring out was I was using dpm 2012 to backup a VM Hosted on a Server 2012.    And and on a NTFS volume it worked fine until I converted one of the volumes to ReFS move my VM there and DPM jobs
    stopped working.
    DPM failed to stop protection for Microsoft Hyper-V \Backup Using Child Partition Snapshot\VMservertest on Someserve.test.test.com. (ID 3116)
    An unexpected error occurred while the job was running. (ID 104 Details: Do not attach the filter to the volume at this time (0x801F000F))
    So after some testing it looks like, If one volume on the target Server is ReFS the protection group fails.   
    Checked technet and I don't see Server 2012 under Supported DPM Agents yet.  
    http://technet.microsoft.com/en-us/library/hh758036
    Has anyone played with the System Center RC Service Pack to see if that fixes the issue.
    Eric (Trying to be truly Wise by leaning form the mistakes of others, my success rate says I need to work harder at that.)

    Hi,
    Please add feedback for that feature and let others vote on it to help prioritize future features.
    http://feedback.azure.com/forums/258995-azure-backup-and-scdpm/filters/new
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread. Regards, Mike J. [MSFT]
    This posting is provided "AS IS" with no warranties, and confers no rights.

  • How to read XML from UNIX file system

    Hi,
    I have this code with read a xml file from windows:
    Document doc = builder.build(new File(C:\\mywork\\src\\Contacts.xml));
    But how can I read the same file from UNIX? (The file is located in "\tmp\Contacts.xml")
    Thanks
    Kenny

    Hi,
    I have this code with read a xml file from windows:
    Document doc = builder.build(new
    File(C:\\mywork\\src\\Contacts.xml));
    But how can I read the same file from UNIX? (The file
    is located in "\tmp\Contacts.xml")
    Thanks
    KennyNo, the file is really located in "/tmp/Contacts.xml". You would writeDocument doc = builder.build(new File("/tmp/Contacts.xml"));Also, your original code would need quotes around the filename too.

Maybe you are looking for

  • Bridge crashed on opening folder with .mpeg

    Does not matter if the Bridge is opened as a standalone or within Premiere or AE it craches as soon as i try to open a folder containing  .mpeg files. The files are captured with Premeire CS4 straight from a HDV camera. Vista 64, Intel i7 940, 12 gig

  • Need advice on laptop CPU and RAM

    Hello, I've just started a Multimedia and Design education, where I will be working with Adobe cc - mostly Photoshop, Illustrator and inDesign. I'm looking for a portable laptop to use while in school, and I have found one with a good IPS display and

  • I WANT TO DELETE TOOL BAR WITH DISTRACTING ADVERTISING

    I DON'T LIKE THE DISTRACTIONS ON THE TOOL BAR

  • Smart forms. How to setup logic for standard text usage.

    Hi All! Could you advise please is there a way, except abap programming, to setup in smart forms that different standard texts values (tcode so10) to be used based on different selected parameters (e.g. customers, vendors etc... ). e.g. For picking l

  • Where do I find the Hibernateconsole ?

    Hi, I typed the following orm setting in my Application.cfc <cfset  this.ormsettings={cfclocation="C:\ColdFusion9\wwwroot\2009\cfcs\qrys",logsql = "true"}> The ORM SQL queries should be logged now. But where do I find the HibernateConsole ? It should