File structure validation in PLSQL

Hello All,
I need few clarification about this if i want to validate the file structure how can i do in plsql, below is the example where my files looks like.
IIORRIDGE
VA1601800700028120000000008+0000123822092013
VA1601800700029040000000008+0000323822092013
VA1601800700030860000000008+0000123822092013
VA1601800767943220000000008+0000123822092013
end;
Please help me
Mahesh

I need few clarification about this if i want to validate the file structure how can i do in plsql
Foe example like this line
VA1601800700028120000000008+0000123822092013
There are totally 44 characters if its more than 44 i want to alert it and if it is less than 44 also i want  to alert it
Like ablve line there will be more than 100000 lines will be der.
The problem of 'validate the file structure' implies that there could be issues with the structure of the file or the records.
And therein lies the problem if you try to use an external table since a table, or external table, is generally expected to have a known number of columns of known datatypes.
So first you need to document the business rules that you want to implement:
1. What are ALL of the validation rules that you need to check for?
  a. what record delimiters are expected (e.g. LF, CRLF, etc)?
  b. do you need to validate that the proper delimiters have been used?
  c. how are the fields of a record delimited? Is it a delimited file (by commas or other?), fixed-width, etc.
  d. what is the field delimiter (if delimited)?
  e. what are the datatypes for each field of the record
  f. can fields be NULL? If so, how is a null field indicated?
  g. do all records have to have the same number of fields? Or is TRAILING NULLCOLS allowed?
2. How do you want to handle all of the errors that you detect?
  a. reject the entire file if there is an error in ANY record?
  b. reject only the records that have an error but accept the ones that do not have errors?
3. How do you plan to allow users access to the problem records?
  a. give them access to the actual BAD file containing the raw data?
  b. give them access to a DB table containing the raw data?
4. How do you plan to reprocess ONLY the bad data once it has been corrected?
  a. reload the entire file?
  b. reload a new file containing the fixed records?
5. How do you plan to allow users to access the validated, good data?
  a. give them access to a DB table?
Sometimes the 'best' solution is to NOT load data into Oracle that isn't already validated.
You could write a standalone Java application to perform this 'file validation' and either produce 'clean' files that you need load into Oracle or the app could validate a file and load the clean data at the same time.
But until you document ALL of the requirements in terms of handling the 'dirty data' it is too early to be talking about the 'best' solution.

Similar Messages

  • Import Manager Usage : Approaches for developing Import file structure and text validations

    Hi Experts,
    We are having 50+ import maps. We have provided option to users to drop the files for data import. Currently Import Manager(7.1 Sp08) does not have capability of pre-import validation of file data such as
    a. file structure - number of columns specific to import map
    b. file text validations - special characters, empty lines, empty cells
    c. Uniqueness of the records in the file
    For this, we are planning to build temporary folder(port specific) in which user drops in the file. We use the custom development to do above mentioned validations and then move the files to actual import ports.
    Please let us know if you have worked on similar requirements and how you have fulfilled such requirements.
    Regards,
    Ganga

    Hi Ganga,
    Assuming you have a well defined xsd and are getting valid xmls from source in the Inbound Port of MDM. Also,you have a Primary key in form of External ID (say).
    So just by making and defining a XSD you get most of what you want in your questions a and b.
    Now if you wish to use PI to drop files in the inbound port then you can build all the validations in PI itself and you would not need Staging table.
    Otherwise,you can have another table (preferably Main table) in the same repository or other dummy repository where records are created on import based on External ID.
    Here you can launch an MDM workflow on import of these records and run assignments to replace unwanted characters and Validations to give error for rejecting some records based on the data quality level desired.Once unwanted characters are removed and data is validated it can be syndicated using a syndication step in the Workflow.So records which fail are not sent and which pass are sent to a outbound port.
    From the outbound port PI or some other job can pick the file from this outbound folder and drop to Inbound folder of the same repository which imports to the required Primary Main table.Here again you have the option to leverage validations in PI and further check if data is fine.
    Once this activity is done you can delete the records from the staging table.
    Thanks,
    Ravi

  • How do I make or recreate a  "valid file structure"

    I am trying to get FCP to get the video from my harddisk (boot volume). I dont have the camera right now, but still need to have a solution to make FCP to read video from harddisk and not be dependent to read from a harddisk inside a camera. I have read abot using several kinds of other peaces of software that imports or in someway making FCP able to edit the video. As a secandary solution thats OK, but I think this should be done by FCP. I use 6.05.
    Is there documentation how a file structure should be - if this is at all doable outside a cam. I tried to put the folder structure on a USB stick to make it "look like the camera", but it is still unreadable.
    There is a error message I understand, but cant use get further "please choose a folder whose directory structure matches supported media". (law-school writen this
    Here is the folder structure down to the .mts
    //AVCHD/BDMV/STREAM/00038.mts
    Alle other files and folders is as they is on camera.
    So whats missing to get a "directory structure that matches supported media". Any suggestions?

    Just copy the contents of the disc as a whole, into a folder on your HD, and follow the instructions in the "HD and Broadcast Formats" to navigate to it in Log & Transfer. You should be using a separate physical drive for media.

  • Reg: consistency check in recordset structure validation

    Hi all,
    I am doing FCC to FCC scenario, i am getting this error. Help me out, it is very urgent for me.
    Sender Adapter v2723 for Party '', Service 'bs_text2text':
    Configured at 2007-08-21 08:00:49 UTC
    History:
    - 2007-08-21 09:29:06 UTC: Retry interval started. Length: 2.000 s
    - 2007-08-21 09:29:06 UTC: Error: Conversion of complete file content to XML format failed around position 0: Exception: ERROR consistency check in recordset structure validation (line no. 87: missing structure(s) in last recordset
    - 2007-08-21 09:29:06 UTC: Processing started
    - 2007-08-21 09:29:04 UTC: Error: Conversion of complete file content to XML format failed around position 0: Exception: ERROR consistency check in recordset structure validation (line no. 87: missing structure(s) in last recordset
    - 2007-08-21 09:29:04 UTC: Processing started
    Regards,
    Ajay.

    Hi Ajay,
    Problem with the file content conversion parameters you specified in the sender file adapter. 
    Do let us know your DT structure and FCC parameters specified in adapter.
    Regards,
    Sumit

  • ERROR consistency check in  recordset structure validation

    Hi,
    I have a file that has the format below. The file is managed with the below content conversion parameters. However, it's not working and the error message that i get is "Conversion of file content to XML failed at position 0: java.lang.Exception: ERROR consistency check in recordset structure validation (line no. 7: missing structure(s) in last recordset". Any idea, what when wrong? Thanks.
    Message Type defined:
    FileHeader            (occurrence = 1)
      ARPostingRecord    (occurrence = 1 to unbounded)
        RecordHeader       (occurrence = 1 to unbounded)
        RecordDetail          (occurrence = 1 to unbounded)
    Source file looks like this:
    01 Field1 Field2...Field-n  
    H  Field1 Field2...Field-n  
    L  Field1 Field2...Field-n  
    L  Field1 Field2...Field-n  
    H  Field1 Field2...Field-n  
    L  Field1 Field2...Field-n  
    L  Field1 Field2...Field-n  
    where:
    01 is the keyfieldvalue for FileHeader
    H is the keyfieldvalue for RecordHeader
    L is the keyfieldvalue for RecordDetail
    Specified parameters:
    FileHeader.fieldSeparator
    FileHeader.endSeparator
    FileHeader.fieldNames
    FileHeader.keyFieldValue
    RecordHeader.fieldSeparator
    RecordHeader.endSeparator
    RecordHeader.fieldNames
    RecordHeader.keyFieldValue
    RecordHeader.missingLastFields
    RecordHeader.keepIncompleteFields
    RecordDetail.fieldSeparator
    RecordDetail.endSeparator
    RecordDetail.fieldNames
    RecordDetail.keyFieldValue
    RecordDetail.missingLastFields
    RecordDetail.keepIncompleteFields
    ignoreRecordsetName
    Recordset Name:
    ARPostingRecord
    Recordset Structure:
    FileHeader,1,RecordHeader,1000000,RecordDetail,1000000

    >ERROR consistency check in recordset structure validation (line no. 7: missing structure(s) in last recordset". Any idea, what when wrong?
    When you are not sure about number of records for RecordHeader and RecordDetail, Please specify wildchar character. Dont give max numbers.  The errors says expecting the specified number for recordset and  since it is missing gives errror.

  • Getting file structure off system

    Im trying to create a program which can read the directory stucture of a system and put it in a list of sum sort so i can process it.
    my aim is to be able to represent the file structure in an xml document
    i am totally lost and would just like to get a bit of a push in the right direction

    The code snippet below shows how to move through all the directories of a specific drive on a computer.
    public void traverse (File root) {
        // get all the files in this location
        File [] files = root.listFiles();
        System.out.println ("Visiting " + root.getPath());
        // move through each of them and look for directories
        for ( int i=0; i < files.length; i++) {
            // if this is a directory then investigate it by calling traverse again
            if ( files.isDirectory() ) {
    traverse (files[i]);
    else {
    System.out.println ("File: " + files[i].getPath());
    You can get the drive letters (on a windows machine) that are valid by calling the static .listRoots method of the File class. Be warned that running the above code on a big drive will take some time!
    Jonathan

  • IPhoto import finder file structure sort by date taken

    Hi! I'm switching from iPhoto11 to Aperture. I've imported my iPhoto Library via the Import menu asking Aperture to Copy the photos.
    File structure in the finder in important for me for further usage so I checked the "Aperture Library" package content and discovered that my 10000 photos had been imported in one unique folder with today's date. Even worse each subfolder has a complex name...
    Is there an option to specify "Import photos into folders devided by taken date" ?
    Thanks

    Messy for you but not for the computer. Both iPhoto and Aperture maintain the files in an order that is appropriate for that application. It may not a format that is understandable to humans. I always use managed and allow Aperture to do what it wants in the way of file storage. If I need anything I ask Aperture to get it for me. It is a DAM, Digital Asset Management, system.
    If you have valid reason to have to access the files without using Aperture then maybe referenced is the solution for you. Personally I figure that is what the computer is there to do so I leave it do its thing.
    Allan

  • Disconnected Use(with external drive)= Corrupted file structure

    Help. Really like the way LR handles RAW. Big problem for me is that my Mac Book Pro HD is too full. I'm using an EHD from OWC(FW800). If I launch LR with the drive unmounted LR starts verifying files and pulling all sorts of garbage into the LR Folder panel. If I reconnect the drive there is no repair and only a fraction of the files show up with their terminal folder- the folder structure on the EHD is flattened to useless. Fatal flaw? or my error?
    Thanks,
    Steve

    Hi Dan- Happy to try and help. It has happen under 2 scenarios, the lrdb files are at MacHD/users/srp/pictures/LightRoom/...irdb in both cases.
    Scenario 1: I had a folder with some duplicates in MAC HD/users/...pictures/My Photos/duplicate files. When I would dismount the EHD and launch LR it would pull the dups on the Mac HD into the DB and reference what looked to be temps for missing files as well with fairly random preview distribution. The EHD file structure is OWC/users/srp/pictures/My Photos/files by date.
    Scenario 2: I ditched the duplicate folders on the MAC HD (they were not referenced in LR before disconnect anyway) and launched LR. It initiates checking for files and photos again and pulled random preview references to odd files replacing the original folder structure. In this case there were no valid references to pull so it was mostly garbage, it did preserve a few "?" files.
    Folders Panels: http://aktlr.smugmug.com/gallery/2668872#141260988.
    Ive included scenario 1 & 2 grabs and an example of a garbage path
    My work around is that I delete the lrdb and replace it with a clean copy, better than re-importing everything.

  • Using Adobe Bridge file structure with iPhoto (latest version)

    I use Adobe Bridge and have all my pics in named folders in Pictures/PICS/Folder Names.  Inside the PICS folder is the iPhoto Library (only a few pics in it).  Is there any way I can use the file structure I have set up with Bridge and iPhoto (latest) simultaneously?  I really dont want to import (copy) all my pics into IPhot because I am pretty sure I will end up with two versions of each.  I havent been able to manage pics manually the way I like to in older versions of iPhoto. 

    Here's some info to help you setup Photoshop for use with iPhoto:
    Using Photoshop or Photoshop Elements as Your Editor of Choice in iPhoto.
    1 - select Photoshop or Photoshop Elememts as your editor of choice in iPhoto's General Preference Section's under the "Edit photo:" menu.
    2 - double click on the thumbnail in iPhoto to open it in Photoshop.  When you're finished editing click on the Save button. If you immediately get the JPEG Options window make your selection (Baseline standard seems to be the most compatible jpeg format) and click on the OK button. Your done. 
    3 - however, if you get the navigation window
    that indicates that  PS wants to save it as a PS formatted file.  You'll need to either select JPEG from the menu and save (top image) or click on the desktop in the Navigation window (bottom image) and save it to the desktop for importing as a new photo.
    This method will let iPhoto know that the photo has been editied and will update the thumbnail file to reflect the edit..
    NOTE: With Photoshop Elements  the Saving File preferences should be configured as shown:
    I also suggest the Maximize PSD File Compatabilty be set to Always.  In PSE’s General preference pane set the Color Picker to Apple as shown:
    Note 1: screenshots are from PSE 10
    Note:  to switch between iPhoto and PS or PSE as the editor of choice Control (right)-click on the thumbnail and select either Edit in iPhoto or Edit in External Editor from the contextual menu. If you use iPhoto to edit more than PSE re-select iPhoto in the iPhoto General preference pane. Then iPhoto will be the default editor and you can use the contextual menu to select PSE for your editor when desired.
    OT

  • How to join 5 different tables using SQL to make it to a flat file structur

    I am trying to load five differnt tables into one flat file structure table without cartesian product.
    I have five different tables Jobplan, Jobtask(JT), Joblabor(JL), Jobmaterial(JM) and Jpsequence(JS) and the target table as has all the five tables as one table.
    The data i have here is something like this.
    jobplan = 1record
    jobtask = 5 records
    joblabor = 2 records
    jobmaterial = 1 record
    jpsequence = 3 records
    The output has to be like this.
    JPNUM     DESCRIPTION     LOCATION     JT_JPNUM     JT_TASK     JL_JPNUM     JL_labor     JM_JPNUM     JM_MATERIAL     JS_JPNUM     JS_SEQUENCE
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     10     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     20     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     30     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     40     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     50     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     1001     Sam     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     1001     Mike     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     1001     Hammer     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     1001     1
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     1001     2
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     1001     3
    Please help me out with this issue.
    Thanks,
    Siva
    Edited by: 931144 on Apr 30, 2012 11:35 AM

    Hope below helps you
    CREATE TABLE JOBPLAN
    ( JPNUM NUMBER,
      DESCRIPTION VARCHAR2(100)
    INSERT INTO JOBPLAN VALUES(1001,'Test Jobplan');
    CREATE TABLE JOBTASK
    ( LOCATION VARCHAR2(10),
      JT_JPNUM NUMBER,
      JT_TASK  NUMBER
    INSERT INTO JOBTASK VALUES('USA',1001,10);
    INSERT INTO JOBTASK VALUES('USA',1001,20);
    INSERT INTO JOBTASK VALUES('USA',1001,30);
    INSERT INTO JOBTASK VALUES('USA',1001,40);
    INSERT INTO JOBTASK VALUES('USA',1001,50);
    CREATE TABLE JOBLABOR
    ( JL_JPNUM NUMBER,
      JL_LABOR VARCHAR2(10)
    INSERT INTO JOBLABOR VALUES(1001,'Sam');
    INSERT INTO JOBLABOR VALUES(1001,'Mike');
    CREATE TABLE JOBMATERIAL
    ( JM_JPNUM    NUMBER,
      JM_MATERIAL VARCHAR2(10)
    INSERT INTO JOBMATERIAL VALUES(1001,'Hammer');
    CREATE TABLE JOBSEQUENCE
    ( JS_JPNUM    NUMBER,
      JS_SEQUENCE NUMBER
    INSERT INTO JOBSEQUENCE VALUES(1001,1);
    INSERT INTO JOBSEQUENCE VALUES(1001,2);
    INSERT INTO JOBSEQUENCE VALUES(1001,3);
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             JT.LOCATION     AS LOCATION    ,
             JT.JT_JPNUM     AS JT_JPNUM    ,
             JT.JT_TASK      AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBTASK JT
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             JL.JL_JPNUM     AS JL_JPNUM    ,
             JL.JL_labor     AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBLABOR JL
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             JM.JM_JPNUM     AS JM_JPNUM    ,
             JM.JM_MATERIAL  AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBMATERIAL JM
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             JS.JS_JPNUM     AS JS_JPNUM    ,
             JS.JS_SEQUENCE  AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBSEQUENCE JS;
         JPNUM DESCRIPTION     LOCATION      JT_JPNUM    JT_TASK   JL_JPNUM JL_LABOR     JM_JPNUM JM_MATERIA   JS_JPNUM JS_SEQUENCE
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        10         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        20         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        30         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        40         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        50         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       1001       Sam        NULL       NULL    NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       1001       Mike       NULL       NULL    NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       1001       Hammer  NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    1001          1
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    1001          2
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    1001          3
    {code}                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Moving a Folder from the iTunes Folder to My Own File Structure

    I'm trying to move a folder from the default iTunes folder to a folder on another drive where I am managing the file structure. This allows me to file music as I would like to on the disk an makes things much easier to find when I need to.
    Recently I've noticed that when I do this I lose some of the song meta data such as the image and rating which is kind of annoying. Is there any way to do this and not lose meta data?

    If the external drive is in NTFS format (which is how most new drives are shipped), the Mac can't write to it. Reformat it with Disk Utility, putting it into Mac OS Extended format.

  • Imported par file to NWDS, jar file missing. Where find par file structure?

    Hi All,
    We want to implement idle session timeout on the portal.  We found a wiki here on the subject.  In order to implement the solution we need to make adjuestments to com.sap.portal.navigation.masthead.par.bak.  When we import this par file to Netweaver Developer Studio, the libraries and jar files are missing.  So obviously it errors then in portal as files are missing.
    So I could add them manually but I don't know what the file structure is on this par file so don't where to add what.  Does anyone know?
    Also when I export the file after manually adding the missing files, will it export all the files or will it behave the same way as the import and leave out the additional files?
    If anyone could help, I'd really appreciate it,
    Kind regards,
    Liz.

    Hi Bala,
    I thought of that just before you responded and it worked!!  Thanks for the response..
    Regards,
    Liz.

  • Need help understanding Time Capsule file structure and how to get it back

    I have the original Time Capsule on which I backup both my Mac Pro and my wife’s Macbook. When Snow Leopard came out, I successfully used the ‘Restore from Time Machine’ feature on my Mac Pro so I know it has worked. However, my wife’s MacBook harddrive died the other day and I was trying to do the ‘Restore from Time Machine’ and all it would find was a backup from April (when I put her in a new larger drive). Time Machine would not find any backup files newer that April. She stated that she had seen the the Time Machine backup notices regularly and as recent as the day the haddrive died (Nov. 23) so I figured that I should have no problem. Here is what I have found in my trouble shooting and what leads to my questions below.
    This is the file structure I found: (note that our ID’s are ‘Denise’ and ‘John’)
    *Time Capsule* (the drive as listed in my Finder window sidebar under ‘shared’)
    >Folder called ‘Time Capsule’ (when logged in as either ‘Denise’ or ‘John’)
    >>Denise Sparsebundle
    >>>Backup of Denise’s iBook (mounted image)
    >>>>Folder called ‘Backups.backupdb’
    >>>>>Folder called ‘Denise’s iBook
    >>>>>>Single folder with old April backup (not the right files)
    >>John Sparsebundle
    >>>Backup of John’s Mac Pro (mounted image)
    >>>>Folder called ‘Backups.backupdb’
    >>>>>Folder called ‘John’s Mac Pro’
    >>>>>>Folders containing all my backup files
    >Folder Called ‘Denise’ (if logged as ‘Denise’)
    >>Denise’s Sparsebundle (a disk image)
    >>>Backup of Denise iBook (the mounted image. Name from old machine)
    >>>>Backups.Backupdb
    >>>>>Denise’s iBook (Contains the backup I need)
    >Folder Called ‘John’ (if logged in as ‘John’)
    >> (empty)
    For some reason, my wife’s backup files are stored within a folder located at the top level of the Time Capsule drive called ‘Denise’, however, mine are within a folder called ‘Time Capsule’ which is at the same level as the ‘Denise’ folder.
    For some reason, when trying to use Time Machine to recover, it bypasses the top level ‘Denise’ folder which contains the correct files and goes into the ‘Time Capsule’ folder and finds the outdated backup files.
    I would assume that both my backup files and my wife’s should be at the same level of the Time Capsule.
    I was eventually able to use Migration Assistant to recover the files after installing a fresh OS and mounting the correct Sparsebundle image.
    So, my question, how do I get this fixed so that if I have to recover a drive with Time Capsule, it will find the correct files.
    Sorry for the long post and thanks in advance for any help.

    John Ormsby wrote:
    What I was trying to determine is why different backups for one particular machine are located at different file structure levels on my Time Capsule and why the most resent one for that machine is at a different level that for my other machine. Also, what is the correct level that both machines backups should be at.
    well John, first can you clarify if you are on 10.4.8 as your profile suggests ? if so, i'm wondering how you can use TM at all because TM was introduced with Leo. if you're not on 10.4.8, please update your profile. if you could, please also indicate details of your machine such as available RAM, processor, etc.
    second, what OS is your wife's machine running ?
    third, i frankly don't know too much about TM's file structure or if there indeed is such a thing as the correct one. however, i do know that it is best to leave TM to do its thing (as long as it's working). FWIW, though off-topic, you may want to have a look at this read for further information.
    last but not least, i see TM backups to my TC only as a part of my backup strategy. the backbone of my strategy is to create bootable clone of my startup disk(s) using e.g. Carbon Copy Cloner on a regular basis. while TM is capable of doing a full system restore, i don't trust it as much as a working clone. i use TM to retrieve a file i accidentally trashed a week ago but you cannot boot from a TM backup if your startup disk goes belly up.
    If I have missed this information in either the FAQs or the Troubleshooting article, I would greatly appreciate being pointed to it .
    i expect you didn't miss anything and i'm sorry you didn't find any help there. perhaps if Pondini (author of the FAQ and troubleshooting user tips) reads this thread, you will get the necessary advice on the file structure and more besides.
    good luck to you !

  • AVCHD File Structure From MAC to PC

    Hello,
    I know this is sort of off topic but I need the expertise of this forum....
    I am the videographer (mac user) for a speaking group and have to film speeches 5-7 minutes in length each week. The problem is how to get the videos to each person without processing on my Mac and uploading to youtube so they can see them.
    I am Using a Canon HF10 and wonder if I could pop out the memory card from the camera put it in a card reader attached to a person's PC and transfer the files that way. I know that I would copy the entire file structure to their computer. I assume the PC person would have to have something that could work with the AVCHD files but I don't know what is on a PC computer these days. The other dilemma is that the speakers HAVE ABSOLUTELY NO COMPUTER skills.
    Anyone have any ideas how to get the video to a viewable state without going through FCE.
    thanks,
    Al

    I agree with Martin and might add that I was in a similar project late last year with several folks who certainly weren't as computer savvy as myself, and YouTube turned out to be the solution. That way I didn't have to deal with the perennial "I couldn't open it" comments you WILL receive. Everyone was able to view YouTube.
    You might try uploading to Vimeo - I think you can actually upload .MTS files directly to their site, if you want to just post your raw footage. You could just get the files in the Finder directly from the file structure on the HF10, copy them to your Mac and upload from there. (I haven't tried this, but Vimeo's site says they accept .MTS files, which is what the HF10 videos will be, in the "STREAM" subfolder)
    The downside is that their free account maxes out at 500MB of uploads per week. That may be enough for you but I don't know. If you pay, the limit is quite a bit higher.

  • File Structure with iTunes on an External Drive

    This is an image of the file structure on my external drive that holds all of my iTunes music:
    Is this correct, or should I have set it up differently? I am asking since it is differnet the the iTunes file structure on my MacBook Pro and some files will get uploaded to my external drive and others will go onto my MBP,s iTunes Music folder. All of my iPhone/iPad apps, Album Artwork, Mobile Applications. I have Book folders on both drives and the both have books on them.
    I should say that I had a major problem a few years ago, which has nothing to do with the above problem, with one of the iTunes updates which disconnected most of my songs from their respective files. I tried a solution a friend sugested, which really messed up the files and now many of the files have different info inside the song files, i.e., Otis Redding as the Artist on a Radiohead track. Or another groups cover for a Beatles album. I have a large library and have gone through slowly and fixed many things, but it is still quite messy and a huge bummer.. As a proud new owner of a Late 2011 MacBook Pro, I would love to be able to have a solution for this as well.
    TIA

    Well I just went through the same thing. I had a Lacie Porsche mobile drive at 80GB and changed to a Lacie Rugged 250. I keep all my music on external never local. Not a fan of loading my laptop. Anyways.... I was getting those same errors. I had copied all the itunes library contents from the Porsche drive to the Rugged drive. It took quite some time to transfer so you knew it was working. Then changed in the preferences of iTunes to save in the Rugged drive. If you trashed or deleted any files from the new drive and just put them in the trash without emptying you get that 501 trashes path. I found I had multiple duplicates of songs with the name of the song followed by a 1.mp4a. I think the library.xml file is the key to a clean transfer of music. I think this file is stored locally in your user folder and then it's used to direct to the alternate drive but from what I've read it's the file with all the playlist information and stuff on it. If you find it and move it over to the mobile drive then hold option and boot up itunes then tell it to look for that file on the new mobile I think you'll have no worries. I'm actually gonna try this myself.
    You don't want to know how it got it working. I physically opened every music file on the rugged drive to make sure the were all where they were supposed to be. All 70'gb's of music. Never doing that again.

Maybe you are looking for