Data storage and read

Hello all,
I got a problem in data storage and read. I used the combination of "Open data storage", "Write data" and "Close data storage" to store some data as an array as shown below.
And used the inverted combination to read data as shown below:
As shown the data file is in tdm form. This works fine in my computer, both the VI and the stand alone application. However when I run the stand alone application in a target pc, the storage file can't be generated!
I don't know if it's the problem of my code or the problem of the target pc, since the target pc has tiny memory card with few drivers installed. I'm wondering if anybody could help me fix this problem. If there's other way I can store and read the file? or if I can make the target pc generate the tdm file.
Thanks!
Chao
Solved!
Go to Solution.

What error is being presented when the file isn't being generated?  Is it an error with permissions?  Or an error with a folder not existing?  Can you manually make a file in the location where you expect the file to be generated?
Unofficial Forum Rules and Guidelines - Hooovahh - LabVIEW Overlord
If 10 out of 10 experts in any field say something is bad, you should probably take their opinion seriously.

Similar Messages

  • Data storage and retrieval

    I am working on a project that requires the storage and retrieval of about 70 unique values per class. There are also about 60 different instances of this class- all with 70 unique values.
    For security reasons, I am reluctant to use a separate database to retrieve and store these values and have opted for Collections, using Hashmaps or Treemaps to store and retrieve. Is this the only way forward?
    Have been working on Java for only 18 months so still a bit green. Any guidance would be appreciated.
    Thanks

    I do need persistant variables and a database is certainly the tidiest option however, the data security required is not only to preserve the data integrity but it is also of significant commercial worth, hence the attraction of going down the hashmaps/treemaps route. The aim is to deploy the software without needing to reference an external or local database.
    I'll do a bit more research into database security before scrapping the idea altogether.
    Thanks
    Kola

  • S6 IPSec Service Data Storage and WiFi Calling

    Hi,
    Can any other S6 owners tell me the storage size of the IPSec Service application.
    Then can you tell me if you have wifi calling enabled.
    I am getting the run around between EE and Samsung as I think there is a data leak in the wifi calling.
    My IPSec Service application is taking up 5.16GB and there is no way to clean this other than a phone reset.
    Cheers
    Ian

    I have the same issue. Only 1.6gb though. I heard this is some sort of memory leak and 5.1.1 should fix this issue. But being EE and a Samsung device, who knows when the update will arrive! 

  • Two questions: Event on a Date Picker and Read Only on a full page?

    Hello,
    I make use of this very informative board to develop an application but i have been unable to find info on these two topics. Sorry if it has been already discussed extensively.
    Using HTMLDB 2.2 for now.
    =>First question:
    I have a date picker item with an event
    onChange="resaStatusChange();"
    in the HTML Form Element Attribute field.
    The event is triggerring if i put a new data in the field but does not trigger if i use the picker.
    Is it a standart behaviour? Is there a way to trigger the event when new data is provided through the date picker?
    => Second question:
    One page in my application has to be used either to only display information or as a mixed displaying/entering data sheet.
    When the context is appropriate to use the page as a simple display (testing both global authorization and data to be displayed), is there a way to put a read only attribute on the full page instead of putting the read only conditions through each and every item of the page?
    Sorry if the questions are not clear and apologize for what should be newbie questions.
    Any link or info will be appreciated.
    Regards,
    Daniel Gureghian

    Hi Daniel,
    I'm not sure about your first question, but I can hopefully give you a helpfull answer for your second one.
    I'm handling it this way:
    var elInput = document.getElementsByTagName('input');
    var elTextArea = document.getElementsByTagName('textarea');
    var elSelect = document.getElementsByTagName('select');
    for (i = 0; i < elInput.length; i++)
      setReadOnly(elInput, true);
    for (i = 0; i < elTextArea.length; i++)
    setReadOnly(elTextArea[i], true);
    for (i = 0; i < elSelect.length; i++)
    setReadOnly(elSelect[i], true);
    function setReadOnly(pThis, pRead)
         if (pThis && pRead)
              pThis.disabled = pRead;
              pThis.style.emptyCells = "show";
              pThis.style.color = "black";
              pThis.style.backgroundColor = "#DDDDDD";
         else if (pThis && !pRead)
              pThis.disabled = pRead;
              pThis.style.emptyCells = "show";
              pThis.style.color = "black";
              pThis.style.backgroundColor = "#FFFFFF";
    I never used it on a whole page, only for regions. So maybe you have to make some changes.
    chrissy

  • Data formatting and reading a CSV file without using Sqlloader

    I am reading a csv file to an Oracle table called sps_dataload. The table is structured based on the record type of the data at the beginning of
    each record in the csv file. But the first two lines of the file are not going to be loaded to the table due to the format.
    Question # 1:
    How can I skip reading the first two lines from my csv file?
    Question # 2:
    There are more fields in the csv file than there are number of columns in my table. I know I can add filler as an option, but then there are
    about 150 odd fields which are comma-separated in the file and my table has 8 columns to load from the file. So, do I really have to use filler
    for 140 times in my script or, there is a better way to do this?
    Question # 3:
    This is more of an extension of my question above. The csv file has fields with block quotes - I know this could be achieved in sql loader when we mention Occassionally enclosed by '"'.
    But can this be doable in the insert as created in the below code?
    I am trying to find the "wrap code" button in my post, but do not see it.
    Heres my file layout -
    PROSPACE SCHEMATIC FILE
    ; Version 2007.7.1
    Project,abc xyz Project,,1,,7,1.5,1.5,1,1,0,,0,1,0,0,0,0,3,1,1,0,1,0,0,0,0,2,3,1,0,1,0,0,0,0,3,3,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
    Subproject,25580-1005303.pst,,102,192,42,12632256,1,1,102,192,42,1,12632256,0,6,1,0,32896,1,0,0,0,0,,,-1,-1,-1,-1,0,0,0,-1,-1,-1,-1,-1,-1,0,0,0,-1,-1,-1,-1,-1,-1,0,0,0,-1,-1,0,1,1,,,,,,1
    Segment, , , 0, 102, 0, 0, 0, 0, 0, 0, 0, 0, 0, , , , , , , , , , , 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, -1, 0, , , 1
    Product,00093097000459,26007,2X4 MF SF SD SOLR,,28.25,9.5,52.3, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.60,0,0,0,0,00,-1,0
    Product,00093097000329,75556,"22""X22"" BZ CM DD 1548",,27,7,27, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.20,0,0,0,0,0,0,0,,345.32
    Product,00093097000336,75557,"22""X46"" BZ CM XD 48133",,27,7.5,51, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.20,0,0,0,0,0,0,0,0
    Product,00093097134833,75621,"22""X22"" BZ CM/YT DD 12828",,27,9,27, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.20,0,0,0,0,0,0,0,,1
    This is my table structure -
    desc sps_dataload;
    File_Name     Varchar2 (50) Not Null,
    Record_Layer Varchar2 (20) Not Null,     
    Level_Id     Varchar2 (20),
    Desc1          Varchar2 (50),
    Desc2          Varchar2 (50),
    Desc3          Varchar2 (50),
    Desc4          Varchar2 (50)
    Heres my code to do this -
    create or replace procedure insert_spsdataloader(p_filepath IN varchar2,
    p_filename IN varchar2,
    p_Totalinserted IN OUT number) as
    v_filename varchar2(30) := p_filename;
    v_filehandle UTL_FILE.FILE_TYPE;
    v_startPos number; --starting position of a field
    v_Pos number; --position of string
    v_lenstring number; --length of string
    v_record_layer varchar2(20);
    v_level_id varchar2(20) := 0;
    v_desc1 varchar2(50);
    v_desc2 varchar2(50);
    v_desc3 varchar2(50);
    v_desc4 varchar2(50);
    v_input_buffer varchar2(1200);
    v_delChar varchar2(1) := ','
    v_str varchar2(255);
    BEGIN
    v_Filehandle :=utl_file.fopen(p_filepath, p_filename, 'r');
    p_Totalinserted := 0;
    LOOP
    BEGIN
    UTL_FILE.GET_LINE(v_filehandle,v_input_buffer);
    EXCEPTION
    WHEN NO_DATA_FOUND THEN
    EXIT;
    END;
    -- this will read the 1st field from the file --
    v_Pos := instr(v_input_buffer,v_delChar,1,1);
    v_lenString := v_Pos - 1;
    v_record_layer := substr(v_input_buffer,1,v_lenString);
    v_startPos := v_Pos + 1;
    -- this will read the 2nd field from the file --
    v_Pos := instr(v_input_buffer,v_delChar,1,2);
    v_lenString := v_Pos - v_startPos;
    v_desc1 := substr(v_input_buffer,v_startPos,v_lenString);
    v_startPos := v_Pos + 1;
    -- this will read the 3rd field from the file --
    v_Pos := instr(v_input_buffer,v_delChar,1,3);
    v_lenString := v_Pos - v_startPos;
    v_desc2 := substr(v_input_buffer,v_startPos,v_lenString);
    v_startPos := v_Pos + 1;
    -- this will read the 4th field from the file --
    v_Pos := instr(v_input_buffer,v_delChar,1,4);
    v_lenString := v_Pos - v_startPos;
    v_desc3 := substr(v_input_buffer,v_startPos,v_lenString);
    v_startPos := v_Pos + 1;
    -- this will read the 5th field from the file --
    v_Pos := instr(v_input_buffer,v_delChar,1,5);
    v_lenString := v_Pos - v_startPos;
    v_desc4 := substr(v_input_buffer,v_startPos,v_lenString);
    v_startPos := v_Pos + 1;
    v_str := 'insert into table sps_dataload values('||v_filename||','||v_record_layer||','||v_level_id||','||v_desc1||','||v_desc2||','||v_desc3||','||v_desc4||')';
    Execute immediate v_str;
    p_Totalinserted := p_Totalinserted + 1;
    commit;
    END LOOP;
    UTL_FILE.FCLOSE(v_filehandle);
    EXCEPTION
    WHEN UTL_FILE.INVALID_OPERATION THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20051, 'sps_dataload: Invalid Operation');
    WHEN UTL_FILE.INVALID_FILEHANDLE THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20052, 'sps_dataload: Invalid File Handle');
    WHEN UTL_FILE.READ_ERROR THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20053, 'sps_dataload: Read Error');
    WHEN UTL_FILE.INVALID_PATH THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20054, 'sps_dataload: Invalid Path');
    WHEN UTL_FILE.INVALID_MODE THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20055, 'sps_dataload: Invalid Mode');
    WHEN UTL_FILE.INTERNAL_ERROR THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20056, 'sps_dataload: Internal Error');
    WHEN VALUE_ERROR THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE_APPLICATION_ERROR(-20057, 'sps_dataload: Value Error');
    WHEN OTHERS THEN
    UTL_FILE.FCLOSE(v_FileHandle);
    RAISE;
    END insert_spsdataloader;
    /

    Justin, thanks. I did happen to change my pl sql procedure using utl_file.get_file and modifying the instr function based on position of ',' in the file, but my procedure is getting really big and too complex to debug. So I got motivated to use external tables or sql loader as plan b.
    As I was reading more about creating an external table as an efficient way and thus believe I can perhaps build an extern table with my varying selection from the file. But I am still unclear if I can construct my external table by choosing different fields in a record based on a record identifier string value (which is the first field of any record). I guess I can, but I am looking for the construct as to how am I going to use the instr function for selecting the field from the file while creating the table.
    PROSPACE SCHEMATIC FILE
    ; Version 2007.7.1
    Project,abc xyz Project,,1,,7,1.5,1.5,1,1,0,,0,1,0,0,0,0,3,1,1,0,1,0,0,0,0,2,3,1,0,1,0,0,0,0,3,3,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
    Subproject,25580-1005303.pst,,102,192,42,12632256,1,1,102,192,42,1,12632256,0,6,1,0,32896,1,0,0,0,0,,,-1,-1,-1,-1,0,0,0,-1,-1,-1,-1,-1,-1,0,0,0,-1,-1,-1,-1,-1,-1,0,0,0,-1,-1,0,1,1,,,,,,1
    Segment, , , 0, 102, 0, 0, 0, 0, 0, 0, 0, 0, 0, , , , , , , , , , , 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, -1, 0, , , 1
    Product,00093097000459,26007,2X4 MF SF SD SOLR,,28.25,9.5,52.3, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.60,0,0,0,0,00,-1,0
    Product,00093097000329,75556,"22""X22"" BZ CM DD 1548",,27,7,27, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.20,0,0,0,0,0,0,0,,345.32
    Product,00093097000336,75557,"22""X46"" BZ CM XD 48133",,27,7.5,51, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.20,0,0,0,0,0,0,0,0
    Product,00093097134833,75621,"22""X22"" BZ CM/YT DD 12828",,27,9,27, 8421504,,0,,xyz INC.,SOLAR,,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,,0,0,0,0,1,000000.20,0,0,0,0,0,0,0,,1For example, if I want to create an external table like this -
    CREATE TABLE extern_sps_dataload
    ( record_layer            VARCHAR2(20),
      attr1                   VARCHAR2(20),
      attr2                   VARCHAR2(20),
      attr3                   VARCHAR2(20),
      attr4                   VARCHAR2(20)
    ORGANIZATION EXTERNAL
    ( TYPE ORACLE_LOADER
      DEFAULT DIRECTORY dataload
      ACCESS PARAMETERS
      ( RECORDS DELIMITED BY NEWLINE
        BADFILE     dataload:'sps_dataload.bad'
        LOGFILE     dataload:'sps_dataload.log'
        DISCARDFILE dataload:'sps_dataload.dis'
        SKIP 2
        VARIABLE 2 FIELDS TERMINATED BY ',' 
        OPTIONALLY ENCLOSED BY '"' LRTRIM
        MISSING FIELD VALUES ARE NULL
        +LOAD WHEN RECORD_LAYER = 'PROJECT' (FIELD2, FIELD3,FIELD7,FIELD9)+
        +LOAD WHEN RECORD_LAYER= 'PRODUCT' (FIELD3,FIELD4,FIELD8,FIELD9)+
        +LOAD WHEN RECORD_LAYER= 'SEGMENT' (FIELD1,FIELD2,FIELD4,FIELD5)+    LOCATION ('sps_dataload.csv')
    REJECT LIMIT UNLIMITED;
    {code}
    While I was reading the external table documentation, I thought I could achieve similar things by using position_spec option, but I am not getting behind its parameters. I have highlighted italics in the code above(from LOAD WHEN....FIELDS....), the part I think I am going to use, but not sure of it's construct.
    Thank you for your help!! Appreciate your thoughts on this..
    Sanders.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Hard disk life with frequent data write and read

    Dear All,
    Can you kindly advise on the effect of frequent write / read on shortening of (ordinary low-cost, 80 to 250G, office desktop PC type) hard disk life.
    File size is IMAQ iamge file (300k byte)
    1000 wite / read per day  for 1 day in a week   (6 days low activity, 1 day busy)

    I'm going to go out on a limb and say that it should be no problem, from a life time point of view, it averages to ~150 a day over a week. Of more concern will be that you will probably start having disk fragmentation issues in a fairly short period, where continuous blocks of disk space become harder to find and the data then has to be spread over numerous locations on the drive. One end result of this is increased drive activity as the heads have to be moved all over to recover a file's data. Drives these days get a pretty good workout, with people storing thousands of MP3 music file, videos, digital camera pictures. Particularly with the MP3's they will have a lot of files that get accessed frequently. The drive, depending on O/S, will also get a lot of "random" activity, particularly if you have an indexing program such as Google desktop, running. Watch the HD light on your machine for a while and you will see it start blinking like crazy every so often when apparently nothing is running.  Of course if the files you are saving are critical they should be backed up regularly to some other drive or media.
    P.M.
    Putnam
    Certified LabVIEW Developer
    Senior Test Engineer
    Currently using LV 6.1-LabVIEW 2012, RT8.5
    LabVIEW Champion

  • Data storage and aggregation levels

    Hi,
    I want to know, in facts and dimensions, where is data stored? Is it on multiple levels of aggregation? Or it does not depend from aggregation and it is stored at the lowest level?
    Thank you.

    hi,
    It is not stored as aggregations,the data is stored as de-normalised format and that data is picked up by BI server and process the data accordingly applying the conditions and coming to aggregations it will push down the functions applied on that to lower level.
    hope it clears your doubt.
    By,
    KK

  • Data storage and Confidentiality

    How long is data stored by FormsCentral?
    Can I request the removal of specific data?
    Will FormsCentral access any of the data I collect?

    The data is removed when you delete your form.
    Randy

  • Data Streaming to HDD and Reading

    Hi,
    I have three channels at 1 KHz to 5 KHz selectable sampling rate. I would like to stream all three channels to disk in binary format with user supplied header / file name. I would like to watch data graph while continuously saving to disk and also need controls to select time duration to save data. I want to use same VI panel to read data with slider bar control. All three channels must show separate graph charts.
    This is like simple data recording and reading application in same window with three channels and binary format.
    Any help is highly appreciated,
    Kishor
    [email protected]

    I would be glad to help, but your posting sounds like you would like us to write the code for you. Have you started the project and have a specific question? You may want to look at the LabVIEW shipping examples. (open labVIEW 6.1 >> help >> find examples >> fundamentals >> file input and output. There are several examples for reading and writting to file. You may also be interested in the DAQ examples to see how to aquire the signal. Your outline for the slider suggests that you also need to know how to make a control work like an indicator and vise versa. This would be done with property nodes or local variables. Look for property examples at find examples >> programatically controlling VIs >> controlling front panel objects.

  • Editable and Read-Only Date Picker have different display formats -- Bug?

    Hi all,
    This is a follow on to a previous posting about getting a Date Picker field to display read-only (Date Picker and Read-Only Condition works on LAST state of element
    Now past that hurdle, I have noticed something about the date picker field when it is displays read-only. Namely, the date format is not the format selected, nor does the read-only date format comply with the format mask in the source section.
    I've modified my small test app to work as intended in order to demonstrate this feature. I also opened it up so everyone can see it <blush>. You can find it here: http://htmldb.oracle.com/pls/otn/f?p=42513: Please navigate to the Update Project tab.
    When a project in Program One through Four is selected (which has a Spec Committed date), the date is displayed DD-MON-YY format. When a project the other Programs (which do not have Spec Committed dates at the time of this writing) are selected, and a date value selected with the date picker, the date is displayed in the anticipated DD-MON-YYYY format.
    Anyone else notice this, or have I just missed something obvious?
    Thanks,
    Petie

    I'm not sure why this happens but to work around it I added this before header process code to the page:
    execute immediate 'alter session set nls_date_format = ''DD-MON-YYYY''';
    Scott

  • Use an external HD with Airport Extreme to access it for data storage from computers  connected to the WiFi network?

    Can I connect My Passport HD into my Airport Extreme's USB port to access it from my computers (27" iMac 2012 now using Yosemite) on that Wifi network for data storage and retrieval purposes?

    Yes, if the AirPort base station's hardware supports 802.11n. The drive needs to be formatted as Mac OS Extended (Journaled), FAT16, or FAT32, and may need to be connected through an USB hub with its own power supply.
    (117777)

  • Data storage in Swing application

    I'm currently developing a book collection application, which allows a user to search/add/delete entries. I thought about data storage and figured that perhaps just using XML files would be best, since there would only be a few hundred records involved. Would it be easiest to use JAXB for this type of thing? Or perhaps an XML database instead of a few XML files? I'd really appreciate any hints or comments.

    My personal opinion: no, keep it simple.
    A simple XML handling is done by both:
    dom4j www.dom4j.org (better documentation), and
    jdom[7b] www.jdom.org (better performance the last time).
    Use for instance dom4j-full.jar with XPath support for element selection.

  • My iphone 4 has no data storage. I have deleted all apps, music and texts and it still says I only have 1.9 GB available. There is nothing in my storage.

    My iphone 4 has no data storage. I have deleted all apps, music and texts and it still says I only have 1.9 GB available. There is nothing in my storage.

    The problem could be that you don't have enough storage on the device itself, not on iCloud.  Go to Settings>General>Usage to see how much "Storage" you have available on the device.  Farther down the list is the available storage on iCloud.
    Also check:
    Go to Settings>iCloud>Storage & Backups>Manage Storage; there, tap the device you need info on and the resulting screen lists Backup Options with which apps store data on iCloud as well.  Tap Show All Apps to get the complete list of apps and MB used for backup storage.
    A device needs many MB of storage in order to perform a backup to iCloud.
    Also see:  http://support.apple.com/kb/ht4847`

  • For the last month or so my icloud won't back up, it keeps telling me that there is too much storage and I need to delete some data. But I have deleted between 3-400 photos and it is still saying the same, can anyone please help

    For the last month or so my icloud won't back up, it keeps telling me that there is too much storage and I need to delete some data. But I have deleted between 3-400 photos and it is still saying the same, can anyone please help

    If they were deleted from your camera roll that should have resulted in a corresponding decrease in the estimated size of your next backup.
    If it's still saying that you don't have enough storage to back up, read through this article: http://support.apple.com/kb/ht4847.  It provides some suggestions for reducing your iCloud storage, such as deleting unneeded email. 
    The other area to look at is your text messages.  These are included in the backup.  If you have lots of photos and videos attached to your messages, that can significantly increase the size of your backup.  If that's the case, deleting these may make a big difference.

  • I have a Time Capsule with 2tb of storage, and an I-Mac with 500Gb. (which is at capacity) Can I use the Time Capsule to store data in addition to backing up the hard drive?

    I have a Time Capsule with 2tb of storage, and an I-Mac with 500Gb. (which is at capacity) Can I use the Time Capsule to store data from my hard drive that I want to access later in addition to backing up the hard drive?

    Can I use the Time Capsule to store data from my hard drive that I want to access later in addition to backing up the hard drive?
    Yes, but if you move the "original" data on your iMac to the Time Capsule....then the "original" data is now on the Time Capsule disk. 
    When...not if...the Time Capsule disk has a problem, you have no backups for this data.
    Perhaps a much safer plan might be to add an external hard drive attached directly to your iMac. Then, move the data to that drive from the iMac.
    Now, Time Machine will back up both your iMac and the external hard drive so you have "originals" on one drive and backups on another. Would that plan work for you?

Maybe you are looking for

  • Unable to install OS 2.0

    Can anyone help me here? I already tried for so many times downloading OS 2.0 but the result is still the same, it always say unable to install OS 2.0. Any idea?

  • NI 5772, using clk_in as an external clock

    Hi.  I'm using NI 5772R with NI PXIe-7966R NI FlexRIO FPGA Module. I'm trying to sample the signal with external clock (CLK_IN). An example file 'Clock Select (FPGA).vi' is successfully compiled. However when running 'Clock Select (host).vi', the dat

  • Trouble with current playback time on progress bar

    First, some terminology. As Adobe Flash Player is preparing to play a video, it displays a spinning icon in the center of the video window.  I have seen web pages which refer to this as the spinning beach ball.  I will refer to this simply as the "sp

  • Wireless to Wireless bridge help needed

    Folks.... I think I am looking for a wireless to wireless bridge. My next door neighbor & I are splitting the cost of a fast Internet connection.  He is transmitting the signal to  me (and to his home) using a WAP54G access point.   I can pick up the

  • Error message when reinstalling Snow Leopard

    hai i have macbook pro 13" 2,4 GHz core 2 duo processor. i want to reinstall SL 10.6.3 from original dvd installer came up with the machine. i tried to do clean install. but each and everytime i got error message "install failed. the installer could