Data Aquisition from HP 54504A

I have an older model HP 54504A Digitizing Oscilloscope, and I need to create a LabVIEW program that can pull data off the scope. I also need to be able to capture an eye-diagram, which will require repetitive sampling to acquire a large number of data points. I was wondering if anyone had a simliar program that I could use as a reference in create mine. Thanks in advance.

Hi Trigeek,
I do not have personal experience with this device, but you might check this site for a driver for your instrument. I'm not sure if there is one for this instrument or not but you might be able to find a driver for a similar instrument that you can modify. Maybe someone else who has experience with this instrument will post an example for you. Hope this helps.
JenK

Similar Messages

  • Data aquisition from spectrophotometer in two analog channels, X-Y plot

    Dear colleagues!
    I am new to LabView and I am only starting to learn the system (I have never been a programmer, I'm a chemist to be precise).
    I have before me a task to interface an old analog scanning spectrophotometer to PC via LabView.
    It has two output analog channels: one is wavelength (two ranges, currenlty let it be 190 to 390 nm; it reads on multimeter as 0.19-0.39 V range) and another one is absorbance (0-1V range). An UV spectrum sholud be an X-Y graph wherein X is wavelength and Y is absorbance.
    I am able to read one channel OR another. I am able to read BOTH channels simultaneously, but how can I plot them AGAINST each other, not against TIME?
    I am using DAQPad-1200 (LPT interface) and LabView 6.1 (that's why I am unable to read most of the VI's posted on the forum ).
    Is such an interfacing even possible given my software and hardware limitations???
    Thank you all in advance!

    Thanks a lot, johnsold!
    I am now able to get something which looks like an UV spectrum! %)
    1. I will post an image (later); however, my primary objective is to get two columns of numbers, not necessary to see an X-Y graph onscreen.
    2. Without inserting arrays like I did LV always tells about error connecting this data line from two channels to save "2-D data to file" option telling me that input is 1D!
    3. I have a manual for the device, it has all of the schematics. I am no electrician though. I will post scanned circuit diagrams then when I manage to scan them.
    Now I see two more opportunities to control the device.
    A. There is a circuit (+12V DC) on the rear of one of the modlues, which, if shorted for at least 500 milliseconds (according to Tracor manual) can initialize automatic scan. How can this be done using Labview and DAQ device? My first guess is to use an analog output and some kind of a relay, either electromagentic or semiconductor one. Is there any other possibility?
    B. In my code I am always asked to write to a file after I initialize the sequence. Can this be done AFTER the sequence is finished?
    C. Finally, is it possible to automatically stop the While loop when the number generated by shift register becomes equal to 190?
    Here are attached a new code and a sample of data.
    Attachments:
    Tracor970A-v1.0.vi ‏53 KB
    UV-test-01.txt ‏15 KB

  • How to visualise continues Data aquisition from a PCI 6602 with 2121 BNC connector board in C++ ?

    Hello everyone,
    We are trying to get on the screen the aquired pulses from our  PCI 6602 with a 2121 BNC connector board, from several devices. We are able to read the data and save without problem, but we cannot look at it while we are measuring. Anybody has an idea to how program this in C++ ? Any suggestion is welcome!
    Thanks for the help

    Hi,
    try to look for some example programs and Tutorials:
    Examples Results
    http://search.ni.com/nisearch/app/main/p/bot/no/ap/tech/lang/en/pg/1/sn/catnav:ex/q/DAQmx%20C%2B%2B/...
    Tutorials Results
    http://search.ni.com/nisearch/app/main/p/bot/no/ap/tech/lang/en/pg/1/sn/catnav:tu/q/DAQmx%20C%2B%2B/...
    You should also have a look at the "C Reference Help" which is installed with the NI DAQmx driver.
    Acquire N Scans (Visual C++ 6.0, CW++, NI-DAQ)
    http://zone.ni.com/devzone/cda/epd/p/id/207
    Continuous Analog Acquisition with Producer Consumer Architecture in C#
    https://decibel.ni.com/content/docs/DOC-4253
    Good Luck!
    Matteo

  • How do you save dynamic data type, from the DAQ assistant, for use in Excel or matlab?

    Currently, I have the following basic VI setup to save Data from my PCI6221 Data Aquisition Card.  The problem I'm having is I keep getting the last iteration of the while loop in the measurement file and that's pretty much it.  When I try to index the data leaving the loop it gives me a 2D array of Data which cannot be input into the "Write to Measurement File" VI.  How would I save this to a useful Data/time step format?  I was wondering of a way to continuously collect the Data and then save it in a large measurement file that I would manipulate in Matlab/excel?  Am I using the wrong type of loop for this application?  I also noticed my Dynamic Data array consists of data, time, timestep and then a vector of the data taken.  Is it possible to just get a vector of the time change per sample alongside the data?    Sorry for the barrage of questions but any help would be greatly appreciated, and thanks in advance!
    -Bryan
    Attachments:
    basic DAQ.vi ‏120 KB

    There is a VI in the Express > Signal Manipulation palette called "From DDT" that lets you convert from the Dynamic Data Type to other data types that are more compatible with operations like File I/O....for instance, you could convert your DDT into a 2D array and use the Write To Spreadsheet File.vi.  Just a thought...
    -D
    Darren Nattinger, CLA
    LabVIEW Artisan and Nugget Penman

  • Resizing window affects data aquisition

    Hi Guys, got a really annoying problem that I've been trying to solve for about 2 weeks now, I'm hoping someone will be able to help me.
    We communicate through a dll to our data aquisition box (custom). Our custom data aquisition box has an array which we read in a timed loop every 20ms (although the array is present for around 50ms), once read we send a command saying that we have read the data.
    Once we've read the data we put it into a queue. The logger then reads from this queue, all works fine on a normal run. However if we resize any window (not just a labview one) this seems to knock out our Timed loop. i.e. we miss 1-2 arrays worth of data (50-100ms). This is strange as we have it in a Highest priority VI running in the "Data aquisition" thread. We also do lots of other things i.e. read from a TDMS file and read/write to global variables and display data on a graph, this doesn't seems to have much of an affect on the timed loops, its just when we resize a window. Obviously we do not want to risk losing critical data, so we need a way of stopping this from happening?
    Its almost as though the VI is running in the User interface thread and is a lower priority than the resize window method.
    Does anyone have any ideas on how we can solve this issue?
    Thanks in advance
    Joe
    We are collecting data at 1Khz and our Dll is thread safe and is set to "run in any thread".
    Using Labview 2009f3

    Hi Guys,
    Thanks for all your replies. Sorry for late reply, I was out the office yesterday.
    Unfortunately i can not post my source code however i have made a small mock up of how the system generally works, to give you an idea.
    josbone: Our dll is classed as thread safe and we have it running as such - i thought this could be the case to start with. I presume because it is classed as "any thread" it could still be running in the UI thread though?
    Tinnitus: I have attached a mock up that shows the main points in the system. I think I got a little confused on my previous reply. The DAQ holds the data there for 50ms and that is our only chance to read the data(more data comes through and our buffer can only hold 50ms worth). Which if the Timed Loop executes at 20ms should be fine.
    DFGray: I got really excited reading your reply - it sounded like it was a great fix, however, unfortunately it did not work
    Hope the mock code helps
    Regards,
    Joe
    Attachments:
    Streaming example.zip ‏90 KB

  • Dynamic data aquisition on USB-6212

    I am using USB-6212 to aquire (AI) and control (DO) from/to 8 UUTs. It works perfectly. What I want to do is dynamically launch my daq vi for each UUT that will acquire AI and DO data for its respective lines.
    For example, I launch UUT-1 daq vi to get AI data from ai0 and DO data to p0.0, p0.1 and p0.2. Then launch UUT-2 to get data from ai1 and data DO to p0.3, p0.4 and p0.5 at different times.
    I want to be able to launch and stop daq at my will.
    Here is the problem. So, I tried launching VIs dynamically, which works ok. But, data aquisition is not working. It works for the first launch of the daq vi. For the other UUTs, it give me this error:
    Error -50103 occurred at DAQmx Start Task.vi:2
    Possible reason(s):
    NI Platform Services:  The specified resource is reserved. The operation could not be completed as specified.
    Task Name: _unnamedTask<15>
    So, my question is, can this be done? Would USB-6212 support dynamically allocating task resources on DAQmx?
    I have attached my main launch vi and daq vi block diagrams. Any help/feedback is appreciated.
    Attachments:
    MainLaunch vi.png ‏35 KB
    LaunchTest vi.png ‏43 KB

    Hi sbk143,
    The 6212 only has a single analog input timing engine, so you cannot run more than a single analog input task at a time.
    What you can do is acquire from all of your AI channels in a single task, then programmatically index out the channels that you are interested in.
    Your approach would work for software-timed DO, so you can control the UUTs independently at different times.
    Best Regards,
    John Passiak

  • Wireless Data Aquisition

    I wanted to begin doing data aquisition wirelessly and was wondering how I would go about it. I am currently using scxi modules for my signal conditioning and a USB Data Acquisition and Control Module for the SCXI. I have read about using a something called a MiniDAT by ViaSat to connect a DAQcard to achieve wireless data acquisition capabilities but I can't find any information on the MiniDAT or if it is even still in production. Any suggestions or info about the MiniDAT would be appreciated. Thanks

    Hello Postemun
    Take a look at the TDK Blu2i module.
    I'm using this in conjunction with LB7.1 apps I've written for PDAs and PCs with bluetooth capability and it is designed as serial cable replacement, although it also has some rudimentary analogue and digital DAQ.
    In cable replacement mode it effectively converts a dumb serial device for wireless operation. Obviously there is a bandwidth issue if you're trying to get large amounts of data out of your DAQ unit ie high sampling rates, so you're limited to lower end serial comms speeds.
    The other issue is distance. How far away is your pc going to be from your wiress DAQ kit? Another alternative is to use 802.11 based serial modems for longer distances, although comms is still limited to se
    rial speeds. This website may be of interest:
    http://datahunter.com/
    Cheers,

  • Data Recovery from an encrypted Time Machine HD

    Does anyone have eny experience with having data retrieved from a time machine hard drive that is encrypted? I've clean installed Mountain Lion and now can't get at my time machine to restore files.
    Apple support recommended I posted here and explore 3rd party recovery servies.
    Please please anyone!

    v8r wrote:
    It was my first time machine back up. Mid way I felt that it is stuck so I powered down my mac book pro. Now its not booting. I gave the details in a simultaneous thread "Gray screen, spinning cog and midway to first back up of time machine". I am assuming that if I have lost all the data on my harddrive, can I at least get some from the incomplete time machine back up.
    well, it might be possible but it won't be in any form that you can retrieve automatically. so your first priority should be to get your computer to boot and after what you can start worrying about getting your data from an incomplete TM backup. I've done that it the past and it works sometimes but it really depends on the situation. I had to do it from terminal too as it didn't work from finder.
    but it need not be that you lost all data on the main drive. boot from the SL install dvd and see if disk utility sees the drive. if so, repair the drive and try booting from it. if you have disk warrior it might help too if disk utility can not fix some errors.
    Thanks,
    v8r.

  • No master data transmitted from ERP to GTS

    We are using ERP 4.70 with the following settings:
    SAP_APPL 470 SP-Level 0031
    PI 2004_1_470 SP-Level 0016
    PI_BASIS 2005_1_620 SP-Level 0017
    SLL_PI 720_470 SP-Level 0007
    together with SAP GTS 7.2 SP-Level 09
    I created the message types /SAPSLL/DEBMAS_SLL and /SAPSLL/MATMAS_SLL manually and activated them.
    I try to send an customer master data to GTS but the partner will not transmitted. In the log protocoll in GTS the error message appears that there was no data collected from table SAPSLL/TCOGVA
    If I try to send a material master from ERP to GTS the RFC connection gets broken with error message:
    RFC error (The transaction has dumped the connection). What I had found out is, that the error appears in the function /SAPSLL/API_1006_SYNCH_MASS.
    Has this something to do maybe with a wrong IDOC type? Does I have to to something in some tables. Do I use the wrong programme? I use /SAPSLL/MATMAS_DISTRIBUTE_R3 and /SAPSLL/DEBMAS_DISTRIBUTE_R3.
    I made absolutely the same customizing settings in another system at a customer and it worked fine there.
    Does anybody has an idea? Thanks to all for the possible help.
    Thanks very much in advance

    Hi Andreas,
    there are several possibilities why this is not working.
    First of all you should check, if your RFC settings are correctly maintained.
    Please bear in mind that the logical systems have to be assigned to logical system groups and that GTS and R/ shall not be in teh same group.
    Please also ensure that the logical system name is the same in both systems, R/3 and GTS.
    Secondly there is also the possibilites that necessary tables are not filled correctly.
    Please run report /SAPSLL/PLUGIN_CHECK_R3
    It could be thate.g. the tables TBD24 and TBD62 are not filled 
    for the relevant message type (in your case /SAPSLL/DEBMAS_SLL and /SAPSLL/MATMAS_SL)
    To update this tables corretly pls. do the following:                                                                               
    - BD53                                                                
    - select message type (/SAPSLL/MATMAS_SLL)and edit it                 
    - select a segment                                                    
    - save                                                                
    - deselect segment again                                              
    - activate                                                            
    - set in trx BD60 the FM to the correct value                         
      (/SAPSLL/MATMAS_DISTRIBUTE_R3)                                                                               
    You can than check with the programm: '/SAPSLL/PLUGIN_CHECK_R3' if    
    all 4 required tables (TBD24, TBD62, TBDA2, TBDME) are filled, or not.
    I hope this helps with the issue.

  • Data import from EBS failed via FDMEE in fdm . Getting error message as "Error connecting to AIF URL.

    FDM Data import from EBS failed via FDMEE after roll back the 11.1.2.3.500 patch . Getting below error message in ERPI Adapter log.
    *** clsGetFinData.fExecuteDataRule @ 2/18/2015 5:36:17 AM ***
    PeriodKey = 5/31/2013 12:00:00 AM
    PriorPeriodKey = 4/30/2013 12:00:00 AM
    Rule Name = 6001
    Execution Mode = FULLREFRESH
    System.Runtime.InteropServices.COMException (0x80040209): Error connecting to AIF URL.
    at Oracle.Erpi.ErpiFdmCommon.ExecuteRule(String userName, String ssoToken, String ruleName, String executionMode, String priorPeriodKey, String periodKey, String& loadId)
    at fdmERPIfinE1.clsGetFinData.fExecuteDataRule(String strERPIUserID, String strDataRuleName, String strExecutionMode, String strPeriodKey, String strPriorPeriodKey)
    Any help Please?
    Thanks

    Hi
    Getting this error in ErpiIntergrator0.log . ODI session ID were not generated in ODI / FDMEE. If I import from FDMEE its importing data from EBS.
    <[ServletContext@809342788[app:AIF module:aif path:/aif spec-version:2.5 version:11.1.2.0]] Servlet failed with Exception
    java.lang.RuntimeException
    at com.hyperion.aif.servlet.FDMRuleServlet.doPost(FDMRuleServlet.java:76)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:301)
    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:119)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:324)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:460)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:163)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3730)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3696)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2273)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2179)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1490)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)

  • Getting data transfered from old HD to new ones, leaving 2 slots filled . .

    Hello,
    *(Below is a thread I started in another forum, but, didn't get my final questions answered (see at bottom of this post in BOLD), and am hoping for further help and suggestions. I know there is a lot to read here, but, I think this information will help other novices, such as myself. Thanks!)*
    I have 2 new HDs (each 1T) uninstalled, and I have my old HD (160 gigs) currently residing in one of the two slots in my G5. Any ideas on the best way/process to get the data transfered from the old HD to the new ones, leaving my two slots filled with the 2 new drives AND my data from the old HD transfered and intact, with a RAID0 setup?
    OR, are there any other suggestions on a good setup with the 2 new HDs??
    I'm open to suggestions.
    I always back-up on external HDs, so, is their a need for me to worry about setting up a RAID0??
    I'm doing video processing for DVD release.
    Should I just max out the 2x1T gig HDs and not worry about a HD failure??
    All Suggestions welcomed!!
    Thanks, Jack
    *FROM THE CUTTERMAN:*
    You are limited with only 2 drive slots. In this scenario what usually works best is a small fast drive (eg Velocirapter or SSD) for OSX/Applications and a large drive for video files. It is not advisable to have the operating system running from a striped array.
    Since you already have the drives, here is what you can do.
    Install one new drive and format the partition you wish to use for OSX. You may not want to use the whole drive for OSX, so make 2 partitions. Be sure that it is GPT-formatted so it can boot.
    Use CarbonCopy Cloner to copy your current OSX partition to the new one.
    Set the startup disk as the new partition.
    Remove the old drive and replace it with the second new one.
    Reboot, and partition/format new drive.
    *Hey Cutterman,*
    Thanks for the advice. That sounds like a good set-up. I'll have to research the process of doing what you've advised, but, should work out ok:
    A. Install one new drive and format the partition you wish to use for OSX. You may not want to use the whole drive for OSX, so make 2 partitions:
    1. Partitions are done under Disc Utilities correct? Never done it before; will take a look.
    B. Be sure that it is GPT-formatted so it can boot.
    2. GPT is also done under Disc Utilities, correct? Here I start to get a bit confused, as I've formatted Extended Journal, SO, the partition with the OSX & Applications is formatted GPT, and the rest of the HD is formatted as Extended Journal?
    C. Use CarbonCopy Cloner to copy your current OSX partition to the new one.
    3. So, I partition the first new HD and use Carbon Copy Cloner (also under Disc Utilities) to copy the OSX & Applications from my original HD to the first new HD, correct?
    D. Set the startup disk as the new partition.
    4. This will be an option when I setup the partition???
    E. Remove the old drive and replace it with the second new one.
    F. Reboot, and partition/format new drive.
    5. QUESTION: So, you recommend not striping the HDs: If I do a RAID 1, which sets up a mirror of the HDs, and since the first new HD is partitioned with the OSX & Applications, AND is set up as the Startup Disk, then the second new HD will also mirror that?? Another QUESTION: Even if the second new HD does not mirror the OSX & Applications, the second HD will only mirror the space/partition not utilized by the OSX & Applications partition, correct?
    Thanks for your help!!
    Jack
    *FROM THE CUTTERMAN:*
    OK I will try to answer your questions. To begin, if you want to mirror the 2 drives then you will need to boot from the SL install DVD or an external (USB/firewire) drive to set it up. IMO for your purposes it is too much hassle and a waste of HD space.
    1) Yes partitions are created in Disk Utility
    2) Yes, you choose a volume Scheme (ie number of partitions) and size them by dragging the separator bar. Then select the planned boot partition and click Options..., then choose GPT. I think it is usually the default but check and make sure. Extended journaled is the usual format.
    3) You need to download Carbon Copy. It is a free tool that will copy the entire image of your current system partition to the new one and make it bootable. Consider making a donation as it is a very useful and frequently updated utility. It is fairly intuitive to use. You can also use the restore feature in Disk Utility but I have had more experience with Carbon Copy.
    4) Once the copy process is finished the new boot partition will show up under the startup disk selections (this tool is in System Preferences).
    *FROM 666Sheep:*
    If may i correct one thing: OP, don't choose GPT (GUID Partition Table). You got PPC Mac (G5), so valid partition type for you is Apple Partition Map (APM).
    GPT is for Intel Macs and you will not able to boot from this kind of partition.
    *FROM THE CUTTERMAN:*
    Thanks for the correction- my bad. No familiarity with non-Intel Macs.
    G5 PowerMac, 2.5 GHz Dual, Dec. 2004, 6.5 GB RAM, 149 GB HD
    *Hi All,*
    *Since this process will come to a head on Tuesday/Wednesday, and I hope to be successful in transferring Old HD data to new HD/s.*
    *A: I'm still a little foggy here, as I thought I'd set up the 2 new T-gig HDs so that if one of them failed, I'd still have the data backed up on the other HD; more ideas concerning this would be reassuring:*
    *WHAT DOES "IMO" mean?*
    *(IMO for your purposes it is too much hassle and a waste of HD space.)*
    *As I understand the points made: After I partition the first new HD, and Carbon Copy data from old HD, I then install second new HD, and this will just act as an overflow from the first partitioned HD, correct? Because, if set up a Raid1/Mirror, the OSX partitioned portion of the first new HD will not copy to the second new HD, and that space on the second new HD will be wasted, correct?*
    *Also: I have to set up RAID1 (if I do not set up RAID0), correct?*
    *RAID 1*
    *From WIKI: RAID 1 mirrors the contents of the disks, making a form of 1:1 ratio real time mirroring. The contents of each disk in the array are identical to that of every other disk in the array. A RAID 1 array requires a minimum of two drives.*
    *Carbon Copy is a free tool that will copy the entire image of your current system partition to the new one and make it bootable.*
    *1. Is my old HD partitioned? Is that done automatically by MAC prior to PC purchase?*
    *2. Do I Carbon Copy the whole old HD or just parts? (eg. OSX, Photo Shop, After Effects, Various Files)*
    *3. I read that it is necessary to DEACTIVATE Photo Shop (I have CS3) prior to doing Carbon Copy. Is that correct? If so, does this also apply to other Adobe applications? Such as After Affects, Illustrator??*
    *Thanks to all for the help!!*
    *G5 PowerMac, 2.5 GHz Dual, Dec. 2004, 6.5 GB RAM, 149 GB HD*

    Confusing Information:
    A1: You also want to make sure the drive you are backing up to is formatted Mac HFS Extended (HFS+) if using Mac OS 8.1 or above.
    A2: On PowerPC Macs, your clone should be partitioned as Apple Partition Map.
    *Q? I thought the new HD (1T) should be formatted in Extended Journaled?*
    B: Also disable Spotlight (in 10.4 only) on your destination drive using Apple menu -> System Preferences -> Spotlight -> Privacy to add the destination drive to the pane.
    *Q? I'm using 10.4.11, so, I should follow these instructions?*
    C: If possible, boot into safe mode to perform the backup (holding the SHIFT key at startup). In addition, you can clone while logged into another administrative user that you don't use at all to avoid further complications of changes which may be happening to your regular user (though don't use Fast User Switching to get into that other user, since that other user is still active when fast user switching is used). Otherwise you are going to be running a backup on a live system which could have changes happening while you are attempting to backup. These may yield an imperfect clone, with uncertain success at recovery. It may be possible that your clone will have its own hardware issues, so make at least two copies.
    *Q? ?????????????*
    *I GOT THE ABOVE INFO. FROM THE FOLLOWING SITE:*
    Making a clone/mirror/duplicate backup
    http://www.macmaps.com/backup.html#SHORTANDEASY
    *ALSO, WHAT DO YOU THINK ABOUT THE ADVICE ON THIS PAGE:*
    http://www.levoltz.com/2010/04/21/how-to-transfer-data-to-your-new-hard-drive/

  • Unable to delete data files from my Ipod if used as data storage...

    Hello everybody
    I bought months ago an Ipod Nano 8GB. I've got an issue when i use it as disk storage i'm not able to sort out. I've checked all the topics but i cannot find any solution.
    Example: I use the Ipod with the option "Manually manage music and videos". I put audio and video files on my Ipod through Itunes, then i put also some files on my Ipod as normal data files through my Mac Finder (Easy, i drag VLC, JPG files etc. from folders to my Ipod). The Ipod/Itunes summary page shows (Correctly) the capacity column colored in blue (Audio), video (Purple) and orange (Data). No probs when i transfer data files to another computer.
    Problem: when i delete the data files from my Ipod (Through finder) the orange column doesn't disappear (I have no problems if i delete audio/video files via Itunes) and although i've actually deleted all the files if i try to add some other data or videos/music files to my Ipod, Itunes says that there is no space available in my device. It's very strange, i check and double check all the options but eventually i have to restore my Ipod all the times and (Pretty annoying) add all my files again.
    It seems that although i sent my files to the trash bin, Itunes (Fully updated) doesn't recognize the change.
    It'd be great if someone can help me out as it's very important for me to be able to use my Ipod as disk storage, thanks a lot!

    It seems that although i sent my files to the trash bin, Itunes (Fully updated) doesn't recognize the change.
    Empty it from the Finder menu or by control-clicking it in the Dock.
    (44613)

  • Data Load from a DSO to a Cube

    Hello,
    I am facing a problem
    Case:
    There is a DSO with following Char/Key Figure:
    Characteristics     ( Value-1 )     ( Value-2 )     ( Value-3 )
    A(Key)          ( A1 )     ( A2 )     ( A3 )
    B(Key)          ( B1 )     ( B2 )     ( B3 )
    C          ( C1 )     ( C1 )     ( C1 )
    D          ( D1 )     ( D1 )     ( D1 )
    E          ( E1 )     ( E1 )     ( E1 )
    F          ( F1 )     ( F1 )     ( F1 )
    KF1(Key Figure)     ( 10 )     ( 20 )     ( 30 )
    I got below data from the DSO to a Cube(structure is below with data) details are below:
    Dimention          ( Values )     
    C          ( C1 )          
    D          ( D1 )          
    E          ( E1 )          
    F          ( F1 )          
    KF1(Key Figure)     ( 50 )
    My Question is why i am not getting KF1 value as 60.
    When i tried with including Key Fields of the DSO, i got all the records in the Cubes.
    Where i am mistaking?
    Can some one correct me please?
    Thanks.
    Avinash.

    Below are the basic steps which we follow in any BI 2004S system:
    1)Create datasource. Here u can set/check the Soucre System fields.
    2)Create Transformation for that datasource. (no more update rules/transfer rules)
    2.1) While creating transformation for DS it will ask you for data target name, so just assign where u want to update ur data.
    DataSource -> Transformation -> Data Target
    Now if you want to load data into data target from Source System Datasource:
    1) Create infopackage for that data source. If you are creating infopackage for new datasources, it will only allow you update upto PSA, all other options u can see as disabled.
    2)Now Create DTP (Data Transfer Process) for that data source.
    3) NOw schdule the Infopackage, once the data is loaded to PSA, you can execute your DTP which will load data to data target.
    If you are loading data from one one data target to other, no need to use PSA, you can directly execute DTP in that case.
    Data Source -> Transformation (IP/DTP) -> Data Target1 -> DTP ->Data Target 2
    Use the below link for detailed example:
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/fc61e12d-0a01-0010-2883-e2fc63ef729b
    Infosources are no more mandatory with BI 7.0, below is the link to scenarios where we use infosources:
    http://help.sap.com/saphelp_nw04s/helpdata/en/44/0243dd8ae1603ae10000000a1553f6/content.htm
    Full or delta depends on your requirement...
    chk the below thread to know better
    difference between the various loads
    hope it helps
    Message was edited by:
            sriram viswanathan

  • Data loading from DSO to Cube

    Hi,
    I have a question,
    In book TBW10 i read about the data load from DSO to InfoCube
    " We feed the change log data to the InfoCube, 10, -10, and 30 add to the correct 30 value"
    My question is cube already have 10 value, if we are sending 10, -10 and 30 Values(delta), the total should be 40 instead of 30.
    Please some one explaine me.
    Thanks

    No, it will not be 40.
    It ll be 30 only.
    Since cube already has 10, so before image ll nullify it by sending -10 and then the correct value in after immage ll be added as 30.
    so it ll be like this 10-10+30 = 30.
    Thank-You.
    Regards,
    Vinod

  • How to get data back from an action ?

    Hello,
    Would it be possible to get data back from an action (out of the question EO_MESSAGE & ET_FAILED_KEY )?
    For example we got an order with order positions and we need a "function" to e.g. count all positions. Because of Performance the function should not be processed each time the order is changed, read or a position is added. Instead the function should be processed only if it was called explicitly.
    Is it possible to create a kind of action which is actually counting all entries and export the number of them?
    How to mark a parameter in is_parameters as exporting?
    Is this just done by (naming-) convention?
    What is the preferred way to have “methods” with returning/exporting values?
    Regards,
    Lorenz

    Hello Lorenz,
    As you have already figured out , the Action API provides you with only the messages and failed keys if any.
    Post action execution , you can always execute a retrieve or retrive by association , to get the latest buffer snapshot , which of course would include the changes that you have made in your action.
    If you want to ensure that users have explict control on execution of your "fucntion", then of course , you should model it as an action on the BO.
    The parameter is_parameters is an IMPORTING parameter. You CANNOT use it to export anything back from the action. For importing ,  you can of course have any structure to use as the is_paramaters , which you model as the action parameter structure which modelling your BO action.
    From an external entity the only way to interact with a BO is by consuming the BO services and you are bound by the BOBF standard interfaces. Any and all data you require needs to be modelled as node attributes ( persistent or transient ) and fetched using the RETRIEVE, RETRIEVE_BY_ASSOCIATION or QUERY services.
    Regards,
    Indranil.

Maybe you are looking for

  • Getting oracle.jbo.RowCreateException

    Hi All, I have created one custom page from which i have to add a record in a custom table. I have created AM,VO,EO Now After running this page I am getting following error on upper side of page: oracle.jbo.RowCreateException: JBO-25017: Error while

  • I upgrated iTunes and now my 3G iPod Nano won't get picked up

    I believe the iTunes upgrade errased a file of the older version which allowed my laptop recognising my iPod. How can I determine which file is this?? How can I get it back?? I've tried to find it and download it for the Apple website, but I haven't

  • BPM "Fork" step question

    Concerning  fork step : Here is the documentation : You use a fork () when you want to continue a process in branches that are independent of each other, for example, to communicate with two systems that are independent of each other. The branches of

  • Annoying Safari problem after Mountain Lion update...

    I've updated my Macbook Pro to Mountain Lion OS X a while ago and now use Safari 6.0 as my main browser. The zoom button (green button with the plus sign) on the top left corner doesn't seem to be working properly, though. Everytime I click on it, th

  • SCJP 6 study guide: error in book?

    Hi, In the example on page 607 in the box "exam watch"; The pre-java 5 example says: List test = new ArrayList(); test.add(43); ...But on page 568 in the section "Autoboxing with Collections" it says in the pre-java 5 example: List myInts = new Array