Report performance and data quality

Hi,
Can someone help give explanations to following questions :
1.) Does BW Report show how current is my data?
2.) What are the reason why the performance of my BW Report is slow?
3.) What are the reason why my BW Report is have missing data?
4.) Why is my BW Report have incorrect data?
5.) Why doesnu2019t my BW Report Data match SAP R/3 Data?
Thanks,
Milind
Please do not raise generic questions across multiple forums
Edited by: Arun Varadarajan on Apr 9, 2010 2:08 AM

Milind,
1.) Does BW Report show how current is my data?
You should be able to see the data currency when you run in the web - which method are you using - BEx or Web...?
2.) What are the reason why the performance of my BW Report is slow?
It could be due to anything - please search the forums for the same on how to identify possible performance bottlenecks
3.) What are the reason why my BW Report is have missing data?
It depends - Missing data loads etc etc
4.) Why is my BW Report have incorrect data?
You should be knowing that...? I can just say that it has incorrect data because ...." The sun rises in the east...".!!! more akin to asking "Why did the chicken cross the road"
5.) Why doesnu2019t my BW Report Data match SAP R/3 Data?
You should ask SAP that question...
Honestly I am not sure what the reason behind such generic questions are.... if you are looking for answers - then you need to be more specific - if these are more like Interview Questions asked to you - I guess you should be able to answer them or ask further questions to clarify the question further....

Similar Messages

  • Data Integrator and Data Quality

    Hi,
    I worked on BI 7.0 and now exploring Business objects. I wanted to know how Data Integrator and Data Quality products relate to SAP BI 7. What is the purpose of using them. Is this integration and quality features not embedded in SAP BI 7.

    BI and Data Services share many integration features, but Data Service provides better connectivity to third party data sources.  The data quality features of Data Services are not available in BI.

  • ODI Data Profiling and Data Quality

    Hi experts,
    Searching about ODI features for data profiling and data quality I found (I think) many ... extensions? for the product, that confuse me. Cause I think there are at least three diferents ways to do data profiling and data quality:
    In first place, I found that ODI has out of the box features for data profiling and data quality, but, acording to the paper, this features are too limited.
    The second way I found was the product Oracle Data Profiling and Oracle Data Quality for Oracle Data Integrator 11gR1 (11.1.1.3.0) that is in the download page of ODI. Acording to the page, this product extends the existing inline Data Quality and Data profiling features of ODI.
    Finally, the third way is Oracle Enterprise Data Quality that is another product that can be integrated to ODI.
    I dont know if I understood good my alternatives. But, in fact, I need a general explanation of what ODI offer to do Data Quality and Data profiling. Can you help me to understand this?
    Very thanks in advance.

    Hi after 11.1.1.3 version of ODI release, oracle no longer supports ODP/ODQ which is a trillium software product and this not owned by oracle. Oracle is recommending the usage OEDQ for quality purposes. It's better if you could spend time on OEDQ rather than trying to learn and implement ODP/ODQ in ODI

  • Oracle Data Profiling and Data Quality

    Hi,
    How to create metabase for Oracle Data Profiling and Data Quality.Is metabase and repository are same.

    Hi,
    You can create a metabase in the Metabase Manager:
    - Expand Control Admin
    - Click on Metabases
    - in the Metabases window, right-click on the white area and select Add...
    - go through the wizard to create your metabase
    This is documented in the ODQ/ODP tutorial (http://www.oracle.com/technology/products/oracle-data-quality/pdf/oracledq_tutorial.pdf) and in the Documentation (in Metabase Manager or Oracle Data Quality go to Help and then Manuals).
    Thanks,
    Julien

  • Report, Form, and Data Component portlets go to error page

    I am able to add Report Form and Data Component portlets to my page, but when I click on the define button it takes me to my error page.
    Have I missed configuring something?

    hello friends,
    thanks for replly,
    i m using command noe after table node in main window, but i want display the data below the main window , it is diplaying  new page as required but only missing data below the main window.
    to be more clear i want to print all the window in first page then only go for next page , so is there any condition or parmeter in main window to go next page only agfter completion of first page.
    dont worry abt points.

  • QUERY PERFORMANCE AND DATA LOADING PERFORMANCE ISSUES

    WHAT ARE  QUERY PERFORMANCE ISSUES WE NEED TO TAKE CARE PLEASE EXPLAIN AND LET ME KNOW T CODES...PLZ URGENT
    WHAT ARE DATALOADING PERFORMANCE ISSUES  WE NEED TO TAKE CARE PLEASE EXPLAIN AND LET ME KNOW T CODES PLZ URGENT
    WILL REWARD FULL POINT S
    REGARDS
    GURU

    BW Back end
    Some Tips -
    1)Identify long-running extraction processes on the source system. Extraction processes are performed by several extraction jobs running on the source system. The run-time of these jobs affects the performance. Use transaction code SM37 — Background Processing Job Management — to analyze the run-times of these jobs. If the run-time of data collection jobs lasts for several hours, schedule these jobs to run more frequently. This way, less data is written into update tables for each run and extraction performance increases.
    2)Identify high run-times for ABAP code, especially for user exits. The quality of any custom ABAP programs used in data extraction affects the extraction performance. Use transaction code SE30 — ABAP/4 Run-time Analysis — and then run the analysis for the transaction code RSA3 — Extractor Checker. The system then records the activities of the extraction program so you can review them to identify time-consuming activities. Eliminate those long-running activities or substitute them with alternative program logic.
    3)Identify expensive SQL statements. If database run-time is high for extraction jobs, use transaction code ST05 — Performance Trace. On this screen, select ALEREMOTE user and then select SQL trace to record the SQL statements. Identify the time-consuming sections from the results. If the data-selection times are high on a particular SQL statement, index the DataSource tables to increase the performance of selection (see no. 6 below). While using ST05, make sure that no other extraction job is running with ALEREMOTE user.
    4)Balance loads by distributing processes onto different servers if possible. If your site uses more than one BW application server, distribute the extraction processes to different servers using transaction code SM59 — Maintain RFC Destination. Load balancing is possible only if the extraction program allows the option
    5)Set optimum parameters for data-packet size. Packet size affects the number of data requests to the database. Set the data-packet size to optimum values for an efficient data-extraction mechanism. To find the optimum value, start with a packet size in the range of 50,000 to 100,000 and gradually increase it. At some point, you will reach the threshold at which increasing packet size further does not provide any performance increase. To set the packet size, use transaction code SBIW — BW IMG Menu — on the source system. To set the data load parameters for flat-file uploads, use transaction code RSCUSTV6 in BW.
    6)Build indexes on DataSource tables based on selection criteria. Indexing DataSource tables improves the extraction performance, because it reduces the read times of those tables.
    7)Execute collection jobs in parallel. Like the Business Content extractors, generic extractors have a number of collection jobs to retrieve relevant data from DataSource tables. Scheduling these collection jobs to run in parallel reduces the total extraction time, and they can be scheduled via transaction code SM37 in the source system.
    8). Break up your data selections for InfoPackages and schedule the portions to run in parallel. This parallel upload mechanism sends different portions of the data to BW at the same time, and as a result the total upload time is reduced. You can schedule InfoPackages in the Administrator Workbench.
    You can upload data from a data target (InfoCube and ODS) to another data target within the BW system. While uploading, you can schedule more than one InfoPackage with different selection options in each one. For example, fiscal year or fiscal year period can be used as selection options. Avoid using parallel uploads for high volumes of data if hardware resources are constrained. Each InfoPacket uses one background process (if scheduled to run in the background) or dialog process (if scheduled to run online) of the application server, and too many processes could overwhelm a slow server.
    9). Building secondary indexes on the tables for the selection fields optimizes these tables for reading, reducing extraction time. If your selection fields are not key fields on the table, primary indexes are not much of a help when accessing data. In this case it is better to create secondary indexes with selection fields on the associated table using ABAP Dictionary to improve better selection performance.
    10)Analyze upload times to the PSA and identify long-running uploads. When you extract the data using PSA method, data is written into PSA tables in the BW system. If your data is on the order of tens of millions, consider partitioning these PSA tables for better performance, but pay attention to the partition sizes. Partitioning PSA tables improves data-load performance because it's faster to insert data into smaller database tables. Partitioning also provides increased performance for maintenance of PSA tables — for example, you can delete a portion of data faster. You can set the size of each partition in the PSA parameters screen, in transaction code SPRO or RSCUSTV6, so that BW creates a new partition automatically when a threshold value is reached.
    11)Debug any routines in the transfer and update rules and eliminate single selects from the routines. Using single selects in custom ABAP routines for selecting data from database tables reduces performance considerably. It is better to use buffers and array operations. When you use buffers or array operations, the system reads data from the database tables and stores it in the memory for manipulation, improving performance. If you do not use buffers or array operations, the whole reading process is performed on the database with many table accesses, and performance deteriorates. Also, extensive use of library transformations in the ABAP code reduces performance; since these transformations are not compiled in advance, they are carried out during run-time.
    12)Before uploading a high volume of transaction data into InfoCubes, activate the number-range buffer for dimension IDs. The number-range buffer is a parameter that identifies the number of sequential dimension IDs stored in the memory. If you increase the number range before high-volume data upload, you reduce the number of reads from the dimension tables and hence increase the upload performance. Do not forget to set the number-range values back to their original values after the upload. Use transaction code SNRO to maintain the number range buffer values for InfoCubes.
    13)Drop the indexes before uploading high-volume data into InfoCubes. Regenerate them after the upload. Indexes on InfoCubes are optimized for reading data from the InfoCubes. If the indexes exist during the upload, BW reads the indexes and tries to insert the records according to the indexes, resulting in poor upload performance. You can automate the dropping and regeneration of the indexes through InfoPackage scheduling. You can drop indexes in the Manage InfoCube screen in the Administrator Workbench.
    14)IDoc (intermediate document) archiving improves the extraction and loading performance and can be applied on both BW and R/3 systems. In addition to IDoc archiving, data archiving is available for InfoCubes and ODS objects.
    Hope it Helps
    Chetan
    @CP..

  • Question on CKM and Data Quality

    As I understand, CKM only supports the check based on db constraints. If I want to have more complicated business logic built to the checking, does Data Quality sound like a good choice. Or other suggestions?
    In my case, I will need to check the data in the source table based on the table data from different sources (both target and source tables). This should be doable through data quality, correct? I am new to ODI. When I first installed the ODI, I didn't choose to install data quality module. I suppose I can install DQ separately and link it back to ODI? Do they share the same master repository?
    Sorry for the naive questions, you help is greatly appreciated.
    -Wei

    Hi,
    My idea is just like:
    for instance a FK validation:
    create function F$_validate_FK (value1 number) return number
    as
    v_return number;
    begin
    select my_fk into v_return from any_table where column_fk = value1 ;
    return v_return ;
    exception
    When NO_DATA_FOUND then
    return -1;
    end;
    And at constraint you will have:
    -1 = (select F$_validate(table1.column_to_be_validated) from dual)
    Any record that have -1 as return will be not valide for the flow.
    The F$ function can be created in a ODI procedure before the interface and dropped at end if you think to be necessary.
    Make any sense?
    (Maybe there are several syntax error in this example, I just write it and did not compilate, just to show the idea)
    Edited by: Cezar Santos on 28/04/2009 10:20

  • Data Services and Data Quality Recommnded Install process

    Hi Experts,
    I have a few questions. We have some groups that have requested Data Quality be implemented along with another request for Data Services to be implemented. I've seen the requested for Data Services to be installed on the desktop, but from what I've read, it appears to be best to install this on the server side to allow for more of a central benefit to all.
    My questions are:
    1. Can Data Services (Server) install X.1 3.2 be installed on the same server as X.I 3.1 SP3 Enterprise?
    2. Is the Data Services (CLIENT) Version dependent on if the Data Services (Server) install is completed? Basically can the u201CData Services Designeru201D be used without the Server install?
    3. Do we require a new License key for this or can I use the Enterprise Server license key?
    4. At this time we are not using this to move data in and out of SAP, just using this to read data that is coming from SAP.
    From what I read, DATA Services comes with the SAP BusinessObjects Data Integrator or SAP BusinessObjects Data Quality Management solutions. Right now it's seems we dont have a need for the SAP Connection supplement, but definetly something we would implement in the near future. What would be the recommended architecture? A new Server with tomcat and cmc (seperate from our current BOBJ Enterprise servers)? or can DataServices be installed on the same?
    Thank you,
    Teresa

    Hi Teresa.
    Hope you are referring to BOE 3.1 (Business Objects Enterprise) and BODS (Business Objects Data Services) installation on the same server machine.
    Am not an expert on BODS installation.
    But this is my observation :
    We had recently tested on a test machine BOE BOXI 3.1 SP3 (full build) installation before upgrade of our BOE system.
    We also have BODS in our environment.
    Which we also wanted to check whether we could keep on the same server.
    So on this test machine, which already has BOXI 3.1 SP3 build, when i installed BODS server installation,
    what we observed was that,
    all the menus of BOE went away
    and only menus of BODS were seen.
    May be BODS installation overwrites/ or uninstalls BOE, if it already exists ?
    I dont know.  Though i could not fine any documentation, saying that we cannot have BODS and BOE on the same server machine. But this is what we observed.
    So we have kept BODS and BOE on 2 different machines running independently and we do not see any problem.
    Cheers
    indu

  • Reports, Dashboard and Data Import Documentation

    Do you know where I can find documentation regarding Reports and Analytics, Dashboard and data import release 18?
    Thanks

    Click on Training and Support link and search for "*Release 18 Resources*".
    or click on this link for the resources
    http://download.oracle.com/docs/cd/E15799_01/homepage.htm

  • Cube Performance and Data Explosion

    Hi Experts,
    One of the parterns developed a data warehouse application. And the DW application has some performance issue:
    when the report get the query of the high level dimensions, the performance is okey, and when the query get the very detail data in the cube, the performance gets bad.
    The aggregations and the detail data are all stored in the cube, and the cube data gets Explosion quite quickly since some detailed transaction data need to be queried and stored in the cube too.
    So, experts, do you have any good suggestion on this issue? or if the may be a better design for the cube? e.g. in DW, the cube only stores the aggregations or summary on coarse grained data and measure, for fine grained data , it can be got in ODS.
    another question, I google the architecture solution for the above issue, and someone said that if the DW is designed in a hypercube, there maybe data explosion issue, but instead of desinging the hypercube, multicube should be used, so I wonder if multicube can solve the data explosion issue, and how to solve it. And if the multicube has better performance than the hypercube or can also solve detail data query.
    Last question, do you have any experience on DW implementation on TB level Data, and any good suggestion for architecture design using Oracle OLAP or Essbase for good performance.
    Thanks,
    Royal.
    Edited by: Royal on 2012-11-4 上午4:01

    You have not asked any specific technical question. In my opinion all Oracle Datawarehouses should use Oracle OLAP option for the Aggregation strategy. Significant improvements in 11.2.0.2 (and later versions) have been made. It has become much easier now to create and maintain dimensions/cubes. On the reporting side, OBIEE 11g now understands OLAP metadata. Other reporting tools can use the CUBE_TABLE views.
    Here are some links that you may find useful.
    Comparing MVs and OLAP... Oracle White paper
    http://www.oracle.com/technetwork/database/bi-datawarehousing/comparison-aw-mv-11g-twp-130903.pdf
    Oracle OLAP Support page
    https://support.oracle.com/CSP/main/article?cmd=show&type=NOT&id=1107593.1
    Three demos done by OLAP Development which explains how OLAP can help in a DW.
    http://download.oracle.com/otndocs/products/warehouse/olap/videos/intro_part_1/OLAP_Features_and_Use_Cases_1.html
    http://download.oracle.com/otndocs/products/warehouse/olap/videos/intro_part_2/OLAP_Features_and_Use_Cases_2.html
    http://download.oracle.com/otndocs/products/warehouse/olap/videos/intro_part_3/OLAP_Features_and_Use_Cases_3.html
    Main OLAP page at Oracle OTN site
    http://www.oracle.com/technetwork/database/options/olap/index.html
    Recommended Releases for Oracle OLAP
    http://www.oracle.com/technetwork/database/options/olap/olap-certification-092987.html
    Accelerating Data Warehouses using OLAP option
    http://www.oracle.com/technetwork/issue-archive/2008/08-may/o38olap-085800.html
    What's new in 11.2.0.2 database OLAP option
    http://docs.oracle.com/cd/E11882_01/olap.112/e17123/whatsnew.htm
    Oracle 11.2 OLAP Documentation (scroll down to OLAP section)
    http://www.oracle.com/pls/db112/portal.portal_db?selected=6&frame=#online_analytical_processing_%28olap%29
    Excel reporting from OLAP using Simba tool. This was developed in partnership with Oracle.
    http://www.simba.com/MDX-Provider-for-Oracle-OLAP.htm
    There is a good demo for Simba Excel tool at:
    http://www.simba.com/demos/MDX-Provider-for-Oracle-OLAP-web-demo.html

  • Performance and data werehouse

    Hello! 
    How enabling data warehouse affects on performance of the SQL server and what recommendations exists about it?
    Thank you!

    Are you referring to the Management Datawarehouse that comes with the Data collector ?
    In that case it does not affect ther server significantly. If you are collecting data from many different servers, I would recommend to place the managementdw database on separate disk and be carefull with the retention period.
    Regards
    Rasmus Glibstrup
    http://blog.sqlguy.dk

  • Problems with Photoshop performance and data transfer speed on iMac

    Two months ago, I started noticing slow performances using Photoshop (above all using clone stamp tool) on my 27" iMac (late 2012). I did the AHT and I found that 8GB of 32GB RAM were broken.
    I removed them but the problem didn't disappered, I also noticed that data transfer speed (both copy and paste from/to internal HD and from CF card/external HD) was really slow.
    I tried many solutions suggested by Apple support, none of them worked out. At the end, I tried uninstalling and re-installing Photoshop: no more problems!!!
    10 days ago, I received a new 8GB RAM module and so I installed it back... suddenly, the problem came back, I tried re-installing again Photoshop but the problem, this time, still persist!
    Does anyone had the same experience? All other CC programs work well (LR, AE, Premiere...)

    yes, it does!
    what seems to be very strange to me is how data trasnfer speed could be affected!
    (just to say, I've already tried reset of SMC and PRAM, I've tried with different accounts and I've also re-installed the OS, next step would be formatting the disk and installing the OS from zero)

  • What's the Difference Between Better Performance and Better Quality

    I am making a DVD of a wedding from a tape and I usually create an image and then transfer it to my PC to burn because my Mac Mini doesn't have a DVD burner. It is telling me that my project has exceded 4.2 GB which I figured would happen but would dread if it did. Well i was wondering what type of quality loss would happen if I used Best Quality? Or is there a way to make a Disc image that I dont know of? Any suggestions would be helpful.

    http://docs.info.apple.com/article.html?artnum=164975
    If you don't have a DL SuperDrive, I don't think you can create a disk image over two hours.
    Well i was wondering what type of quality loss would happen if I used Best Quality?
    Best Quality is just that, best quality encoding. I use it for every project.

  • MAC Mini Performance and display quality help needed!

    I am new to the mac, having used pcs for what feels like hundreds of years.
    I am enjoying the experience, but my MAC mini is so slow. Its a 2014 i5 2.6 Ghz with 8GB of RAM.
    For example starting safari takes 20+ seconds from clicking the icon to being able to type a web address.
    Also the display is not brilliant the text is blurred. I am using a LG HD Monitor running at 1920 x 1080 and 60 Hz.
    I have a Dell PC with very similar specs connected to the same monitor and it does not suffer blurred display.
    Thanks for your help

    EtreCheck version: 2.2 (132)
    Report generated 5/4/15, 11:26 AM
    Download EtreCheck from http://etresoft.com/etrecheck
    Click the [Click for support] links for help with non-Apple products.
    Click the [Click for details] links for more information about that line.
    Hardware Information: ℹ️
        Mac mini (Late 2014) (Technical Specifications)
        Mac mini - model: Macmini7,1
        1 2.6 GHz Intel Core i5 CPU: 2-core
        8 GB RAM Not upgradeable
            BANK 0/DIMM0
                4 GB DDR3 1600 MHz ok
            BANK 1/DIMM0
                4 GB DDR3 1600 MHz ok
        Bluetooth: Good - Handoff/Airdrop2 supported
        Wireless:  en1: 802.11 a/b/g/n/ac
    Video Information: ℹ️
        Intel Iris
            IPS237 1920 x 1080
    System Software: ℹ️
        OS X 10.10.3 (14D136) - Time since boot: 0:9:48
    Disk Information: ℹ️
        APPLE HDD HTS541010A9E662 disk0 : (1 TB)
            EFI (disk0s1) <not mounted> : 210 MB
            Recovery HD (disk0s3) <not mounted>  [Recovery]: 650 MB
            Macintosh HD (disk1) / : 998.98 GB (641.30 GB free)
                Core Storage: disk0s2 999.35 GB Online
    USB Information: ℹ️
        Apple, Inc. IR Receiver
        Apple Inc. BRCM20702 Hub
            Apple Inc. Bluetooth USB Host Controller
        Microsoft Microsoft® LifeCam Cinema(TM)
        Hercules Rocksmith USB Guitar Adapter
    Thunderbolt Information: ℹ️
        Apple Inc. thunderbolt_bus
    Gatekeeper: ℹ️
        Anywhere
    Kernel Extensions: ℹ️
            /Library/Application Support/VirtualBox
        [loaded]    org.virtualbox.kext.VBoxDrv (4.3.26) [Click for support]
        [loaded]    org.virtualbox.kext.VBoxNetAdp (4.3.26) [Click for support]
        [loaded]    org.virtualbox.kext.VBoxNetFlt (4.3.26) [Click for support]
        [loaded]    org.virtualbox.kext.VBoxUSB (4.3.26) [Click for support]
            /Library/Extensions
        [loaded]    com.sophos.kext.sav (9.2.50 - SDK 10.8) [Click for support]
        [loaded]    com.sophos.nke.swi (9.2.50 - SDK 10.8) [Click for support]
            /System/Library/Extensions
        [not loaded]    com.motorola-mobility.driver.MotMobileUSB (1.2.2 - SDK 10.5) [Click for support]
            /System/Library/Extensions/MotMobileUSB.kext/Contents/PlugIns
        [not loaded]    com.motorola-mobility.driver.MotMobileMS (1.0.0 - SDK 10.5) [Click for support]
        [not loaded]    com.motorola-mobility.driver.MotMobileMTP (1.2.2 - SDK 10.5) [Click for support]
        [not loaded]    com.motorola-mobility.driver.MotMobileUSBLAN (1.2.2 - SDK 10.5) [Click for support]
        [not loaded]    com.motorola-mobility.driver.MotMobileUSBLANMerge (1.2.2 - SDK 10.5) [Click for support]
        [not loaded]    com.motorola-mobility.driver.MotMobileUSBSwch (1.2.2 - SDK 10.5) [Click for support]
    Launch Agents: ℹ️
        [not loaded]    com.adobe.AAM.Updater-1.0.plist [Click for support]
        [loaded]    com.mokafive.m5tray.plist [Click for support]
        [failed]    com.motorola.MDMUpdater.plist [Click for support] [Click for details]
        [running]    com.motorola.motohelper.plist [Click for support]
        [loaded]    com.motorola.motohelperUpdater.plist [Click for support]
        [loaded]    com.oracle.java.Java-Updater.plist [Click for support]
        [running]    com.sophos.uiserver.plist [Click for support]
        [failed]    net.juniper.pulsetray.plist [Click for support] [Click for details]
    Launch Daemons: ℹ️
        [loaded]    com.adobe.fpsaud.plist [Click for support]
        [loaded]    com.ea.origin.ESHelper.plist [Click for support]
        [loaded]    com.microsoft.office.licensing.helper.plist [Click for support]
        [running]    com.motorola-mobility.mmcfgd.plist [Click for support]
        [loaded]    com.oracle.java.Helper-Tool.plist [Click for support]
        [loaded]    com.oracle.java.JavaUpdateHelper.plist [Click for support]
        [running]    com.sophos.common.servicemanager.plist [Click for support]
        [loaded]    jp.co.canon.IJNetworkToolHelper.plist [Click for support]
        [failed]    net.juniper.AccessService.plist [Click for support]
        [not loaded]    net.juniper.UninstallPulse.plist [Click for support]
        [not loaded]    org.virtualbox.startup.plist [Click for support]
    User Launch Agents: ℹ️
        [loaded]    com.google.keystone.agent.plist [Click for support]
        [failed]    com.lastpass.LastPassHelper.plist [Click for support] [Click for details]
        [running]    com.spotify.webhelper.plist [Click for support]
        [loaded]    com.valvesoftware.steamclean.plist [Click for support]
        [not loaded]    org.virtualbox.vboxwebsrv.plist [Click for support]
    User Login Items: ℹ️
        iTunesHelper    Application  (/Applications/iTunes.app/Contents/MacOS/iTunesHelper.app)
        Dropbox    Application  (/Applications/Dropbox.app)
        Google Chrome    Application  (/Applications/Google Chrome.app)
        Canon IJ Network Scanner Selector2    Application Hidden (/Library/Printers/Canon/IJScanner/Utilities/Canon IJ Network Scanner Selector2.app)
        Skype    Application Hidden (/Applications/Skype.app)
    Internet Plug-ins: ℹ️
        FlashPlayer-10.6: Version: 17.0.0.169 - SDK 10.6 [Click for support]
        EPPEX Plugin: Version: 10.0 [Click for support]
        Flash Player: Version: 17.0.0.169 - SDK 10.6 [Click for support]
        Default Browser: Version: 600 - SDK 10.10
        QuickTime Plugin: Version: 7.7.3
        SharePointBrowserPlugin: Version: 14.4.9 - SDK 10.6 [Click for support]
        nplastpass: Version: 3.1.89 - SDK 10.10 [Click for support]
        Silverlight: Version: 5.1.30514.0 - SDK 10.6 [Click for support]
        MeetingJoinPlugin: Version: Unknown - SDK 10.6 [Click for support]
        JavaAppletPlugin: Version: Java 8 Update 40 Check version
    Safari Extensions: ℹ️
        Evernote Web Clipper
        PasswordBox
    3rd Party Preference Panes: ℹ️
        Flash Player  [Click for support]
        FUSE for OS X (OSXFUSE)  [Click for support]
        Java  [Click for support]
        MusicManager  [Click for support]
    Time Machine: ℹ️
        Time Machine not configured!
    Top Processes by CPU: ℹ️
             4%    WindowServer
             1%    fontd
             1%    loginwindow
             1%    Google Chrome Helper(13)
             0%    taskgated
    Top Processes by Memory: ℹ️
        909 MB    Google Chrome Helper(13)
        674 MB    kernel_task
        467 MB    mdworker(12)
        401 MB    com.apple.WebKit.WebContent(5)
        180 MB    SophosScanD
    Virtual Memory Information: ℹ️
        1.24 GB    Free RAM
        6.76 GB    Used RAM
        0 B    Swap Used
    Diagnostics Information: ℹ️
        May 4, 2015, 10:08:53 AM    /Users/[redacted]/Library/Logs/DiagnosticReports/Google Chrome_2015-05-04-100853_[redacted].crash
        May 2, 2015, 07:47:11 AM    /Users/[redacted]/Library/Logs/DiagnosticReports/EvernoteHelper_2015-05-02-0747 11_[redacted].crash
        May 4, 2015, 11:13:54 AM    Self test - passed
        May 4, 2015, 10:02:56 AM    /Library/Logs/DiagnosticReports/Microsoft AutoUpdate_2015-05-04-100256_[redacted].cpu_resource.diag [Click for details]
        May 2, 2015, 09:13:07 AM    /Library/Logs/DiagnosticReports/Microsoft AutoUpdate_2015-05-02-091307_[redacted].cpu_resource.diag [Click for details]

  • Hard drive performance and data throughput

    I am using my macbook pro for work primarily and part of that entails creating/restoring images of other macs. I've had the best luck with SuperDuper however the process is still VERY slow. For instance at this moment with no other applications open other than S.D. and firefox the copy speed is under 5MB/s from my MBP to an iMAC via fire wire.
    I am looking for suggestions to increase the performance/IO in the hopes to speed up the process. When purchasing this system the 7200rpm drive was not an option (15") which is unfortunate. I realize that both hard drives in the operation will cause the variable in speed but I want the sending drive as fast as possible.
    My thoughts right now are to purchase a 7200rpm external drive to store backup images and also send from. This would cut out any possible IO on the drive that my mac is performing to run the operating system. Another thought was to upgrade my mac to a 7200rpm drive and use the current 5400rpm drive as the storage for images...in the hopes that it would still provide an increase in restoration speed since it wouldn't be running OSX on it.
    Any thoughts or ideas? Experiences? My MBP has the 5400rpm I believe and 2GB of ram.
    Thanks

    I'll try and explain a bit better. I'm not restoring
    the same image to different types of macs. I create
    images of OTHER macs using my macbook pro to perform
    the process as well as store the backup image.
    Thanks for the clarification. I do that too, but when I do I use my Mac Pro to clone a Mac via Target Disk Mode to an external FireWire 800 drive.
    it helps but its a usb2 enclosure with a somewhat
    older hard drive that is only 30Gb. I am looking at
    purchasing a firewire 800 external drive but I will
    see how this other unit works for now since we
    already have it.
    Part of your throughput problem may be the overhead issues with USB 2. FireWire uses its own chipset so is more independent of the CPU, and FireWire can sustain high-speed transfers at a higher level. USB is CPU-bound and is more vulnerable to CPU demands from other apps or background processes or other USB devices. So even though USB 2 has a higher theoretical peak (480Mbps), FireWire (400Mbps) actually does better in the real world.
    About USB 2 vs. FireWire 400 performance
    I'm not sure if FireWire 800 would help because your slowest drive in the chain may not be fast enough to take advantage.

Maybe you are looking for

  • Not sure if i have enough usage for iphone

    Hi, just bought an iphone 3gs - and im worried about it lol i decided to connect to 3 and use the lowest plan ($39cap) that doesn't have much internet bandwidth (150mb). Will 150mb of data be enough to safely use my iphone with out exceeding its usag

  • How to set the page size

    Hi, How to set the page size of a B tree node in berkeley db java edition? thanks.

  • Firmeware

    Firmeware password reset on Apple macbook pro IOS 8.1 / update Mavericks -- douple tuning Microssoft win 7 Firmeware MISS in German Grad Zeichen Kennword delete --- dont come IN   moment based firmeware!!!??? my Numbers!? pleace check          C0****

  • How to access "removable devices" on 'new This pc to allow upload of pics from cell

    how do I access the "removable devices" on the former my computer, now this pc to u/l pics from my cell to 'puter?

  • Go-Pro video cameras, Tulsa Hills

    What kind of go-pro video cameras do you have at the Tulsa Hills location and how much are they?