How fastest is the storage in flash to compensate a less powerful processor?

Hi, i have a MacPro mid 2012 with a 12 core Intel Xeon E5645 processor and i want to buy the new MacPro. This new MacPro starts with a 6 core processor, but it has a faster storage. How fastest is the storage in flash to compensate a less powerful processor?

Intel® Xeon® Processor E5645
(12M Cache, 2.40 GHz, 5.86 GT/s Intel® QPI)
http://ark.intel.com/products/48768/Intel-Xeon-Processor-E5645-12M-Cache-2_40-GH z-5_86-GTs-Intel-QPI
Firstly, those are/were the slowest clocked cpu offered and probably do hold you back, esp when there are cpu you could drop in that offer 1GHZ more per core.  You could easily just throw in a pair of 3.4GHz 6-core now, right?
All We Know About Maximizing CPU Related Performance (  1 2 3 4 5 6 ... Last Page)
Mac Pro 5,1 - E5620 to W5590 CPUs
Hex 3.2 update from 2010 quad 2.8 photos new! CPU to use! (  1 2 3 4 5 6 ... Last Page)  
... and there are more, but gives you something to consider.
Also user reports  AccelerateYourMac

Similar Messages

  • How to change the storage location in Lightroom 3.6.?

    How to change the storage location in Lightroom 3.6.?
    I use my internal hard drive NTFS of my MacBook Pro  as my storage location of Lightroom 3.6. being installed in Snow Leopard now but the data of my photos get too big now. I have two fast and big external hard drives eSATA now which I want to use for my data/photos. I know how to copy the data from one hard drive to another but I don't know how to change the settings in connection with Lightroom easily. I still want to watch all fotos on my MacBook Pro but don't want to save any fotos on my internal harddrive any more. Isn't it possible to have the programm Lightroom on my internal harddrive but the data on two times on two different extermal hard drives? If I look at Lightroom - settings of catalogue - I see that the internal harddrive location is my storage location now but I my two external harddrives don't appear as folders. Why? How to change that so that the external hard drives appear in Lightroom and I can watch all my fotos immediately (sometimes using external hard drive 1 and sometimes using external hard drive 2)?
    My external hard drives are in HFS+ but I use the software of Paragon so that there shouldn't be any problem to read and write the data (I use also Final Cut Express so my external hard drives are HFS+ and not NTFS).
    I want to impert many data from CDs - should I copy the fotos first to my external hard drive and then import it into Lightroom or is it easier to import the data directly from the CD in Lightroom so that I don't have to name the folders two times?
    I see that I have to sort the fotos from the CDs in new different folders and an used to do that in Finder of my MacBookPro. In Lightroom I don't want have to click each of hundreds of fotos to be able to put it into the right folders ... But how to import only a part of a CDs quickly into the right (newly created) folders?
    Kind regards, Karin

    Would you really leave the names of photos like DSC05271.ARW and DSC05271.jpeg etc.?
    What if I name them f.i. like ...
    2013-09-11_Austria_1_mountain.jpeg
    2013-09-11_Austria_2_mountain.jpeg
    2013-09-11_Austria_3_valley.jpeg
    I would NOT do this. I would not change the file name.
    I would add keywords and other metadata to the photos to identify the content of the photos, and from that point on, I would use keywords and other metdata to search for these photos.
    However, let's look at the big picture. I think there are three different methods of organizing
    Organize via keywords and other metadata, and not via operating system constructs such as file name and folder name
    Organize via operating system constructs such as file name and folder name, and not via keywords and other metadata
    Some combination of 1 and 2
    You can choose any one of the three methods above, the choice is yours. I am a very strong believer that for most people (and it sounds like you are one of those people), method 1 is the best method. However, not everyone agrees.
    There are drawbacks to using file names to identify your photos. One big drawback is that you must type the information into each photo's name, and this is more tedious than if you wanted to assign the keyword "mountain" to the photos, you can assign the keyword to multiple photos at once. Also, if you mis-spell the information in the file name on any photo, you have just made the photo much harder to find; whereas in Lightroom, the keyword can be assigned via a mouse click (so you can't mis-spell it), or by typing the first few letters and Lightroom (auto-fill) will find the correct keyword name. So using keywords is a much simpler way to go.
    .. is this much too complicated to do that in Lightroom or even dangerous to got lost of photos?
    It is very simple to add keywords, as I explained for this case. Create keywords "Austria", "Mountain", "Valley"; then select the desired photos, and then click on the check box next to the keyword name. It is much more complicated in my opinion to rename the photos in this manner.
    There is less danger in doing this via keywords than the danger doing this via file names, because of potential mis-spellings.
    If I spend a lot of time in editing photos (in Lightroom or Photoshop) I don't want that I loose those photos f.i. later if Lightroom won't be the software I use regarding the photos (in 10, 20 years?!)
    Lightroom will OPTIONALLY write your keywords and other metadata to the files (or sidecar files in the case of RAW). Every photographic application I know of (and I'm sure those in the future) can read the keywords and other metadata. So if you ever want to switch to some other software, your keywords and other metadata are available. you can turn this option on via Edit->Catalog Settings->Metadata, check Automatically Write Changes to XMP.
    So I have to create two folders - one with the originals and one with the edited exported photos?
    No, you do not have to create two folders. You don't have to export everything, but if you do, you can put the exports wherever you want, including putting them in the same folder as the originals, with different name.
    Why not edit them in Photoshop and save them in the second folder which I import later in Lightroom? Because it takes longer?
    Wow, you really need to think about: Why are you using Lightroom?
    Could you explain this to me? Why are you using Lightroom?
    If your goal is to edit the photos in Photoshop, and then put them in folders with whatever custom name you want to give them, then what is the benefit of using Lightroom? You can do all of this WITHOUT Lightroom. You keep explaining that your intended goal is to use Photoshop and folders/file names, and this is a goal which AVOIDs all benefits of Lightroom. I can't understand what you think the benefits of using Lightroom are for you.
    It is starting to sound to me like Lightroom is not the right software for you.
    Some people want to see a movie on their TV and are able to watch a DVD (not all!). Flickr seems to be a solution for movies in a big size (many GBs) to share them with friends (?)
    You haven't mentioned this before, and you should check the rules on Flickr regarding how long a video you can upload.

  • Where and how to find the storage locatation for the consignment stock

    where and how to find the storage locatation for the consignment stock (customer stock )  for more than one storage location .
    table :msku
    material no    -   werks - batch
    xxx                   sg11  - 200352ac
    table : mchb
    material no    -   werks  -lgort - batch
    xxx                   sg11   sg10   200352ac
    xxx                   sg11   gs11   200352ac
    note : each stock location having same batch no in different storage and my question how to find which link for msku to mchb .

    Hi,
    Try the table MARD: Storage Location Data for Material, where in the field LABST will give the stock field, against a given material/plant/Storage Loc.
    Regards,
    JLN

  • Can any one tell me how to stream the images in flash.

    Hello Friends,
    i have created a dynamic page flip book with xml, i want to
    zoom the images in swf loaded, with streaming, so that user can
    preview the image while a good quality image is loading. I wanna
    know that how to load a part of large image or swf.
    Image effect like :-
    http://www.zoomify.com/
    Can any one tell me how to stream the images in flash.

    On Thu, 17 Apr 2008 05:53:27 +0000 (UTC), "FlashDavil"
    <[email protected]> wrote:
    >I wanna know that how to load a
    >part of large image or swf.
    You show a preview and use loadMovie() to load a big file.
    You can
    create a preloader (progress bar) showing how much part is
    loaded,
    while user sees the small image. When you achieve 100% tou
    show that
    big image. What's more?
    > Can any one tell me how to stream the images in flash.
    What are you talking about is not a "streaming", but
    "progressive
    download". Images "streaming" is impossible.

  • HT201238 How to downgrade the storage capacity?  step by step please:-)

    How to downgrade the storage capacity?  step by step please:-)

    Is this document detailed enough?
    See:  iCloud storage upgrades and downgrades - Apple Support
    Scroll down to "Downgrade your iCloud storage plan"

  • How to decrease the storage in macbook air

    how to decrease the storage in macbook air ?

    you mean decrease all the files filling your Air?
    See here for answer about the OTHER which is taking up space:
    http://pondini.org/TM/30.html
    and here:
    http://pondini.org/OSX/DiskSpace.html
    See Kappys excellent note on the rest of “other” files taking up your space:
    What is "Other" and What Can I Do About It?
    In the case of a Macbook Air or Macbook Pro Retina with ‘limited’ storage on the SSD, this distinction becomes more important in that in an ever rapidly increasing file-size world, you keep vital large media files, pics, video, PDF collections, music off your SSD and archived on external storage, for sake of the necessary room for your system to have free space to operate, store future applications and general workspace.  You should never be put in the position of considering “deleting things” on your macbook SSD in order to ‘make space’.
    Professionals who create and import very large amounts of data have almost no change in the available space on their computers internal HD because they are constantly archiving data to arrays of external or networked HD.
    Or in the case of the consumer this means you keep folders for large imported or created data and you ritually offload and archive this data for safekeeping, not only to safeguard the data in case your macbook has a HD crash, or gets stolen, but importantly in keeping the ‘breathing room’ open for your computer to operate, expand, create files, add applications, for your APPS to create temp files, and for general operation.

  • How to Edit the Text in Flash?

    Hi there,
    I want to know that how to change the text on the Flash
    (.Fla)
    I have .Fla and want to edit it with the step by step help.
    Currently, I want to edit the text on the home page.
    http://www.mtkinetics.com/ayc/index.html
    Any help will be highly appreciated.
    Regards
    bushib

    1. click on the text. in the properties panel it will say
    static text, dynamic text or input text.
    2. if it's static text, click on the text tool and click
    inside your text. then edit. if it's not static text, you can do
    the same thing, but the text will be changed by actionscript during
    run-time IF text was originally assigned by actionscript.

  • How to connect external storage devices (flash drive) with it

    hello dear. I am using this hp webOS-3.0.5 device for last 8 months. It is really a wonderful device to have. But I have some queries regarding it's usage. First can we attach any usb device(external storage devices (flash drive) with it? Second is there any other way to use internet on it without wi-fi connection? Thanks
    Post relates to: HP TouchPad (WiFi)

    I'm afraid the answer is No to both questions.  
    External storage can be acheived easily using the free 50GB Box.com account included with each touchpad.  This does rely on a good wifi source however, which leads to your other question.
    The wifi touchpad needs a wifi signal to go online.  Modern cellphones can act as "wifi hotspots" and can translate a cellular data connection to wifi.  (I do this with my Pre3 and a free program called "FreeTether", a homebrew app).
    WyreNut
    I am a Volunteer here, not employed by HP.
    You too can become an HP Expert! Details HERE!
    If my post has helped you, click the Kudos Thumbs up!
    If it solved your issue, Click the "Accept as Solution" button so others can benefit from the question you asked!

  • I have an iPhone4s and my storage has reduced unexplainably due to apps. How can recoup the storage (besides deleting apps because I deleted a lot of my apps already.)

    My phone's storage space is depleting rapidly.  It is mainly apps taking up this space, but I honestly don't have a lot of apps on my phone.  The apps should only take about .6 gigs and they are taking up almost 5 gigs or storage. The apps that are taking up the most room are Instagram and Pinterest, which are internet sights, so to me it doesn't make sense why they are taking up all of the storage room if they are backed up to the internet.  I know I really don't know a lot about this stuff, but it would really help if someone could explain to me what is going on and how I can fix the problem without completely deleting the apps if that is possible.

    Alright, super weird, but looks like it's fixed.  I should note that before I tried anything, I deleted enough stuff so that I had plenty of space left for updates.  I held both buttons down until the phone turned off and then turned it back on.  Still the same issue.  Maybe half an hour later as I'm thinking about how I'll have to try to restore the phone, the screen lights back up on it's own and shows the lock screen...like someone had pressed the home button, even though I pressed nothing.  I unlocked the phone to see if I had a missed call or text or something and noticed that the "stuck" apps had updated and everything was working again. #applemagic

  • How to specify the storage access key for a ResourceFile?

    The azure batch tutorial shows how to put program file into a public container in a storage account and let azure batch to download them to TVMs and run.
    In real world scenario, if I don't want to use a public container or a shared access signature and want the azure batch to use a access key to access the container where my task program file is located, is it possible? How to do it?

    I see that you are conversant with the issues here but for other readers let me provide a quick review:
    The properties of a task (ICloudTask/CloudTask) include a collection of ResourceFile instances. ResourceFile instances
    map blobs in Azure Storage to local files in the Container/VM/Guest-OS.  Azure Batch copies the files from storage into the VM before the task runs and it uses the SAS (and other data) in the ResourceFile to do so.
    The ICloudTask/CloudTask.FilesToStage collection exposes the object model's mechanism for customizable file staging.
     The collection accepts instances of IFileStagingProvider which ultimately are invoked to create/augment the ResourceFile collection on the task.
    A default implementation is provided: FileToStage.
    An instance of FileToStage maps a file local to the client library to a file ultimately in the VM (indirecting through
    blob storage/SAS).  When instances of FileToStage are added to the CloudTask.FilesToStage the following occurs on Commit()/AddTask:
    A container is created in the given storage account.  The name is constructed to avoid collisions.
    The container is given a restricted SharedAccessBlobPolicy.
    All of the local files specified are uploaded to that container
    An SAS for each blob is created
    (24hr expiry)
    and a ResourceFile is constructed for each FileToStage
    The ResourceFile for each FileToStage is added to the CloudTask.ResourceFiles collection.
    FileToStage and the FilesToStage collection are intended to assist the customers that either want a shortcut around the issues of blob containers and SAS or want to control the file staging process via a custom implementation of IFileStagingProvider.
    When using the default implementation FileToStage to stage local files, care should be taken to monitor the number of containers created and the storage cost implications.
    Your concerns about SAS based methods are not directly addressed by the default implementation.  I would only note that SAS values can be re-used across tasks and jobs so the existing implementation can be used to get local data into storage and usable
    SAS values.  However, you already have these sorts of features implemented it seems and as you point out, there is the problem of SAS expiry. 
    daryl

  • How to get the storage details during 5 min interval

    Hi,
    I used below java code to collect the storage details. But i didn't get any output. Without filter condition it display all the records available in table. How to get the
    last five min data. Based on this how to change the tablequery.? I used partition key too. that also didnt work.
    1) How to use the tick count in this example.
    CloudStorageAccount httpAcc = CloudStorageAccount.parse(storageConnectionString);
    CloudTableClient tableClient = httpAcc.createCloudTableClient();
    CloudTable table = tableClient.getTableReference("$MetricsHourPrimaryTransactionsTable");
    Date date = new Date();
    SimpleDateFormat sdf = new SimpleDateFormat("yyyy-MM-dd"); 
    TimeZone utc = TimeZone.getTimeZone("UTC");
    sdf.setTimeZone(utc);
    String upperBound = TableQuery.generateFilterCondition("TIMESTAMP",QueryComparisons.LESS_THAN,sdf.parse("2015-02-04")); // // sdf.format(cal.getTime())
    System.out.println("call lowertime ->"+upperBound);
    String lowerBound = TableQuery.generateFilterCondition("TIMESTAMP",QueryComparisons.GREATER_THAN,sdf.parse("2015-02-01"));
    System.out.println("call uppertime ->"+lowerBound);
    String filter = TableQuery.combineFilters(upperBound, Operators.AND, lowerBound);
    //String startTick = TableQuery.generateFilterCondition("TIMESTAMP",QueryComparisons.GREATER_THAN,sdf.parse("2015-09-03"));
    //System.out.println("the startTick==="+startTick);
    //String tick = TableQuery.generateFilterCondition("PartitionKey",QueryComparisons.GREATER_THAN,startTick);
    TableQuery<DynamicTableEntity> query = TableQuery.from(DynamicTableEntity.class).where(filter).select(new String[] {"Availability","TotalRequests","PercentSuccess"});
    for (DynamicTableEntity ent : table.execute(query))
    EntityProperty countProp = ent.getProperties().get("Availability");
    if (countProp == null) 
    throw new IllegalArgumentException("Invalid entity, Count property not found!");
    System.out.println(countProp.getValueAsString());
    EntityProperty countProp1 = ent.getProperties().get("TotalRequests");
    System.out.println(countProp1.getValueAsString());
    EntityProperty countProp2 = ent.getProperties().get("PercentSuccess");
    System.out.println(countProp2.getValueAsString());
    Regards,
    Rathidevi
    rathidevi

    checkout this document for secure web application
    http://help.adobe.com/en_US/flex/using/WS2db454920e96a9e51e63e3d11c0bf6167e-7fff.html    

  • How to change the storage location using BAPI_OUTB_DELIVERY_CHANGE

    Hi !
    I want to do batch split in the delivery using BAPI_OUTB_DELIVERY_CHANGE.
    Can anyone tell me how to pass/change the storage location of each batch item.
    Is there any other BAPI that can do the batch split and populate the storage location also for the split batches?
    Regards,
    Firoz.

    Hi all,
    BAPI_OUTB_DELIVERY_CHANGE  can be used to do batch split and updating storage loaction against each item of an outbond delivey.
    I have done that in the folllowing way:
    1 > Firstly i have updated the storage location for each delivery item using 'BAPI_OUTB_DELIVERY_CHANGE' passing some mininal parameters.
       Fetch the item details from LIPS table based on the outbound delivery and pass the corresponding fields to item_data, item_control  and item_data_spl parameters and passed into intenal table li_lips.
    Loop at li_lips inti lw_lips.
        lw_item_data-deliv_numb           = lw_lips-vbeln.
        lw_item_data-deliv_item             = lw_lips-posnr.
        lw_item_data-material                = lw_lips-matnr.
        lw_item_data-fact_unit_nom      = lw_lips-umvkz.
        lw_item_data-fact_unit_denom  = lw_lips-umvkn.
        lw_item_data-base_uom            = lw_lips-meins.
        lw_item_data-sales_unit            = lw_lips-vrkme.
        lw_item_control-deliv_numb      = lw_lips-vbeln.
        lw_item_control-deliv_item        = lw_lips-posnr.
        lw_item_data_spl-deliv_numb   = lw_lips-vbeln.
        lw_item_data_spl-deliv_item     = lw_lips-posnr.
        lw_item_data_spl-pick_denial   = 'X'.
        lw_item_data_spl-stge_loc       = v_lgort.
    "(This would be your Storage Location which you want to be updated)
    Appending work areas into internal table to pass as parameter
        APPEND lw_item_data_lgort     TO li_item_data_lgort .
    Appending work areas into internal table to pass as parameter
        APPEND lw_item_control_lgort  TO li_item_control_lgort.
    Appending work areas into internal table to pass as parameter
        APPEND lw_item_data_spl_lgort TO li_item_data_spl_lgort.
    endloop.
    Passing the delivery no in the work area of header data
      lw_header_data-deliv_numb           = v_delivery_no.
      lw_header_control-deliv_numb       = v_delivery_no.
      lw_header_tech_control-upd_ind   = 'U'.
    Calling BAPI to change the Storage location
      CALL FUNCTION 'BAPI_OUTB_DELIVERY_CHANGE'
        EXPORTING
          header_data       = lw_header_data
          header_control   = lw_header_control
          delivery               = v_delivery_no
          techn_control     = lw_header_tech_control
        TABLES
          item_data           = li_item_data
          item_control       = li_item_control
          return                 = li_return_change
          item_data_spl    = li_item_data_spl.
    Calling BAPI to committ the task
        CALL FUNCTION 'BAPI_TRANSACTION_COMMIT' .
    2> Then i have used the same BAPI again 'BAPI_OUTB_DELIVERY_CHANGE' to do the batch split/update batch only(if required) and change the Actual Delivered Quantity.
    Here you have to pass the same thing along with actual delivery qauntity and different batches to do batch split.
    Here it is assumed that Batch numbers, actual delivered quantity are coming from an internal table i_lqua.
    Looping through Internal table to get Bin details
      LOOP AT i_lqua INTO w_lqua .
    Clearing work area before use
        CLEAR : lw_item_data, lw_lips, lw_item_control, lw_item_data_spl.
    Reading table comparing metrial number
        READ TABLE li_lips INTO lw_lips
        WITH KEY matnr = w_lqua-matnr BINARY SEARCH.
    If read is successful, passing values from table
        IF sy-subrc EQ 0.
    Passing the delivery details into Item level table
          lw_item_data-deliv_numb              = lw_lips-vbeln.
          lw_item_data-deliv_item                = lw_lips-posnr.
          lw_item_data-material                   = lw_lips-matnr.
          lw_item_data-batch                       = w_lqua-charg.
          lw_item_data-dlv_qty                    = w_lqua-verme.
          lw_item_data-dlv_qty_imunit         = w_lqua-verme.
          lw_item_data-base_uom               = w_lqua-meins.
          lw_item_data-hieraritem                = lw_lips-posnr.
          lw_item_data-usehieritm                = 1.
          lw_item_data-fact_unit_nom          = lw_lips-umvkz.
          lw_item_data-fact_unit_denom      = lw_lips-umvkn.
          lw_item_data-sales_unit                 = lw_lips-vrkme.
          lw_item_control-deliv_numb           = lw_lips-vbeln.
          lw_item_control-deliv_item             = lw_lips-posnr.
          lw_item_control-chg_delqty           = 'X'.
          lw_item_data_spl-deliv_numb        = lw_lips-vbeln.
          lw_item_data_spl-deliv_item          = lw_lips-posnr.
          lw_item_data_spl-stge_loc            = w_lqua-lgort.
          lw_item_data_spl-pick_denial        = 'X'
    Appending work area into internal table to pass as parameter
          APPEND  lw_item_data TO  li_item_data.
    Appending work area into internal table to pass as parameter
          APPEND lw_item_control TO li_item_control.
    Appending work area into internal table to pass as parameter
          APPEND lw_item_data_spl TO li_item_data_spl.
    Clearing work areas after use
          CLEAR : lw_item_data, w_lqua,lw_item_data_spl,lw_item_control,
                        lw_vbpok, lw_lips.
        ENDIF.
      ENDLOOP.
    Passing the delivery no in the work area of header data
      lw_header_data-deliv_numb           = v_delivery_no.
      lw_header_control-deliv_numb       = v_delivery_no.
      lw_header_tech_control-upd_ind   = 'U'.
    Calling BAPI to change the Batch/Batch-Split/Delivery Quantity
      CALL FUNCTION 'BAPI_OUTB_DELIVERY_CHANGE'
        EXPORTING
          header_data           = lw_header_data
          header_control       = lw_header_control
          delivery                   = v_delivery_no
          techn_control         = lw_header_tech_control
        TABLES
          item_data                = li_item_data
          item_control            = li_item_control
          return                     = li_return_change
          item_data_spl         = li_item_data_spl.
    Calling BAPI to committ the task
        CALL FUNCTION 'BAPI_TRANSACTION_COMMIT' .
    This is the only way which i found as better way to do the Batch split and updating storage location togetherly.
    I hope this code will help you.

  • How to understand the storage systems?

    How to get more and more or complete understanding of the storage systems like understanding the hardware components involved,best configuration for the hardware,best practices.
    Is there any book for the storage knowledge for DBA's.

    LearningOracle,
    I shall second to Sybrand's suggestion for the Scaling Oracle 8i book. That's an awesome book , covering alot of stuff , for both oracle and operating systems. Its available for download from here,
    http://www.scaleabilities.co.uk/book/scalingOracle8i.pdf
    In addition to this, I would say that you start reading the books/notes which are there for SAN and other systems. There is no dedicated book for storage systems for DBAs AFAIK or I should say ,there is none which is very popular. If you are interested in managing the operating system for ORacle prospective, OU offers a course for System Admins for managing oracle on linux, that ha coverage for storage as well. That should be helpful,
    http://education.oracle.com/pls/web_prod-plq-dad/show_desc.redirect?dc=D46592GC10&p_org_id=1001&lang=US
    HTH
    Aman....

  • How to delete the storage

    i created a Storage, and have not any VM or any service build on this Storage. 
    but i click delete, it show the following message. 
    "Storage account xxxx has container(s) which have an active image and/or disk artifacts. Ensure those artifacts are removed from the image repository before
    deleting this storage account."
    how can i do?  and if i have not any usage in this storage, Microsoft will charge to me? 
    Thanks!

    Hi,Base on my experience, I guess something may be not deleted in your storage account.
    Please try those step firstly :
    The procedure to remove the disk image from the blob repository is:
    Go to the Windows Azure Management Portal->Go to Virtual Machines -> "Disks" or "Images" tab -> select the images/disks that you have located in the Storage Account you want to delete -> then hit the "Delete disk" or "delete
    image" command. Back these up if you need before hitting the delete.
    After that, the storage account can be deleted (http://stackoverflow.com/questions/10969012/how-do-i-delete-an-azure-storage-account-containing-a-leased-blob).
    If it doesn't work, I guess you may need contact support via
    https://support.microsoft.com/oas/default.aspx?gprid=17024&st=1&wfxredirect=1&sd=gn&wa=wsignin1.0 mentioned by Tabrez at: 
    http://social.msdn.microsoft.com/Forums/en-US/7a2acffb-6419-4a3d-8fb4-c8258790a2bb/why-cant-i-sign-up-for-the-1month-free-trial .Please try it.
    Any question, please let me know free.
    Regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to decrease the latency with flash+http+camera video?

    I build a HTTP server between live camera and flash player. The flash player connect http server with NetStream. But i find there is a long latency and NetStream doesn't decode frames when the buffersize reach a special value. So i want to know, how to decrease the latency in those environment?
    And why the frame cannot be decoded when the buffersize reachs a special value. I find the buffersize*framerate is about 64. Can i low this value?

    Would you mind reposting this question over on the Flash Professional forums?  This forum is primarily for end users, the Pro forums will get you in touch with a wider developer audience.
    Thanks,
    Chris

Maybe you are looking for

  • MRP - Problem in scheduling of MRP created schedule lines

    Hi, I am working with single level BOM of FG and trying to generate Schedule lines for RM as per requirement coming from FG - BOM explosion during MRP run -  Material master setting of RM is  as below - MRP type - PD, Lot Size - Ex, Procurement type

  • BASIS Abbreviation

    Hi All, While surfing some SAP groups, I came across the full form of BASIS. I was wondering whether it is authentic or just someone throwing a prank. As per them BASIS stands for : <b>Business Application Software Integration System</b> Seems to goo

  • What is needed on non-weblogic client to make soap request

    Hi, I am trying to make a soap request from a client that is not weblogic to a weblogic6.1 server. I was under the impression that all I needed was the client.jar created by wsgen. But when I try to access the service through the client I get the bel

  • 100% scaling problem with datagrid

    I have a situation where I have an AdvancedDataGrid in an hbox . I have a checkbox called scrollTable that changes the width of the AdvancedDataGrid <mx:HBox height="100%" id="gridWrap" width="100%" > <mx:AdvancedDataGrid if(scrollTable.selected){   

  • Where are the html pages located?

    Can anyone let me know where the html pages are located? I have to send a webpage via internet explorer to a sharepoint server.