Report Script Performance Issues

Essbase Nation,
We have a report script that extracts a full 12 months worth of history in 7 minutes. The script that is used to extract the period dimension is as follows:
<Link (<Descendants("Dec YTD") And <Lev("Period",0))
The line above is then changed to pull just one month of data, and now the report script runs for 8 hours.
Please advise as to why the difference in performance.
Thank you.

ID 581459.1:
Goal
How to optimize Hyperion Essbase Report Scripts?
Solution
To optimize your Report follow the suggested guidelines below:
1. Decrease the amount of Dynamic Calcs in your outline. If you have to, make it dynamic calc and store.
2. Use the <Sparse command at the beginning of the report script.
3. Use the <Column command for the dense dimensions instead of using the Page command. The order of the dense dimensions in the Column command should
be the same as the order of the dense dimension in the outline. (Ex. <Column (D1, D2)).
4. Use the <Row command for the sparse dimensions. The order of the sparse dimensions in the Row command should be in the opposite order of the sparse
dimension in the outline. (Ex. <Row (S3, S2, S1)). This is commonly called sparse bottom up method.
5. If the user does not want to use the <Column command for the dense dimensions, then the dense dimensions should be placed at the end of the <Row command.
(Ex. <Row (S3, S2, S1, D1, D2)).
6. Do not use the Page command, use the Column command instead.

Similar Messages

  • Report Script- Performance Issue

    Hi,
    I ran this report script and it is taking around 2 hours to complete. Is there any possiblity to better tune this script. Please advice me where else can we better tune this.
    Thanks,
    UB.

    ID 581459.1:
    Goal
    How to optimize Hyperion Essbase Report Scripts?
    Solution
    To optimize your Report follow the suggested guidelines below:
    1. Decrease the amount of Dynamic Calcs in your outline. If you have to, make it dynamic calc and store.
    2. Use the <Sparse command at the beginning of the report script.
    3. Use the <Column command for the dense dimensions instead of using the Page command. The order of the dense dimensions in the Column command should
    be the same as the order of the dense dimension in the outline. (Ex. <Column (D1, D2)).
    4. Use the <Row command for the sparse dimensions. The order of the sparse dimensions in the Row command should be in the opposite order of the sparse
    dimension in the outline. (Ex. <Row (S3, S2, S1)). This is commonly called sparse bottom up method.
    5. If the user does not want to use the <Column command for the dense dimensions, then the dense dimensions should be placed at the end of the <Row command.
    (Ex. <Row (S3, S2, S1, D1, D2)).
    6. Do not use the Page command, use the Column command instead.

  • How should I report forum performance issues?

    The forums rely heavily on the caching features of browsers to improve the speed of page rendering. Performance of these forums should greatly improve after a few pages because more and more of the images, css and javascript is cached in the browser. As a consequence, when reporting forums performance issues the report should include some information on the state of the browser cache to determine whether the issue is a browser issue or a server issue. Such detailed information is generally not available from just watching the browser screen, but needs to come from specialized tools such as performance monitor plugins and recording proxies.
    The preferred report method for performance issues is to use the speed reporting features build into or available as a plugin for a browser for both the page you want to report a problem with and several refence pages in the site. Detailed instructions are listed below separated out for different browsers. If possible, please use Firefox for submitting the report because it provides an export format that can be read back electronically.
    Known performance issues
    The performance issues with any screen with a Rich Text Editor, such as the Reply window and the compose Private Message window have been acknowleged and improvements are being implemented.
    Mozilla Firefox (preferred)
    Warning: it is currently not recommended to generate a speed report when logged in. The speed report has enough detail for somebody else to hijack your session and impersonate you on the forums. If you really must report while logged in, make sure you log out your browser after generating the speed report and wait at least 4 hours before posting.
    Install the Firebug plugin
    Install the NetExport 0.6 extension for Firebug
    Enable all Firebug panels
    Switch to the "Net" panel in Firebug
    Click on this link
    Export the data from the Firebug Net panel
    Click on this link
    Export the data from the Firebug Net panel
    Browse to the page where you are experiencing the performance problem.
    Export the data from the Firebug Net panel
    Click on this link
    Export the data from the Firebug Net panel
    Click on this link
    Export the data from the Firebug Net panel
    Browse to the page where you are experiencing the performance problem.
    Export the data from the Firebug Net panel
    When you report a performance problem please attach the 6 exports from the Firebug Net panel and an explanation of how you are experiencing the issues (for instance how much slower it is then normal) and include a description of your internet connection (dial-up, dsl, cable etc.) and the country from where you are connecting. If you have non-standard tweaks to your Firefox configuration (such as pipelining enabled) or are running any plugins please include that information in your report as well.
    Google Chrome
    Open the Developer Tools (Ctrl-Shift-J)
    Navigate to the resources tab
    Enable resource tracking.
    Click on this link
    Export the resource loading data.
    Reset the data by disabling and enabling resource tracking
    Click on this link
    Export the data
    Reset the data by disabling and enabling resource tracking
    Navigate to the page where you experience the performance problem
    Export the data
    Reset the data by disabling and enabling resource tracking
    Click on this link
    Export the data
    Reset the data by disabling and enabling resource tracking
    Click on this link
    Export the data
    Reset the data by disabling and enabling resource tracking
    Navigate to the page where you experience the performance problem
    Export the data
    Since Google Chrome does not have an export format for the Resource Tracking information best current practice is to take a screenshot and note the hover details for any resource with a tail that is longer then 25% of the total load time. When you report a performance problem please attach the screenshots and an explanation of how you are experiencing the issues (for instance how much slower it is then normal)  and include a description of your internet connection (dial-up, dsl, cable etc.) and the country from where you are connecting.
    Apple Safari
    The Apple Safari Web Inspector has a Resources panel similar to the Resources panel in the Google Chrome developer tools.To get there, follow these steps:
    Show the menu bar.
    Go to preferences
    Go to the Advanced Tab
    Check “Show  Develop menu in menu bar”.
    From the Develop menu select “Show Web  Inspector”. 
    Collecting the performance information and exporting works exactly the same as in Google Chrome. Please refer to the instructions for Google Chrome.
    Microsoft Internet Explorer
    IE does not have native features to analyze web traffic. No plugins have been found that produce the required information (please let us know if we missed any). For now, please reproduce the issue with Firefox, Chrome or Safari.
    Please note that due to the reliance on Javascript for the interactive effects the performance of these forums will be much better on MS IE 8 then on previous versions of MS IE.

    Hi
    It works, check once again...
    regards
    Swami

  • BW BCS cube(0bcs_vc10 ) Report huge performance issue

    Hi Masters,
    I am working out for a solution for BW report developed in 0bcs_vc10 virtual cube.
    Some of the querys is taking more 15 to 20 minutes to execute the report.
    This is huge performance issue. We are using BW 3.5, and report devloped in bex and published thru portal. Any one faced similar problem please advise how you tackle this issue. Please give the detail analysis approach how you resolved this issue.
    Current service pack we are using is
    SAP_BW 350 0016 SAPKW35016
    FINBASIS 300 0012 SAPK-30012INFINBASIS
    BI_CONT 353 0008 SAPKIBIFP8
    SEM-BW 400 0012 SAPKGS4012
    Best of Luck
    Chris
    BW BCS cube(0bcs_vc10 ) Report huge performance issue

    Ravi,
    I already did that, it is not helping me much for the performance. Reports are taking 15 t0 20 minutes. I wanted any body in this forum have the same issue how
    they resolved it.
    Regards,
    Chris

  • SQL Services 2012 Reporting Services Performance Issue - PowerView

    Power view Reports are loading very slow while opening in SharePoint 2013, it is taking more than 15 secs. It is development environment, maximum 10 users , no traffic at all but still not sure why it is taking such long time.
    We have 2 servers in SharePoint farm one is for SharePoint and other is for SQL. I have gone through the logs in reporting database, attached the same below. Can you please help me what we can say from the sheet attached ,whether it is slow or fast. Or where
    we are having issue.
    SQL server version is SQL 2012 SP2.
    SharePoint 2013 is RTM.
    Gone through the below blogs but no luck.
    http://blogs.msdn.com/b/psssql/archive/2013/07/29/tracking-down-power-view-performance-problems.aspx
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/4ed01ff4-139a-4eb3-9e2e-df12a9c316ff/ssrs-2008-r2-and-sharepoint-2010-performance-problems
    Thanks.
    Thanks, Ram Ch

    Hi Ram Ch,
    According to your description, your have performance issue when running your Power View report. Right?
    In this scenario, based on your screenshot, it takes long time on data retrieval. How is the performance when executing the query in SQL Server Management Studio? Since you mention there's no traffic at all and 15 seconds will not cause query
    time out, we suggest you optimize the query for retrieving data. Please refer to links below:
    Troubleshooting Reports: Report Performance
    Please share some detail information about the data query if possible. Thanks.
    Best Regards,
    Simon Hou

  • Crystal Reports 11g Performance Issue

    We just upgraded our database to Oracle 11g from Oracle 10g and we are having significant performance issues with Crystal Reports.
    Our DEV and TEST environments are on 11g and are very slow to connect to the database and attach to specific tables.  It is unusable.  Our PROD environment is still 10g and works fine.  
    We  have tested with both the lastest version -  Crystal Reports 2008 V1  and  Crystal Reports XI R2 SP6.  We have also tested on several different machines.
    We are using Oracle 10g ODBC drivers.
    Does anyone have any recommendations?

    You could also try our Data direct drivers we have available on our WEB site using this [link|https://smpdl.sap-ag.de/~sapidp/012002523100008666562008E/cr_datadirect53_win32.zip].
    Those drivers are the most recent and do support 11g. It also has a wired driver that doesn't require the client to be installed.
    Also, highly recommended that when you do update the Oracle client to uninstall 10 first. There have been issues with CR and Oracle mixing the Oracle dependencies and causing problems.
    Thank you
    Don

  • How to maximize report script performance

    I have a report script that takes a while to run. Basically the script is zooming into 3 sparse dimensions and 1 dense dimension. <BR><BR><BR>The dense dimension has 732 stored members.<BR><BR>The sparse dimensions have 92, 280 and 576 members respectively. <BR><BR>when I did the zoom bottom in excel, it only took me about 4 minutes total to zoom in through the whole thing. <BR><BR>But when I did it in the report script, it ran for about half an hour and then I had to killed it. <BR><BR>I tried to change the order of the rows in the report script to match the zoom order in my excel retrieval, but that didn't help. <BR><BR>i remember from the bootcamp that the report script zoom in order should follow the hourglass or something. i can't remember exactly which dimension should go first though, was it the biggest dense and then the smallest sparse and then the biggest sparse dimensions? <BR><BR>here is how the script looks like:<BR><BR>{Rowrepeat}<BR>{SUPFEED}<BR>{Pagelength 10000}<BR>{Supemptyrows}<BR>{Supzerorows}<BR>{supcolheading}<BR><BR><ROW ("REV AND COST CENTERS", ENTITY, ACCOUNTS, ALLOCATION)<BR><BR><LINK (<descendants ("local currency entities") or <descendants (historical) <BR>or <children(celerops) OR <CHILDREN (CRAOPS) OR <DESC (ONLINEOPS) OR <DESC (PARACELOPS) OR <DESC (DSCIEOPS) OR <DESC (DSERVOPS) OR <DESC (EXCAPOPS) or <CHILDREN (GENERAL) OR AND <DIMBOT (ENTITY))<BR><BR><LINK (<descendants (statistical) or <descendants ("net income") or <DESCendants ("pre-assessment expenses") AND <LEV(Accounts,0))<BR><BR><LINK (<DESC("SGA") OR <DESC ("R&D AND ENG.") OR <DESC ("OTHER CC") AND <LEV ("REV AND COST CENTERS",0))<BR><LINK (<descendants ("direct") or <descendants ("indirect") or <descendants ("manual indirect") <BR>or <children ("allocation") and <lev(allocation,0))<BR><BR>!<BR><BR>any suggestions would be appreciated. <BR><BR>thx.

    <p>For improving the retreival performance put the most densedimension on top followed by least dense dimension members,followedby most queried sparse dimension members followed by least sparsedimension members.....For improving calcualtion performance theorderof sparse dimension members changes while that of densedimension members remain same.</p><p> </p><p>Other thing which you can do is to increase the size ofretreival buffer.</p>

  • ALI Scripting - Performance Issues

    We are using ALI scripting to raise events so that other portlets in the page can listen to the events and also passing some data to other portlets using "PTPortlet.setSessionPref" scripting API. In my local testing, the scripting is taking just 0.3 seconds, however when we deployed this code to common portal which has few other portlets, the scripting is taking 5 seconds. Does any one know if there are any known best practices around ALI scripting to avoid performance issues especially to transfer events between portlets?
    Thanks
    Sampath

    Hi, I would like to provide additional information on the performance issue we are facing with ALI scripting. In the code highlighted below, we are using “setSessionPref” methods to set values in session variables using ALI scripting APIs and raising events for IPC.
    var prefName = 'selectionString';
    var prefValue = xmlFile;
    var gpConfigValue = "<%=gpPromptsConfigID%>";
    PTPortlet.setSessionPref(prefName,prefValue);
    PTPortlet.setSessionPref('gpFormObj',frmObj.name);
    PTPortlet.setSessionPref('gpPromptsConfigID', gpConfigValue);
    myportlet$$PORTLET_ID$$.raiseEvent('onSelectionSubmitFormObj',false);
    myportlet$$PORTLET_ID$$.raiseEvent('onSelectionSubmit', false);
    This code does not take lot of time(less than a second) to execute in my local machine however, it takes lof of time (around 5 seconds) in integrated development environment.
    After lot of debugging, I observed that if I change “PTPortalContext.GET_SESSION_PREFS_URL” to point to http://localhost:7001 instead of integrated development environemnt, the processing time is considerably reduced. This code below is inserted automatically by ALUI in every portal page. Does any one know the signficance of the below code automaitcally inserted by ALUI and how it can impact IPC.
    // Define PTPortalContext for CSAPI
    PTPortalContext = new Object();
    PTPortalContext.GET_SESSION_PREFS_URL = 'http://dcdev.pg.com/portal/server.pt?space=SessionPrefs&control=SessionPrefs&action=getprefs';
    PTPortalContext.SET_SESSION_PREFS_URL = 'http://dcdev.pg.com/portal/server.pt?space=SessionPrefs&control=SessionPrefs&action=setprefs';
    Thanks
    Sampath

  • Crystal Report 9 performance issue wehn we move old DB to new DB server

    We are using Crystal Reports 9 version, recently we move to new database server and reports data source are pointing to Old database server and old database server is no longer in the network.
    , Now we have issue with the performance, if the we call the reports from web application, it is talking long to retrieve the report at least 1 min.. , if we manually update the report data source pointing to new server and the reports are pulling in 1 second, Is there any easy solution to update all the report pointing to new server, we have 3000 reports, we can not use manual process, it will take long time and it is also issue with the future if we change the server.
    Thanks!
    Ram

    The 1 minute timeout is due to ODBC driver trying to connect, it's the default timeout you are seeing. There is not tool Cr has to migrate/update database connection info. You can develop your own though or search the web for a third party tool, I'm sure there are lots out there.

  • How to solve the report generating performance issue in BI server

    ->I HAVE USED ONLY ONE PIVOT IN MY REPORT QUERY AND THE REPORT QUERY IS FULLY DEPEND ON ONE FACT TABLE
    ->IN THE REPORT I AM SHOWING DATA BY MORE THAN 150 COLUMNS.
    ->I HAVE USED REPORT FUNCTIONS FOR 3 FIELDS IN THE REPORT.ALL THREE REPORT FUNCTIONS JUST ADD TWO COLUMN FIELD AND MAKE IT IN TO ONE COLUMN FILED
    ->REPORT QUERY INDIVIDUALLY RUNNING BY 2.40 SECONDS.WHEN I USED THAT TO GENERATE A REPORT IT TAKES TO RUN MORE THAN 8 MINS.WHY THIS DELAY?WILL YOU EXPLAIN
    Edited by: 873091 on Jul 18, 2011 3:50 AM

    Hello Dude,
    So from your post I understand that there is a report that basically takes 8 minutes to retrieve the data, but when you get the logical sql from the cache, and run it against the database, it only takes less than 20 seconds to retrieve the same data?
    This should never happen. There might be a delay of more 20-40 seconds more than the time taken against the database to load the page elements etc. Is this happening for only this particular report or all reports ? Are you running the query against the same instance's database or a different one?
    Try to re-boot all the BI services and run the report again, if the issue still exists, enable caching and try.
    Assign points if helpful.
    Thanks,
    -Amith.

  • Report Painter Performance Issue

    Helloo experts,
    I have a report painter report which is pulling the data from a custom table (which has 32million records). Its taking a lot of time .
    Is there any method to improve its performance ?

    Hi,
    Please check report painter selecting records using primary key or secondary index. If not please create secondary index for custom table (after consultation with basis, due to creating the secondary index will increase the load on server of custom table with 32 millions records)
    If not possible. extract the resultset from custom table into another table and make query for report painter on that. and update the new table on daily basis or hourly basis.
    aRs

  • Crystal Report / Subreport - Performance Issue

    I am having problems with a Crystal Report / Subreport with performance. I am using Crystal Version 11, directly in Crystal Developer.  There are no other programs involved.  I am linking to a SQL Server database using ODBC.
    I have narrowed the problem down to this:
    The main report has 4 tables.  See diagram using the following link.   http://screencast.com/t/TA9YYlwwl7
    The subreport has 1 table, this table has > 2 Million records in it.
    The main report links to the subreport with one field:
    Main report field SAMPLE.PATIENT = subreport field Z_RESULT_HISTORY.PHIN
    When I set the subreport linking within Crystal   it automatically generates the following in the record selection for the subreport:
    {Z_RESULT_HISTORY.PHIN} = {?Pm-SAMPLE.PATIENT}
    The problem is that the report execution time is dependent on the field that I am using for the record selection in the main report.
    Case I   works lightening fast:
    There are 16 records returned on the Main Report   each one of these has about 1-5 records returned on the subreport.
    {SAMPLE.PATIENT} = "MOUSEMICKEY"  and {SAMPLE.SAMPLE_NUMBER} < 200
    Case II   brutally slow   there are 51 records in the main report that qualify   with a few records each in the subreport.  By  brutally slow    I mean a few minutes:
    {BATCH.NAME} = "HEP_ARCH-20090420-1"
    In this case, I can see in the bottom right of the Crystal Preview window, that it is reading through all 2M records in the Z_RESULT_HISTORY table
    Case III   brutally slow - a couple minutes
    {BATCH_OBJECTS.OBJECT_ID} = 111
    This returns 1 record on the main report and 0 records on the subreport.  Yet I can see it reading through ALL 2 Million records before the report is displayed.
    What I can t understand is why the field used for record selection on the MAIN report is affecting the speed of the execution of the subreport.  I need the main report to be selected by BATCH.NAME   yet I can t figure out what I can change to make the report run fast.  When I record select the main report by SAMPLE.PATIENT, I do NOT see the subreport reading all 2M reocrds - the report preview is returned in less than 1 second.
    Any help would be much appreciated.

    Lindell - your response was very helpful.  I was able to create a SQL Command on the subreport and change the subreport links on the main report to use the paramters in the SQL Command.  The report is now very fast - even when there are lots of detail records on the main report.  It is properly executing the query and not reading all 2M records into memory for each subreport.
    I'm still totally confused as to why Crystal was misbehaving so poorly and changing how the subreport queried the database when the only change was the fields used for record selection on the main report.  It really looks like a bug to me, but maybe someone can still enlighten me.
    thanks again so much.  We are in Parallel testing for a production rollout - and the users are MUCH happier now!  (which of course makes me much happier!)
    Susan
    PS I was not aware of the SQL Command - had never used it before.  Took me a little while to figure it out and how to do the linking - but it is very powerful.  Thanks again.

  • Sap Crystal Reports 2013 Performance Issues

    Hi,
    We are planing to do the report migration to sap crystal report 2013. now we have crystal report xi Developer licensed version and we download the 30 days evaluation(trial) version of sap crystal report 2013 only for testing. we have 1 report that created in crystal report xi version it runs within 5 mins for 2 lakhs records and same report we are testing on sap crystal report 2013 evaluation(trial) version with same connection and same report parameters it took 10 mins for 2 lakhs records.every report take more time in sap crystal report 2013 evaluation(trial) version as compared to crystal report xi Developer licensed version.
    Why this happened? is it issue of evaluation(trial) version of sap crystal report 2013?
    Kind Regards,
    Ganesh

    Hi Jamie,
    what database are you using?
    Oracle
    what specific db version?
    Oracle  11g
    what connection method?
    Oracle Server
    Above same thing we r using for crystal report xi licensed version and sap crystal report 2013 evaluation(trial) version but now issues is that report takes 45 min in sap crystal report 2013 evaluation(trial) version for 2 lac records,but same report takes only 5 min to run in crystal report xi licensed version. is this issue of licensed bcoz now we are run report on sap crystal report 2013 evaluation(trial) version.

  • Essbase calc script performance issues

    Hi,
    I have essbase 9.3 running on Sun solaris 4 CPU, 16 GB server. The calc script "calc all" takes ~3 hrs to complete.
    This is the calc script.
    /ESS_LOCALE English_UnitedStates.US-ASCII@Binary
    SET UPDATECALC OFF;
    SET CALCPARALLEL 4;
    SET CALCTASKDIMS 2;
    CALC ALL;
    We don't have to calc all dim, but even if we
    But even with specific dim we get the same timing. Below is the script
    SET UPDATECALC OFF;
    SET CALCPARALLEL 4;
    SET CALCTASKDIMS 2;
    FIX ("Y2009", "Actual");
    CALC DIM("Data Source","Intercompany","LegalEntity","Site","Department","Entity");
    ENDFIX
    The ess00001.ind is 700 Mb and ess00001.pag is 2.1 GB.
    In Admin services, this is what I see for caches
    1) Index cache size is 1 GB for this DB
    2) Index cache current value is 1gb
    3) Datafile cache setting is 1.5 GB
    4) Datafile cache current value is 0 (?? not sure why??)
    5) Data cache setting 4.1 GB
    6) Index page setting 8 kb
    please help ...
    Thanks
    Moe

    Moe,
    I'm guessing you inherited this thing, else you would know why the cache settings are what they are, but here are some thoughts:
    Caches:
    3) Datafile cache setting is 1.5 GB
    4) Datafile cache current value is 0 (?? not sure why??)You're running the database in Buffered I/O, so the data file cache is ignored.
    1) Index cache size is 1 GB for this DB
    2) Index cache current value is 1gb You have consumed all of the cache -- I'm a little confused, as you state your .ind file to be 700 megabytes -- generally the index cache consumption doesn't go beyond the .ind file size. When you look at your hit ratio statistics in EAS, does it show a 1 against the index cache? If yes, then you don't need to look any further as that's as good as it's going to get.
    5) Data cache setting 4.1 GBUnless you're using MEMSCALINGFACTOR, I don't think Essbase is actually addressing all of the memory you've assigned. What are you showing as actually used? In any case, having a data cache almost twice as big as the .pag files is a waste as it's way too large.
    Easy, off the cuff suggestions without knowing more about your db:
    1) Try AGG instead of CALC DIM for sparse dimensions.
    2) Try turning off (yes, turning off, you'd be surprised) parallel calc, and benchmark it. It will probably be slower, but it's nice to know.
    3) Dimension order? Modified hourglass?
    4) Tried defragmenting the database and benchmarking the performance?
    5) What is your block size? Big? Small?
    6) I think you are not calculating your Accounts/Measures dimension in your calc? If you are, and it's dense, could you make those Accounts dynamic calc -- dropping a dimension from the calc can be huge.
    I'm sure there will be other suggestions -- these are the easiest.
    Regards,
    Cameron Lackpour

  • Calc script & performance issues

    Hi All we have a calc script which used to take only 10 mint every day. But today it is taking long time 4 hrs,,,stil running..if i cancel that calc operation what is the impact on database?. Earlier all users are hapy with speed...but suddenly every one got pissed off with speed...it is taking long time to retrieve data..Quickly what are the parameters I need to check?thanks in adavance..

    <p>If you are using committed access then you can safely cancel thecalculation. All data will be reverted back to what it was beforecalculation. However if you are using uncommitted access it isrecommended not to cancel any running operation.</p><p>If you want to eliminate fragmentation, just export your level 0data and import again, and then do a calc all.</p><p>Doing so will remove any fragmentation. It is recommended to dothis once in  a while like 2 months to get rid offragmentation.</p>

Maybe you are looking for

  • Error when ejb deployment in WLS 9.1 clustered environment

    Hello, We have a WLS 9.1 cluster consisting of one admin server and two managed servers. we are deploying ejb jar,It is succeed on the first managed server but fail on the second managed server. The admin console and for the second managed server con

  • How to change country of Salesperson in Oracle Sales Online

    Hi, In our database i have one sales resource from say France. Who has nearly 20 opportunities under him. Now the business requires that he is transfered to UK office. As a effect all the existing opportunities as well new opportunites should be list

  • DataGrid in a DataGrid - Flash Builder 4 Beta 2

    My dear friends, hi... There is a problem here: I have a Flex and WCF webservice project... One part of my project is look like this: I wonder what kind of controls should I use for this part? Something like Grids, Lists or something like Repeater in

  • HT1977 Old App purchases are not appearing in itunes 11

    I have recently upgraded from iTunes 10.x to iTunes 11 and my old purchased Apps that were not on my ipad but on my computer are not appearing in itunes 11 to put back on my ipad?

  • [SOLVED] USB storage devices do not appear as /dev/ block devices

    Hello. I'm experiencing the exact same problem described in this thread and this thread I'm running linux-ck-ivybridge. Both of these threads are marked solved but I don't get their solutions. Could someone help me understand what all this means? I t