Crystal Reports performance is too slow

Dear SDNers,
I have designed a crystal report which is fetching data from a Z developed function module. From crystal report side, used filters.But while executing report, it is calling the function module multiple times.
due to this performance is very bad. Why report is calling mutiple times of the function module. Do need to modify from crystal report side or from function module side. Please clarify.
Regards,
Venkat

A similar issue was seen a year ago It was regarding a function module call being executed multiple times from  CR4Ent tool. It involved the usage of sub reports inside that report and the issue was generic for any function module used for testing.
At that time, the issue was resolved by upgrading to the latest available patch of CR4Ent and also by applying the latest patch at SAP R3 end.
If you are able to post the exact support package and patch level of CR4Ent and also for the SAP R3 system, then someone can tell whether its the latest or not.
-Prathamesh

Similar Messages

  • Sales & Order report performance is too slow!

    Hi All,
    Sales report is prety slow.It took 90% in OLAP time. I tried RSRT all the possible ways & also created OLAP fill query level also but No result. Please help me.
    Thanks
    Vasu.

    Hi Vasu,
    CAN You please refer below link
    http://wiki.sdn.sap.com/wiki/display/BI/HowtoImproveQueryPerformance-A+Checklist
    it tells explains how you can improve your query performace.
    Also as i told since you are FETCHING DATA direct from MASTER data report will be bit slow so try to get data directly from cube and may be u can use filters also.
    Hope this may help you.
    Regards
    Nilesh

  • Performance is too slow on SQL Azure box

    Hi,
    Performance is too slow on SQL Azure box (Located in Europe)
    Below query returns 500,000 rows in 18 Min. on SQL Azure box (connected via SSMS, located in India)
    SELECT * FROM TABLE_1
    Whereas, on local server it returns 500,000 rows in (30 sec.)
    SQL Azure configuration:
    Service Tier/Performance Level : Premium/P1
    DTU       : 100
    MAX DB Size : 500GB     
    Max Worker Threads : 200          
    Max Sessions     : 2400
    Benchmark Transaction Rate      : 105 transactions per second
    Predictability : Best
    Any suggestion would be highly appreciated.
    Thanks,

    Hello,
    Can you please explain in a little more detail the scenario you testing? Are you comparing a SQL Database in Europe against a SQL Database in India? Or a SQL Database with a local, on-premise SQL Server installation?
    In case of the first scenario, the roundtrip latency for the connection to the datacenter might play a role. 
    If you are comparing to a local installation, please note that you might be running against completely different hardware specifications and without network delay, resulting in very different results.
    In both cases you can use the below blog post to assess the resource utilization of the SQL Database during the operation:
    http://azure.microsoft.com/blog/2014/09/11/azure-sql-database-introduces-new-near-real-time-performance-metrics/
    If the DB utilizes up to 100% you might have to consider to upgrade to a higher performance level to achieve the throughput you are looking for.
    Thanks,
    Jan 

  • IR Report found 1 million record with blob files performance is too slow!

    we are using
    oracle apex 4.2.x
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - Production
    mod_plsql with Apache
    Hardware: HP proliant ML350P
    OS: WINDOWS 2008 R2
    customized content management system developed in apex.when open the IR report have 1 ml rows found and each rows have blob(<5MB as pdf/tiff/bmp/jpg) it will be raising rows in future! but the searching performance is very slow!
    how can increasing the performance?
    how can showing progressing status to user while searching progress going on IR report itself?
    Thanx,
    Ram

    It's impossible to make definitive recommendations on performance improvement based on the limited information provided (in particular the absence of APEX debug traces and SQL execution plans), and lacking knowledge of the application  requirements and access to real data.
    As noted above, this is mainly a matter of data model and application design rather than a problem with APEX.
    Based on what has been made available on apex.oracle.com, taking action on the following points may improve performance.
    I have concerns about the data model. The multiple DMS_TOPMGT_MASTER.NWM_DOC_LVL_0x_COD_NUM columns are indications of incomplete normalization, and the use of the DMS_TOPMGT_DETAILS table hints at an EAV model. Look at normalizing the model so that the WM_DOC_LVL_0x_COD_NUM relationship data can be retrieved using a single join rather than multiple scalar subqueries. Store 1:1 document attributes as column values in DMS_TOPMGT_MASTER rather than rows in DMS_TOPMGT_DETAILS.
    There are no statistics on any of the application tables. Make sure statistics are gathered and kept up to date to enable the optimizer to determine correct execution plans.
    There are no indexes on any of the FK columns or search columns. Create indexes on FK columns to improve join performance, and on searched columns to improve search performance.
    More than 50% of the columns in the report query are hidden and not apparently used anywhere in the report. Why is this? A number of these columns are retrieved using scalar subqueries, which will adversely impact performance in a query processing 1 million+ rows. Remove any unnecessary columns from the report query.
    A number of functions are applied to columns in the report query. These will incur processing time for the functions themselves and context switching overhead in the case of the non-kernel dbms_lob.get_length calls. Remove these function calls from the query and replace them with alternative processing that will not impact query performance, particularly the use of APEX column attributes that will only apply transformations to values that are actually displayed, rather than to all rows processed in the query.
    Remove to_char calls from date columns and format them using date format masks in column attributes.
    Remove decode/case switches. Replace this logic using Display as Text (based on LOV, escape special characters) display types based on appropriate LOVs.
    Remove the dbms_lob.get_length calls. Instead add a file length column to the table, compute the file size when files are added/modified using your application or a trigger, and use this as the BLOB column in the query.
    Searching using the Search Field text box in the APEX interactive report Search Bar generates query like:
    select
    from
      (select
      from
        (...your report query...)
      ) r
      where ((instr(upper("NWM_DOC_REF_NO"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("NWM_DOC_DESC"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("SECTION_NAME"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("CODE_TYPE"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("REF_NUMBER_INDEX"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("DATE_INDEX"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("SUBJECT_INDEX"), upper(:apxws_search_string_1)) > 0
      or instr(upper("NWM_DOC_SERIEL"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("NWM_DOC_DESCRIPTION"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("NWM_DOC_STATUS"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("MIME_TYPE"), upper(:APXWS_SEARCH_STRING_1)) > 0
      or instr(upper("NWM_DOC_FILE_BINARY"), upper(:APXWS_SEARCH_STRING_1)) > 0 ))
      ) r
    where
      rownum <= to_number(:APXWS_MAX_ROW_CNT)
    This will clearly never make use of any available indexes on your table. If you only want users to be able to search using values from 3 columns then remove the Search Field from the Search Bar and only allow users to create explicit filters on those columns. It may then be possible for the optimizer to push the resulting simple predicates down into the inlined report query to make use of indexes on the searched column.
    I have created a copy of your search page on page 33 of your app and created an After Regions page process that will create Debug entries containing the complete IR query and bind variables used so they can be extracted for easier performance analysis and tuning outside of APEX. You can copy this to your local app and modify the page and region ID parameters as required.

  • Crystal Report Performance for dbf files.

    We have a report which was designed 5 -6 years ago. This report has 4 linked word doc and dbf file as datasource. This report also as 3 subreports. The size of field in dbf is 80 chars and couple of field are memo field. The report performance was excellent before we migrated the crystall report to 2008. After CR2008 the system changed and it is suddenly really slow. We have not change our reports so much it should have an influence on performance. When the user presses the preview button on printing tool window the control is transferred to Crystal. Something has happened inside black box of Crystal ( IMO ).   the dll we have are crdb_p2bxbse.dll 12.00.0000.0549 . The issues seems to be of xbase driver (not possible to use latest version of crdb_p2bxbse.dll and dbase files with memo fields).

    Hi Kamlesh,
    Odd that the word doc is opened before the RPT, I would think that the RPT would need to be opened first so it sees that the doc also needs to be opened. Once it's been loaded then the connection can be closed, CR embeds the DOC in the RPT so the original is no longer required.
    Also, you should upgrade to Service Pack 3, it appears you are still using the original release. SP1 is required first but then you should be able to skip SP2 and install SP3.
    You did not say what earlier version of Cr you were using? After CR 8.5 we went to full UNICODE support at which time they completely re-built the report designer and removed the database engines from the EXE and made them separate dll's now. OLE objecting also changed, you can use a formula and database field to point to linked objects now so they can be refreshed any time. Previously they were only refreshed when the report was opened.
    You may want to see if linking them using a database field would speed up the process. Other than that I can't suggest anything else as a work around.
    Thank you
    Don

  • Crystal Reports 9 over network: slow initialization

    We have a VC++ application that uses Crystal Reports RDC (actually 8.5, and trying to change to 9.2). By design our application is installed on a file server and users execute it directly from the server, without copying or install any component in the local drive. We just register the minumum Crystal Reports DLLs needed to point directly to the location of this DLLs in the server. So we don't copy any Crystal Reports Runtime DLLs to c:\
    At this time, with Crystal Reports 8.5, the minimum runtime DLLs on the server needed to run our application and print reports have a size of 8Mb, and the first time our application tries to open a report it hangs for about 5 seconds (with FileMon I see that Crystal Reports is loading its runtime DLLs). But now I'm trying to change it to use Crystal Reports 9 (due to unsolved CR8.5 limitations). With this version our minimum required runtime DLLs have a size of almost 12Mb, and the load time of this DLLs become almost 10 seconds (too big).
    I did tests with Crystal Reports XI and it was even worse. If I copy and register the DLLs in the local drive it loads very fast, but we need to make it to work directly from the server without copying the files to c:\
    Is there any way to reduce the loading time of this runtime DLLs without having to copy them to the local drive? Thanks

    Well, I think the problem you had with CR XI is that you can not just copy the dlls. You have to use the merge modules and create a deployment project using a utility such as InstallShield, Wise, MS Visual Installer, etc.
    As far as the issue at hand.
    Here is a basic list of possible issues that may improve the performance:
    Report Design Issues
    u2022     Linked OLE object inserted into a report is not where the report expects it to be. If this is the case, the report will attempt to locate the object, potentially consuming large amounts of time.
    u2022     The subreport option "Re-import when opening" is enabled (right click the subreport(s), choose format subreport, look at the subreport tab). While this is a great design time feature, it is a time consuming process and should be used judiciously.
    u2022     Specific printer is set for the report and the printer does not exist. Try the "No printer" option (File | Page setup).
    u2022     Also, see the Article u201CHow Printer Driver Options Affect a Reportu201D regarding printers and Crystal Reports;
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/a09051e9-721e-2b10-11b6-f9c65c64ef29
    u2022     The number of subreports the report contains and in which section the subreports are located will impact report performance. Minimize the number of subreports used, or avoid using subreports if possible. Subreports are reports within a report, and if there is a subreport in a detail section, the subreport will run as many times as there are records, leading to long report processing times. Incorrect use of subreports is often the biggest factor why a report takes a long time to process.
    u2022     Use of "Page N of M", or "TotalPageCount" special fields. When the special field "Page N of M" or "TotalPageCount" is used on a report, it will have to generate each page of the report before it displays the first page. This will cause the report to take more time to display the first page of the report
    u2022     Graphics, while a powerful feature of Crystal Reports, need to be used with care. When displaying a report using the Web Forms report viewer, for each chart or image on the report page, the report engine must create a temporary JPG file and the browser must make a separate request to the web server to retrieve it. The Web Forms viewer also periodically searches through the images folder for old images and deletes them. This disk access and creation of temporary files will reduce report performance.
    u2022     Remove any unused object from the report. This includes database tables, formulas and parameters and running total formulas. Even if these objects are not used in a report, the report engine will attempt to evaluate the objects, thus affecting performance.
    u2022     Suppress unnecessary report sections. Even if a report section is not used, the report engine will attempt to evaluate the section, thus affecting performance.
    u2022     If summaries are used in the report, use conditional formulas instead of running totals when ever possible.
    u2022     Whenever possible, limit records through Record selection Formula, not suppression.
    u2022     Use SQL expressions to convert fields to be used in record selection instead of using formula functions. Since SQL Expression Fields are added to the SELECT clause of the SQL Query the SQL expressions are evaluated server side.
    For example, if you need to concatenate 2 fields together, instead of using a formula, create a SQL Expression Field.
    u2022     Using one command table or Stored Procedure or a Table View as the datasource can be faster if you returns only the desired data set.
    u2022     Perform grouping on the database server. This applies if you only need to return the summary to your report but not the details. It will be faster as less data will be returned to the reports.
    Report Options
    u2022     u201CVerify on First Refreshu201D option (File | Report Options). This option forces the report to verify that no structural changes were made to the database. There may be instance when this is necessary, but once again, the option should be used only if really needed. Often, disabling this option will improve report performance significantly.
    u2022     u201CVerify Stored Procedure on First Refreshu201D option (File | Report Options). Essentially the same function as above, however this option will only verify stored procedures.
    If the above does not help, perhaps that is the best performance you will get over the network and going to a web based solution may be worth considering(?). Let's see if anyone else can provide any help.
    Ludek

  • How to interpret or calculate Crystal Report Performance Timing?

    HI,
    I have been trying to interpret the Performance Information/Performance Timing of a Crystal Report. When I access this information from Crystal designer menu I see several fields but I don't know exactly which one gives you the total execution time in miliseconds or how do you calculate the total time using the fields.
    Help to interpret the following information will be appreciate. Example:
    MainReport.rpt
    Open Document:   0 ms  
    Run the Database Query:   703 ms  
    Read Database Records:   92 ms  
    Format First Page:   949 ms  
    Number of pages formatted:   2   
    Average time to format a page:   474 ms  
    Number of page starts generated:   13   
    Average time to generate a page start:   13 ms  
    Subreport.rpt
    Run the Database Query:   4 ms   For all instances
    Read Database Records:   2 ms   For all instances
    Number of page starts generated:   3   
    Average time to generate a page start:   0 ms   For all instances
    Number of subreport instances formatted:   1   
    Time to format all subreport instances:   38 ms
    Thanks!

    Have you seen this post:  [Needing to monitor the report run time and produce the dates/time on my rpt;

  • Bumblebee performance is too slow

    Hi everyone,
    This is my second post in this forum and it has been only 2 days that I met Arch. Before, I was using Ubuntu for 3 years, but due to low performance in my PC, unfortunately, I decided to say good bye which was hard to say.
    Now, I am trying to have the same setup as my previous laptop and Bumblebee was one of them. I followed the instructions here: https://wiki.archlinux.org/index.php/Bumblebee.
    It seems that bumblebee is installed and working, however, the FPS is too slow:
    $ optirun glxspheres64 -info
    Polygons in scene: 62464
    Visual ID of window: 0x20
    Context is Direct
    OpenGL Renderer: GeForce GT 520MX/PCIe/SSE2
    0.023848 frames/sec - 0.021114 Mpixels/sec
    Without optirun I get:
    $ glxspheres64 -info
    Polygons in scene: 62464
    Visual ID of window: 0x20
    Context is Direct
    OpenGL Renderer: Mesa DRI Intel(R) Sandybridge Mobile
    0.033237 frames/sec - 0.029426 Mpixels/sec
    0.029968 frames/sec - 0.026533 Mpixels/sec
    This is impossible as I was getting very good results before.
    I am wondering if I did something wrong, or missed anything.
    Just for information, system specs:
    Intel i7 2670QM 2.2 GHZ
    4 GB RAM
    1 GB GeForce GT 520MX
    512 MB Intel Graphics
    I played 0ad with and without optirun and the performance was good in both of them, but not sure if it switches the video cards itself.
    I also have bbswitch installed.
    Any help would be appreciated. Thank you.
    Last edited by wakeup12 (2014-09-27 14:29:19)

    Hello,
    Can you please explain in a little more detail the scenario you testing? Are you comparing a SQL Database in Europe against a SQL Database in India? Or a SQL Database with a local, on-premise SQL Server installation?
    In case of the first scenario, the roundtrip latency for the connection to the datacenter might play a role. 
    If you are comparing to a local installation, please note that you might be running against completely different hardware specifications and without network delay, resulting in very different results.
    In both cases you can use the below blog post to assess the resource utilization of the SQL Database during the operation:
    http://azure.microsoft.com/blog/2014/09/11/azure-sql-database-introduces-new-near-real-time-performance-metrics/
    If the DB utilizes up to 100% you might have to consider to upgrade to a higher performance level to achieve the throughput you are looking for.
    Thanks,
    Jan 

  • Print Preview using crystal report in SAP B1 slow.

    Dear all,
    I face another problem. Print preview using crystal report in SAP B1 is very slow. Although i have create a query view in SQL server but it's still slow for my user. First time i create using table in crystal report. just drag the table. It's take more than 10 minutes. After that i move all the query to view in SQL server , it's take 1-2 minutes to print preview some time more.. But i think that it's still slow. Does any one know how to print preview fast. Just like in PLD.
    Thanks in advance
    bodhi86

    Hello,
    How do you answered to this question ?
    Malika
    Edited by: Malika Sanoune on May 13, 2011 1:48 PM

  • Crystal Report 10.5 very slow (30 minutes) to display

    crystal report 10.5 in vs 2008 display at extremely very slow speed in development server and production server. i have loaded a dataset as the datasource for the report. i also checked the code in debugging mode to know where the delay is but found out it is not from the code as the code executes to the end in less than a minute but the browser keeps loading for several minute before display.
    please kindly help to resolve this issue.

    With CR 2008:
    For OLE DB then use:
    MS SQL 2005 - OLE DB Provider
    MS SQL 2008 - SQL Native 10
    For ODBC then use:
    MS SQL 2005 - SQL Native
    MS SQL 2008 - SQL Native 10
    Once you have a report created, you can use the following code at runtime:
    Public Sub New()
            'This call is required by the Windows Form Designer.
            InitializeComponent()
            'Add any initialization after the InitializeComponent() call
            'Create an instance of the strongly-typed report object
            crReportDocument = New CrystalReport1()
            'Setup the connection information structure to be used
            'to log onto the datasource for the report.
            crConnectionInfo = New ConnectionInfo()
            With crConnectionInfo
                .ServerName = "escalade"    'physical server name
                .DatabaseName = "Pubs"
                .UserID = "sa"
                .Password = "admin"
            End With
            'Get the table information from the report
            crDatabase = crReportDocument.Database
            crTables = crDatabase.Tables
            'Loop through all tables in the report and apply the connection
            'information for each table.
            For Each crTable In crTables
                crTableLogOnInfo = crTable.LogOnInfo
                crTableLogOnInfo.ConnectionInfo = crConnectionInfo
                crTable.ApplyLogOnInfo(crTableLogOnInfo)
            Next
            'Set the viewer to the report object to be previewed.
            CrystalReportViewer1.ReportSource = crReportDocument
    End Sub
    And be careful with subreports. If you use the same connection type as for the main report the subreport should piggy back on the main connection. If you use a different connection, you'll have to use code to connect the subreports:
    //loop through all the sections to find all the report objects
    foreach (CrystalDecisions.CrystalReports.Engine.Section crSection in crSections)
        crReportObjects = crSection.ReportObjects;
        //loop through all the report objects to find all the subreports
        foreach (CrystalDecisions.CrystalReports.Engine.ReportObject crReportObject in crReportObjects)
            if (crReportObject.Kind == ReportObjectKind.SubreportObject)
                //you will need to typecast the reportobject to a subreport
                //object once you find it
                crSubreportObject = (CrystalDecisions.CrystalReports.Engine.SubreportObject)crReportObject;
                string mysubname = crSubreportObject.SubreportName.ToString();
                //open the subreport object
                crSubreportDocument = crSubreportObject.OpenSubreport(crSubreportObject.SubreportName);
                //set the database and tables objects to work with the subreport
                crDatabase = crSubreportDocument.Database;
                crTables = crDatabase.Tables;
                tableIndex = 0;
                bool SecureDB;
                //loop through all the tables in the subreport and
                //set up the connection info and apply it to the tables
                foreach (CrystalDecisions.CrystalReports.Engine.Table crTable in crTables)
                {   // check if integrated security is enabled
                    SecureDB = crSubreportDocument.Database.Tables[tableIndex].LogOnInfo.ConnectionInfo.IntegratedSecurity;
                    string TableName = crTable.Name.ToString();
                    tableIndex++;
                    crConnectioninfo.ServerName = "VMW2K8CRSE2K8";
                    //if (!SecureDB)
                        crConnectioninfo.UserID = "sa";
                        crConnectioninfo.Password = "PW";
                        crConnectioninfo.DatabaseName = "QTYLTD";
                        crTableLogOnInfo = crTable.LogOnInfo;
                        crTableLogOnInfo.ConnectionInfo = crConnectioninfo;
                        crTable.ApplyLogOnInfo(crTableLogOnInfo);
    rpt.VerifyDatabase();
    For more info:
    SAP Crystal Reports .NET SDK Developer Guide
    SAP Crystal Reports .NET API Guide
    http://scn.sap.com/docs/DOC-21967
    http://scn.sap.com/docs/DOC-21953
    (Not sure if I gave you the last two links already.
    - Ludek

  • FIM Reporting Resume-FIMReportingInitialSync too slow

    I installed FIM Reporting services , and followed the document step by step.
    I started the "Start-FIMReportingInitialSync" script  , it ran for some time and stopped due to an error.The error reads
    "The client has been disconnected from the server.Please call ManagementGroup.Reconnect() to reestablish the connection."
    This issue as i understand might have occured due to a connectivity issue.
    Hence i used the "Resume-FIMReportingInitialSync" to start the initial synce.but this script is too slow and even slower than the "Start-FIMReportingInitialSync".
    Can anybody suggest how to speed up the Initial sync process.
    Also, if anybody can tell me where are the data stored moved by the script "Start-FIMReportingInitialSync"
    Regards
    shakti

    Any Update on the above issue.
    shakti

  • Crystal Report Query taking too much time

    Hi,
    We are developing one report based on SQL Server 2008 in Crystal Report. There are around 50,000 valid combination in database. Based on dynamic filter we need to bring few records in report. Since these filters are at report level, and crystal report is using microcube, it is taking more than 15 mins to execute.
    Is there any option to fetch record based on filter applied at report level.
    Regards
    Baby

    HI,
    First of all , thank you very much.
    Since having cascading prompt, we never thought in this way.
    Details:-
    For our report we have 4 prompts.
    1. category->family-brand (cascading- mandatory)
    2. season(madatory)
    3.collection (madatory)
    4.owner(not mandatory)
    previously we set all these filters at record level.
    Now we set season and collection at query level and brand, owner at report level. Report only query for selected season and collection only.
    Thanks once again.
    Regards
    Baby

  • Crystal Reports - Problem with too long BEx Query technical variable names?

    Hello!
    I have Crystal Reports 2008 SP3 and the SAP BO Integration Kit XI 3.1 SP3 installed and want to access a BEx query on a BW system within my Crystal Report.
    The problem is that the name of the BEx variable is not displayed correctly (just "[]" is displayed for the variable name) as a parameter within Crystal Reports when I import the query. Furthermore I can't navigate through the fields of the query within Crystal Reports.
    The technical names of the BEx query-variables have a length of 12 characters.
    The interessting thing is that if I change the technical names of the BEx query variables to a maximum length of 8 characters everything works fine.
    I use the SAP toolbar within Crystal Reports 2008 to create the report and have the "MDX-driver"-setting selected.
    Has anyone ever had the same problem? Any suggestions?
    I don't want to change the technical names of the variables of my queries...
    thanks,
    Dominik

    Hi!
    Just fixed the problem: The "Transports" for the Integration Kit had to be imported at the BW-System
    regards

  • Camileo X Sports - SD card performance issue - too slow

    Hello,
    can anybody help me please.....when i tun on the cam and make the FIRST record it always shows me that the speed card is too slow, but i have sandisk 64gb extreme with 48mbs.....but it only dissapears at the first record and not the following.
    what can i do....i have also the newest firmware.

    yup! I had the same problem at first, i have the Samy EVO class 10 32G SDHC.
    It's usually brilliant, actually it works fine with the 50 fps, 30 fps and even the 720p 120 fps.
    However i found the problem easier than most people simply due to sheer coincidence.
    I left the screen on while recording and right before i stopped recording (btw it never froze in my case it just stopped recording) and the screen displayed "low speed card"! what do you know!
    I researched the issue and it turns out the Samsun EVO was a bit slower than what the camera required at 60 fps. at 50 fps it barely stopped recording once.
    SanDisk Micro SDHC Extreme Class 3 32GB SDSDQXN-032G-G46A
    This card is a relatively cheap one, it is a UHS class 3 which mean it's an Ultra High Speed card. The class 3 is a great ranking. It's like 10 Euros more expensive than the Samsun and from what i've read works great with the GoPro hero 4 while shooting 4K & 1080p at 120 fps so it should hold up very well.
    It was never the upgrade or downgrade of the software. Lucky i left the screen on!

  • Crystal Report performance - subreports or sql command?

    Hello,
    bit of quandary, I've an existing report that contains six subreports all sharing the same parameter to the main report, a 'projectID' field.
    The main report approximately returns 300 records/projects and each of the 6 subreports are passed the corresponding 'projectID' field 300 times.
    Now, I've succesfully incorporated the main report query and corresponding subreports into a TSQL command containing a number of subqueries in the Select clause:
    SELECT   PROJECT.PROJECTID, PROJECT.TITLE, PROJECT.REFERENCE, PROJECTSTATUS.PROJECTSTATUS, PROJECT.INPUT_DATE
    ,(SELECT max(INPUT_DATE) FROM V_PROJECT_NOTE AS VPN
    WHERE PROJECT.PROJECTID = VPN.PROJECTID) AS LastHeadline
    ,(SELECT
    MAX(CASE
         WHEN MODIFIED_DATE IS NULL THEN INPUT_DATE
         WHEN INPUT_DATE > MODIFIED_DATE THEN INPUT_DATE
              ELSE MODIFIED_DATE
    END) FROM ISSUE WHERE PROJECT.PROJECTID = ISSUE.PROJECTID) AS LastNewIssue
    ,(SELECT max(ISSUENOTE.INPUT_DATE) FROM ISSUE
    INNER JOIN ISSUENOTE ON ISSUE.ISSUEID = ISSUENOTE.ISSUEID
    WHERE PROJECT.PROJECTID = ISSUE.PROJECTID ) AS LastIssueNote 
    ,(SELECT max(modified_date) FROM V_PROJECT_RISK_LAST_AMMENDED AS VPR
    WHERE PROJECT.PROJECTID = VPR.PROJECTID) AS LastRiskLogUpdate
    ,(SELECT max(INPUT_DATE) FROM PROJECT_CHECKLIST AS PC
    WHERE PROJECT.PROJECTID = PC.PROJECTID) AS LastChecklistUpdate
    ,(SELECT max(input_date) FROM V_PROJECT_ACTIVITY_LAST_AMMENDED AS VPA
    WHERE PROJECT.PROJECTID = VPA.PROJECTID) AS LastGANTTUpdate 
    ,V_PROJECT_KEYCONTACT.USERNAME AS KeyContact
    ,(SELECT     USERPROFILE.USERNAME AS Supervisor
    FROM         USERPROFILE INNER JOIN
                          PROJECT_MEMBER AS SUPERVISOR ON USERPROFILE.USERID = SUPERVISOR.USERID RIGHT OUTER JOIN
                          PROJECT_MEMBER AS KEYCONTACT ON SUPERVISOR.PROJECT_MEMBERID = KEYCONTACT.PARENTID
    WHERE PROJECT.PROJECTID = KEYCONTACT.PROJECTID
    AND KEYCONTACT.KEY_CONTACT =1) AS Supervisor
    FROM         PROJECT INNER JOIN
                          PROJECTSTATUS ON PROJECT.PROJECTSTATUSID = PROJECTSTATUS.PROJECTSTATUSID LEFT OUTER JOIN
                          V_PROJECT_KEYCONTACT ON PROJECT.PROJECTID = V_PROJECT_KEYCONTACT.PROJECTID
    WHERE V_PROJECT_KEYCONTACT.USERNAME IN ('aaa','bbb','ccc')
    AND (PROJECTSTATUS.PROJECTSTATUS IN ('111', '222', '333'))
    AND ((PROJECT.TITLE NOT LIKE 'xxx%') AND (PROJECT.TITLE NOT LIKE 'yyy%') AND (PROJECT.TITLE NOT LIKE 'zzz%'))
    ORDER BY V_PROJECT_KEYCONTACT.USERNAME, PROJECT.INPUT_DATE
    Now, I've run both SQL Server Profiler and looked at the Performance Information in Crystal and the SQL command method is an order of magnitude less efficient!!
    I may have to post my query on a TSQL forum but was wondering if anyone on here can possible shed any light on it?
    Thanks in advance,
    Dom

    Dom,
    I assume that the last edit was to remove a few curse words...
    The reality is that we can give a good explanation as to why the sub-report version would execute faster that the command version w/o seeing the report (with the sub-reports included).
    Looking at the SQL, I can tell that it's a fairly expensive query, firing several sub-queries for every row returned by the outer query. That said, I sounds like the same thing is taking place with the sub-report version too... So I would think that the performance would be similar.
    The only thing that I can think of, and this is just a guess... With the sub-report version, each all queries are being run independently and then combined locally by CR, whereas the command version, everything is combined. The more complex the query, the harder the optimizer has to work to come up with the best execution plan... and the greater the likelihood that that the resulting plan isn't as optimized as it could be.
    That said, I'd still think the command version would be faster over all. Are you seeing real differences between the two reports, in the amount of time that it takes from refresh to final render?
    Either way, I think moving all of those sub-queries from the select list (where they are executing once for every row) down to the FROM area (where they only have to execute once) should increase the speed dramatically and surpass the sub-report version by a decent margin.
    Give this version of your SQL a test drive and see if it yields an improvement.
    SELECT  
    PROJECT.PROJECTID,
    PROJECT.TITLE,
    PROJECT.REFERENCE,
    PROJECTSTATUS.PROJECTSTATUS,
    PROJECT.INPUT_DATE,
    VPN.LastHeadline,
    ISSUE.LastNewIssue,
    ISSUE_NOTE.LastIssueNote,
    VPR.LastRiskLogUpdate,
    PC.LastChecklistUpdate,
    VPA.LastGANTTUpdate,
    V_PROJECT_KEYCONTACT.USERNAME AS KeyContact,
    Supervisor.Supervisor
    FROM PROJECT
    INNER JOIN PROJECTSTATUS ON PROJECT.PROJECTSTATUSID = PROJECTSTATUS.PROJECTSTATUSID
    LEFT OUTER JOIN V_PROJECT_KEYCONTACT ON PROJECT.PROJECTID = V_PROJECT_KEYCONTACT.PROJECTID
    LEFT OUTER JOIN (
         SELECT PROJECTID,
         max(INPUT_DATE) AS LastHeadline
         FROM V_PROJECT_NOTE
         GROUP BYPROJECTID) AS VPN ON PROJECT.PROJECTID = VPN.PROJECTID
    LEFT OUTER JOIN (
         SELECT PROJECTID,
         MAX(CASE
              WHEN MODIFIED_DATE IS NULL THEN INPUT_DATE
              WHEN INPUT_DATE > MODIFIED_DATE THEN INPUT_DATE
              ELSE MODIFIED_DATE END) AS LastNewIssue
         FROM ISSUE
         GROUP BY PROJECTID) AS ISSUE ON PROJECT.PROJECTID = ISSUE.PROJECTID
    LEFT OUTER JOIN (
         SELECT ISSUE.PROJECTID,
         max(ISSUENOTE.INPUT_DATE) AS LastIssueNote
         FROM ISSUE
         INNER JOIN ISSUENOTE ON ISSUE.ISSUEID = ISSUENOTE.ISSUEID
         GROUP BY ISSUE.PROJECTID) AS ISSUE_NOTE ON PROJECT.PROJECTID = ISSUE_NOTE.PROJECTID
    LEFT OUTER JOIN (
         SELECT PROJECTID,
         max(modified_date) AS LastRiskLogUpdate
         FROM V_PROJECT_RISK_LAST_AMMENDED
         GROUP BY PROJECTID) AS VPR ON PROJECT.PROJECTID = VPR.PROJECTID
    LEFT OUTER JOIN (
         SELECT PROJECTID,
         max(INPUT_DATE) AS LastChecklistUpdate
         FROM PROJECT_CHECKLIST
         GROUP BY PROJECTID) AS PC ON PROJECT.PROJECTID = PC.PROJECTID
    LEFT OUTER JOIN (
         SELECT PROJECTID,
         max(input_date) AS LastGANTTUpdate
         FROM V_PROJECT_ACTIVITY_LAST_AMMENDED
         GROUP BY PROJECTID) AS VPA ON PROJECT.PROJECTID = VPA.PROJECTID
    LEFT OUTER JOIN (
         SELECT PROJECTID,
         USERPROFILE.USERNAME AS Supervisor
         FROM USERPROFILE
         INNER JOIN PROJECT_MEMBER AS SUPERVISOR ON USERPROFILE.USERID = SUPERVISOR.USERID RIGHT
         OUTER JOIN PROJECT_MEMBER AS KEYCONTACT ON SUPERVISOR.PROJECT_MEMBERID = KEYCONTACT.PARENTID
         WHERE KEYCONTACT.KEY_CONTACT =1) AS Supervisor ON PROJECT.PROJECTID = Supervisor.PROJECTID
    WHERE V_PROJECT_KEYCONTACT.USERNAME IN ('aaa','bbb','ccc')
    AND (PROJECTSTATUS.PROJECTSTATUS IN ('111', '222', '333'))
    AND ((PROJECT.TITLE NOT LIKE 'xxx%')
    AND (PROJECT.TITLE NOT LIKE 'yyy%')
    AND (PROJECT.TITLE NOT LIKE 'zzz%'))
    ORDER BY V_PROJECT_KEYCONTACT.USERNAME, PROJECT.INPUT_DATE
    HTH,
    Jason

Maybe you are looking for

  • Filter in mail does not work

    All of a sudden, filtering in Mail does not work. If I type a word that I KNOW appears in several emals, in the Search Mailbox field in the top right corner, it comes up empty. That happens even if I type the word "and". I have the All button selecte

  • Managing session using cookies..Need urgent help

    Hi, For my app thats running in Tomcat, I need to get the user authentication by using a cookie created by a perl application. Is that possible? I know the cookies name and the field names. Pls suggest me. Thnx in advance. Guhan

  • New to Firefox 8, how do i arrange bookmarks a b c .....

    New to Firefox 8, how do arrange bookmarks, alpha order , thanx

  • BBP_EXTREQ_TRANSFER fails if SMQ2 or SMQ1 Queues have errors

    We are using the standard functionality to send Plant Maintenance Orders from R/3 to SRM.  Our current landscape is SRM 5.0 and R/3 4.6C.  When the queues SMQ1 (Outbound R/3 Queue) or SMQ2 (Inbound SRM) Queues have other errors, not related to extern

  • Problem about printer

    sir,      my printer is hp printer .In which i got a problem of ''E'' THIS sign is stand for error for ink or paper jam .BUT there is no problem of ink an paper jam in that . when i replace the old cartrage with new one then the error sign gone out .