Large Report Model SMDL files

Hello! I am generating and modifying the report model programmatically and it comes out as big
as 260MB - for over 500 entities with hundreds of attributes in each of them.
When the report model is being uploaded into Report Manager, the ReportingServicesService.exe spikes up to 4GB in memory usage.
Of course, sometimes such an amount of memory is not available and the upload fails with OutOfMemory exception
Also, I cannot open the report model in the BIDS - I get Insufficient Memory exception.
Why would a 260MB model require 4GB of memory?
How can I reduce the size of my model and keep all my entities?
Thanks a lot!!

Hi aedna,
Generally, we always see the Insufficient Memory exception when Visual Studio truly runs out of memory. And if we upload 260MB file to Report Manager, the ReportingServicesService.exe spikes up to 4GB in memory usage is a normal phenomenon.  Microsoft
ASP.NET imposes a maximum size for items that are posted to the server. By default, this limit is 4 megabytes (MB). If you upload or publish a file that exceeds this limit to a report server, you receive an HTTP exception. In this case, you can modify the
default by increasing the value of the maxRequestLength element in the Machine.config file.
Besides, a report model is a collection of entities and their relationships that reflects the real-world relationships between business functions and processes. If we want to keep all entities, I think we cannot reduce the model size.
Thank you for your understanding.
Regards,
Katherine Xiong
Katherine Xiong
TechNet Community Support

Similar Messages

  • SSRS adhoc (2005/2008) - Need to create Report Model (.smdl file) for adhoc using Stored procedure

    Hi All,
    I need to create Report Models using stored procedures. But whenever I am going to create data sources for same, I am just getting tables and views to use. Please see the screenshot below:
    Is there any way out to create data source based on stored procedure? Please help.
    Thanks in Advance
    Regards
    Kumud

    Hi Kumud,
    As per my understanding, it is not support to create a Datasource View (DSV) using stored procedures.
    In a DataSource View, only views, tables or named queries can be used. We can use stored procedures in a query. No parameterized queries, parameterized stored procedures or parameterized UDFs can be used in a DSV named query. You can try to using below approach
    to achieve what you are require:
    To build views in our relational source and add the views to the datasource view to proceed with modelling.
    There is a similar issue, you can refer to it.
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/22207c21-03c7-4e5a-bb67-0372f29220a3/sql-server-reporting-services-2008-creating-report-models-using-stored-procedures
    Regards,
    Alisa Tang
    Alisa Tang
    TechNet Community Support

  • Unable to show all fields from a table\view within a SSRS Report Model for user selection

    I recently added a new view with a simple SELECT (with 3 fields) statement to our existing Report Model DSV. I then went ahead to the report model (.smdl) file proceeded with the auto-generation of the new attributes which of course ran with no issues.
    I was seeing the three new fields as attributes. But when I went within the existing Perspective to add the newly added fields so that the users see them, one of the field is missing. It is just a VARCHAR field. What would it cause this field to not show up
    within a Perspective?
    Any help is greatly appreciated.Thanks,
    Nimesh
    NM

    Hi NM,
    I am trying to involve someone more familiar with this topic for a further look at this issue. Sometime delay might be expected from the job transferring. Your patience is greatly appreciated.
    Thank you for your understanding and support.
    Regards,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Apex+Cocoon PDF Printing issue  on printing Large Reports

    Hello All,
    I m in a real hard situation with PDF printing in apex.. The application manages to push the report into PDF with lesser record reports but for Large reports, it says File corrupted when i open the downloaded PDF file..
    I use Apex+Coocoon and fop printing. I found from some forums as this issue was with Java Heap size being lesser and asked me to increase the size.. so i increased the size from 128MB to 1024MB and my server settings allows me to set 1550MB at the maximum..
    If i cross this limit the TOMCAT server stops from running...
    But all of the above was not much fruitful, since it throws again the same errors if report rows crosses 6000+ rows, can any one help me out, do i need to check any other settings to make my large reports printable in PDF.
    My application needs to print around *20000+ rows each report may have at least 20 columns* to get printed...
    Much awaited for your assistance
    Edited by: mohansanjee on Sep 8, 2010 3:58 AM
    Edited by: mohansanjee on Sep 8, 2010 3:59 AM
    Edited by: mohansanjee on Sep 8, 2010 4:00 AM

    Matt,
    On this note, it probably also makes sense to mention that with APEX 3.1 we introduce a PL/SQL API that gives you much broader control over what is generated how and the method it's made available to the client. You'll get the ability to retrieve the PDF in a BLOB directly on the server, store in your own tables for later download or send via email attachment. You'll also get the ability to write your own download process using after-submit page processes. And you can have combinations of your own XML with your own stylesheets, or combine report queries with customer stylesheets at runtime, or just simply retrieve the results of pre-defined report queries / report layouts in BLOBs.
    I'll talk about this much more on my Blog once 3.1 is out, and will provide whitepapers, how-to's etc
    Regards,
    Marc

  • Report model - number of values selected in DDL param

    Good afternoon,
    I have created a report model that references Teradata views, and I am using this model to populate several datasets in a Report Builder report.  One dataset returns actual results, and the other datasets are used to populate available values for multivalued
    parameters.
    The issue I'm encountering is that some of my parameters return a large number (tens or hundreds of thousands) of selectable values.  When clicking "select all", I'm getting an error, since the concatenated list of values exceeds
    the length limit.  However, since I'm working with a report model, I cannot use T-SQL workarounds such as custom "<ALL>" values/handling.  I also need to limit my result set using these values as parameters, not filters, since I want
    the intensive processing handled in the DB server, not the app server.
    Is there any possible way to accommodate this?  As an example, I have a large (200k+) list of products, belonging to about 7,000 brands.  I want both my product number and brand lists as multivalue parameters...I might want to run
    the report for all products and all brands (returning 200k+ rows), or choose "select all" in the products list, but only pick a few brands (or vice versa).  Cascading parameters (with the first parameter limiting the next set of values)
    is not an option since there are many of these parameters that have too many values for the allowed limit (i.e. even my brand list generated the error when I choose "select all" and only picked a few products).
    I have searched literally dozens of forums and tech sites, all to no avail thus far.  Any assistance will be greatly appreciated...thank you!
    -Chris

    I have a SSRS Report with 4 multi-select parameters.
    The customer has requested that they would like the drill-through to open in a new window.
    I am using Java Script for that purpose.
    However, there is a limit on how long a URL may be.
    Is there a way in a SSRS report to limit the number of values in a multi-select parameter selected?
    Thanks.
    One workaround is to add a internal parameter which will count the number of selected values and will allow the navigation only if the count falls below the threshold.
    For this you can use an expression like below for setting the internal parameter
    Len(Join(Parameters!MultiValuedParameter.Value,","))-Replace(Join(Parameters!MultiValuedParameter.Value,","),",","")
    Then use expression like below for the jump to url/report property
    IIf(Val(Parameters!HiddenParameter.Value)<=<your thresholdvalue>,<report url>,Nothing)
    You can also add a notification textbox on top which will show a message like "Report navigation not possible due to too many values selected" and keep it hidden by default. Then based on above parameter value you can make it visible
    ie make hidden property like below
    IIf(Val(Parameters!HiddenParameter.Value)<=<your thresholdvalue>,False,True)
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My Wiki User Page
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • REP-0002: Unable to retrieve a string from the Report Builder message file;

    Hi,
    I've a custom report in which i need to display a large string, of more than 4000 characters. To cater to this requirement i've written a formula column which displays string upto 4k characters and for a string of size beyond 4k i am calling a procedure which uses a temp table having a clob field.
    For a small string the report runs fine but whenever a large string requirement comes into the picture, said procedure gets triggered and i get REP-0002: Unable to retrieve a string from the Report Builder message file.
    OPP log for the same gives an output as under:
    Output type: PDF
    [9/21/10 2:17:12 PM] [UNEXPECTED] [388056:RT14415347] java.io.FileNotFoundException: /app/soft/test1udp/appsoft/inst/apps/e180r_erptest01/logs/appl/conc/out/o14415347.out (No such file or directory)
         at java.io.FileInputStream.open(Native Method)
         at java.io.FileInputStream.<init>(FileInputStream.java:106)
         at oracle.apps.fnd.cp.opp.XMLPublisherProcessor.process(XMLPublisherProcessor.java:241)
         at oracle.apps.fnd.cp.opp.OPPRequestThread.run(OPPRequestThread.java:172)
    Report Builder 10g
    BI Publisher's version : 4.5.0
    RDBMS Version : 10.2.0.4.0
    Oracle Applications Version : 12.0.6
    Searched for the same on metalink and found Article ID 862644.1, other than applying patch and upgrading version of BI Publisher is there any other workaround to display a comma seperated string as long as 60k characters, If any please suggest.
    Thanks,
    Bhoomika

    Metalink <Note:1076529.6> extracts
    Problem Description
    When running a *.REP generated report file you get errors:
    REP-00002 Unable to retrieve string from message file
    REP-01439 Cannot compile program since this is a runtime report
    Running the same report using the *.RDF file is successful.
    You are running the report with a stored procedure,
    OR you are running against an Oracle instance other than the one developed on,
    Or, you recently upgraded your Oracle Server.
    Solution Description
    1) Check that the user running the report has execute permissions on any stored
    SQL procedures called. <Note:1049458.6>
    2) Run the report as an .RDF not an .REP , that is remove or rename the .REP
    version of the report. <Note:1012546.7>
    3) Compile the report against the same instance it is being run against.
    <Note:1071209.6> and <Note:1049620.6>

  • How to convert Report Model to Report Designer DataSet

    To all,
    I have a two part question. One is I am experiencing the error at the end of this question when migrating SQL Server 2005 SSRS Native Reports, the report uses a Report Model, to SQL Server 2008 SSRS Native. I bring the Report and Report Model to SQL Server 2008 SSRS and get the error at the end of the question below. Background and the 2nd question are below.
    I have a bunch of Report Builder based Reports. I need to convert them to VS BIDS based Report Designer Reports with Shared Datasources and MDX based DataSets.
    How can I grab the MDX from the Report Builder Report Model so I can put it into the Report Designer DataSet Query Definition?
    What is the easiest way to make this migration between these Report development environments.
    I am able to open the Report.RDL file in VS BIDS and can see the entire UI layer. The DataSet I see here is the one that is related to the error at the end of the question. The RD/DataSet holds:
    <SemanticQuery xmlns="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling" xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:rb="http://schemas.microsoft.com/sqlserver/2004/11/reportbuilder" xmlns:qd="http://schemas.microsoft.com/sqlserver/2004/11/semanticquerydesign">
      <Hierarchies>
        <Hierarchy>
          <BaseEntity>
            <!--Fact Messages-->
            <EntityID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Entity_MeasureGroup_DW_Text_Fact_Messages</EntityID>
          </BaseEntity>
          <Groupings>
            <Grouping Name="Date">
              <Expression Name="Date">
                <Path>
                  <RolePathItem>
                    <!--Dim Date-->
                    <RoleID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Role_MeasureGroupDimension_Entity_MeasureGroup_DW_Text_Fact_Messages_Entity_Dimension_Dim_Date_Entity_Dimension_DW_Text_Dim_Date</RoleID>
                  </RolePathItem>
                </Path>
                <AttributeRef>
                  <!--Full Date-->
                  <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Hierarchy_Dim_Date.Full_Date</AttributeID>
                </AttributeRef>
              </Expression>
            </Grouping>
            <Grouping Name="Hour">
              <Expression Name="Hour">
                <Path>
                  <RolePathItem>
                    <!--Dim Time-->
                    <RoleID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Role_MeasureGroupDimension_Entity_MeasureGroup_DW_Text_Fact_Messages_Entity_Dimension_Dim_Time_Entity_Dimension_DW_Text_Dim_Time</RoleID>
                  </RolePathItem>
                </Path>
                <AttributeRef>
                  <!--Hour24-->
                  <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Hierarchy_Dim_Time.Hour24</AttributeID>
                </AttributeRef>
              </Expression>
            </Grouping>
            <Grouping Name="Year">
              <Expression Name="Year">
                <Path>
                  <RolePathItem>
                    <!--Dim Date-->
                    <RoleID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Role_MeasureGroupDimension_Entity_MeasureGroup_DW_Text_Fact_Messages_Entity_Dimension_Dim_Date_Entity_Dimension_DW_Text_Dim_Date</RoleID>
                  </RolePathItem>
                </Path>
                <AttributeRef>
                  <!--Year-->
                  <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Hierarchy_Dim_Date.Year</AttributeID>
                </AttributeRef>
              </Expression>
            </Grouping>
          </Groupings>
          <Filter>
            <Expression Name="expr1">
              <Function>
                <FunctionName>And</FunctionName>
                <Arguments>
                  <Expression>
                    <Function>
                      <FunctionName>And</FunctionName>
                      <Arguments>
                        <Expression>
                          <Function>
                            <FunctionName>GreaterThanOrEquals</FunctionName>
                            <Arguments>
                              <Expression>
                                <Path>
                                  <RolePathItem>
                                    <!--Dim Date-->
                                    <RoleID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Role_MeasureGroupDimension_Entity_MeasureGroup_DW_Text_Fact_Messages_Entity_Dimension_Dim_Date_Entity_Dimension_DW_Text_Dim_Date</RoleID>
                                  </RolePathItem>
                                </Path>
                                <AttributeRef>
                                  <!--Date-->
                                  <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Hierarchy_Dim_Date.Date</AttributeID>
                                </AttributeRef>
                              </Expression>
                              <Expression>
                                <Function>
                                  <FunctionName>DateAdd</FunctionName>
                                  <Arguments>
                                    <Expression>
                                      <Literal>
                                        <DataType>String</DataType>
                                        <Value>Day</Value>
                                      </Literal>
                                    </Expression>
                                    <Expression>
                                      <Function>
                                        <FunctionName>Negate</FunctionName>
                                        <Arguments>
                                          <Expression>
                                            <ParameterRef>
                                              <ParameterName>Date  in last  (n)  days</ParameterName>
                                            </ParameterRef>
                                          </Expression>
                                        </Arguments>
                                      </Function>
                                    </Expression>
                                    <Expression>
                                      <Function>
                                        <FunctionName>Today</FunctionName>
                                      </Function>
                                    </Expression>
                                  </Arguments>
                                </Function>
                              </Expression>
                            </Arguments>
                          </Function>
                        </Expression>
                        <Expression>
                          <Function>
                            <FunctionName>LessThan</FunctionName>
                            <Arguments>
                              <Expression>
                                <Path>
                                  <RolePathItem>
                                    <!--Dim Date-->
                                    <RoleID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Role_MeasureGroupDimension_Entity_MeasureGroup_DW_Text_Fact_Messages_Entity_Dimension_Dim_Date_Entity_Dimension_DW_Text_Dim_Date</RoleID>
                                  </RolePathItem>
                                </Path>
                                <AttributeRef>
                                  <!--Date-->
                                  <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Hierarchy_Dim_Date.Date</AttributeID>
                                </AttributeRef>
                              </Expression>
                              <Expression>
                                <Function>
                                  <FunctionName>Today</FunctionName>
                                </Function>
                              </Expression>
                            </Arguments>
                          </Function>
                        </Expression>
                      </Arguments>
                    </Function>
                    <CustomProperties>
                      <CustomProperty Name="qd:FilterCondition" />
                    </CustomProperties>
                  </Expression>
                  <Expression>
                    <Function>
                      <FunctionName>Equals</FunctionName>
                      <Arguments>
                        <Expression>
                          <Path>
                            <RolePathItem>
                              <!--Dim Volume-->
                              <RoleID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Role_MeasureGroupDimension_Entity_MeasureGroup_DW_Text_Fact_Messages_Entity_Dimension_Dim_Volume_Entity_Dimension_DW_Text_Dim_Volume</RoleID>
                            </RolePathItem>
                          </Path>
                          <AttributeRef>
                            <!--Description-->
                            <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Hierarchy_Dim_Volume.Description</AttributeID>
                          </AttributeRef>
                        </Expression>
                        <Expression>
                          <Literal>
                            <DataType>String</DataType>
                            <Value>Volume</Value>
                          </Literal>
                        </Expression>
                      </Arguments>
                    </Function>
                    <CustomProperties>
                      <CustomProperty Name="qd:FilterCondition" />
                    </CustomProperties>
                  </Expression>
                </Arguments>
              </Function>
              <CustomProperties>
                <CustomProperty Name="qd:Filter" />
                <CustomProperty Name="qd:ContextEntityID">
                  <Value xsi:type="xsd:string">http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling:Entity_MeasureGroup_DW_Text_Fact_Messages</Value>
                </CustomProperty>
                <CustomProperty Name="qd:AutoChangeBaseEntity" />
                <CustomProperty Name="qd:Design">
                  <Value xsi:type="xsd:string">expr2</Value>
                </CustomProperty>
              </CustomProperties>
            </Expression>
          </Filter>
        </Hierarchy>
      </Hierarchies>
      <MeasureGroups>
        <MeasureGroup>
          <BaseEntity>
            <!--Fact Messages-->
            <EntityID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Entity_MeasureGroup_DW_Text_Fact_Messages</EntityID>
          </BaseEntity>
          <Measures>
            <Expression Name="Volume">
              <AttributeRef>
                <!--Volume-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Volume</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Auto %">
              <AttributeRef>
                <!--Automated_Pct-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Automated_Pct</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Send Now %">
              <AttributeRef>
                <!--Send_Now_Pct-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Send_Now_Pct</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Use %">
              <AttributeRef>
                <!--Use_Pct-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Use_Pct</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Stock Ans %">
              <AttributeRef>
                <!--StockAns_Pct-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_StockAns_Pct</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Charged">
              <AttributeRef>
                <!--Charged_Pct-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Charged_Pct</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Billed">
              <AttributeRef>
                <!--Billed_Pct-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Billed_Pct</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Avg TTA (s)">
              <AttributeRef>
                <!--Avg_TT_Answer-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Avg_TT_Answer</AttributeID>
              </AttributeRef>
            </Expression>
            <Expression Name="Avg TTC (s)">
              <AttributeRef>
                <!--Avg_Time_To_Customer-->
                <AttributeID xmlns:np="http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling">np:Attribute_Measure_DW_Text_Avg_Time_To_Customer</AttributeID>
              </AttributeRef>
            </Expression>
          </Measures>
          <SubtotalSets>
            <SubtotalSet>
              <SubtotalGroupings>
                <GroupingName>Year</GroupingName>
              </SubtotalGroupings>
              <SubtotalMeasures>
                <MeasureName>Volume</MeasureName>
                <MeasureName>Auto %</MeasureName>
                <MeasureName>Send Now %</MeasureName>
                <MeasureName>Use %</MeasureName>
                <MeasureName>Stock Ans %</MeasureName>
                <MeasureName>Charged</MeasureName>
                <MeasureName>Billed</MeasureName>
                <MeasureName>Avg TTA (s)</MeasureName>
                <MeasureName>Avg TTC (s)</MeasureName>
              </SubtotalMeasures>
            </SubtotalSet>
            <SubtotalSet>
              <SubtotalGroupings>
                <GroupingName>Date</GroupingName>
                <GroupingName>Year</GroupingName>
              </SubtotalGroupings>
              <SubtotalMeasures>
                <MeasureName>Volume</MeasureName>
                <MeasureName>Auto %</MeasureName>
                <MeasureName>Send Now %</MeasureName>
                <MeasureName>Use %</MeasureName>
                <MeasureName>Stock Ans %</MeasureName>
                <MeasureName>Charged</MeasureName>
                <MeasureName>Billed</MeasureName>
                <MeasureName>Avg TTA (s)</MeasureName>
                <MeasureName>Avg TTC (s)</MeasureName>
              </SubtotalMeasures>
            </SubtotalSet>
          </SubtotalSets>
        </MeasureGroup>
      </MeasureGroups>
      <Parameters>
        <Parameter Name="Date  in last  (n)  days">
          <DataType>Integer</DataType>
          <Expression>
            <Literal>
              <DataType>Integer</DataType>
              <Value>14</Value>
            </Literal>
          </Expression>
        </Parameter>
      </Parameters>
      <CustomProperties>
        <CustomProperty Name="qd:PerspectiveID">
          <Value xsi:type="xsd:string">http://schemas.microsoft.com/sqlserver/2004/10/semanticmodeling/udmmodeling:Perspective_Cube_DW_Text</Value>
        </CustomProperty>
      </CustomProperties>
    </SemanticQuery>
    I have created a Shared Datasource that points the the correct Analysis Services cubes that the original Report Model pointed at.
    However, I get this error when running the report:
    An error occurred during client rendering.
    An error has occurred during report processing.
    Semantic query compilation failed: e EmptySemanticQuery The SemanticQuery does not contain any Groupings or MeasureGroups. SemanticQuery must contain at least one of these elements. (SemanticQuery '').

    Actually the error above is from running the report on SSRS 2008 Native. Does anyone know why I get this error and how it may be resolved.
    Then I try to bring the report into BIDS RD and create a SDS and DataSet. The Semantic Query above is from the RD DataSet Query window.
    I get the following error when running the report in BIDS RD:
    An error occured during local processing.
    An error has occured during report processing.
    Query execution failed for dataset 'dataSet'.
    The SemanticQuery element at line 7, column 351 (namespace http://schema.microsoft.com/sqlserver/2004/10semanticmodeling) cannot appear under Envelope/Body/Execute/Command.
    So I have two errors with two scenarios. I'd like to know how to resolve one of them.
    I'd also like to know how to get the MDX from the RB/RM report so I can put it into the VS BIDS Version and get that to work properly.
    Thanks,
    Eric

  • Best approach for batch mode generation of large report via SSRS

    I have revised this question now that I've determined that we *will* be using the Enterprise edition of SQL Server 2012.
    I have a report (that I will deploy to an SSRS 2012 server) which can vary greatly in size. When the report is expected to be small (which we can determine ahead of time by means that are not relevant to this post), it will be rendered by a .NET 4.5 web
    application using a ReportViewer control (in remote mode).
    When the report will be large, we want the web app to "trigger" the report server to generate the report and deposit it as a PDF file or Excel file in a folder on a file share, and then send an e-mail notification to a configured user. Because
    the reports can be very large, I would prefer not to stream them to the web app and from there to the file share. We have experienced browser timeouts in the past, and so I would prefer to have the SSRS server deposit the generated report directly to the file
    share.
    About 8 parameters have to be passed to the stored procedure which serves as the report’s data source. 
    The database where the stored procedure resides will soon be upgraded to SQL Server 2012.
    I am fairly new to reporting services and wanted to ask what approach would be best suited for accomplishing the above-stated goals.
    From what I've read so far, I know that one possible approach would be to use data-driven subscriptions but the question then becomes:  after inserting report parameters into the subscription table, how does my web app trigger report generation?
    Also, I want to deposit the report in a file share and send an e-mail notification to the user. Do I have to create two subscriptions? There is a configuration option when setting up e-mail notification to embed a hyperlink in the e-mail rather than actually
    include the report in the message. If I select this option, do I give up the ability to specify the destination folder?

    @JoeSchmoe115:  Thanks for your suggestion.  Using a web service or WCF service would still mean dealing with potential timeouts for very large reports (granted you can adjust those timeouts).    
    I think an approach using a console app could work, but rather than having the web app spawn it, an alternative would be to do the following:  (1) have the web app insert report generation requests (including all necessary report parameters) into
    a database table, and (2) have a console app, periodically invoked by Windows Task Scheduler, which reads that database table and then uses the SSRS API as you describe to generate the report, save the byte stream to a file, and finally send
    the e-mail notification to the user.  
    mkedwards
    What you've described will work but saving report generation requests in separate table may not be necessary. I don't have much experience with web services but can they just spawn C# console app with needed parameters and exit without waiting for console
    app to complete?
    The positive side of using database to store report generation requests is that you would have complete log of all requests. Your console app can then update database with completion status.

  • Oracle 10g Discoverer Reports & export to xls fails for large reports

    Hi ,
    We have following configurations:
    1: RHEL 5.4
    2: Discoverer :Version 10.1.2.48.18
    3: Oracle10g Apps Version : Version 10.1.2.0.2
    Issue:
    Most of small reports works fine ....but when large discoverer reports are executed the page
    keeps on refreshing for 15-20 hours but no output ....same for export to xls ......
    But same reports works fine in oracle9i for same data voulme....
    Observations:
    When on linux with top command the processes are monitored its observed that discoverer process
    dis51ws dies for large reports after 1-2 minutes ...& the page keeps on refreshing but no output....
    for 1-2 minutes it consumes 50-80% cpu utilisation then process disappears & cpu 80% idle ...
    It seems that as 10g Apps is installed on RHEL 5.4 ...non certfified OS causing an issue...
    Can any one adds more inputs in this regards......
    we have checked logs : below are log details :
    Below logs gives "Logkeys: exceptions discoiv.servlet_exceptions" for this report ...
    1:
    OC4J~OC4J_BI_Forms~default_island~1:
    10/04/11 12:11:29 Oracle Application Server Containers for J2EE 10g (10.1.2.0.2) initialized
    10/04/11 12:11:59 Using oracle.reports.util.EnvironmentGlobal class
    10/04/11 14:23:37 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:23:38 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114240115278
    10/04/11 14:25:42 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:25:42 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114254315439
    10/04/11 14:28:53 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:28:54 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114285615691
    10/04/11 14:29:13 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:29:13 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114291315728
    10/04/11 14:32:35 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:32:35 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114323615949
    10/04/11 14:32:48 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:32:48 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114324815982
    10/04/11 14:34:44 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:34:44 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114344616128
    10/04/11 14:34:55 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:34:55 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114345516155
    10/04/11 14:36:25 Tutalii: /oracle10gas/app/oracle10g/discoverer/lib/discoverer5.jar archive
    2:
    Discoverer~SessionServer~12
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Active Eul: EULADMIN
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    DCSCORBAInterface::Delete called
    DCSCORBAInterface destructor called
    DCSCORBAInterface::Delete called
    DCSCORBAInterface destructor called
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    3:
    application.log
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Finding async request action forward
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Calling externalize
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_TOOL: dvtb xk-1ml versionzw1.0w kyxdvtbyxbisltyxbicho vzwwjyxjbisltyxjdvtby
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_VIEW: dv xk-1ml versionzw1.0w kyxdv bazw0w cszw25wyxpc vzw1wjyxjdvy
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_VIEW: lc
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_HS: dvhs
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_DS: dv_ds &qls_z!2=New GL Report.Clndr Id&arq=false&qls_x!14=Sheet 1.Closing Balance (Credit)&qls_x!3=New GL Report.Voucher No&qls_z!3=New GL Report.Frm Prd&fm=xml&qls_z!4=New GL Report.To Prd&qls_x!2=New GL Report.Prd Desc&qls_x!11=Sheet 1.Debit Amount&qls_x!4=New GL Report.Gl Voucher No&qls_x!9=Sheet 1.Opening Balance Debit&tss_s!0=New GL Report.Acct Id,lh,group,false&qls_x!1=New GL Report.Acct Sdesc&qls_z!1=New GL Report.Loc Desc&qls_x!10=Sheet 1.Opening Balance (Credit)&qls_x!12=Sheet 1.Credit Amount&aow=false&qls_x!7=New GL Report.Shrt Code&qls_x!13=Sheet 1.Closing Balance (Debit)&qls_x!6=New GL Report.Pmt Rct Dt&qls_x!8=New GL Report.Dtl Nrtn&sss=true&qls_x!5=New GL Report.Voucher Dt&qls_x!0=New GL Report.Acct Id&qls_z!0=New GL Report.Loc Id&ddsver=1
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.ViewerModelImpl.externalize [EXTERN_STATE]: $wksht$%24dv_ds%24%26qls_z%212%3DNew+GL+Report.Clndr+Id%26arq%3Dfalse%26qls_x%2114%3DSheet+1.Closing+Balance+%28Credit%29%26qls_x%213%3DNew+GL+Report.Voucher+No%26qls_z%213%3DNew+GL+Report.Frm+Prd%26fm%3Dxml%26qls_z%214%3DNew+GL+Report.To+Prd%26qls_x%212%3DNew+GL+Report.Prd+Desc%26qls_x%2111%3DSheet+1.Debit+Amount%26qls_x%214%3DNew+GL+Report.Gl+Voucher+No%26qls_x%219%3DSheet+1.Opening+Balance+Debit%26tss_s%210%3DNew+GL+Report.Acct+Id%2Clh%2Cgroup%2Cfalse%26qls_x%211%3DNew+GL+Report.Acct+Sdesc%26qls_z%211%3DNew+GL+Report.Loc+Desc%26qls_x%2110%3DSheet+1.Opening+Balance+%28Credit%29%26qls_x%2112%3DSheet+1.Credit+Amount%26aow%3Dfalse%26qls_x%217%3DNew+GL+Report.Shrt+Code%26qls_x%2113%3DSheet+1.Closing+Balance+%28Debit%29%26qls_x%216%3DNew+GL+Report.Pmt+Rct+Dt%26qls_x%218%3DNew+GL+Report.Dtl+Nrtn%26sss%3Dtrue%26qls_x%215%3DNew+GL+Report.Voucher+Dt%26qls_x%210%3DNew+GL+Report.Acct+Id%26qls_z%210%3DNew+GL+Report.Loc+Id%26ddsver%3D1%24dv%24xk-1ml+versionzw1.0w+kyxdv+bazw0w+cszw25wyxpc+vzw1wjyxjdvy%24wd%24false%24lc%24%24dvtb%24xk-1ml+versionzw1.0w+kyxdvtbyxbisltyxbicho+vzwwjyxjbisltyxjdvtby%24wvs%241101%24dvhs%24$cn$&vct=svd&cnk=cf_a101$ap$%26df%3D%26l%3D%26s%3D%26nc%3D%26dl%3D$expl$&sp=&node=&event=focus&state=(117)&root=63&wbt=2$prid$NEW_GL_REPORT%2F31$ctyp$viewer
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Storing state
    10/04/11 14:37:25 discoverer: [INFO] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Externalize Perf: 8ms
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Saving Attributes
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Saving ApplicationRequest
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Checking for SSO mode
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Saving errors, messages
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Returning final forward: "/ExportProgress.uix" redirect: "false"
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute ----------------------- End Request --------------------------
    10/04/11 14:37:25 discoverer: [INFO] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Total Request Time in AppCtrl: 18
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] org.apache.struts.action.RequestProcessor.processForwardConfig processForwardConfig(ForwardConfig[name=long running operation,path=/ExportProgress.uix,redirect=false,contextRelative=false])
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.view.DiscovererPageBroker.isCacheable Setting cacheable to: true
    4: XML log:
    log2010041114285615691.xml
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>184</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:29:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Timer started.]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>162</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:29:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="TRACE"></MSG_TYPE><MSG_LEVEL>5</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Return DCSModelInterface::SendReceiveData(kScheduleInterface, inTable, outTable)
    outTable = DCITable
         Length=0
    ]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcsmdli.cpp</FILE_NAME> <LINE_NUMBER>259</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:29:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[DCSModelInterface::SendReceiveData(kScheduleInterface)]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS </MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcsmdli.cpp</FILE_NAME> <LINE_NUMBER>226</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> <LOG_SIZE>0</LOG_SIZE> <EXTRA_INFO><MethodEnd duration="0.1" sizeChange="0" >
    real 0m0.1s
    user 0m0.900s
    sys 0m0.109s
    </MethodEnd></EXTRA_INFO> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:37:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Timer stopped.]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>184</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:37:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:37:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Timer started.]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>162</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:37:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    Regards,

    Hi ,
    as per Note: 466697.1 ....its an memory error....& need to increase MaxVirtualDiskMem and MaxVirtualHeapMem
    But we already have slowly increased MaxVirtualDiskMem and MaxVirtualHeapMem to below very high values ...but the issue remains same.......
    as per note we are getting Logkeys: exceptions discoiv.servlet_exceptions error but
    after that we are not getting below error ............
    Unexpected error in state machine: java.lang.OutOfMemoryError
    Hence I presume that its different issue rather than memmory....
    Below are pref.txt values..........
    CacheFlushPercentage          = 25          # Percent of cache flushed if the cache is full. valid values 0 - 100%.
    MaxVirtualDiskMem          = 9294967296     # Maximum amount of disk memory allowed for the data cache. Should be greater than or equal to MaxVirtualHeapMem
    MaxVirtualHeapMem          = 4294967296     # Maximum amount of heap memory allowed for the data cache.
    QueryBehavior = 0          # Action to take after opening a workbook (0 = Run Query Automatically, 1 = Don't Run Query, 2 = Ask for Confirmation)
    Also we have raised an SR & as per SR .............all below settings are tried as per SR except for applying a recent patch....
    ====================================================================
    Discoverer performance is largely determined by how well the database
    has been designed and tuned for queries.
    A well-designed database will perform significantly better than a poorly
    designed database.
    Workbook design can also affect query performance.
    1. Apply latest Discoverer patch as documented in <<Note:237607.1>>
    'ALERT: Required and Recommended Patch Levels For All Discoverer Version'.
    2. Increase the maximum JVM heap memory:
    In general, the default values for the minimum heap (-Xms) memory and
    maximum heap (-Xmx) memory are sufficient.
    However, if your organization consistently runs large Discoverer queries
    then you may benefit from increasing the maximum heap memory from the
    default values.
    This is recomended if your users are typically running large queries via
    Discoverer Viewer.
    Increasing the JVM memory can help to avoid "java.lang.OutOfMemoryError"
    in Discoverer Viewer:
    Please see <<Note 563960.1>>, Best Practice: Configuring The
    OC4J_BI_Forms JVM For
    Discoverer Viewer/Portlet 10g Performance And Stability for specific
    details.
    3. Disable Query Prediction:
    Query Prediction provides an estimate of the time required to retrieve
    the information in a query.
    The Query Prediction appears before the query and consumes time.
    Edit the <oracle_home>/discoverer/util/pref.txt on the middle-tier
    server and set:
    QPPEnable=0
    Also set:
    QPPObtainCostMethod = 0
    4. Uncheck the 'Enable fantrap detection' checkbox.
    When the box is checked, every query generated by Discoverer is
    interrogated. Discoverer will detect a fan trap and rewrite the query to
    ensure that the aggregation is done at the correct level.
    Please refer to <<Note:210553.1>>, Oracle BI Discoverer: Fan Trap
    Resolution - Correct Results
    Everytime for more information on fantraps.
    To disable, in Plus go to Tools -> Options -> Advanced -> Fan Trap
    settings.
    In Viewer go to Preferences and uncheck the box.
    5. Disable Materialized Views/Summaries
    In pref.txt add parameter:
    MaterializedViewRedirectionBehavior = 0
    Value equal to 0 ensures that Materialized View (MV) Redirection is
    always performed when MVs are available.
    6. Improve query performance by optimizing the SQL.
    In pref.txt modify/add following parameters:
    SQLFlatten = 0
    SQLItemTrim = 0
    SQLJoinTrim = 0
    UseOptimizerHints = 0
    If value of SQLFlatten is 1 then Discoverer will merge inline views in
    the query SQL where ever possible.
    In case of SQLItemTrim, Discoverer will remove unused folder items from
    the query SQL where possible and for SQLJoinTrim Discoverer will remove
    unnecessary joins from the query where possible.
    UseOptimizerHints will add Optimizer hints to SQL if set >
    0.Unnecessarily making Discoverer perform these checks consumes
    resources and rather than increasing the performance may reduce the same. So unless you feel these checks have to be performed depending on
    requirements, assign zero to these parameters.
    7. When Discoverer builds a query, Discoverer makes a database security
    check to confirm that the user has access to the tables referenced in
    the folders. Avoiding this check can save time.
    So in pref.txt underDatabase section add:
    ObjectsAlwaysAccessible = 1
    8. Whenever you are creating conditions, ensure that you match the Case.
    This in turn can reduce the time Discoverer spends on changing the Case
    and matching.
    For example:
    option Upper(Department) in (Upper('VIDEO SALES')
    9. Ensure that summaries are refreshed periodically in Discoverer
    Administrator.
    10. Increase the amount of memory available for the Discoverer data cache. Please refer to <<Note 245752.1>>, Explaining Oracle BI Discoverer
    Session Memory Management
    And Server Cache Settings.
    11. Performance may be enhanced by enabling OracleAS Web Cache.
    ==================================================================
    Thanks for your reply....................
    Regards,

  • Adhoc Reporting using Data Sources instead of Report Model

    Hi Gurus,
    I'm working on SSRS 2008. I'm trying to create Adhoc Report using Report Builder 3.0. To select Data Source, both *.rsds and *.smdl are listed. I selected a Data Source which points to a SQL database. When I add New datasets, I'm able to add the fields and
    it automatically detect relationships (Joins) using Primary Key and Foreign Key defined in the tables. I'm also able to edit the query. I ran the report and it ran successfully.
    My question is I couldn't find any documentation regarding using Data Source for Adhoc Reporting. MS SQL documentation and forums insists that Report Model is required for Adhoc Reporting.
    Please let me know if someone tried to create Adhoc Report using Data Source instead of Report Model. While I understand there are more benefits using Report Model (define entities/attributes, show/hide entity/attributes, organize entities/attributes etc.),
    I just want to know if this is possible.
    Are there any known limitation using this method?
    Thanks In Advance!

    Hi Ereswaran,
    Yes, we can create ad hoc report using report using data source instead of Report Model. Ad hoc reports typically are used to answer a specific question at appoint in time, so consequently people don’t spend a lot of time arranging and formatting the data
    in a precise way. Other report development tools are available in Reporting Services to support more sophisticated reporting requirements.
    Using Report Builder, we can build a simple report that display data form a single data source as defined by a report model. This tool is most useful for business users who need access to information but lack the skills to write a SQL query. If we have full
    experience of SQL query, of course we can use data source instead of Report Model.
    If you have any questions, please feel free to ask.
    Regards,
    Charlie Liao
    Charlie Liao
    TechNet Community Support

  • Data Model best Practices for Large Data Models

    We are currently rolling out Hyperion IR 11.1.x and are trying to establish best practces for BQYs and how to display these models to our end users.
    So far, we have created an OCE file that limits the selectable tables to only those that are within the model.
    Then, we created a BQY that brings in the tables to a data model, created metatopics for the main tables and integrated the descriptions via lookups in the meta topics.
    This seems to be ok, however, anytime I try to add items to a query, as soon as i add columns from different tables, the app freezes up, hogs a bunch of memory and then closes itself.
    Obviously, this isnt' acceptable to be given to our end users like this, so i'm asking for suggestions.
    Are there settings I can change to get around this memory sucking issue? Do I need to use a smaller model?
    and in general, how are you all deploying this tool to your users? Our users are accustomed to a pre built data model so they can just click and add the fields they want and hit submit. How do I get close to that ideal with this tool?
    thanks for any help/advice.

    I answered my own question. in the case of the large data model, the tool by default was attempting to calculate every possible join path to get from Table A to Table B (even though there is a direct join between them).
    in the data model options, I changed the join setting to use the join path with the least number of topics. This skipped the extraneous steps and allowed me to proceed as normal.
    hope this helps anyone else who may bump into this issue.

  • Error moving through large report

    Post Author: dbronstein
    CA Forum: General
    We have an ASP.NET 2.0 app with Crystal XI R2.  When we try to move to the end of a large report, we get this message:
    Failed to retrieve data from the database. Details: &#91;Database Vendor Code: 17 &#93; Error in File C:\WINDOWS\TEMP\annproc {72FCA892-D200-4116-9D3E-BA22AEE7DD45}.rpt: Failed to retrieve data from the database. Details: &#91;Database Vendor Code: 17 &#93;
    Small reports work fine.  We get the same error when we try to export the report to excel. Does anyone have any ideas for a workaround?

    I believe KBase [1200141 - How to set the focus of the .NET Windows CrystalReportViewer so the report preview scrolls without clicking|http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/oss_notes_boj/sdn_oss_boj_erq/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/scn_bosap/notes.do]. Unfortunately as I type this, there is something wrong with the search and I keep getting 403 errors. Below is the contents of the kbase:
    *Symptom *
    In Crystal Reports XI and XI R2 the report page view for a .NET Windows application will not scroll up/down with the mouse wheel.
    *Cause *
    This is the default behavior of the CrystalReportViewer control for .NET Windows applications but can be changed with some custom code.
    *Resolution *
    In order to fix this problem you must set the focus of the Windows application to the main report tab in the CrystalReportViewer control.
    This action must be done in the parent form's Activated event. Below is some sample code that demonstrates how this can be accomplished.
    // This is a member variable which allows us to set the focus on the first time the parent form is activated.
    // The form is activated quite often and we only want to focus change to happen once.
    bool m_bFirstTimeActivated = true;
    private void Form1_Activated(object sender, System.EventArgs e)
       if (m_bFirstTimeActivated)
       // The CrystalReportViewer control is made up of 5 controls. The control in the first position is the PageView
       // control and is the one that needs to be modified. The PageView contains a TabControl that allows us to
       // have the main report and all subreports displayed as other tabs.
       // We set the focus to the first tab in the collection so scrolling works automatically.
       System.Windows.Forms.TabControl oTabControl = (System.Windows.Forms.TabControl) crystalReportViewer1.Controls[0].Controls[0];
       oTabControl.TabPages[0].Focus(); 
       m_bFirstTimeActivated = false;
    Ludek
    Edited by: Ludek Uher on Jul 5, 2010 11:26 AM

  • Timout error when schedule a large report

    I received a timeout error when I scheduled a large report. To refresh and save it as an excel file, and then send to a file location. I can do that with a small report. Could I change the setting somewhere to increase the bach process time? Thanks!
    Failed
    External Destination:     File copy the instance with the filename: "Inventory Usage STE2008-11-20-11-48-52.xls" to the directory: "//osh_fs/share/Inventory Usage Report/STE" .
    Creation Time:     11/20/2008 11:48 AM
    Start Time:     11/20/2008 11:48 AM
    End Time:     11/20/2008 12:08 PM
    Server Used:     oshbo1.Web_IntelligenceJobServer
    Error Message:     Transport error: No response from server, timeout exceeded.

    Hi,
    For scheduling a lrage webi report:
    Go into the CCM, and then Stop the Input, Output, WebI Job Server.
    In the Command line of WebI Job Server :
    Before -restart add " -requestTimeout 7200000"
    i.e: ......procwebi -requestTimeout 7200000 -restart.........
    In the Input FRS:
    Do the same as above.
    ........ -requestTimeout 7200000 -restart......
    In the Output Server do the same as well: add -requestTimeout 7200000 before -restart.
    Start all the servers.
    -Shreyash.

  • How do I make a flexible report model

    we are migrating all our reports from cognos to ssrs. most of our cognos reports are using catalog which is equivalent with a report model. in cognos, we don't define any relationship between the tables on the catalog files to make it flexible tochanges
    on the underlying database tables. all the relationships are defined in the cognos reports. i find that you can't do this in ssrs. ssrsstrictly enforce you to define the relationship in the model. how do i make a report model to be flexible?

    Hi Blusil,
    I don’t know much about Cognos, but for Report Model in Reporting Services I have something to say. A report model is a user-friendly description of an underlying database, with pre-established data relationships and auto-generated queries. Report models
    can be used as data sources for reports created in Report Designer and Report Builder 3.0.
    We can set up connection information, define a view of the data source, define a report model, and publish the model to a report server. Then we can use the report model to build reports using Report Builder. For more details about Creating a Report Model,
    please see the following tutorial:
    http://technet.microsoft.com/en-us/library/ms167048(v=sql.105).aspx
    Hope this helps.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Report Model with mulitple data sources

    Hi,
    We have requirement to develop report model in 2008 having multiple data sources which includes SQL SERVER 2012 and Oracle database so end user can all the relevant data in one place to run in report builder. So please advise how this can be achieved.
    One option I can think of pulling all the data into one database through SSIS and then build a smdl on it.
    I heard it might be possible to build Tabular cube and then use Report builder on it not sure which is the best option unless any other solution.
    Thanks,

    You can use SSIS to collect data from multiple sources (SQL Server, Oracle, etc.) and store it in SQL Server database. Then you can use SSAS to build a cube on that database. With SSRS/reportbuilder you can create reports on the cube or on the database...
    Please mark the post as answered if it answers your question | My SSIS Blog:
    http://microsoft-ssis.blogspot.com |
    Twitter

Maybe you are looking for