Date to Date Report For BSo Cube

Hi All,
We are plan to create one bso cube. we want to build time dimension include dates,I want to load data and generate reports date wise.I know its full difficult to get the reports date wise.please help me out.

You could use relative names for the MEMBERS and then change the aliases each month:
Member CM-0 Alias Sep 2012
Member CM-1 Alias Aug 2012
Member CM- ...
You would have to reload all data each month but you could load to the alias names so the input files would not have to change.
Then just build your reports with the member names and have Essbase respond to the query with the Alias.

Similar Messages

  • Showing master data value in report for which no record in Infoprovider

    Hi,
    We have the following reporting requirement. The master data values , e.g. 0Customer should appear in the report for which no sales has taken place, i.e no record in Infocube.
    We have defined a multicube using Infoprovider & 0Customer Infoobject, but the report seems to show only values from InfoCube, not the values which are not there in infoprovider but are present in infoobject.
    Please suggest.
    Thanks & Regards
    Madhu
    00919830957798

    Madhu,
    There is a how to paper on how to do this. It available in the BW InfoIndex under Multiproviders and is called Slow Moving Inventory Scenario.
    https://service.sap.com/~sapidb/011000358700008387152002
    This document details out the process of reporting on master data values for which no transactions ahve been recorded. This is achieved by creating a MultiProvider on a Cube and InfoObject and then using teh record count property.
    Cheers
    Aneesh

  • Is it possib to create inline report for a cube if wedon't have any measure

    Hi,
    Is it Possible to create an Inline report based on an application Rowsource even when we do not have any measure associated to that Rowsource in the current cube?
    Please help...

    We have a case where for any specific dates in fiscal dimension or any particular product, user want to list whether its a special event or not.
    To handle this we have created an uploadable report where user enter for which date there is a special event.We created a custom rowsource having columns for given dimension in cube and did the cube mapping.We also created inline report for the same. But not able to see data in that report.

  • Searching for Data in the Report for a non-selected column in display.

    Hello,
    I have an interactive report as given below:
    select user, dept_name, sal, description, title from hr;
    Now, in the report display options I remove the title and description from the display as you know title and description are 2000 characters each.
    If I show then it takes the whole page to display one record and with this option I can search any text in the Title or Description columns.
    Now, I have hidden these two columns and when I search I do not see any records.
    Can any one tell me how to achieve this behaviour.
    What I am looking is, I do not show the column 'TITLE' or 'DESCRIPTION' and user show be able to search for any important data inside these columns.
    Please let me know how can we do this.
    Please point me to an already published article or solution or any help on this type.
    Appreciate your help on this.
    thanks
    Edited by: user648778 on Sep 18, 2008 6:13 PM

    As far as I can see this will not work. You could use a trick and make a query like this:
    SELECT USER, dept_name, sal,
              apex_item.hidden (10, mgrdescription || title)
           || '<img src="#IMAGE_PREFIX#some_nice_icon.gif">' title_description
      FROM hrThis would not display the title and description but an image and it would give you a possibility to filter on those.
    Denes Kubicek
    http://deneskubicek.blogspot.com/
    http://www.opal-consulting.de/training
    http://apex.oracle.com/pls/otn/f?p=31517:1
    -------------------------------------------------------------------

  • Updating metadata/data in ASO reporting cube from BSO

    Hi Experts,
    I am creating an ASO reporting cube and would like to know the best way to automate the update of metadata and data from a BSO cube without the use of a partition? This is my first time working with ASO. Any insight or examples would be much appreciated. Please advise.
    Thanks in advance!

    Thank you for the responses!
    The dimensions will be the same. I will be exporting level0 data then loading it into the ASO cube. I believe I should be able to use the same load rule that I am using for the BSO cube, correct? Also, I am planning on only loading the 'Final' versions of 'Budget' and 'Forecast'. In my data export I believe I can put a condition to only export the 'Final' version, please correct me if I'm wrong?
    One last thing, I would like to keep 2 years of 'Actuals' and 1 year of 'Budget' and 'Forecast'. At year end I would to clear previous 'Budget' and 'Forecast' and 2nd prior year after the current 'Budget' and 'Forecast' are closed. Can this be handled in a separate script? Please advise.
    Thanks again for the responses.

  • Sharepoint 2013 Reporting Services & OLAP Cubes for Data Modeling.

    I've been using PowerPivot & PowerView in Excel 2013 Pro for some time now so am now eager to get set up with Sharepoint 2013 Reporting Services.
    Before set up Reporting Services  I have just one question to resolve.
    What are the benefits/differences of using a normal flat table set up, compared to an OLAP cube?
    Should I base my Data Model on an OLAP Cube or just Connect to tables in my SQL 2012 database?
    I realize that OLAP Cubes aggregate data making it faster to return results, but am unclear if this is needed with Data Modeling for Sharepoint 2013.
    Many thanks,
    Mike

    So yes, PV is an in-memory cube. When data is loaded from the data source, it's cached in memory, and stored (compressed) in the Excel file. (also, same concept for SSAS Tabular mode... loads from source, cached in mem, but also stored (compressed) in data
    files, in the event that the server reboots, or something similar).
    As far as performance, tabular uses memory, but has a shorter load process (no ETL, no cube processing)... OLAP/MDX uses less memory, by requiring ETL and cube processing... technically tabular uses column compression, so the memory consumption will be based
    on the type of data (numeric data is GREAT, text not as much)... but the decision to use OLAP (MDX)/TAB (DAX) is just dependent on the type of load and your needs... both platforms CAN do realtime queries (ROLAP in multidimensional, or DirectQuery for tabular),
    or can use their processed/in-memory cache (MOLAP in multidimensional, xVelocity for tabular) to process queries.
    if you have a cube, there's no need to reinvent the wheel (especially since there's no way to convert/import the BIDS/SSDT project from MDX to DAX). If you have SSAS 2012 SP1 CU4 or later, you can connect PV (from Excel OR from within SP) directly to the
    MDX cube.
    Generally, the benefit of PP is for the power users who can build models quickly and easily (without needing to talk to the BI dept)... SharePoint lets those people share the reports with a team... if it's worthy of including in an enterprise warehouse,
    it gets handed off to the BI folks who vet the process and calculations... but by that time, the business has received value from the self-service (Excel) and team (SharePoint) analytics... and the BI team has less effort since the PP model includes data sources
    and calculations - aside from verifying the sources and calculations, BI can just port the effort into the existing enterprise ETL / warehouse / cubes / reports... shorter dev cycle.
    I'll be speaking on this very topic (done so several times already) this weekend in Chicago at SharePoint Saturday!
    http://www.spschicagosuburbs.com/Pages/Sessions.aspx
    Scott Brickey
    MCTS, MCPD, MCITP
    www.sbrickey.com
    Strategic Data Systems - for all your SharePoint needs

  • Data is not getting cleared in BSO cube.

    Hi Gurus,
    I am facing issue with one of my BSO cubes(Target) which is in transparent partition with the ASO cube(Source). Earlier the data was double the input and when i cleared using maxl script the data is halfed but as per my script values should be zeros. I need the values in the BSO cube to be zeros but not able to make it and the values halfed are present and i tried different ways. Please help me how to resolve this issue.
    script used is very simple in this way:
    FIX(Account, Year)
    Cleardata "Scenario";
    ENDFIX
    Thanks in Advance.
    regards,
    902970.

    Hi Brian,
    Thanks for your quick reply, Actually we load the input to a static scenario and then the data is copied to the relevant scenario(Actual/Forecast) and need this to be cleared.
    Scenario Dimension:
    Scenario(Label only)
    Static Scenario(~)
    Plan(~)
    Actual Scenario(~)
    Forecast Scenario(~)
    Thanks,
    902970.

  • How to get start date of the period for a given date from cube

    I have a situation where i need to find the start day of the period for a given date. is there a way to know that. i want to use that in my report. i enter the date from my report(i have date parameter), depends on the date, i want to display the start day
    of the period. how can i write expression for that in my report?
    ram

    Hi ramprasad74,
    According to your description, you are using Analysis Services as a data source for the report, the cube has hierarchy: Fyear, FQuarter, FPeriod, fweek, Fdate. You want to add a date parameter to the report, after you changed value of the parameter, the
    report will return the first day of FPeriod.
    To achieve your goal, we need to add a parameter to the report, then use the parameter in mdx query. For detail information, please refer to the following steps:
    In the Report Data pane, right-click on a dataset created from a SQL Server Analysis Services data source type, and then click Query. The MDX query designer opens in Design mode.
    On the toolbar, click Design to toggle to Query mode.
    On the MDX query designer toolbar, click Query Parameters symbol. The Query Parameters dialog box opens.
    In the Parameter column, click <Enter Parameter>, and then type the name of a parameter.
    In the Dimension column, choose a value from the drop-down list.
    In the Hierarchy column, choose a value from the drop-down list.
    In the Default column, from the drop-down list, select a single value.
    Click OK. 
    In query designer dialog box, type the mdx query like below:
    with member [Measures].[FirstChild]
    as
    [Date].[Fiscal].currentmember.parent.firstchild.name
    select {[Measures].[FirstChild]} on 0,
    [Date].[Fiscal].[Date].members on 1
    from
    ( SELECT ( STRTOSET(@ParameterName, CONSTRAINED) ) on 0
    from
    [Cube]
    Here are relevant threads you can reference:
    https://social.msdn.microsoft.com/forums/sqlserver/en-US/c7146ac3-40ea-4d53-b321-c707aebbd405/how-to-pass-date-parameter-to-mdx-query
    https://social.msdn.microsoft.com/forums/sqlserver/en-US/fd12a865-bc90-4a65-af42-ce38a8cfa29b/pass-date-time-parameter-to-mdx-query-ssrs
    If you have any more questions, please feel free to ask.
    Thanks,
    Wendy Fu
    If you have any feedback on our support, please click
    here.

  • Loading BPC data back to BW for reporting.

    Hi All,
           We are in NW BPC 7.5 and the tool is not matching up to the drill-down reporting capabilities of BW so we need to pull the plan data back to BW for Plan/Actual reporting. I am interested in knowing about the best practice to achieve this.
    1. Flat Files dump and loads - not acceptable as it beats the purpose of NW based BPC.
    2. BW ETL - I understand the technical names of BPC Application Cube change after optimiztion and between Dev Q and Prod so ETL would need constant changes.
    Any other options / recommendations ?
    Regards
    Anupam

    I agree with Ethan's reply as being the preferred route.  But regarding your original question and the option of using ETL.  Keep in mind that the only way the InfoCube might change when full optimization is run is IF there are more than 13 dimensions in the application.  Otherwise, BW assigns each characteristic to its own BW dimension (of which there are possible 16 with 3 being predefined).  So if you have an application with less than 13 dimensions and you feel this will be stable, then you can probably safely go ahead and build an ETL process directly against the BPC cube without risk that it will be impacted by optimization.
    Best regards,
    [Jeffrey Holdeman|http://wiki.sdn.sap.com/wiki/display/profile/Jeffrey+Holdeman]
    SAP Labs, LLC
    BusinessObjects Division
    Americas Applications Regional Implementation Group (RIG)

  • Taking too long for Loaded data visible in Reporting

    Hi experts,
    I have a cost center cube, which takes very very long to show the  " Loaded data is not yet visible in Reporting " icon for the request. I donot have aggregate on the cube and no other requests are Red /yellow. What is the reason for this weird nature. I checked the consistency check, there are no over lap /errors for the request.
    Any help is appreciated.
    thanks in adavnce
    D Bret

    Hi Dave,
    Try to run the RSRV for the Cube and try to refresh the data in the cube.
    RSRV --> Combined test --> transaction data --> Status of data in the info cube.
    If you find any errors after running this check,repair it and then refresh the cube's data.
    Hope this helps.
    Thanks & Regards,
    Suchitra.V

  • Performance Tuning Data Load for ASO cube

    Hi,
    Anyone can help how to fine tune data load on ASO cube.
    We have ASO cube which load around 110 million records from a total of 20 data files.
    18 of the data files has 4 million records each and the last two has around 18 million records.
    On average, to load 4 million records it took 130 seconds.
    The data file has 157 data column representing period dimension.
    With BSO cube, sorting the data file normally help. But with ASO, it does not seem to have
    any impact. Any suggestion how to improve the data load performance for ASO cube?
    Thanks,
    Lian

    Yes TimG it sure looks identical - except for the last BSO reference.
    Well nevermind as long as those that count remember where the words come from.
    To the Original Poster and to 960127 (come on create a profile already will you?):
    The sort order WILL matter IF you are using a compression dimension. In this case the compression dimension acts just like a BSO Dense dimension. If you load part of it in one record then when the next record comes along it has to be added to the already existing part. The ASO "load buffer" is really a file named <dbname.dat> that is built in your temp tablespace.
    The most recent x records that can fit in the ASO cache are still retained on the disk drive in the cache. So if the record is still there it will not have to be reread from the disk drive. So you could (instead of sorting) create an ASO cache as large as your final dat file. Then the record would already still be on the disk.
    BUT WAIT BEFORE YOU GO RAISING YOUR ASO CACHE. All operating systems use memory mapped IO therefore even if it is not in the cache it will likely still be in on the disk in "Standby" memory (the dark blue memory as seen in Resource Monitor) this will continue until the system runs out of "Free" memory (light blue in resource monitor).
    So in conclusion if your system still has Free memory there is no need (in a data load) to increase your ASO cache. And if you are out of Free memory then all you will do is slow down the other applications running on your system by increasing ASO Cache during a data load - so don't do it.
    Finally, if you have enough memory so that the entire data file fits in StandBY + Free memory then don't bother to sort it first. But if you do not have enough then sort it.
    Of course you have 20 data files so I hope that you do not have compression members spread out amongst these files!!!
    Finally, you did not say if you were using parallel load threads. If you need to have 20 files read up on having parrallel load buffers and parallel load scripts. that will make it faster.
    But if you do not really need 20 files and just broke them up to load parallel then create one single file and raise your DLTHREADSPREPARE and DLTHREADSWRITE settings. Heck these will help even if you do go parallel and really help if you don't but still keep 20 separate files.

  • MB5B Report for Date Wise Stock and Value

    Hi,
    I am Taking MB5B Report for Date Wise Stock and Value.
    But I have one doubt all stocks is coming or not in these report like Unrestrected Stock,Return Stock.Blocked stock,Transist stock,Restrected Stock,qty Inspection Stock.
    I have another Doubt in these report three Stock type indicaters are there like these.
    1.Storage Location / Batch Stock
    2.valuated Stock
    3.Special Stock.
    But i have one doubt what is defferent these
    1.Storage Location / Batch Stock
    2.valuated Stock

    Hi Prasad,
    Yes MB5B report consider the Unrestricted, Quality, Blocked, Transit stock and restricted stock. Not sure about Return Stock.
    If you select the Storage location/Batch stock radio button then the system will display all the possible stock from the storage location and the corresponding batch also.
    If you select Valuated stock radio button then system will show only the valuated stock not the Non-valuated stock. Because Non-valuated material type is available in SAP system will not show those stock suppose if you select the Valuated stock radio button.
    Regards
    Karthick

  • I need the Log Report for the Data which i am uploading from SAP R/3.

    Hi All,
    I am BI 7.0 Platform with Support Patch 20.
    I need the Log Report for the Data which i am uploading from SAP R/3.
    I extract the DATA from R/3 into BI 7.0 DSO where I am mapping the GL Accounts with the FS Item.   In the Transformation i have return a routine on the FS Item InfObject . I am checking the Gl code into Z table for the FS Item .
    I capture the FS item from the Z table then update this FS item to Infobject FS item.
    Now i  need to stop the Data upload if i do not find the GL code in the Z table, and generate report for all GL code for which the FS item is not maintained in the Z table.
    Please suggest.
    Regards
    nilesh

    Hi.
    Add a field that you will use to identify if the GL account of the record was found in the Z table or not. Fx, create ZFOUND with length 1 and no text.
    In your routine, when you do the lookup, populate ZFOUND with X when you found a match (sy-subrc = 0) and leave it blank if you don't find a match. Now create a report filtering on ZFOUND = <blank> and output the GL accounts. Those will be the ones not existing in the Z table, but coming in from your transactions.
    Regards
    Jacob

  • How to create a report with data using the Crystal Reports for Java SDK

    Hi,
    How do I create a report with data that can be displayed via the Crystal Report for Java SDK and the Viewers API?
    I am writing my own report designer, and would like to use the Crystal Runtime Engine to display my report in DHTML, PDF, and Excel formats.  I can create my own report through the following code snippet:
    ReportClientDocument boReportClientDocument = new ReportClientDocument();
    boReportClientDocument.newDocument();
    However, I cannot find a way to add data elements to the report without specifying an RPT file.  Is this possible?  I seems like it is since the Eclipse Plug In allows you to specify your database parameters when creating an RPT file.
    is there a way to do this through these packages?
    com.crystaldecisions.sdk.occa.report.data
    com.crystaldecisions.sdk.occa.report.definition
    Am I forced to create a RPT file for the different table and column structures I have? 
    Thank you in advance for any insights.
    Ted Jenney

    Hi Rameez,
    After working through the example code some more, and doing some more research, I remain unable to populate a report with my own data and view the report in a browser.  I realize this is a long post, but there are multiple errors I am receiving, and these are the seemingly essential ones that I am hitting.
    Modeling the Sample code from Create_Report_From_Scratch.zip to add a database table, using the following code:
    <%@ page import="com.crystaldecisions.sdk.occa.report.application.*"%>
    <%@ page import="com.crystaldecisions.sdk.occa.report.data.*"%>
    <%@ page import="com.crystaldecisions.sdk.occa.report.document.*"%>
    <%@ page import="com.crystaldecisions.sdk.occa.report.definition.*"%>
    <%@ page import="com.crystaldecisions.sdk.occa.report.lib.*" %>
    <%@ page import = "com.crystaldecisions.report.web.viewer.*"%>
    <%
    try { 
                ReportClientDocument rcd = new ReportClientDocument();
                rcd.newDocument();
    // Setup the DB connection
                String database_dll = "Sqlsrv32.dll";
                String db = "qa_start_2012";
                String dsn = "SQL Server";
                String userName = "sa";
                String pwd = "sa";
                // Create the DB connection
                ConnectionInfo oConnectionInfo = new ConnectionInfo();
                PropertyBag oPropertyBag1 = oConnectionInfo.getAttributes();
                // Set new table logon properties
                PropertyBag oPropertyBag2 = new PropertyBag();
                oPropertyBag2.put("DSN", dsn);
                oPropertyBag2.put("Data Source", db);
                // Set the connection info objects members
                // 1. Pass the Logon Properties to the main PropertyBag
                // 2. Set the Server Description to the new **System DSN**
                oPropertyBag1.put(PropertyBagHelper.CONNINFO_CRQE_LOGONPROPERTIES, oPropertyBag2);
                oPropertyBag1.put(PropertyBagHelper.CONNINFO_CRQE_SERVERDESCRIPTION, dsn);
                oPropertyBag1.put("Database DLL", database_dll);
                oConnectionInfo.setAttributes(oPropertyBag1);
                oConnectionInfo.setUserName(userName);
                oConnectionInfo.setPassword(pwd);
                // The Kind of connectionInfos is CRQE (Crystal Reports Query Engine).
                oConnectionInfo.setKind(ConnectionInfoKind.CRQE);
    // Add a Database table
              String tableName = "Building";
                Table oTable = new Table();
                oTable.setName(tableName);
                oTable.setConnectionInfo(oConnectionInfo);
                rcd.getDatabaseController().addTable(oTable, null);
        catch(ReportSDKException RsdkEx) {
                out.println(RsdkEx);  
        catch (Exception ex) {
              out.println(ex);  
    %>
    Throws the exception
    com.crystaldecisions.sdk.occa.report.lib.ReportSDKException: java.lang.NullPointerException---- Error code:-2147467259 Error code name:failed
    There was other sample code on SDN which suggested the following - adding the table after calling table.setDataFields() as in:
              String tableName = "Building";
                String fieldname = "Building_Name";
                Table oTable = new Table();
                oTable.setName(tableName);
                oTable.setAlias(tableName);
                oTable.setQualifiedName(tableName);
                oTable.setDescription(tableName) ;
                Fields fields = new Fields();
                DBField field = new DBField();
                field.setDescription(fieldname);
                field.setHeadingText(fieldname);
                field.setName(fieldname);
                field.setType(FieldValueType.stringField);
                field.setLength(40);
                fields.add(field);
                oTable.setDataFields(fields);
                oTable.setConnectionInfo(oConnectionInfo);
                rcd.getDatabaseController().addTable(oTable, null);
    This code succeeds, but it is not clear how to add that database field to a section.  If I attempt to call the following:
    FieldObject oFieldObject = new FieldObject();
                oFieldObject.setDataSourceName(field.getFormulaForm());
                oFieldObject.setFieldValueType(field.getType());
                // Now add it to the section
                oFieldObject.setLeft(3120);
                oFieldObject.setTop(120);
                oFieldObject.setWidth(1911);
                oFieldObject.setHeight(226);
                rcd.getReportDefController().getReportObjectController().add(oFieldObject, rcd.getReportDefController().getReportDefinition().getDetailArea().getSections().getSection(0), -1);
    Then I get an error (which is not unexpected)
    com.crystaldecisions.sdk.occa.report.lib.ReportDefControllerException: The field was not found.---- Error code:-2147213283 Error code name:invalidFieldObject
    How do I add one of the table.SetDataFields()  to my report to be displayed?
    Are there any other pointers or suggestions you may have?
    Thank you

  • AP Invoice Aging Report for past date

    hi guys,
    I am running R12.
    my requirement is to run AP invoice aging report for past dates (same like AP Trial Balance Report), but the seeded invoice aging report has no option to do so.
    would appreciate some help
    thanks in advance
    Ravi

    Hi Senthil,
    I am running R12.
    my requirement is to run AP invoice aging report for past dates (same like AP Trial Balance Report
    if u have any solution to this , please suggest me how to resolve this requierement.
    thanks in advance
    Ravi
    mail:[email protected]

Maybe you are looking for